Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
12 changes: 6 additions & 6 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -102,7 +102,7 @@ Here we will keep track of the latest AI Game Development Tools, including LLM,
| [Lumina-T2X](https://github.com/Alpha-VLLM/Lumina-T2X) | Lumina-T2X is a unified framework for Text to Any Modality Generation. |[arXiv](https://arxiv.org/abs/2405.05945) | | Tool |
| [MetaGPT](https://github.com/geekan/MetaGPT) | The Multi-Agent Framework | | | Tool |
| [MiniCPM-2B](https://github.com/OpenBMB/MiniCPM) | An end-side LLM outperforms Llama2-13B. | | | Tool |
| [MiniGPT-4](https://github.com/Vision-CAIR/MiniGPT-4) | Enhancing Vision-language Understanding with Advanced Large Language Models. |[arXiv](https://arxiv.org/abs/2304.10592) | | Tool |
| [MiniGPT-4](https://github.com/Vision-CAIR/MiniGPT-4) | EenhancingVision-language Understanding with Advanced Large Language Models. |[arXiv](https://arxiv.org/abs/2304.10592) | | Tool |
| [MiniGPT-5](https://github.com/eric-ai-lab/MiniGPT-5) | Interleaved Vision-and-Language Generation via Generative Vokens. |[arXiv](https://arxiv.org/abs/2310.02239) | | Tool |
| [Mixtral 8x7B](https://mistral.ai/news/mixtral-of-experts/) | A high quality Sparse Mixture-of-Experts. |[arXiv](https://arxiv.org/abs/2401.04088) | | Tool |
| [Mistral 7B](https://mistral.ai/news/announcing-mistral-7b/) | The best 7B model to date, Apache 2.0. | | | Tool |
Expand Down Expand Up @@ -595,9 +595,9 @@ Here we will keep track of the latest AI Game Development Tools, including LLM,
| [CogVideo](https://models.aminer.cn/cogvideo/) | Generate Videos from Text Descriptions. | | | Video |
| [CogVideoX](https://github.com/THUDM/CogVideo) | CogVideoX is an open-source version of the video generation model, which is homologous to 清影. | | | Video |
| [CogVLM](https://github.com/THUDM/CogVLM) | CogVLM is a powerful open-source visual language model (VLM). | | | Visual |
| [CoNR](https://github.com/megvii-research/CoNR) | Genarate vivid dancing videos from hand-drawn anime character sheets(ACS). |[arXiv](https://arxiv.org/abs/2207.05378) | | Video |
| [CoNR](https://github.com/megvii-research/CoNR) | Generates vivid dancing videos from hand-drawn anime character sheets(ACS). |[arXiv](https://arxiv.org/abs/2207.05378) | | Video |
| [Decohere](https://www.decohere.ai/) | Create what can't be filmed. | | | Video |
| [Descript](https://www.descript.com/) | Descript is the simple, powerful , and fun way to edit. | | | Video |
| [Descript](https://www.descript.com/) | Descript is the simple, powerful, and fun way to edit. | | | Video |
| [Diffutoon](https://github.com/modelscope/DiffSynth-Studio) | High-Resolution Editable Toon Shading via Diffusion Models. |[arXiv](https://arxiv.org/abs/2401.16224) | | Video |
| [dolphin](https://github.com/kaleido-lab/dolphin) | General video interaction platform based on LLMs. | | | Video |
| [DomoAI](https://domoai.app/) | Amplify Your Creativity with DomoAI. | | | Video |
Expand Down Expand Up @@ -668,7 +668,7 @@ Here we will keep track of the latest AI Game Development Tools, including LLM,
| [Stable Video Diffusion](https://github.com/Stability-AI/generative-models) | Stable Video Diffusion (SVD) Image-to-Video. | | | Video |
| [StoryDiffusion](https://github.com/HVision-NKU/StoryDiffusion) | Consistent Self-Attention for Long-Range Image and Video Generation. |[arXiv](https://arxiv.org/abs/2405.01434) | | Video |
| [StreamingT2V](https://github.com/Picsart-AI-Research/StreamingT2V) | Consistent, Dynamic, and Extendable Long Video Generation from Text. |[arXiv](https://arxiv.org/abs/2403.14773) | | Video |
| [StyleCrafter](https://gongyeliu.github.io/StyleCrafter.github.io/) | nhancing Stylized Text-to-Video Generation with Style Adapter. |[arXiv](https://arxiv.org/abs/2312.00330) | | Video |
| [StyleCrafter](https://gongyeliu.github.io/StyleCrafter.github.io/) | enhancingStylized Text-to-Video Generation with Style Adapter. |[arXiv](https://arxiv.org/abs/2312.00330) | | Video |
| [TATS](https://songweige.github.io/projects/tats/index.html) | Long Video Generation with Time-Agnostic VQGAN and Time-Sensitive Transformer. | | | Video |
| [Text2Video-Zero](https://github.com/Picsart-AI-Research/Text2Video-Zero) | Text-to-Image Diffusion Models are Zero-Shot Video Generators. |[arXiv](https://arxiv.org/abs/2303.13439) | | Video |
| [TF-T2V](https://tf-t2v.github.io/) | A Recipe for Scaling up Text-to-Video Generation with Text-free Videos. |[arXiv](https://arxiv.org/abs/2312.15770) | | Video |
Expand All @@ -689,7 +689,7 @@ Here we will keep track of the latest AI Game Development Tools, including LLM,
| [VideoFactory](https://arxiv.org/abs/2305.10874) | Swap Attention in Spatiotemporal Diffusions for Text-to-Video Generation. | | | Video |
| [VideoGen](https://videogen.github.io/VideoGen/) | A Reference-Guided Latent Diffusion Approach for High Definition Text-to-Video Generation. |[arXiv](https://arxiv.org/abs/2309.00398) | | Video |
| [VideoLCM](https://arxiv.org/abs/2312.09109) | Video Latent Consistency Model. |[arXiv](https://arxiv.org/abs/2312.09109) | | Video |
| [Video LDMs](https://research.nvidia.com/labs/toronto-ai/VideoLDM/) | Align your Latents: High- resolution Video Synthesis with Latent Diffusion Models. |[arXiv](https://arxiv.org/abs/2304.08818) | | Video |
| [Video LDMs](https://research.nvidia.com/labs/toronto-ai/VideoLDM/) | Align your Latents: High-resolution Video Synthesis with Latent Diffusion Models. |[arXiv](https://arxiv.org/abs/2304.08818) | | Video |
| [Video-LLaVA](https://github.com/PKU-YuanGroup/Video-LLaVA) | Learning United Visual Representation by Alignment Before Projection. |[arXiv](https://arxiv.org/abs/2311.10122) | | Video |
| [VideoMamba](https://github.com/OpenGVLab/VideoMamba) | State Space Model for Efficient Video Understanding. |[arXiv](https://arxiv.org/abs/2403.06977) | | Video |
| [Video-of-Thought](https://github.com/scofield7419/Video-of-Thought) | Video-of-Thought: Step-by-Step Video Reasoning from Perception to Cognition. | | | Video |
Expand Down Expand Up @@ -759,7 +759,7 @@ Here we will keep track of the latest AI Game Development Tools, including LLM,
| [Mubert](https://mubert.com/) | AI Generative Music. | | | Music |
| [MuseNet](https://openai.com/research/musenet) | A deep neural network that can generate 4-minute musical compositions with 10 different instruments, and can combine styles from country to Mozart to the Beatles. | | | Music |
| [MusicGen](https://github.com/facebookresearch/audiocraft) | Simple and Controllable Music Generation. | [arXiv](https://arxiv.org/abs/2306.05284) | | Music |
| [MusicLDM](https://musicldm.github.io/) | Enhancing Novelty in Text-to-Music Generation Using Beat-Synchronous Mixup Strategies. | [arXiv](https://arxiv.org/abs/2308.01546) | | Music |
| [MusicLDM](https://musicldm.github.io/) | EenhancingNovelty in Text-to-Music Generation Using Beat-Synchronous Mixup Strategies. | [arXiv](https://arxiv.org/abs/2308.01546) | | Music |
| [MusicLM](https://google-research.github.io/seanet/musiclm/examples/) | Generating Music From Text. | [arXiv](https://arxiv.org/abs/2301.11325) | | Music |
| [Riffusion App](https://github.com/riffusion/riffusion-app) | Riffusion is an app for real-time music generation with stable diffusion. | | | Music |
| [Sonauto](https://sonauto.ai/Home) | Sonauto is an AI music editor that turns prompts, lyrics, or melodies into full songs in any style. | | | Music |
Expand Down