Skip to content

Conversation

@Blank-z0
Copy link

@Blank-z0 Blank-z0 commented Mar 9, 2025

Good work! Could you please include our ICLR2025 work Dynamic-LLaVA? TL;DR: Dynamic-LLaVA is the first MLLM acceleration framework that simultaneously sparsifies both vision and language contexts while integrating inference efficiency optimization across different MLLM inference modes into a unified framework. In practice, Dynamic-LLaVA can achieve additional inference efficiency throughout the entire generation process, with negligible understanding and generation ability degradation or even performance gains compared to the full-context inference baselines. GitHub: https://github.com/Osilly/dynamic_llava

Good work! Could you please include our ICLR2025 work Dynamic-LLaVA?
TL;DR: Dynamic-LLaVA is the first MLLM acceleration framework that simultaneously sparsifies both vision and language contexts while integrating inference efficiency optimization across different MLLM inference modes into a unified framework. In practice, Dynamic-LLaVA can achieve additional inference efficiency throughout the entire generation process, with negligible understanding and generation ability degradation or even performance gains compared to the full-context inference baselines.
GitHub: https://github.com/Osilly/dynamic_llava
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant