Skip to content

Commit 1f7b446

Browse files
mudlergithub-actions[bot]
authored andcommitted
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
1 parent 1e5b913 commit 1f7b446

File tree

1 file changed

+41
-0
lines changed

1 file changed

+41
-0
lines changed

gallery/index.yaml

Lines changed: 41 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23150,3 +23150,44 @@
2315023150
- filename: financial-gpt-oss-20b-q8.i1-Q4_K_M.gguf
2315123151
sha256: 14586673de2a769f88bd51f88464b9b1f73d3ad986fa878b2e0c1473f1c1fc59
2315223152
uri: huggingface://mradermacher/financial-gpt-oss-20b-q8-i1-GGUF/financial-gpt-oss-20b-q8.i1-Q4_K_M.gguf
23153+
- !!merge <<: *qwen3
23154+
name: "deepwerewolf-qwen3-8b-grpo-agentic-chinese"
23155+
urls:
23156+
- https://huggingface.co/mradermacher/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese-GGUF
23157+
description: |
23158+
**Model Name**: Qwen3-8B
23159+
**Repository**: [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B)
23160+
**Base Model**: Qwen/Qwen3-8B-Base
23161+
**License**: Apache 2.0
23162+
23163+
### 🔍 Overview
23164+
Qwen3-8B is a state-of-the-art 8.2-billion-parameter causal language model from Alibaba's Qwen series. It excels in reasoning, instruction-following, agent capabilities, and multilingual tasks. The model uniquely supports **seamless switching between thinking mode** (for complex logic, math, and coding) and **non-thinking mode** (for fast, general-purpose dialogue) — all within a single model.
23165+
23166+
### ✨ Key Features
23167+
- **Dual-mode inference**: Toggle between deep reasoning (thinking) and efficient response generation (non-thinking).
23168+
- **Advanced reasoning**: Outperforms prior models in math, code generation, and logical reasoning.
23169+
- **Agent-ready**: Built-in tool calling and integration capabilities with Qwen-Agent.
23170+
- **Long context support**: Natively handles up to **32,768 tokens**, extendable to **131,072 tokens** via YaRN RoPE scaling.
23171+
- **Multilingual**: Supports over 100 languages with strong translation and instruction-following abilities.
23172+
23173+
### ⚙️ Usage
23174+
- Use with `transformers`, `vLLM`, `SGLang`, `llama.cpp`, or Ollama.
23175+
- Enable `enable_thinking=True` for reasoning tasks (e.g., math, coding), or `False` for speed.
23176+
- Supports dynamic mode switching via `/think` and `/no_think` in prompts.
23177+
23178+
### 📚 Reference
23179+
- [Technical Report (arXiv)](https://arxiv.org/abs/2505.09388)
23180+
- [Official Blog](https://qwenlm.github.io/blog/qwen3/)
23181+
- [Documentation](https://qwen.readthedocs.io/en/latest/)
23182+
23183+
> **Ideal for**: Research, agentic AI, multilingual applications, and high-accuracy reasoning tasks.
23184+
23185+
---
23186+
*Note: The model served here is the original, unquantized Qwen3-8B. For quantized versions (GGUF), see community repos like `mradermacher/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese-GGUF`, which are based on this base model.*
23187+
overrides:
23188+
parameters:
23189+
model: DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese.Q4_K_M.gguf
23190+
files:
23191+
- filename: DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese.Q4_K_M.gguf
23192+
sha256: 32a341badc695d9e8bc1bdae92c67b81295d6e3cfd8e901a508f323718db5141
23193+
uri: huggingface://mradermacher/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese-GGUF/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese.Q4_K_M.gguf

0 commit comments

Comments
 (0)