|
23150 | 23150 | - filename: financial-gpt-oss-20b-q8.i1-Q4_K_M.gguf |
23151 | 23151 | sha256: 14586673de2a769f88bd51f88464b9b1f73d3ad986fa878b2e0c1473f1c1fc59 |
23152 | 23152 | uri: huggingface://mradermacher/financial-gpt-oss-20b-q8-i1-GGUF/financial-gpt-oss-20b-q8.i1-Q4_K_M.gguf |
| 23153 | +- !!merge <<: *qwen3 |
| 23154 | + name: "deepwerewolf-qwen3-8b-grpo-agentic-chinese" |
| 23155 | + urls: |
| 23156 | + - https://huggingface.co/mradermacher/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese-GGUF |
| 23157 | + description: | |
| 23158 | + **Model Name**: Qwen3-8B |
| 23159 | + **Repository**: [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) |
| 23160 | + **Base Model**: Qwen/Qwen3-8B-Base |
| 23161 | + **License**: Apache 2.0 |
| 23162 | + |
| 23163 | + ### 🔍 Overview |
| 23164 | + Qwen3-8B is a state-of-the-art 8.2-billion-parameter causal language model from Alibaba's Qwen series. It excels in reasoning, instruction-following, agent capabilities, and multilingual tasks. The model uniquely supports **seamless switching between thinking mode** (for complex logic, math, and coding) and **non-thinking mode** (for fast, general-purpose dialogue) — all within a single model. |
| 23165 | + |
| 23166 | + ### ✨ Key Features |
| 23167 | + - **Dual-mode inference**: Toggle between deep reasoning (thinking) and efficient response generation (non-thinking). |
| 23168 | + - **Advanced reasoning**: Outperforms prior models in math, code generation, and logical reasoning. |
| 23169 | + - **Agent-ready**: Built-in tool calling and integration capabilities with Qwen-Agent. |
| 23170 | + - **Long context support**: Natively handles up to **32,768 tokens**, extendable to **131,072 tokens** via YaRN RoPE scaling. |
| 23171 | + - **Multilingual**: Supports over 100 languages with strong translation and instruction-following abilities. |
| 23172 | + |
| 23173 | + ### ⚙️ Usage |
| 23174 | + - Use with `transformers`, `vLLM`, `SGLang`, `llama.cpp`, or Ollama. |
| 23175 | + - Enable `enable_thinking=True` for reasoning tasks (e.g., math, coding), or `False` for speed. |
| 23176 | + - Supports dynamic mode switching via `/think` and `/no_think` in prompts. |
| 23177 | + |
| 23178 | + ### 📚 Reference |
| 23179 | + - [Technical Report (arXiv)](https://arxiv.org/abs/2505.09388) |
| 23180 | + - [Official Blog](https://qwenlm.github.io/blog/qwen3/) |
| 23181 | + - [Documentation](https://qwen.readthedocs.io/en/latest/) |
| 23182 | + |
| 23183 | + > **Ideal for**: Research, agentic AI, multilingual applications, and high-accuracy reasoning tasks. |
| 23184 | + |
| 23185 | + --- |
| 23186 | + *Note: The model served here is the original, unquantized Qwen3-8B. For quantized versions (GGUF), see community repos like `mradermacher/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese-GGUF`, which are based on this base model.* |
| 23187 | + overrides: |
| 23188 | + parameters: |
| 23189 | + model: DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese.Q4_K_M.gguf |
| 23190 | + files: |
| 23191 | + - filename: DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese.Q4_K_M.gguf |
| 23192 | + sha256: 32a341badc695d9e8bc1bdae92c67b81295d6e3cfd8e901a508f323718db5141 |
| 23193 | + uri: huggingface://mradermacher/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese-GGUF/DeepWereWolf-Qwen3-8B-Grpo-Agentic-Chinese.Q4_K_M.gguf |
0 commit comments