Skip to content

Commit e94e725

Browse files
authored
chore(model gallery): add alibaba-nlp_tongyi-deepresearch-30b-a3b (#6295)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
1 parent e4ac7b1 commit e94e725

File tree

2 files changed

+60
-0
lines changed

2 files changed

+60
-0
lines changed

gallery/index.yaml

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2623,6 +2623,21 @@
26232623
- filename: Qwen3-Stargate-SG1-Uncensored-Abliterated-8B.i1-Q4_K_M.gguf
26242624
sha256: 31ec697ccebbd7928c49714b8a0ec8be747be0f7c1ad71627967d2f8fe376990
26252625
uri: huggingface://mradermacher/Qwen3-Stargate-SG1-Uncensored-Abliterated-8B-i1-GGUF/Qwen3-Stargate-SG1-Uncensored-Abliterated-8B.i1-Q4_K_M.gguf
2626+
- !!merge <<: *qwen3
2627+
url: "github:mudler/LocalAI/gallery/qwen3-deepresearch.yaml@master"
2628+
name: "alibaba-nlp_tongyi-deepresearch-30b-a3b"
2629+
urls:
2630+
- https://huggingface.co/Alibaba-NLP/Tongyi-DeepResearch-30B-A3B
2631+
- https://huggingface.co/bartowski/Alibaba-NLP_Tongyi-DeepResearch-30B-A3B-GGUF
2632+
description: |
2633+
We present Tongyi DeepResearch, an agentic large language model featuring 30 billion total parameters, with only 3 billion activated per token. Developed by Tongyi Lab, the model is specifically designed for long-horizon, deep information-seeking tasks. Tongyi-DeepResearch demonstrates state-of-the-art performance across a range of agentic search benchmarks, including Humanity's Last Exam, BrowserComp, BrowserComp-ZH, WebWalkerQA, GAIA, xbench-DeepSearch and FRAMES.
2634+
overrides:
2635+
parameters:
2636+
model: Alibaba-NLP_Tongyi-DeepResearch-30B-A3B-Q4_K_M.gguf
2637+
files:
2638+
- filename: Alibaba-NLP_Tongyi-DeepResearch-30B-A3B-Q4_K_M.gguf
2639+
sha256: 1afefb3b369ea2de191f24fe8ea22cbbb7b412357902f27bd81d693dde35c2d9
2640+
uri: huggingface://bartowski/Alibaba-NLP_Tongyi-DeepResearch-30B-A3B-GGUF/Alibaba-NLP_Tongyi-DeepResearch-30B-A3B-Q4_K_M.gguf
26262641
- &gemma3
26272642
url: "github:mudler/LocalAI/gallery/gemma.yaml@master"
26282643
name: "gemma-3-27b-it"

gallery/qwen3-deepresearch.yaml

Lines changed: 45 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,45 @@
1+
---
2+
name: "qwen3"
3+
4+
config_file: |
5+
mmap: true
6+
backend: "llama-cpp"
7+
template:
8+
chat_message: |
9+
<|im_start|>{{if eq .RoleName "tool" }}user{{else}}{{ .RoleName }}{{end}}
10+
{{ if eq .RoleName "tool" -}}
11+
<tool_response>
12+
{{ end -}}
13+
{{ if .Content -}}
14+
{{.Content }}
15+
{{ end -}}
16+
{{ if eq .RoleName "tool" -}}
17+
</tool_response>
18+
{{ end -}}
19+
{{ if .FunctionCall -}}
20+
<tool_call>
21+
{{toJson .FunctionCall}}
22+
</tool_call>
23+
{{ end -}}<|im_end|>
24+
function: |
25+
<|im_start|>system
26+
You are a function calling AI model. You are provided with functions to execute. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools:
27+
{{range .Functions}}
28+
{'type': 'function', 'function': {'name': '{{.Name}}', 'description': '{{.Description}}', 'parameters': {{toJson .Parameters}} }}
29+
{{end}}
30+
For each function call return a json object with function name and arguments
31+
<|im_end|>
32+
{{.Input -}}
33+
<|im_start|>assistant
34+
chat: |
35+
{{.Input -}}
36+
<|im_start|>assistant
37+
completion: |
38+
{{.Input}}
39+
context_size: 8192
40+
f16: true
41+
stopwords:
42+
- '<|im_end|>'
43+
- '<dummy32000>'
44+
- '</s>'
45+
- '<|endoftext|>'

0 commit comments

Comments
 (0)