Skip to content

Commit 0e9082f

Browse files
author
LittleMouse
committed
[update] support qwen3.5 & add model file
1 parent 721cd50 commit 0e9082f

17 files changed

Lines changed: 1714 additions & 173 deletions

File tree

ext_components/openai-api/SConstruct

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ _SDK_PATH = os.path.normpath(
1212

1313
env["GIT_REPO_LISTS"]["openai-api"] = {
1414
"url": "https://github.com/ZHEQIUSHUI/openai-api.cpp.git",
15-
"commit": "127f3e75054a733ca1b903ff1f8ec600b90cb6b0",
15+
"commit": "3d4be086e95cee55bfd8e983b20f0758b6f82182",
1616
"path": str(Path(_SDK_PATH) / "github_source" / "openai-api.cpp"),
1717
}
1818

Lines changed: 48 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,48 @@
1+
{
2+
"mode": "Qwen3.5-2B-ax650",
3+
"type": "llm",
4+
"homepage": "https://huggingface.co/AXERA-TECH/Qwen3.5-2B-AX650-C128-P1152-CTX2047",
5+
"capabilities": [
6+
"text_generation",
7+
"chat"
8+
],
9+
"input_type": [
10+
"llm.utf-8",
11+
"llm.utf-8.stream",
12+
"llm.chat_completion",
13+
"llm.chat_completion.stream"
14+
],
15+
"output_type": [
16+
"llm.utf-8",
17+
"llm.utf-8.stream"
18+
],
19+
"mode_param": {
20+
"system_prompt": "you are a helpful assistant.",
21+
"model_name": "AXERA-TECH/Qwen3.5-2B",
22+
"url_tokenizer_model": "qwen3_5_tokenizer.txt",
23+
"tokenizer_type": "Qwen3_5VL",
24+
"post_config_path": "post_config.json",
25+
"template_filename_axmodel": "qwen3_5_text_p128_l%d_together.axmodel",
26+
"axmodel_num": 24,
27+
"full_attention_interval": 4,
28+
"filename_post_axmodel": "qwen3_5_text_post.axmodel",
29+
"filename_tokens_embed": "model.embed_tokens.weight.bfloat16.bin",
30+
"tokens_embed_num": 248320,
31+
"tokens_embed_size": 2048,
32+
"b_use_mmap_load_embed": true,
33+
"b_use_mmap_load_layer": true,
34+
"vlm_type": "Qwen3VL",
35+
"filename_image_encoder_axmodel": "qwen3_5_vision.axmodel",
36+
"vision_patch_size": 16,
37+
"vision_width": 384,
38+
"vision_height": 384,
39+
"vision_temporal_patch_size": 2,
40+
"vision_spatial_merge_size": 2,
41+
"vision_fps": 1,
42+
"vision_tokens_per_second": 1,
43+
"vision_cache_dir": "vision_cache",
44+
"devices": [
45+
0
46+
]
47+
}
48+
}

projects/llm_framework/main_llm2/src/main.cpp

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -146,6 +146,7 @@ class llm_task {
146146
CONFIG_AUTO_SET(file_body["mode_param"], b_bos);
147147
CONFIG_AUTO_SET(file_body["mode_param"], b_eos);
148148

149+
CONFIG_AUTO_SET(file_body["mode_param"], full_attention_interval);
149150
CONFIG_AUTO_SET(file_body["mode_param"], filename_tokens_embed);
150151
CONFIG_AUTO_SET(file_body["mode_param"], tokens_embed_num);
151152
CONFIG_AUTO_SET(file_body["mode_param"], tokens_embed_size);
@@ -169,6 +170,16 @@ class llm_task {
169170
CONFIG_AUTO_SET(file_body["mode_param"], kv_cache_size);
170171
CONFIG_AUTO_SET(file_body["mode_param"], b_use_mmap_load_layer);
171172

173+
CONFIG_AUTO_SET(file_body["mode_param"], vision_cache_dir);
174+
175+
CONFIG_AUTO_SET(file_body["mode_param"], vision_width);
176+
CONFIG_AUTO_SET(file_body["mode_param"], vision_height);
177+
CONFIG_AUTO_SET(file_body["mode_param"], vision_temporal_patch_size);
178+
CONFIG_AUTO_SET(file_body["mode_param"], vision_spatial_merge_size);
179+
CONFIG_AUTO_SET(file_body["mode_param"], vision_patch_size);
180+
CONFIG_AUTO_SET(file_body["mode_param"], vision_fps);
181+
CONFIG_AUTO_SET(file_body["mode_param"], vision_tokens_per_second);
182+
172183
mode_config_.template_filename_axmodel = base_model + mode_config_.template_filename_axmodel;
173184
mode_config_.filename_post_axmodel = base_model + mode_config_.filename_post_axmodel;
174185
mode_config_.filename_tokens_embed = base_model + mode_config_.filename_tokens_embed;

0 commit comments

Comments
 (0)