diff --git a/src/config.ts b/src/config.ts index 2e7430d1..e464bab7 100644 --- a/src/config.ts +++ b/src/config.ts @@ -855,8 +855,8 @@ export const prebuiltAppConfig: AppConfig = { }, // Qwen-2 { - model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q4f16_1-MLC", - model_id: "Qwen2-0.5B-Instruct-q4f16_1-MLC", + model: "https://huggingface.co/mlc-ai/Qwen2.5-0.5B-Instruct-q4f16_1-MLC", + model_id: "Qwen2.5-0.5B-Instruct-q4f16_1-MLC", model_lib: modelLibURLPrefix + modelVersion + @@ -867,9 +867,22 @@ export const prebuiltAppConfig: AppConfig = { context_window_size: 4096, }, }, + { + model: "https://huggingface.co/mlc-ai/Qwen2.5-0.5B-Instruct-q4f32_1-MLC", + model_id: "Qwen2.5-0.5B-Instruct-q4f32_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-0.5B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 1060.2, + overrides: { + context_window_size: 4096, + }, + }, { model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q0f16-MLC", - model_id: "Qwen2-0.5B-Instruct-q0f16-MLC", + model_id: "Qwen2.5-0.5B-Instruct-q0f16-MLC", model_lib: modelLibURLPrefix + modelVersion + @@ -881,8 +894,8 @@ export const prebuiltAppConfig: AppConfig = { }, }, { - model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q0f32-MLC", - model_id: "Qwen2-0.5B-Instruct-q0f32-MLC", + model: "https://huggingface.co/mlc-ai/Qwen2.5-0.5B-Instruct-q0f32-MLC", + model_id: "Qwen2.5-0.5B-Instruct-q0f32-MLC", model_lib: modelLibURLPrefix + modelVersion + @@ -894,8 +907,8 @@ export const prebuiltAppConfig: AppConfig = { }, }, { - model: "https://huggingface.co/mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC", - model_id: "Qwen2-1.5B-Instruct-q4f16_1-MLC", + model: "https://huggingface.co/mlc-ai/Qwen2.5-1.5B-Instruct-q4f16_1-MLC", + model_id: "Qwen2.5-1.5B-Instruct-q4f16_1-MLC", model_lib: modelLibURLPrefix + modelVersion + @@ -907,8 +920,8 @@ export const prebuiltAppConfig: AppConfig = { }, }, { - model: "https://huggingface.co/mlc-ai/Qwen2-1.5B-Instruct-q4f32_1-MLC", - model_id: "Qwen2-1.5B-Instruct-q4f32_1-MLC", + model: "https://huggingface.co/mlc-ai/Qwen2.5-1.5B-Instruct-q4f32_1-MLC", + model_id: "Qwen2.5-1.5B-Instruct-q4f32_1-MLC", model_lib: modelLibURLPrefix + modelVersion + @@ -920,8 +933,34 @@ export const prebuiltAppConfig: AppConfig = { }, }, { - model: "https://huggingface.co/mlc-ai/Qwen2-7B-Instruct-q4f16_1-MLC", - model_id: "Qwen2-7B-Instruct-q4f16_1-MLC", + model: "https://huggingface.co/mlc-ai/Qwen2.5-3B-Instruct-q4f16_1-MLC", + model_id: "Qwen2.5-3B-Instruct-q4f16_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2.5-3B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 2504.76, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2.5-3B-Instruct-q4f32_1-MLC", + model_id: "Qwen2.5-3B-Instruct-q4f32_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2.5-3B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 2893.64, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2.5-7B-Instruct-q4f16_1-MLC", + model_id: "Qwen2.5-7B-Instruct-q4f16_1-MLC", model_lib: modelLibURLPrefix + modelVersion + @@ -933,8 +972,65 @@ export const prebuiltAppConfig: AppConfig = { }, }, { - model: "https://huggingface.co/mlc-ai/Qwen2-7B-Instruct-q4f32_1-MLC", - model_id: "Qwen2-7B-Instruct-q4f32_1-MLC", + model: "https://huggingface.co/mlc-ai/Qwen2.5-7B-Instruct-q4f32_1-MLC", + model_id: "Qwen2.5-7B-Instruct-q4f32_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-7B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: false, + vram_required_MB: 5900.09, + overrides: { + context_window_size: 4096, + }, + }, + // Qwen2.5-Coder + { + model: + "https://huggingface.co/mlc-ai/Qwen2.5-Coder-1.5B-Instruct-q4f16_1-MLC", + model_id: "Qwen2.5-Coder-1.5B-Instruct-q4f16_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-1.5B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: false, + vram_required_MB: 5106.67, + overrides: { + context_window_size: 4096, + }, + }, + { + model: + "https://huggingface.co/mlc-ai/Qwen2.5-Coder-1.5B-Instruct-q4f32_1-MLC", + model_id: "Qwen2.5-Coder-1.5B-Instruct-q4f32_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-1.5B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: false, + vram_required_MB: 5900.09, + overrides: { + context_window_size: 4096, + }, + }, + { + model: + "https://huggingface.co/mlc-ai/Qwen2.5-Coder-7B-Instruct-q4f16_1-MLC", + model_id: "Qwen2.5-Coder-7B-Instruct-q4f16_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-7B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: false, + vram_required_MB: 5106.67, + overrides: { + context_window_size: 4096, + }, + }, + { + model: + "https://huggingface.co/mlc-ai/Qwen2.5-Coder-7B-Instruct-q4f32_1-MLC", + model_id: "Qwen2.5-Coder-7B-Instruct-q4f32_1-MLC", model_lib: modelLibURLPrefix + modelVersion + @@ -1186,6 +1282,98 @@ export const prebuiltAppConfig: AppConfig = { context_window_size: 4096, }, }, + // Qwen-2 + { + model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q4f16_1-MLC", + model_id: "Qwen2-0.5B-Instruct-q4f16_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-0.5B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 944.62, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q0f16-MLC", + model_id: "Qwen2-0.5B-Instruct-q0f16-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-0.5B-Instruct-q0f16-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 1624.12, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2-0.5B-Instruct-q0f32-MLC", + model_id: "Qwen2-0.5B-Instruct-q0f32-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-0.5B-Instruct-q0f32-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 2654.75, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC", + model_id: "Qwen2-1.5B-Instruct-q4f16_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-1.5B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 1629.75, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2-1.5B-Instruct-q4f32_1-MLC", + model_id: "Qwen2-1.5B-Instruct-q4f32_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-1.5B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: true, + vram_required_MB: 1888.97, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2-7B-Instruct-q4f16_1-MLC", + model_id: "Qwen2-7B-Instruct-q4f16_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-7B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: false, + vram_required_MB: 5106.67, + overrides: { + context_window_size: 4096, + }, + }, + { + model: "https://huggingface.co/mlc-ai/Qwen2-7B-Instruct-q4f32_1-MLC", + model_id: "Qwen2-7B-Instruct-q4f32_1-MLC", + model_lib: + modelLibURLPrefix + + modelVersion + + "/Qwen2-7B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm", + low_resource_required: false, + vram_required_MB: 5900.09, + overrides: { + context_window_size: 4096, + }, + }, // Llama-3 { model: "https://huggingface.co/mlc-ai/Llama-3-8B-Instruct-q4f32_1-MLC",