You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
it should correctly log to cortex logs (logs/cortex.log, logs/cortex-cli.log)
cortex ps should return server status and running models (or no model loaded)
cortex stop should stop server
Model Pulling
Pulling a model should pull .gguf and model.yml file
Model download progress should appear as download bars for each file
Model download progress should be accurate (%, total time, download size, speed)
cortex.so
it should pull by built in model_ID
pull by model_ID should recommend default variant at the top (set in HF model.yml)
it should pull by built-in model_id:variant
huggingface.co
it should pull by HF repo/model ID
it should pull by full HF url (ending in .gguf)
Interrupted Download
it should allow user to interrupt / stop download
pulling again after interruption should accurately calculates remainder of model file size neeed to be downloaded (Found unfinished download! Additional XGB needs to be downloaded)
it should allow to continue downloading the remainder after interruption
Model Management
it should list downloaded models
it should get a local model
it should update model parameters in model.yaml
it should delete a model
it should import models with model_id and model_path
Model Running
cortex run <cortexso model> - if no local models detected, shows pull model menu
cortex run - if local model detected, runs the local model
cortex run - if multiple local models detected, shows list of local models (from multiple model sources eg cortexso, HF authors) for users to select (via regex search)
cortex run <invalid model id> should return gracefully Model not found!
run should autostart server
cortex run <model> starts interactive chat (by default)
cortex run <model> -d runs in detached mode
cortex models start <model>
terminate StdIn or exit() should exit interactive chat
Hardware Detection / Acceleration [WIP, no need to QA]
it should auto offload max ngl
it should correctly detect available GPUs
it should gracefully detect missing dependencies/drivers
CPU Extension (e.g. AVX-2, noAVX, AVX-512)
GPU Acceleration (e.g. CUDA11, CUDA12, Vulkan, sycl, etc)
Uninstallation / Reinstallation
it should uninstall 2 binaries (cortex and cortex-server)
it should uninstall with 2 options to delete or not delete data folder
it should gracefully uninstall when server is still running
uninstalling should not leave any dangling files
uninstalling should not leave any dangling processes
it should reinstall without having conflict issues with existing cortex data folders
--
2. API QA
Checklist for each endpoint
Upon cortex start, API page is displayed at localhost:port endpoint
Endpoints should support the parameters stated in API reference (towards OpenAI Compatibility)
QA details:
Version:
v1.0.1-227
OS (select one)
1. Manual QA (CLI)
Installation
/usr/local/bin
]Data/Folder structures
cortex.so/model_name/variants/
, with .gguf and model.yml filehuggingface.co/author/model_name
with .gguf and model.yml filemodel
,author_repo_id
,branch_name
,path_to_model_yaml
(view via SQL)Cortex Update
cortex update
should update from ~3-5 versions ago to latest (+3 to 5 bump)cortex update
should update from the previous version to latest (+1 bump)cortex update -v 1.x.x-xxx
should update from the previous version to specified versioncortex update
should update from previous stable version to latestOverall / App Shell
cortex
orcortex -h
displays help commandsEngines
Server
cortex start
should start server and output localhost URL & port numbercortex start
can be configured with parameters (port, logLevel [WIP]) https://cortex.so/docs/cli/start/cortex ps
should return server status and running models (or no model loaded)cortex stop
should stop serverModel Pulling
cortex.so
huggingface.co
Interrupted Download
Found unfinished download! Additional XGB needs to be downloaded
)Model Management
Model Running
cortex run <cortexso model>
- if no local models detected, showspull
model menucortex run
- if local model detected, runs the local modelcortex run
- if multiple local models detected, shows list of local models (from multiple model sources eg cortexso, HF authors) for users to select (via regex search)cortex run <invalid model id>
should return gracefullyModel not found!
cortex run <model>
starts interactive chat (by default)cortex run <model> -d
runs in detached modecortex models start <model>
exit()
should exit interactive chatHardware Detection / Acceleration [WIP, no need to QA]
CPU Extension (e.g. AVX-2, noAVX, AVX-512)
GPU Acceleration (e.g. CUDA11, CUDA12, Vulkan, sycl, etc)
Uninstallation / Reinstallation
--
2. API QA
Checklist for each endpoint
cortex start
, API page is displayed at localhost:port endpointEndpoints
Chat Completions
v1/chat/completions
Engines
/v1/engines
/v1/engines/{name}
/v1/engines/install/{name}
v1/engines/{name}/default
v1/engines/{name}/default
v1/engines/{name}/load
v1/engines/{name}/load
v1/engines/{name}/update
/v1/engines/install/{name}
Pulling Models
/v1/models/pull
starts download (websockets)websockets /events
emitted/v1/models/pull
(websockets)websockets /events
stoppedv1/models/import
Running Models
v1/models
/v1/models/start
/v1/models/stop
/v1/models/{id}
/v1/models/{id}
/v1/models/{model}
updates model.yaml paramsServer
/healthz
/processManager/destroy
Test list for reference:
The text was updated successfully, but these errors were encountered: