We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cortex ps
GET http://localhost:39281/inferences/server/models
/v1
/models
/system
Response:
{ "data": [ { "engine": "cortex.llamacpp", "id": "llama3.2-1b-instruct", "model_size": 123, "object": "model", "ram": 123, "start_time": 123, "vram": 123, } ], "object": "list" }
GET /v1/endpoint
Body:
{ "key": "value" }
Response
200 { } Error { }
The text was updated successfully, but these errors were encountered:
vansangpfiev
No branches or pull requests
Goal
cortex ps
to show active models and consumed resources (Ram/vram usage and how long its been running)GET http://localhost:39281/inferences/server/models
/v1
APIDiscussion / Success Criteria: @louis-jan
cortex ps
/models
(per model or all models?)/system
Tasklist
Current
Response:
Future: API / CLI
API
1. Feature
Body:
Response
User request
The text was updated successfully, but these errors were encountered: