Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: cortex run flakiness #1272

Closed
0xSage opened this issue Sep 19, 2024 · 2 comments
Closed

bug: cortex run flakiness #1272

0xSage opened this issue Sep 19, 2024 · 2 comments
Assignees
Labels
category: model running Inference ux, handling context/parameters, runtime type: bug Something isn't working
Milestone

Comments

@0xSage
Copy link
Contributor

0xSage commented Sep 19, 2024

Problem Statement

  1. cortex run mistral
Model loaded!
Inorder to exit, type `exit()`
  1. Chat a bit and quit interactive shell
  2. Run it again
  3. sudo cortex run mistral
Starting server ...
20240919 15:54:32.723277 UTC 33913 INFO  Host: 127.0.0.1 Port: 3928
 - main.cc:32

Server started

Model loaded!
Inorder to exit, type `exit()`
  1. Model loads & chats just fine

Expected

Expected: the same stdout experience as the first time I invoked cortex run
But I got: "Starting Server..." for all subsequent invocations

Questions

What could be causing this inconsistency in the run step?

OS: Mac

@0xSage 0xSage added good first issue Good for newcomers category: model running Inference ux, handling context/parameters, runtime P3: nice to have Nice to have feature labels Sep 19, 2024
@0xSage 0xSage changed the title idea: when loading models, stdout "loading model into v/ram" bug: cortex run flakiness Sep 19, 2024
@0xSage 0xSage added type: bug Something isn't working P0: critical Mission critical and removed P3: nice to have Nice to have feature good first issue Good for newcomers labels Sep 19, 2024
@0xSage 0xSage moved this to Need Investigation in Jan & Cortex Sep 19, 2024
@0xSage 0xSage removed the P0: critical Mission critical label Sep 19, 2024
@vansangpfiev vansangpfiev self-assigned this Sep 20, 2024
@vansangpfiev
Copy link
Contributor

cortex run is the command chain which includes:

  • cortex engines install
  • cortex pull
  • cortex start
  • cortex models start
  • cortex chat
    In this situation, the cortex models start doesn't work correctly. The CLI doesn't check if the model was loaded before starting it, and the engine doesn't ignore it when starting it again.
    The issue will be fixed on the CLI side first, and then on the engines side.

@0xSage
Copy link
Contributor Author

0xSage commented Sep 21, 2024

seems fixed on v75, nice job!

@0xSage 0xSage closed this as completed Sep 21, 2024
@github-project-automation github-project-automation bot moved this from Need Investigation to Completed in Jan & Cortex Sep 21, 2024
@gabrielle-ong gabrielle-ong added this to the v1.0.0 milestone Oct 3, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
category: model running Inference ux, handling context/parameters, runtime type: bug Something isn't working
Projects
Archived in project
Development

No branches or pull requests

3 participants