Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add the deepseek as the model provider to run it locally #2948

Open
yasir23 opened this issue Jan 29, 2025 · 2 comments
Open

Add the deepseek as the model provider to run it locally #2948

yasir23 opened this issue Jan 29, 2025 · 2 comments
Labels
enhancement New feature or request

Comments

@yasir23
Copy link

yasir23 commented Jan 29, 2025

Deepseek model to AI agent
Running local models like r1 (7b, 8b,14b and etc)

@yasir23 yasir23 added the enhancement New feature or request label Jan 29, 2025
@yasir23
Copy link
Author

yasir23 commented Jan 29, 2025

From Ollama local models

@bucurdavid
Copy link

You can run any model through ollama, but you should setup your current model in the .env

OLLAMA_SERVER_URL=   # Default: localhost:11434        
OLLAMA_MODEL= deepseek-r1:7b                      # Default: llama3.2
SMALL_OLLAMA_MODEL= deepseek-r1:7b            # Default: llama3.2
MEDIUM_OLLAMA_MODEL=  deepseek-r1:7b          # Default: hermes3
LARGE_OLLAMA_MODEL=   deepseek-r1:7b          # Default: hermes3:70b

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants