Description
Hello,
I am having trouble running specific models from lm studio.
Not sure if I am doing anything wrong if its bugged.
This command works, it picks whatever model I have loaded in lm studio. But it incorrectly says that it is using this when I run it with verbose LiteLLM completion() model= gpt-4o; provider = openai
uvx wdoc --llms_api_bases='{"model": "http://127.0.0.1:1234/v1"}' --path="~/to/paper" --filetype="auto" --task=summary
This gives me the following error:
AttributeError: module 'litellm' has no attribute 'model_by_providers'
uvx wdoc --llms_api_bases='{"model": "http://127.0.0.1:1234/v1"}' --modelname="lm_studio/deepseek-r1-distill-qwen-1.5b" --path="~/to/paper" --filetype="auto" --task=summary
I have tried different providers, but according to https://docs.litellm.ai/docs/providers/lm_studio it should be correct.
Got any ideas?