Skip to content

Conversation

yanxi0830
Copy link
Contributor

@yanxi0830 yanxi0830 commented Nov 19, 2024

TL;DR

  • adding llama-stack-client inference chat_completion to quickly test a client example w/o getting apps

Test

Non-Streaming

llama-stack-client inference chat_completion --message "hello, what model are you"
image

Streaming

llama-stack-client inference chat_completion --message "hello, what model are you" --stream
image

@yanxi0830 yanxi0830 marked this pull request as ready for review November 19, 2024 23:01
@ashwinb
Copy link
Contributor

ashwinb commented Nov 19, 2024

chat-completion instead of chat_completion maybe? we should probably standardize with the URLs I think.

@yanxi0830 yanxi0830 merged commit da419c6 into main Nov 19, 2024
3 checks passed
@yanxi0830 yanxi0830 deleted the chat_completion_cli branch November 19, 2024 23:23
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants