Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat(api_server): add options for users to control output #1613

Closed

Conversation

aarnphm
Copy link
Contributor

@aarnphm aarnphm commented Nov 9, 2023

cc @WoosukKwon when you have bandwidth. This allows users to have more fine tuned control on how the API server yield back the text. Default behaviour are still the same, but it will address #1612

Signed-off-by: Aaron 29749331+aarnphm@users.noreply.github.com

Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com>
Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com>
@aarnphm aarnphm changed the title feat(api_server): add options for uers to fine tune prompt output feat(api_server): add options for users to control output Nov 9, 2023
Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com>
@Tostino
Copy link
Contributor

Tostino commented Nov 14, 2023

@aarnphm I implemented somewhat similar functionality for the OpenAI API chat endpoint (here: #1493) for completing a model response. I used return_full_response as the parameter name. I did not add it for the generation endpoint there though, so I figured we should have consistent implementations of this type of feature throughout.

@aarnphm
Copy link
Contributor Author

aarnphm commented Nov 14, 2023

@Tostino I believe for OpenAI there is a echo feature right?

@Tostino
Copy link
Contributor

Tostino commented Nov 14, 2023

Correct, for their completions (legacy) API they support echo. Not for their chat completions API.

The current support looks like:

if request.echo:
        # We do not support echo since the vLLM engine does not
        # currently support getting the logprobs of prompt tokens.
        return create_error_response(HTTPStatus.BAD_REQUEST,
                                     "echo is not currently supported")

@simon-mo
Copy link
Collaborator

Currently, we are minimizing the functionality of api server. API server is not designed for production and mostly for demo purpose.

If you can add echo to OpenAI server that would be great. I believe we are support prompt log prob now.

@aarnphm
Copy link
Contributor Author

aarnphm commented Nov 15, 2023

Thats fair. I will close this PR then and will leave it to the OpenAI to implement echo support.

@aarnphm aarnphm closed this Nov 15, 2023
@aarnphm aarnphm deleted the feat/api-server-echo-prompt branch November 15, 2023 21:51
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants