Skip to content

Support speculative decoding with llama.cpp #240

Closed
@kerthcet

Description

@kerthcet

What would you like to be added:

We have supported vllm, since llama.cpp adds this feature, we should support it as well, see ggml-org/llama.cpp#10455

Why is this needed:

Completion requirements:

This enhancement requires the following artifacts:

  • Design doc
  • API change
  • Docs update

The artifacts should be linked in subsequent comments.

Metadata

Metadata

Assignees

Labels

featureCategorizes issue or PR as related to a new feature.help wantedExtra attention is neededneeds-priorityIndicates a PR lacks a label and requires one.needs-triageIndicates an issue or PR lacks a label and requires one.

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions