Support starcoder family architectures (1B/3B/7B/13B) · Issue #3076 · ggerganov/llama.cpp #362
Labels
llm
Large Language Models
llm-applications
Topics related to practical applications of Large Language Models in various fields
llm-evaluation
Evaluating Large Language Models performance and behavior through human-written evaluation sets
llm-inference-engines
Software to run inference on large language models
llm-serving-optimisations
Tips, tricks and tools to speedup inference of large language models
Models
LLM and ML model repos and links
Previously, it wasn't recommended to incorporate non-llama architectures into llama.cpp. However, in light of the recent addition of the Falcon architecture (see Pull Request #2717), it might be worth reconsidering this stance.
One distinguishing feature of Starcoder is its ability to provide a complete series of models ranging from 1B to 13B. This capability can prove highly beneficial for speculative decoding and making coding models available for edge devices (e.g., M1/M2 Macs).
I can contribute the PR if it matches llama.cpp's roadmap.
Suggested labels
{ "key": "LLM-Applications", "value": "Practical applications of Large Language Models, such as edge device coding models and speculative decoding" } { "key": "Multimodal-LM", "value": "LLMs that combine modes such as text and image recognition" }
The text was updated successfully, but these errors were encountered: