You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Merge branch 'master' of github.com:ggerganov/llama.cpp into grammar-example
* 'master' of github.com:ggerganov/llama.cpp: (24 commits)
convert : fix Baichuan2 models by using vocab size in config.json (ggml-org#3299)
readme : add project status link
ggml : fix build after ggml-org#3329
llm : add Refact model (ggml-org#3329)
sync : ggml (conv 1d + 2d updates, UB fixes) (ggml-org#3468)
finetune : readme fix typo (ggml-org#3465)
ggml : add RISC-V Vector Support for K-Quants and improved the existing intrinsics (ggml-org#3453)
main : consistent prefix/suffix coloring (ggml-org#3425)
llama : fix session saving/loading (ggml-org#3400)
llama : expose model's rope_freq_scale in the API (ggml-org#3418)
metal : alibi for arbitrary number of heads (ggml-org#3426)
cmake : make LLAMA_NATIVE flag actually use the instructions supported by the processor (ggml-org#3273)
Work on the BPE tokenizer (ggml-org#3252)
convert : fix vocab size when not defined in hparams (ggml-org#3421)
cmake : increase minimum version for add_link_options (ggml-org#3444)
CLBlast: Add broadcast support for matrix multiplication (ggml-org#3402)
gguf : add BERT, MPT, and GPT-J arch info (ggml-org#3408)
gguf : general usability improvements (ggml-org#3409)
cmake : make CUDA flags more similar to the Makefile (ggml-org#3420)
finetune : fixggml-org#3404 (ggml-org#3437)
...
0 commit comments