Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RPC is broken due to change of interface in llama.cpp main repository (rpc : early register backend devices #11262) #1914

Open
j-lag opened this issue Jan 30, 2025 · 0 comments

Comments

@j-lag
Copy link

j-lag commented Jan 30, 2025

RPC does not work in llama-cpp-python since this commit in llama.cpp

Revision: 667d72846c06b2cf4f7c8a4265e210991a49706b
Author: Radoslav Gerganov <rgerganov@gmail.com>
Date: 17/01/2025 09:57:09
Message:
rpc : early register backend devices (#11262)
Early register RPC devices and do not propagate RPC specifics in the
llama model structures.

This commit change the way rpc_servers are passed to llama.cpp, (in llama.cpp\common\arg.cpp) it is not a parameter anymore (params.rpc_servers = value;) but need to call add_rpc_devices(value);

To reproduce:

from llama_cpp import Llama
llm = Llama(
    model_path=MODEL,
	rpc_servers=RPC_NODES,
	n_ctx = 4096,
	flash_attn = False,
	n_gpu_layers = 99,
    verbose=True
)

--> should init with RPC offloading

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant