How to chain RemoteRunnable clients to local llm server (hosted using langserve)? #28647
Open
5 tasks done
Labels
Ɑ: core
Related to langchain-core
Checked other resources
Example Code
On the server side, I used HuggingFacePipeline to load a local model
Client side, use RemoteRunnable to connect, although it can successfully invoke an input string, it failed to be applied in LLMChain
Error Message and Stack Trace (if applicable)
site-packages/langserve/client.py:448, in RemoteRunnable.batch(self, inputs, config, return_exceptions, **kwargs)
439 def batch(
440 self,
441 inputs: List[Input],
(...)
445 **kwargs: Any,
446 ) -> List[Output]:
447 if kwargs:
--> 448 raise NotImplementedError(f"kwargs not implemented yet. Got {kwargs}")
449 return self._batch_with_config(
450 self._batch, inputs, config, return_exceptions=return_exceptions
451 )
NotImplementedError: kwargs not implemented yet. Got {'stop': None}
Description
I try to use langserve to start a server and use RemoteRunnable as clients to communicate with it. This is helpful to try multiple time without worry about client failure, because restart a client is way faster than reload a llm model. Although, I can do simple llm.invoke using RemoteRunnable, but I cannot use any Chain classes, e.g. LLMChain, SimpleSequentialChain, SequentialChain.
System Info
System Information
Package Information
Other Dependencies
The text was updated successfully, but these errors were encountered: