Skip to content

v0.9.6

Compare
Choose a tag to compare
@paulguerrie paulguerrie released this 13 Dec 18:12
· 3737 commits to main since this release

What's Changed

Highlights

CogVLM

Inference server users can now run CogVLM for a fully self hosted, multimodal LLM. See the example here.

Slim Docker Images

For use cases that do not need Core Model functionality (e.g. CLIP), there are -slim docker images available which include fewer dependancies and are much smaller.

  • roboflow/roboflow-inference-server-cpu-slim
  • roboflow/roboflow-inference-server-gpu-slim

Breaking Changes

Infer API Update

The infer() method of Roboflow models now returns an InferenceResponse object instead of raw model output. This means that using models in application logic should feel similar to using models via the HTTP interface. In practice, programs that used the following pattern

...
model = get_roboflow_model(...)
results = model.infer(...)
results = model.make_response(...)
...

should be updated to

...
model = get_roboflow_model(...)
results = model.infer(...)
...

New Contributors

Full Changelog: v0.9.5...v0.9.6