Skip to content

Cost-efficient and pluggable Infrastructure components for GenAI inference

License

Notifications You must be signed in to change notification settings

TheCodeWrangler/aibrix

 
 

Repository files navigation

AIBrix

Welcome to AIBrix, an open-source initiative designed to provide essential building blocks to construct scalable GenAI inference infrastructure. AIBrix delivers a cloud-native solution optimized for deploying, managing, and scaling large language model (LLM) inference, tailored specifically to enterprise needs.

| Documentation | Blog | White Paper | Twitter/X | Developer Slack |

Key Features

The initial release includes the following key features:

  • High-Density LoRA Management: Streamlined support for lightweight, low-rank adaptations of models.
  • LLM Gateway and Routing: Efficiently manage and direct traffic across multiple models and replicas.
  • LLM App-Tailored Autoscaler: Dynamically scale inference resources based on real-time demand.
  • Unified AI Runtime: A versatile sidecar enabling metric standardization, model downloading, and management.
  • Distributed Inference: Scalable architecture to handle large workloads across multiple nodes.
  • Distributed KV Cache: Enables high-capacity, cross-engine KV reuse.
  • Cost-efficient Heterogeneous Serving: Enables mixed GPU inference to reduce costs with SLO guarantees.
  • GPU Hardware Failure Detection: Proactive detection of GPU hardware issues.

Architecture

aibrix-architecture-v1

Quick Start

To get started with AIBrix, clone this repository and follow the setup instructions in the documentation. Our comprehensive guide will help you configure and deploy your first LLM infrastructure seamlessly.

# Local Testing
git clone https://github.com/vllm-project/aibrix.git
cd aibrix

# Install nightly aibrix dependencies
kubectl create -k config/dependency

# Install nightly aibrix components
kubectl create -k config/default

Install stable distribution

# Install component dependencies
kubectl create -k "github.com/vllm-project/aibrix/config/dependency?ref=v0.2.0"

# Install aibrix components
kubectl create -k "github.com/vllm-project/aibrix/config/overlays/release?ref=v0.2.0"

Documentation

For detailed documentation on installation, configuration, and usage, please visit our documentation page.

Contributing

We welcome contributions from the community! Check out our contributing guidelines to see how you can make a difference.

Slack Channel: #aibrix

License

AIBrix is licensed under the APACHE 2.0 License.

Support

If you have any questions or encounter any issues, please submit an issue on our GitHub issues page.

Thank you for choosing AIBrix for your GenAI infrastructure needs!

About

Cost-efficient and pluggable Infrastructure components for GenAI inference

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 46.4%
  • Go 29.4%
  • Python 21.6%
  • Makefile 1.5%
  • Other 1.1%