Heavily inspired by Neuro-sama
Unlike the other AI driven VTuber open source projects, アイリ VTuber was built with many support of Web technologies such as WebGPU, WebAudio, Web Workers, WebAssembly, WebSocket, etc. from the first day.
This means that アイリ VTuber is capable to run on modern browsers and devices, and even on mobile devices (already done with PWA support), this brought a lot of possibilities for us (the developers) to build and extend the power of アイリ VTuber to the next level, while still left the flexibilities for users to enable features that requires TCP connections or other non-Web technologies such as connect to voice channel to Discord, or playing Minecraft, Factorio with you and your friends.
Note
We are still in the early stage of development where we are seeking out talented developers to join us and help us to make アイリ VTuber a reality.
It's ok if you are not familiar with Vue.js, TypeScript, and devtools that required for this project, you can join us as an artist, designer, or even help us to launch our first live stream.
Even you are a big fan of React or Svelte, even Solid, we welcome you, you can open a sub-directory to add features that you want to see in アイリ VTuber, or would like to experiment with.
Fields (and related projects) that we are looking for:
- Live2D modeller
- VRM modeller
- VRChat avatar designer
- Computer Vision
- Reinforcement Learning
- Speech Recognition
- Speech Synthesis
- ONNX Runtime
- Transformers.js
- vLLM
- WebGPU
- Three.js
- WebXR (checkout the another project we have under @moeru-ai organization)
If you are interested in, why not introduce yourself here? Would like to join part of us to build Airi?
Capable of
- Brain
- Play Minecraft
- Play Factorio (WIP, but PoC and demo available)
- Chat in Telegram
- Chat in Discord
- Ears
- Audio input from browser
- Audio input from Discord
- Client side speech recognition
- Client side talking detection
- Mouth
- ElevenLabs voice synthesis
- Body
- VRM support
- Control VRM model
- VRM model animations
- Auto blink
- Auto look at
- Idle eye movement
- Live2D support
- Control Live2D model
- Live2D model animations
- Auto blink
- Auto look at
- Idle eye movement
- VRM support
pnpm i
pnpm dev
Supported the following LLM API Providers (powered by xsai)
- OpenRouter
- vLLM
- SGLang
- Ollama
- Google Gemini
- OpenAI
- Azure OpenAI API (PR welcome)
- Anthropic Claude
- AWS Claude (PR welcome)
- DeepSeek
- Qwen
- xAI
- Groq
- Mistral
- Cloudflare Workers AI
- Together.ai
- Fireworks.ai
- Novita
- Zhipu
- SiliconFlow
- Stepfun
- Baichuan
- Minimax
- Moonshot AI
- Tencent Cloud
- Sparks (PR welcome)
- Volcano Engine (PR welcome)
unspeech
: Universal endpoint proxy server for/audio/transcriptions
and/audio/speech
, like LiteLLM but for any ASR and TTShfup
: tools to help on deploying, bundling to HuggingFace Spaces@proj-airi/drizzle-duckdb-wasm
: Drizzle ORM driver for DuckDB WASM@proj-airi/duckdb-wasm
: Easy to use wrapper for@duckdb/duckdb-wasm
@proj-airi/lobe-icons
: Iconify JSON bundle for amazing AI & LLM icons from lobe-icons, support Tailwind and UnoCSS@proj-airi/elevenlabs
: TypeScript definitions for ElevenLabs API- Airi Factorio: Allow Airi to play Factorio
- Factorio RCON API: RESTful API wrapper for Factorio headless server console
autorio
: Factorio automation library- `tstl-plugin-reload-factorio-mod: Reload Factorio mod when developing
- 🥺 SAD: Documentation and notes for self-host and browser running LLMs
%%{ init: { 'flowchart': { 'curve': 'catmullRom' } } }%%
flowchart TD
Core("Core")
Unspeech["unspeech"]
DBDriver["@proj-airi/drizzle-duckdb-wasm"]
MemoryDriver["[WIP] Memory Alaya"]
DB1["@proj-airi/duckdb-wasm"]
ICONS["@proj-airi/lobe-icons"]
UI("@proj-airi/stage-ui")
Stage("Stage")
F_AGENT("Factorio Agent")
F_API["Factorio RCON API"]
F_MOD1["autorio"]
SVRT["@proj-airi/server-runtime"]
MC_AGENT("Minecraft Agent")
XSAI["xsai"]
subgraph Airi
DB1 --> DBDriver --> MemoryDriver --> Memory --> Core
ICONS --> UI --> Stage --> Core
Core --> STT
Core --> SVRT
end
STT --> |Speaking|Unspeech
SVRT --> |Playing Factorio|F_AGENT
SVRT --> |Playing Minecraft|MC_AGENT
subgraph Factorio Agent
F_AGENT --> F_API -..- factorio-server
subgraph factorio-server-wrapper
subgraph factorio-server
F_MOD1
end
end
end
subgraph Minecraft Agent
MC_AGENT --> Mineflayer -..- minecraft-server
subgraph factorio-server-wrapper
subgraph factorio-server
F_MOD1
end
end
end
XSAI --> Core
XSAI --> F_AGENT
XSAI --> MC_AGENT
%%{ init: { 'flowchart': { 'curve': 'catmullRom' } } }%%
flowchart TD
subgraph deploy&bundle
direction LR
HFUP["hfup"]
HF[/"HuggingFace Spaces"\]
HFUP -...- UI -...-> HF
HFUP -...- whisper-webgpu -...-> HF
HFUP -...- moonshine-web -...-> HF
end
- SugarcaneDefender/z-waif: Great at gaming, autonomous, and prompt engineering
- semperai/amica: Great at VRM, WebXR
- elizaOS/eliza: Great examples and software engineering on how to integrate agent into various of systems and APIs
- ardha27/AI-Waifu-Vtuber: Great about Twitch API integrations
- InsanityLabs/AIVTuber: Nice UI and UX
- IRedDragonICY/vixevia
- t41372/Open-LLM-VTuber
- PeterH0323/Streamer-Sales
- pixiv/ChatVRM
- josephrocca/ChatVRM-js: A JS conversion/adaptation of parts of the ChatVRM (TypeScript) code for standalone use in OpenCharacters and elsewhere
- Design of UI and style was inspired by Cookard, UNBEATABLE, and Sensei! I like you so much!, and artworks of Ayame by Mercedes Bazan with Wish by Mercedes Bazan
- mallorbc/whisper_mic
xsai
: Implemented a decent amount of packages to interact with LLMs and models, like Vercel AI SDK but a lot more smaller.