Highlights
- Pro
Stars
Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞
Delivers efficient, stable, and secure data distribution and acceleration powered by P2P technology, with an optional content‑addressable filesystem that accelerates OCI container launch.
A Persistent Key-Value Store designed for Streaming processing
ToplingDB is a cloud native LSM Key-Value Store with searchable compression algo and distributed compaction
DeepRec is a high-performance recommendation deep learning framework based on TensorFlow. It is hosted in incubation in LF AI & Data Foundation.
A Cheat Sheet 📜 to revise Python syntax. Particularly useful for solving Data Structure and Algorithmic problems with Python.
Examples for Recommenders - easy to train and deploy on accelerated infrastructure.
A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.
A high-performance distributed file system designed to address the challenges of AI training and inference workloads.
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation
A guidance language for controlling large language models.
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
Supercharge Your LLM with the Fastest KV Cache Layer
An experimentation platform for LLM inference optimisation
16-fold memory access reduction with nearly no loss
This is the implementation repository of our OSDI'23 paper: SMART: A High-Performance Adaptive Radix Tree for Disaggregated Memory.
[Start here!] Flow-IPC - Modern C++ toolkit for high-speed inter-process communication (IPC)
A Flexible Framework for Experiencing Heterogeneous LLM Inference/Fine-tune Optimizations
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
SGLang is a high-performance serving framework for large language models and multimodal models.


