Related Products
|
||||||
About
LMCache is an open source Knowledge Delivery Network (KDN) designed as a caching layer for large language model serving that accelerates inference by reusing KV (key-value) caches across repeated or overlapping computations. It enables fast prompt caching, allowing LLMs to “prefill” recurring text only once and then reuse those stored KV caches, even in non-prefix positions, across multiple serving instances. This approach reduces time to first token, saves GPU cycles, and increases throughput in scenarios such as multi-round question answering or retrieval augmented generation. LMCache supports KV cache offloading (moving cache from GPU to CPU or disk), cache sharing across instances, and disaggregated prefill, which separates the prefill and decoding phases for resource efficiency. It is compatible with inference engines like vLLM and TGI and supports compressed storage, blending techniques to merge caches, and multiple backend storage options.
|
About
OrbitDB is a serverless, distributed, peer-to-peer database that utilizes IPFS for data storage and Libp2p Pubsub for automatic synchronization across peers. It employs Merkle-CRDTs to ensure conflict-free database writes and merges, making it suitable for decentralized applications, blockchain integrations, and local-first web apps. OrbitDB offers various database types tailored to different use cases: 'events' for immutable append-only logs, 'documents' for JSON document storage indexed by a specified key, 'keyvalue' for traditional key-value pairs, and 'keyvalue-indexed' for LevelDB-indexed key-value data. All these databases are built atop OpLog, an immutable, cryptographically verifiable, operation-based CRDT structure. The JavaScript implementation supports both browser and Node.js environments, with a Go version maintained by the Berty project.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
AI engineers and infrastructure teams looking for a tool to lower latency, reduce compute cost, and scale throughput
|
Audience
Individuals requiring a solution to automatically sync databases with peers
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
Free
Free Version
Free Trial
|
Pricing
Free
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationLMCache
United States
lmcache.ai/
|
Company InformationOrbitDB
United States
orbitdb.org
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
|
|
|
|||||
|
|
||||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
Go
JSON
JavaScript
LevelDB
Node.js
|
||||||
|
|
|