Skip to content

Commit 2310c61

Browse files
Update README.md
1 parent bf78300 commit 2310c61

1 file changed

Lines changed: 2 additions & 2 deletions

File tree

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ Reliable and Efficient Semantic Prompt Caching
2323

2424

2525

26-
Semantic caching reduces LLM latency and cost by returning cached model responses for semantically similar prompts (not just exact matches). **vCache** is the first verified semantic cache that **guarantees user-defined error rate bounds**. vCache replaces static thresholds with **online-learned, embedding-specific decision boundaries**—no manual fine-tuning required. This enables reliable cached response reuse across any embedding model or workload.
26+
vCache is the first semantic prompt cache that guarantees user-defined error rate bounds. Semantic caching reduces LLM latency and cost by returning cached model responses for semantically similar prompts (not just exact matches). vCache replaces static thresholds with **online-learned, embedding-specific decision boundaries**—no manual fine-tuning required. This enables reliable cached response reuse across any embedding model or workload.
2727

2828
> 💳
2929
> **Cost & Latency Optimization** <br>
@@ -192,4 +192,4 @@ If you use vCache for your research, please cite our [paper](https://arxiv.org/a
192192
journal={arXiv preprint arXiv:2502.03771},
193193
year={2025}
194194
}
195-
```
195+
```

0 commit comments

Comments
 (0)