You are here
Memory-Efficient KV Cache Optimization for Large Language Model Inference at the Edge
| Title | Memory-Efficient KV Cache Optimization for Large Language Model Inference at the Edge |
| Publication Type | Conference Paper |
| Year of Publication | 2026 |
| Authors | Zhang, C., H. Tan, H. Pan, Y. Xu, H. Du, L. Zhang, and X. Fu |
| Conference Name | IEEE INFOCOM 2026 |
| Date Published | 05/2026 |
| Publisher | IEEE |
| Conference Location | Tokyo, Japan |