[LLM] InfiniGen: Efficient Generative Inference of LLMs with Dynamic KV Cache Management (OSDI 2024)
Similar Tracks
[Compression] The FastLanes Compression Layout: Decoding 100 Billion Integers per Second (VLDB 2023)
Data Lakehouse Systems for Data Science
[Fault Tolerance] Exploiting Nil-Externality for Fast Replicated Storage (SOSP 2021)
Data Lakehouse Systems for Data Science
[OS] An Empirical Study of Rust-for-Linux: The Success, Dissatisfaction, and Compromise (ATC 2024)
Data Lakehouse Systems for Data Science
[LLM Serving] Llumnix: Dynamic Scheduling for Large Language Model Serving (OSDI 2024)
Data Lakehouse Systems for Data Science
[KV store] BonsaiKV: Towards Fast, Scalable, and Persistent Key-Value Stores (VLDB 2024)
Data Lakehouse Systems for Data Science
Machine Learning Tutorial | Machine Learning Basics | Machine Learning Algorithms | Simplilearn
Simplilearn
Pertemuan Teknis Laboratorium dan Penyelenggara Uji Profisiensi Tahun 2024
Komite Akreditasi Nasional
[KV store] FluidKV: Seamlessly Bridging the Gap between Indexing Performance and Memory-Footprint
Data Lakehouse Systems for Data Science