Dipika Sikka

Dipika Sikka's contributions

LLM Compressor v0.10.0 is here
Article

LLM Compressor v0.10: Faster compression with distributed GPTQ

Kyle Sayers +2

LLM Compressor v0.10 introduces Distributed Data Parallel (DDP) for faster compression, memory management, and advanced quantization formats. Make model compression workflows more efficient for large language models.

Featured image for LLM Compressor 0.7.0 release blog.
Article

LLM Compressor 0.7.0 release recap

Dipika Sikka +3

LLM Compressor 0.7.0 brings Hadamard transforms for better accuracy, mixed-precision FP4/FP8, and calibration-free block quantization for efficient compression.