-
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper • 2402.17764 • Published • 627 -
Atom: Low-bit Quantization for Efficient and Accurate LLM Serving
Paper • 2310.19102 • Published • 11 -
AMSP: Super-Scaling LLM Training via Advanced Model States Partitioning
Paper • 2311.00257 • Published • 10 -
BiLLM: Pushing the Limit of Post-Training Quantization for LLMs
Paper • 2402.04291 • Published • 50
Spencer Presley
swpresley
AI & ML interests
Transformers, LLMs, Neural Networks, Hierarchical Classification.
Recent Activity
liked
a model
9 days ago
dphn/Dolphin3.0-R1-Mistral-24B
liked
a model
9 days ago
ArliAI/gpt-oss-20b-Derestricted
liked
a model
9 days ago
nvidia/Nemotron-Orchestrator-8B
Organizations
None yet