Inference Optimized Checkpoints (with Model Optimizer) Collection A collection of generative models quantized and optimized for inference with TensorRT Model Optimizer. • 45 items • Updated 4 days ago • 62
Gemma 3 QAT Collection Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory • 15 items • Updated Jul 10 • 210
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 626