TorchAO: A PyTorch-Native Shortcut To Smaller, Faster Models TorchAO is PyTorch's native toolkit for model efficiency: it unifies post-training quantization (PTQ), quantization-aware training (QAT), float8 (FP8) training, and structured sparsity in one coherent... deep learning FP8 model efficiency open source PyTorch QAT quantization sparsity TorchAO
BitNet: 1-bit LLMs Land With Practical Inference on CPUs and GPUs BitNet from Microsoft Research is the official C++ inference stack for native 1-bit large language models, centered on BitNet b1.58. The repo ships fast, lossless ternary kernels for CPUs, a CUDA W2A8... 1-bit LLM BitNet CPU GGUF GPU inference llama.cpp quantization T-MAC