TorchAO: A PyTorch-Native Shortcut To Smaller, Faster Models TorchAO is PyTorch's native toolkit for model efficiency: it unifies post-training quantization (PTQ), quantization-aware training (QAT), float8 (FP8) training, and structured sparsity in one coherent... deep learning FP8 model efficiency open source PyTorch QAT quantization sparsity TorchAO