Cpt_acl_findings

🚀 Continual Quantization-Aware Pre-Training: When to transition from 16-bit to 1.58-bit pre-training for BitNet language models?, has been accepted to the ACL 2025 Findings!