Posted inAI News
NVIDIA just achieved the impossible: 4-bit training for large language models
NVIDIA has officially trained a 12B-parameter language model on 10 trillion tokens — entirely in 4-bit precision (NVFP4). This is the first-ever stable large-scale 4-bit pretraining run without accuracy loss…









