Hi, I’m Wei

Welcome to my blog, where I share notes on tech, AI, and more.

Why Formatting Matters?

Introduction: What’s the Deal with Precision? Why does switching from FP32 to FP8 turn a coherent chatbot into a babbling mess? Floating-point precision isn’t just a nerdy detail—it’s the heartbeat of large-scale AI. In a world of trillion-parameter models, formats like FP32, BF16, and FP8 decide if training converges, inference flies, or your GPU melts. I’ve been digging into this lately, especially with models like DeepSeek V3 pushing low-precision boundaries. Here’s what I found: precision shapes everything from compute costs to output quality. Let’s break it down—starting with the basics, then diving into training, inference, and beyond. ...

February 22, 2025 · 1 min