Sky-T1: A Cost-Effective Open Source Reasoning AI Model
NovaSky introduces Sky-T1-32B-Preview, an open source reasoning AI model trained for under $450, showcasing affordable AI development.

Big news from the world of AI: NovaSky, a bunch of brainy folks from UC Berkeley’s Sky Computing Lab, just dropped the Sky-T1-32B-Preview. And let me tell you, this isn’t just another model—it’s a game-changer. It goes toe-to-toe with some of OpenAI’s earlier creations on various benchmarks, all while keeping the training costs under $450. (Yeah, you read that right.) Plus, they’re not just giving away the model; they’re throwing in the dataset and training code too, making it a full open-source buffet.
This is huge. Sky-T1-32B-Preview proves you don’t need to sell a kidney to develop top-notch reasoning AI. By leaning heavily on synthetic data, they’ve slashed costs dramatically. For context, Writer’s Palmyra X 004, which also used synthetic data, set them back $700,000. Ouch.
Why should you care? Well, models like Sky-T1 are the meticulous fact-checkers of the AI world, especially in tricky fields like physics and math. They might not be the fastest, but when it comes to accuracy, they’re like that one friend who always has to be right (but actually is).
Here’s how they did it: NovaSky started with Alibaba’s QwQ-32B-Preview for the initial data, then polished it up with OpenAI’s GPT-4o-mini. Training this 32-billion-parameter beast took about 19 hours on 8 Nvidia H100 GPUs. It aced the MATH500 and LiveCodeBench challenges, though it’s still playing catch-up with OpenAI’s o1 on the GPQA-Diamond test—a brutal exam covering advanced science topics.
What’s next? NovaSky’s not resting on their laurels. They’re on a mission to push open-source models further, making them smarter, faster, and even more reliable. Because in the end, who doesn’t love a good underdog story?