Company Updates & Technology Articles
March 25, 2026

EgoVerse: An open-source recipe for human-to-robot transfer Consortium with Georgia Tech, Stanford, Meta, and others introduces EgoVerse, a new foundation for scalable robot learning
Read more
March 20, 2026

Voice Showdown is a preference-based benchmark for voice AI models, built on real human speech to measure how models perform across languages and real-world conversations.
Read more
March 18, 2026

Read more
March 17, 2026

Read more
March 16, 2026

Scale AI and Universal Robots (UR) announce a partnership at NVIDIA GTC to integrate the Physical AI Data Engine into UR industrial robots, enabling scalable, real-world AI deployment.
Read more
March 9, 2026

Scale Labs is Scale’s new research hub studying how advanced AI systems behave in real-world environments. Building on the work of SEAL, the lab focuses on evaluation, agentic and multimodal systems, post-training methods, enterprise deployment, and AI risk and oversight infrastructure, with research spanning frontier capability measurement, long-horizon agent behavior, and national-scale safety evaluation.
Read more
March 4, 2026

As LLMs and coding agents take on professional coding work, evaluations must assess them like junior engineers: by how they investigate a system, gather evidence, and explain what they’re observing. SWE Atlas, the first evaluation suite of its kind, does just that.
Read more
February 27, 2026

Scale RL Environments provide simulated, high-fidelity systems where AI agents can learn through trial and error across real workflows. By pairing realistic data universes with structured trajectories and process-level verification, teams can train, evaluate, and iterate on agent behavior safely, producing measurable improvements on public benchmarks while integrating directly into existing agent pipelines.
Read more
February 4, 2026

Scale AI partners with Webster University to launch a technical writing certificate advancing AI workforce skills.
Read more
February 3, 2026

The emergence of Moltbook offer an early glimpse into AI agents operating collectively at scale. These systems demonstrate how interaction between autonomous agents can give rise to emergent behaviors that are not attributable to any single model or prompt, complicating traditional, model-level approaches to AI safety. Viewed through the AI Risk Matrix, such agent collectives point to a distinct and underdeveloped category of system-level risk with implications for evaluation, red teaming, and governance.
Read more