The ‘truth serum’ for AI: OpenAI’s new method for training models to confess their mistakes

OpenAI researchers have introduced a novel method that acts as a “truth serum” for large language models (LLMs), compelling them to self-report their own misbehavior, hallucinations and policy violations. This technique, “confessions,” addresses a growing concern in enterprise AI: Models can be dishonest, overstating their confidence or covering up the shortcuts they take to arrive …

Scientists develop a glasses-free 3D system with a little help from AI

Watching 3D movies and TV shows is a fun and exciting experience, where images leap out of the screen. To get this effect, you usually have to wear a special pair of glasses. But that could soon be a thing of the past as scientists have developed a new display system that delivers a realistic …

Semantic Regexes: Auto-Interpreting LLM Features with a Structured Language

Automated interpretability aims to translate large language model (LLM) features into human understandable descriptions. However, these natural language feature descriptions are often vague, inconsistent, and require manual relabeling. In response, we introduce semantic regexes, structured language descriptions of LLM features. By combining primitives that capture linguistic and semantic feature patterns with modifiers for contextualization, composition, …

Gemini 3 Pro scores 69% trust in blinded testing up from 16% for Gemini 2.5: The case for evaluating AI on real-world trust, not academic benchmarks

Just a few short weeks ago, Google debuted its Gemini 3 model, claiming it scored a leadership position in multiple AI benchmarks. But the challenge with vendor-provided benchmarks is that they are just that — vendor-provided. A new vendor-neutral evaluation from Prolific, however, puts Gemini 3 at the top of the leaderboard. This isn’t on …

Guided learning helps previously ‘untrainable’ neural networks learn effectively

Even networks long considered “untrainable” can learn effectively with a bit of a helping hand. Researchers at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have shown that a brief period of alignment between neural networks, a method they call guidance, can dramatically improve the performance of architectures previously thought unsuitable for modern tasks.

SV9YWsUS3WxDADpgEOpN9HscFXXOFmP Ck2kz6GeKWA

Okay… Z-Image Is GENUINELY Revolutionary. The Fact That We Can Upscale to THIS Quality THIS Coherently Without ControlNET Tiles?!?! The Fact That There’s THIS Much Prompt Understanding with Hyper-Detailed Prompts??? The Fact That the Images Look THIS Beautiful with Only 6B Parameters?!?! THIS Is IT.

The first image of every set is the upscaled one, and the second image of every set is the original one. And upscaling is so SIMPLE—just Ultimate SD Upscaled with 0.20 denoise in ComfyUI. I am genuinely FLOORED by this. So many prompts I wanted to try (especially the building one) that no other model …