Human-centric photo dataset aims to help spot AI biases responsibly

A database of more than 10,000 human images to evaluate biases in artificial intelligence (AI) models for human-centric computer vision is presented in Nature this week. The Fair Human-Centric Image Benchmark (FHIBE), developed by Sony AI, is an ethically sourced, consent-based dataset that can be used to evaluate human-centric computer vision tasks to identify and …

Adapting Self-Supervised Representations as a Latent Space for Efficient Generation

We introduce Representation Tokenizer (RepTok), a generative modeling framework that represents an image using a single continuous latent token obtained from self-supervised vision transformers. Building on a pre-trained SSL encoder, we fine-tune only the semantic token embedding and pair it with a generative decoder trained jointly using a standard flow matching objective. This adaptation enriches …

10I8DAvXCQEN1RpTTC0ZyaQ

Supercharging the ML and AI Development Experience at Netflix

Supercharging the ML and AI Development Experience at Netflix with Metaflow Shashank Srikanth, Romain Cledat Metaflow — a framework we started and open-sourced in 2019 — now powers a wide range of ML and AI systems across Netflix and at many other companies. It is well loved by users for helping them take their ML/AI workflows from prototype to production, allowing …

Screenshot 2025 11 04 at 101509AM

Iterate faster with Amazon Bedrock AgentCore Runtime direct code deployment

Amazon Bedrock AgentCore is an agentic platform for building, deploying, and operating effective agents securely at scale. Amazon Bedrock AgentCore Runtime is a fully managed service of Bedrock AgentCore, which provides low latency serverless environments to deploy agents and tools. It provides session isolation, supports multiple agent frameworks including popular open-source frameworks, and handles multimodal …

Databricks research reveals that building better AI judges isn’t just a technical concern, it’s a people problem

The intelligence of AI models isn’t what’s blocking enterprise deployments. It’s the inability to define and measure quality in the first place. That’s where AI judges are now playing an increasingly important role. In AI evaluation, a “judge” is an AI system that scores outputs from another AI system.  Judge Builder is Databricks’ framework for …