ml 16297 image001

Cost-effective document classification using the Amazon Titan Multimodal Embeddings Model

Organizations across industries want to categorize and extract insights from high volumes of documents of different formats. Manually processing these documents to classify and extract information remains expensive, error prone, and difficult to scale. Advances in generative artificial intelligence (AI) have given rise to intelligent document processing (IDP) solutions that can automate the document classification, …

1 23J2Zh5.max 1000x1000 1

Introducing multimodal and structured data embedding support in BigQuery

Embeddings represent real-world objects, like entities, text, images, or videos as an array of numbers (a.k.a vectors) that machine learning models can easily process. Embeddings are the building blocks of many ML applications such as semantic search, recommendations, clustering, outlier detection, named entity extraction, and more. Last year, we introduced support for text embeddings in …

The future of application delivery starts with modernization

IDC estimates that 750 million cloud native will be built by 2025. Where and how these applications are deployed will impact time to market and value realization. The reality is that application landscapes are complex, and they challenge enterprises to maintain and modernize existing infrastructure, while delivering new cloud-native features. Three in four executives reported …

ML 14896 veoneer blog image 1

Build an active learning pipeline for automatic annotation of images with AWS services

This blog post is co-written with Caroline Chung from Veoneer. Veoneer is a global automotive electronics company and a world leader in automotive electronic safety systems. They offer best-in-class restraint control systems and have delivered over 1 billion electronic control units and crash sensors to car manufacturers globally. The company continues to build on a …

1 EMFU for Gemma Models on Cloud TPU.max 1000x1000 1

Performance deep dive of Gemma on Google Cloud

Earlier this year we announced Gemma, an open weights model family built to enable developers to rapidly experiment with, adapt, and productionize on Google Cloud. Gemma models can run on your laptop, workstation, or on Google Cloud through either Vertex AI or Google Kubernetes Engine (GKE) using your choice of Cloud GPUs or Cloud TPUs. …

12Azzgos2lGtQFbg3EGuAa11A

User-Centered Machine Learning for Visual Search

AuthorsDimitrios Lymperopoulos, Head of Machine Learning, Palantir Ben Radford, Product Manager, Palantir In our previous blog post, we introduced User-Centered Machine Learning (UCML). The core UCML workflow enables end users to rapidly adapt cutting-edge Computer Vision (CV) capabilities to their specific and evolving missions, based on their feedback. It combines the power of research-grade detection models …

The Making of VES: the Cosmos Microservice for Netflix Video Encoding

Liwei Guo, Vinicius Carvalho, Anush Moorthy, Aditya Mavlankar, Lishan Zhu This is the second post in a multi-part series from Netflix. See here for Part 1 which provides an overview of our efforts in rebuilding the Netflix video processing pipeline with microservices. This blog dives into the details of building our Video Encoding Service (VES), and …

How the Masters uses watsonx to manage its AI lifecycle

At the Masters®, storied tradition meets state-of-the-art technology. Through a partnership spanning more than 25 years, IBM has helped the Augusta National Golf Club capture, analyze, distribute and use data to bring fans closer to the action, culminating in the AI-powered Masters digital experience and mobile app. Now, whether they’re lining the fairways or watching …

ML 16514 image001

Knowledge Bases for Amazon Bedrock now supports custom prompts for the RetrieveAndGenerate API and configuration of the maximum number of retrieved results

With Knowledge Bases for Amazon Bedrock, you can securely connect foundation models (FMs) in Amazon Bedrock to your company data for Retrieval Augmented Generation (RAG). Access to additional data helps the model generate more relevant, context-specific, and accurate responses without retraining the FMs. In this post, we discuss two new features of Knowledge Bases for …

Powering the next generation of AI startups with Google Cloud

Today, many of the most exciting innovations in generative AI are coming from fast-growing startups.  We’re proud that the majority of these companies are building on Google Cloud, using our open and optimized stack for AI; broad choice of models and compute infrastructure; and many important resources and routes-to-market tailored for startups. In fact more …