ml16490 rag

Super charge your LLMs with RAG at scale using AWS Glue for Apache Spark

Large language models (LLMs) are very large deep-learning models that are pre-trained on vast amounts of data. LLMs are incredibly flexible. One model can perform completely different tasks such as answering questions, summarizing documents, translating languages, and completing sentences. LLMs have the potential to revolutionize content creation and the way people use search engines and …

Adapting model risk management for financial institutions in the generative AI era

Generative AI (gen AI) promises to usher in an era of transformation for quality, accessibility, efficiency, and compliance in the financial services industry. As with any new technology, it also introduces new complexities and risks. Striking a balance between harnessing its potential and mitigating its risks will be crucial for the adoption of gen AI …

From accessibility upgrades to a custom cat-food bowl, this mobile 3D printer can autonomously add features to a room

Researchers created MobiPrint, a mobile 3D printer that can automatically measure a room and print objects onto the floor. The team’s graphic interface lets users design objects in a space that the robot has mapped out. The prototype, which the team built on a modified consumer vacuum robot, can add a range of objects to …

OpenAI unveils sCM, a new model that generates video media 50 times faster than current diffusion models

Two experts with the OpenAI team have developed a new kind of continuous-time consistency model (sCM) that they claim can generate video media 50 times faster than models currently in use. Cheng Lu and Yang Song have published a paper describing their new model on the arXiv preprint server. They have also posted an introductory …

CtrlSynth: Controllable Image-Text Synthesis for Data-Efficient Multimodal Learning

Pretraining robust vision or multimodal foundation models (e.g., CLIP) relies on large-scale datasets that may be noisy, potentially misaligned, and have long-tail distributions. Previous works have shown promising results in augmenting datasets by generating synthetic samples. However, they only support domain-specific ad hoc use cases (e.g., either image or text only, but not both), and …

ML 17337 image001

Unlocking generative AI for enterprises: How SnapLogic powers their low-code Agent Creator using Amazon Bedrock

This post is cowritten with Greg Benson, Aaron Kesler and David Dellsperger from SnapLogic. The landscape of enterprise application development is undergoing a seismic shift with the advent of generative AI. SnapLogic, a leader in generative integration and automation, has introduced the industry’s first low-code generative AI development platform, Agent Creator, designed to democratize AI …

1 Choosing the right metric GPU Utilizat.max 1000x1000 1

Save on GPUs: Smarter autoscaling for your GKE inferencing workloads

While LLM models deliver immense value for an increasing number of use cases, running LLM inference workloads can be costly. If you’re taking advantage of the latest open models and infrastructure, autoscaling can help you optimize your costs — ensuring you’re meeting customer demand while only paying for the AI accelerators you need. As a …