faang

A Small-Scale System for Autoregressive Program Synthesis Enabling Controlled Experimentation

What research can be pursued with small models trained to complete true programs? Typically, researchers study program synthesis via large…

4 hours ago

Scaling LLM Post-Training at Netflix

Baolin Li, Lingyi Liu, Binh Tang, Shaojing LiIntroductionPre-training gives Large Language Models (LLMs) broad linguistic ability and general world knowledge, but…

4 hours ago

Customize AI agent browsing with proxies, profiles, and extensions in Amazon Bedrock AgentCore Browser

AI agents that browse the web need more than basic page navigation. Our customers tell us they need agents that…

4 hours ago

Mapping the Design Space of User Experience for Computer Use Agents

Large language model (LLM)-based computer use agents execute user commands by interacting with available UI elements, but little is known…

1 day ago

Introducing PFCS Forward

Introducing PFCS Forward: Extending IL5/IL6 Authorization from Cloud to EdgeIntegrated systems that solve meaningful problems for commanders and their warfighting requirements…

1 day ago

Automating RDS Postgres to Aurora Postgres Migration

Ram Srivasta Kannan, Wale Akintayo, Jay Bharadwaj, John Crimmins, Shengwei Wang, Zhitao ZhuIntroductionIn 2024, the Online Data Stores team at Netflix…

1 day ago

AI meets HR: Transforming talent acquisition with Amazon Bedrock

Organizations face significant challenges in making their recruitment processes more efficient while maintaining fair hiring practices. By using AI to…

1 day ago

NVIDIA Nemotron 3 Nano 30B MoE model is now available in Amazon SageMaker JumpStart

Today we’re excited to announce that the NVIDIA Nemotron 3 Nano 30B model with  3B active parameters is now generally…

2 days ago

Build financial resilience with AI-powered tabletop exercises on Google Cloud

In the financial sector, resilience isn't optional. Recent cloud outages have shown us exactly how fast critical data can disappear.…

2 days ago

Parallel Track Transformers: Enabling Fast GPU Inference with Reduced Synchronization

Efficient large-scale inference of transformer-based large language models (LLMs) remains a fundamental systems challenge, frequently requiring multi-GPU parallelism to meet…

3 days ago