Machine learning masters massive data sets: Algorithm breaks the exabyte barrier
A machine-learning algorithm demonstrated the capability to process data that exceeds a computer’s available memory by identifying a massive data set’s key features and dividing them into manageable batches that don’t choke computer hardware. Developed at Los Alamos National Laboratory, the algorithm set a world record for factorizing huge data sets during a test run on Oak Ridge National Laboratory’s Summit, the world’s fifth-fastest supercomputer.
Filling gaps in data sets or identifying outliers—that's the domain of the machine learning algorithm TabPFN, developed by a team led by Prof. Dr. Frank Hutter from the University of Freiburg. This artificial intelligence (AI) uses learning methods inspired by large language models. TabPFN learns causal relationships from synthetic data…
Soon, researchers may be able to create movies of their favorite protein or virus better and faster than ever before. Researchers at the Department of Energy's SLAC National Accelerator Laboratory have pioneered a new machine learning method—called X-RAI (X-Ray single particle imaging with Amortized Inference)—that can "look" at millions of…
The world of artificial intelligence (AI) has recently seen significant advancements in generative models, a type of machine-learning algorithm that "learns" patterns from sets of data in order to generate new, similar sets of data. Generative models are often used for things like drawing images and natural language generation—a famous…