Machine learning masters massive data sets: Algorithm breaks the exabyte barrier
A machine-learning algorithm demonstrated the capability to process data that exceeds a computer’s available memory by identifying a massive data set’s key features and dividing them into manageable batches that don’t choke computer hardware. Developed at Los Alamos National Laboratory, the algorithm set a world record for factorizing huge data sets during a test run on Oak Ridge National Laboratory’s Summit, the world’s fifth-fastest supercomputer.
Filling gaps in data sets or identifying outliers—that's the domain of the machine learning algorithm TabPFN, developed by a team led by Prof. Dr. Frank Hutter from the University of Freiburg. This artificial intelligence (AI) uses learning methods inspired by large language models. TabPFN learns causal relationships from synthetic data…
Soon, researchers may be able to create movies of their favorite protein or virus better and faster than ever before. Researchers at the Department of Energy's SLAC National Accelerator Laboratory have pioneered a new machine learning method—called X-RAI (X-Ray single particle imaging with Amortized Inference)—that can "look" at millions of…
The field of machine learning is traditionally divided into two main categories: "supervised" and "unsupervised" learning. In supervised learning, algorithms are trained on labeled data, where each input is paired with its corresponding output, providing the algorithm with clear guidance. In contrast, unsupervised learning relies solely on input data, requiring…