Categories: FAANG

Regularized Training of Nearest Neighbor Language Models

Including memory banks in a natural language processing architecture increases model capacity by equipping it with additional data at inference time. In this paper, we build upon kNN-LM, which uses a pre-trained language model together with an exhaustive kNN search through the training data (memory bank) to achieve state-of-the-art results. We investigate whether we can improve the kNN-LM performance by instead training a LM with the knowledge that we will be using a kNN post-hoc. We achieved significant improvement using our method on language modeling tasks on WIKI-2 and WIKI-103. The main…
AI Generated Robotic Content

Recent Posts

The 2025 Machine Learning Toolbox: Top Libraries and Tools for Practitioners

2024 was the year machine learning (ML) and artificial intelligence (AI) went mainstream, affecting peoples'…

22 hours ago

A Complete Introduction to Using BERT Models

Overview This post is divided into five parts; they are: • Why BERT Matters •…

22 hours ago

Accelerate video Q&A workflows using Amazon Bedrock Knowledge Bases, Amazon Transcribe, and thoughtful UX design

Organizations are often inundated with video and audio content that contains valuable insights. However, extracting…

22 hours ago

Elon Musk Ally Tells Staff ‘AI-First’ Is the Future of Key Government Agency

Sources say the former Tesla engineer now in charge of the Technology Transformation Services wants…

23 hours ago

User-friendly system can help developers build more efficient simulations and AI models

The neural network artificial intelligence models used in applications like medical image processing and speech…

23 hours ago

The AI paradox: How tomorrow’s cutting-edge tools can become dangerous cyber threats (and what to do to prepare)

AI agents will bring enterprises to the next level, but the same applies to related…

2 days ago