Categories: FAANG

TiC-LM: A Web-Scale Benchmark for Time-Continual LLM Pretraining

This paper was accepted to the ACL 2025 main conference as an oral presentation.
This paper was accepted at the Scalable Continual Learning for Lifelong Foundation Models (SCLLFM) Workshop at NeurIPS 2024.
Large Language Models (LLMs) trained on historical web data inevitably become outdated. We investigate evaluation strategies and update methods for LLMs as new data becomes available. We introduce a web-scale dataset for time-continual pretraining of LLMs derived from 114 dumps of Common Crawl (CC) – orders of magnitude larger than previous continual language modeling benchmarks. We also…
AI Generated Robotic Content

Recent Posts

RES4LYF nodes really do make a difference with Wan 2.2

submitted by /u/Hearmeman98 [link] [comments]

17 hours ago

7 Matplotlib Tricks to Better Visualize Your Machine Learning Models

Visualizing model performance is an essential piece of the machine learning workflow puzzle.

17 hours ago

Introducing Gemma 3 270M: The compact model for hyper-efficient AI

Today, we're adding a new, highly specialized tool to the Gemma 3 toolkit: Gemma 3…

17 hours ago

Investigating Intersectional Bias in Large Language Models using Confidence Disparities in Coreference Resolution

Large language models (LLMs) have achieved impressive performance, leading to their widespread adoption as decision-support…

17 hours ago

Scalable intelligent document processing using Amazon Bedrock Data Automation

Intelligent document processing (IDP) is a technology to automate the extraction, analysis, and interpretation of…

17 hours ago

How Keeta processes 11 million financial transactions per second with Spanner

Keeta Network is a layer‑1 blockchain that unifies transactions across different blockchains and payment systems,…

17 hours ago