Adaptive Data Collection via Autoregressive Generation

Published on ● Video Link: https://www.youtube.com/watch?v=X1eA5sQQ-7g



Duration: 0:00
140 views
0


Hongseok Namkoong (Columbia University)
https://simons.berkeley.edu/talks/hongseok-namkoong-coumbia-university-2024-11-15
Domain Adaptation and Related Areas

Real-world decision-making requires grappling with a perpetual lack of data as environments change; intelligent agents must comprehend uncertainty and actively gather information to resolve it. We propose a new framework for learning adaptive data collection algorithms from massive historical data, which we demonstrate in a cold-start recommendation problem. First, we use historical data to pretrain an autoregressive model to predict a sequence of repeated feedback/rewards (e.g., responses to news articles shown to different users over time). In learning to make accurate predictions, the model implicitly learns an informed prior based on rich action features (e.g., article headlines) and how to sharpen beliefs as more rewards are gathered (e.g., clicks as each article is recommended). At decision-time, we autoregressively sample (impute) an imagined sequence of rewards for each action, and choose the action with the largest average imputed reward. Far from a heuristic, our approach is an implementation of Thompson sampling (with a learned prior), a prominent active exploration algorithm. We prove our pretraining loss directly controls online decision-making performance, and we demonstrate our framework on a news recommendation task where we integrate end-to-end fine-tuning of a pretrained language model to process news article headline text to improve performance.




Other Videos By Simons Institute for the Theory of Computing


2024-11-19Talk by Mahdi Soltanolkotabi (University of Southern California)
2024-11-18Some Easy Optimization Problems Have the Overlap-Gap Property
2024-11-18Understanding Contrastive Learning and Self-training
2024-11-18Revisiting Scalarization in Multi-Task Learning
2024-11-18Beyond Decoding: Meta-Generation Algorithms for Large Language Models (Remote Talk)
2024-11-18Omnipredicting Single-Index Models with Multi-Index Models
2024-11-17The Truth About Your Lying Calibrated Forecaster: How to Design Truthful Calibration Measures
2024-11-17A Discrepancy-Based Theory of Adaptation
2024-11-17Bypassing the Impossibility of Online Learning Thresholds: Unbounded Losses and Transductive Priors
2024-11-17Learning from Dynamics
2024-11-14Adaptive Data Collection via Autoregressive Generation
2024-11-13Toward Understanding the Extrapolation of Nonlinear Models to Unseen Domains or the Whole Domain
2024-11-13Language-guided Adaptation
2024-11-13On Spurious Associations and LLM Alignment
2024-11-13Causally motivated robustness to shortcut learning
2024-11-13Talk by Zachary Lipton
2024-11-12Distribution shift in ecological data: generalization vs. specialization,
2024-11-12Transfer learning via local convergence rates of the nonparametric least squares estimator
2024-11-12Transfer learning for weak-to-strong generalization
2024-11-12User-level and federated local differential privacy
2024-11-11The Evolution of Domain Transfer in the Era of Pre-trained Language Models