Workshop
|
Fri 17:00
|
Knowledge and Skill Acquisition through Language Model Pre-training and Instruction-tuning
Xi Victoria Lin
|
|
Workshop
|
|
Unsupervised Adversarial Detection without Extra Model: Training Loss Should Change
|
|
Workshop
|
|
Beyond Scale: the Diversity Coefficient as a Data Quality Metric Demonstrates LLMs are Pre-trained on Formally Diverse Data
Alycia Lee · Brando Miranda · Sanmi Koyejo
|
|
Poster
|
Wed 17:00
|
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Guangxuan Xiao · Ji Lin · Mickael Seznec · Hao Wu · Julien Demouth · Song Han
|
|
Workshop
|
|
Training Diffusion Models with Reinforcement Learning
Kevin Black · Michael Janner · Yilun Du · Ilya Kostrikov · Sergey Levine
|
|
Workshop
|
|
Evidence of Meaning in Language Models Trained on Programs
Charles Jin · Martin Rinard
|
|
Workshop
|
|
Can Chatbots “Understand”? Evidence of Meaning in Language Models Trained on Programs
Charles Jin · Martin Rinard
|
|
Workshop
|
Sat 12:01
|
Using Megatron to Train Large Language Models (Deepak Narayanan, Microsoft Research)
|
|
Workshop
|
|
Training Diffusion Models with Reinforcement Learning
Kevin Black · Michael Janner · Yilun Du · Ilya Kostrikov · Sergey Levine
|
|
Workshop
|
Sat 12:40
|
Training Large Language Models on Cerebras Wafer-Scale Clusters AI (Natalia Vassilieva, Cerebras)
|
|
Workshop
|
|
Improving Training of Likelihood-based Generative Models with Gaussian Homotopy
Ba-Hien Tran · Giulio Franzese · Pietro Michiardi · Maurizio Filippone
|
|
Workshop
|
|
Learning and Leveraging Verifiers to Improve Planning Capabilities of Pre-trained Language Models
Daman Arora · Subbarao Kambhampati
|
|