Apple is presenting new research at the annual conference on Neural Information Processing Systems (NeurIPS), which takes place in person in Vancouver, Canada, from December 10 - 15. We are proud to again sponsor the multi-track interdisciplinary conference, which brings together the scientific and industrial research communities surrounding Machine Learning. Below is an overview of Apple’s participation at NeurIPS 2024.
Schedule
Stop by Apple's booth (#323 in West Hall A) during exhibition hours (all times Pacific):
- Tuesday, December 10: 12:00 PM - 8:00 PM
- Wednesday, December 11: 9:00 AM - 5:00 PM
- Thursday, December 12: 9:00 AM - 4:00 PM
Tuesday, December 10
- Women in Machine Learning (WiML) Workshop and Roundtable
- 11:00 AM - 12:30 PM, West Meeting Rooms 211-214
- Samy Bengio, Natalie Mackraz and Eleonora Gualdoni are representing Apple at the WiML roundtables.
- Black in AI Workshop
- 1:30 - 2:30 PM, West Meeting Rooms 208-209
- Barry Theobald and Josh Gardner are representing Apple during the workshop Mentorship Feedback Session.
- LatinX in AI Workshop
- 3:10 - 4:00 PM, West Meeting Rooms 202-204
- Samy Bengio, Erdrin Azemi and Lauren Araujo are representing Apple during the workshop mentorship hour.
Wednesday, December 11
- Queer in AI Workshop
- 10:00 AM - 5:00 PM, West Meeting Rooms 202-204
- Dataset Decomposition: Faster LLM Training with Variable Sequence Length Curriculum
- 11:00 AM - 2:00 PM, Poster Session 1 East
- Hadi Pour Ansari, Chun-Liang Li, Rick Chang, Pavan Kumar Anasosalu Vasu, Cem Koc, Vaishaal Shankar, Oncel Tuzel
- Private Stochastic Convex Optimization with Heavy Tails: Near-Optimality from Simple Reductions
- 11:00 AM - 2:00 PM, Poster Session 1 West
- Hilal Asi, Daogao Liu, Kevin Tian
- GENOT: Entropic (Gromov) Wasserstein Flow Matching with Applications to Single-Cell Genomics
- 4:30 PM - 7:30 PM, Poster Session 2 East
- Dominik Klein, Theo Uscidda, Fabian Theis, Marco Cuturi
- How Far Can Transformers Reason? The Globality Barrier and Inductive Scratchpad
- 4:30 PM - 7:30 PM, Poster Session 2 East
- Emmanuel Abbe, Samy Bengio, Aryo Lotfi, Colin Sandon, Omid Saremi
- How JEPA Avoids Noisy Features: The Implicit Bias of Deep Linear Self Distillation Networks
- 4:30 PM - 7:30 PM, Poster Session 2 East
- Etai Littwin, Omid Saremi, Madhu Advani, Chen Huang, Preetum Nakkiran, Josh Susskind, Vimal Thilak
- Progressive Entropic Optimal Transport Solvers
- 4:30 PM - 7:30 PM, Poster Session 2 West
- Parnian Kassraie, Aram Alexandre Pooladian, Michal Klein, James Thornton, Jonathan Niles-Weed, Marco Cuturi Cameto
Thursday, December 12
- 4M-21: An Any-to-Any Vision Model for Tens of Tasks and Modalities
- 11:00 AM - 2:00 PM, Poster Session 3 East
- Roman Bachmann, Oguzhan Kar, David Mizrahi, Ali Garjani, Mingfei Gao, David Griffiths, Jimmy Hu, Afshin Dehghan, Amir Zamir
- Instance-Optimal Private Density Estimation in the Wasserstein Distance
- 11:00 AM - 2:00 PM, Poster Session 3 West
- Vitaly Feldman, Audra McMillan, Satchit Sivakumar, Kunal Talwar
- PFL-Research: Simulation Framework for Accelerating Research in Private Federated Learning
- 11:00 AM - 2:00 PM, Poster Session 3 West
- Filip Granqvist, Congzheng Song, Aine Cahill, Rogier van Dalen, Martin Pelikan, Yi Sheng Chan, Xiaojun Feng, Natarajan Krishnaswami, Vojta Jina, Mona Chitnis
- Transformation-Invariant Learning and Theoretical Guarantees for OOD Generalization
- 4:30 PM - 7:30 PM, Poster Session 4 West
- Omar Montasser, Han Shao, Emmanuel Abbe
- When is Multicalibration Post-Processing Necessary?
- 4:30 PM - 7:30 PM, Poster Session 4
- Dutch Hansen, Siddartha Devic, Preetum Nakkiran, Vatsal Sharan
Friday, December 13
- Aggregate-and-Adapt Natural Language Prompts for Downstream Generalization of CLIP
- 11:00 AM - 2:00 PM, Poster Session 5 East
- Chen Huang, Skyler Seto, Samira Abnar, David Grangier, Navdeep Jaitly, Josh Susskind
- Grounding of Multimodal Large Language Models in Action Spaces
- 11:00 AM - 2:00 PM, Poster Session 5 East
- Andrew Szot, Bogdan Mazoure, Harsh Agrawal, Devon Hjelm, Zsolt Kira, Alexander Toshev
- Kaleido Diffusion: Improving Conditional Diffusion Models with Auto-Regressive Latent Modeling
- 11:00 AM - 2:00 PM, Poster Session 5 East
- Jiatao Gu, Ying Shen, Shuangfei Zhai, Yizhe Zhang, Navdeep Jaitly, Josh Susskind
- Learning Elastic Costs to Shape Monge Displacements
- 11:00 AM - 2:00 PM, Poster Session 5 West
- Michal Klein, Aram Alexandre Pooladian, Pierre Ablin, Eugene Ndiaye, Jonathan Niles Weed, Marco Cuturi
- Private and Personalized Frequency Estimation in a Federated Setting
- 11:00 AM - 2:00 PM, Poster Session 5 West
- Amrith Setlur, Vitaly Feldman, Kunal Talwar
- Private Online Learning via Lazy Algorithms
- 11:00 AM - 2:00 PM, Poster Session 5 West
- Hilal Asi, Daogao Liu, Tomer Koren, Kunal Talwar
- Strategic Linear Contextual Bandits
- 11:00 AM - 2:00 PM, Poster Session 5 West
- Thomas Kleine Buening, Aadi Saha, Christos Dimitrakakis, Haifeng Xu
- DataComp-LM: In Search of the Next Generation of Training Sets for Language Models
- 4:30 PM - 7:30 PM, Poster Session 6 West
- Jeffrey Li, Alex Fang, Georgios Smyrnis, Matt Jordan, Maor Igvi, Hadi Pour Ansari, Fartash Faghri, Alaaeldin Mohamed Elnouby Ali, Alexander Toshev, Alex Dimakis, Hanlin Zhang, Hritik Bansal, Igor Vasiljevic, Jean Mercat, Jenia Jitsev, Kushal Arora, Mayee Chen, Niklas Muenninghoff, Luca Soldaini, Pang Wei Koh, Reinhard Heckel, Rui Xin, Samir Gadre, Rulin Shao, Sarah Pratt, Saurabh Garg, Sedrick Keh, Suchin Gururangan, Sunny Sanyal, Yonatan Bitton, Thomas Kollar, Mitchell Wortsman, Etash Guha, Amro Abbas, Cheng-Yu Hsieh, Dhruba Ghosh, Gabriel Ilharco, Giannis Daras, Kalyani Marathe, Joshua Gardner, Marianna Nezhurina, Achal Dave, Yair Carmon, Ludwig Schmidt, Vaishaal Shankar
- Faster Algorithms for User-Level Private Stochastic Convex Optimization
- 4:30 PM - 7:30 PM, Poster Session 6 West
- Hilal Asi, Daogao Liu, Andrew Lowy
- Learning Spatially-Aware Language and Audio Embeddings
- 4:30 PM - 7:30 PM, Poster Session 6 East
- Bhavika Devnani, Skyler Seto, Zak Aldeneh, Alessandro Toso, Elena Menyaylenko, Barry Theobald, Jonathan Sheaffer, Miguel Sarabia del Castillo
- ODGEN: Domain-Specific Object Detection Data Generation with Diffusion Models
- 4:30 PM - 7:30 PM, Poster Session 6 East
- JingYuan Zhu, Shiyu Li, Andy Liu, Ping Huang, Jiulong Shan, Huimin Ma, Jian Yuan
Saturday, December 14
- The Fourth Workshop on Efficient Natural Language and Speech Processing (ENLSP)
- 8:10 AM - 5:00 PM, West Meeting Room 301
-
- Navdeep Jaitly will be giving a keynote talk on Hardware-aware Algorithms for Language Modeling
- 11:00 AM - 11:30 AM
-
- Towards Low-Bit Communication for Tensor Parallel LLM Inference
- 12:30 PM - 1:30 PM
- Harry Dong, Tyler Johnson, Minsik Cho, Emad Soroush
-
- Scaling Smart: Accelerating Large Language Model Pre-Training with Small Model Initialization
- 1:30 PM - 2:30 PM
- Mohammad Samragh Razlighi, Iman Mirzadeh, Keivan Alizadeh Vahid, Fartash Faghri, Minsik Cho, Moin Nabi, Devang Naik, Mehrdad Farajtabar
-
- Duo-LLMs: A Framework for Studying Adaptive Computation in Large Language Models
- 4:30 PM - 5:30 PM
- Keivan Alizadeh Vahid, Iman Mirzadeh, Mohammad Sekhavat, Minsik Cho, Dmitry Belenko, Frank Sun, Hooman Shahrokhi, Moin Nabi, Mehrdad Farajtabar
-
- Speculative Streaming: Fast LLM Inference without Auxiliary Models
- 4:30 PM - 5:30 PM, West Exhibition Hall A
- Nikhil Bhendawade, Irina Belousova, Qichen Fu, Henry Mason, Mohammad Rastegari, Mahyar Najibikohnehshahri
-
- Computational Bottlenecks of Training Small-Scale Large Language Models
- 4:30 PM - 5:30 PM
- Saleh Ashkboos, Iman Mirzadeh, Keivan Alizadeh, Mohammad Hossein Sekhavat, Moin Nabi, Mehrdad Farajtabar, Fartash Faghri
- Unifying Representations in Neural Models
- 8:15 AM - 6:00 PM, West Exhibition Hall C, B3
-
- Marco Cuturi will be giving an invited talk during the workshop.
- 2:00 PM - 2:30 PM
- UniReps Workshop 2024
- 8:15 AM - 5:00 PM, East Exhibition Hall B, C
-
- Learning Functions on Symmetric Matrices and Point Clouds via Lightweight Invariant Features
- 3:30 PM - 5:00 PM
- Ben Blum-Smith, Teresa Huang, Marco Cuturi, Soledad Villar
- AIM-FM: Advancements In Medical Foundation Models
- 8:20 AM - 5:30 PM, East Ballroom A, B
-
- Promoting Cross-Modal Representations to Improve Multimodal Foundation Models for Physiological Signals
- 11:20 AM - 12:00 PM, and 5:10 PM - 5:30 PM
- Ching Fang, Chris Sandino, Behrooz Mahasseni, Juri Minxha, Hadi Pour Ansari, Erdrin Azemi, Ali Moin, Ellen Zippi
- Adaptive Foundation Models (AFM)
- 8:30 AM - 5:30 PM, West Exhibition Hall A
-
- Device-Directed Speech Detection for Follow-up Conversations Using Large Language Models
- 10:45 AM - 11:45 AM
- Oggi Rudovic, Pranay Dighe, Yi Su, Vineet Garg, Sameer Dharur, Xiaochuan Niu, Ahmed Hussen Abdelaziz, Saurabh Adya, Ahmed Tewfik
- Mathematics of Modern Machine Learning (M3L)
- 8:50 AM - 5:00 PM, Meeting Rooms 1-3 + South Foyer
-
- Classifier-Free Guidance is a Predictor-Corrector
- 10:30 AM - 10:45 AM
- Arwen Bradley, Preetum Nakkiran
-
- Classifier-Free Guidance is a Predictor-Corrector
- 11:15 AM - 12:15 PM
- Arwen Bradley, Preetum Nakkiran
- Algorithmic Fairness Through the Lens of Metrics and Evaluation
- 9:00 AM - 5:30 PM, West Meeting Rooms 111-112
-
- Evaluating Gender Bias Transfer Between Pre-trained and Prompt-Adapted Language Models
- 2:50 PM - 2:55 PM
- Niv Sivakumar, Natalie Mackraz, Samira Khorshidi, Krishna Patel, Barry Theobald, Luca Zappella, Nick Apostoloff
-
- Evaluating Gender Bias Transfer Between Pre-trained and Prompt-Adapted Language Models
- 2:55 PM - 3:30 PM
- Niv Sivakumar, Natalie Mackraz, Samira Khorshidi, Krishna Patel, Barry Theobald, Luca Zappella, Nick Apostoloff
- Self-Supervised Learning - Theory and Practice
- 9:00 AM - 5:30 PM, Meeting Rooms 202-204
-
- Enhancing JEPAs with Spatial Conditioning: Robust and Efficient Representation Learning
- 12:30 PM - 1:50 PM
- Etai Littwin, Vimal Thilak, Anand Gopalakrishnan
- Workshop on Attributing Model Behavior at Scale (ATTRIB)
- 9:00 AM - 5:00 PM, Meeting Rooms 205 - 207
-
- Towards Data-Centric RLHF: Simple Metrics for Preference Dataset Comparison
- 12:30 PM - 2:30 PM
- Judy Hanwen Shen, Archit Sharma, Jun Qin
-
- Understanding Compute-Parameter Trade-offs in Sparse Mixture-of-Expert Language Models
- 3:00 PM - 4:30 PM
- Harshay Shah, Samira Abnar, Vimal Thilak, Dan Busbridge, Alaaeldin Mohamed Elnouby Ali, Josh Susskind
Sunday, December 15
- Time Series in the Age of Large Models
- 8:15 AM - 5:20 PM, Meeting Rooms 220 - 222
-
- Leveraging Periodicity for Robustness with Multi-modal Mood Pattern Models
- 9:30 AM - 10:30 AM , and 1:00 PM - 2:00 PM
- Jaya Narain, Jenny Sun, Oussama Elachqar, Haraldur Hallgrimsson, Feng Zhu, Shirley Ren
-
- Towards Time-Series Reasoning with LLMs
- 9:35 AM - 10:35 AM, and 1:00 PM - 2:00 PM
- Winnie Chow, Lauren Gardiner, Haraldur Hallgrimsson, Maxwell A. Xu, Shirley Ren
-
- Towards Time-Series Reasoning with LLMs
- 2:35 PM - 2:47 PM
- Winnie Chow, Lauren Gardiner, Haraldur Hallgrimsson, Maxwell A. Xu, Shirley Ren
- Workshop on Federated Foundation Models (FL@FM)
- 8:15 AM - 5:15 PM, East Wing, Meeting Rooms 8 & 15
-
- PFL-Research: Simulation Framework for Accelerating Research in Private Federated Learning
- 9:00 AM - 10:00 AM
- Filip Granqvist, Congzheng Song, Aine Cahill, Rogier van Dalen, Martin Pelikan, Yi Sheng Chan, Xiaojun Feng, Natarajan Krishnaswami, Vojta Jina, Mona Chitnis
-
- Momentum Approximation in Asynchronous Private Federated Learning
- 11:00 AM - 12:30 PM
- Tao Yu, Congzheng Song, Jianyu Wang, Mona Chitnis
- Foundation Model Interventions (MINT)
- 8:45 AM - 5:00 PM, Meeting Rooms 121 - 122
-
- Do LLMs Internally "Know" When They Follow Instructions?
- 1:00 PM - 2:00 PM
- Juyeon Heo, Christina Heinze-Deml, Shirley Ren, Oussama Elachqar, Udhay Nallasamy, Andy Miller, Jaya Narain
- Evaluating Evaluations: Examining Best Practices for Measuring Broader Impacts of Generative AI (EvalEval)
- 9:00 AM - 5:30 PM, East Meeting Room 16
-
- Fairness Dynamics During Training
- 12:30 PM - 2:30 PM
- Krishna Patel, Niv Sivakumar, Barry Theobald, Luca Zappella, Nick Apostoloff
- Safe Generative AI
- 9:00 AM - 5:15 PM, East Exhibition Hall A
-
- Do LLMs Estimate Uncertainty Well in Instruction-Following?
- 11:40 AM - 1:00 PM
- Juyeon Heo, Miao Xiong, Christina Heinze-Deml, Jaya Narain
-
- Efficient and Effective Uncertainty Quantification in LLMs
- 3:00 PM - 5:00 PM
- Miao Xiong, Andrea Santilli, Michael Kirchhof, Adam Golinski, Sinead Williamson
- Workshop on Machine Learning and Compression
- 9:00 AM - 5:30 PM, Meeting Rooms 211 - 214
-
- Do Compressed LLMs Forget Knowledge? An Experimental Study with Practical Implications
- 2:20 PM - 3:50 PM
- Scott Hoang, Minsik Cho, Thomas Merth, Atlas Wang, Mohammad Rastegari, Devang Naik
Demos
MLX
MLX is a flexible array framework that is optimized for Apple silicon and brought to you by Apple machine learning research. It enables training and inference of arbitrarily complex models on Apple silicon powered devices with great brevity and flexibility. The demo presents an example of fine-tuning of a 7B parameter LLM on an iPhone, image generation using a large diffusion model on an iPad and text generation using a number of large language models on an M2 Ultra. This demo will be hosted during exhibition booth hours Tuesday through Thursday. Learn more about MLX here.
MobileCLIP: Real-Time Image-Text Models
MobileCLIP is a family of mobile-friendly image-text models with hybrid CNN/Transformer architectures. In combination, these models attain the best accuracy-latency tradeoff. MobileCLIP-B obtains state-of-the-art results. This demo will be hosted during exhibition booth hours Tuesday through Thursday. Learn more about MobileCLIP here.
All conference attendees are invited to visit our booth to experience these demos in person.
Accepted Papers
Links to papers with ◊ will be added after the conference, as they become available
4M-21: An Any-to-Any Vision Model for Tens of Tasks and Modalities
Roman Bachmann, Oguzhan Kar, David Mizrahi, Ali Garjani, Mingfei Gao, David Griffiths, Jimmy Hu, Afshin Dehghan, Amir Zamir
Aggregate-and-Adapt Natural Language Prompts for Downstream Generalization of CLIP
Chen Huang, Skyler Seto, Samira Abnar, David Grangier, Navdeep Jaitly, Josh Susskind
DataComp-LM: In search of the Next Generation of Training Sets for Language Models
Jeffrey Li, Alex Fang, Georgios Smyrnis, Matt Jordan, Maor Igvi, Hadi Pour Ansari, Fartash Faghri, Alaaeldin Mohamed Elnouby Ali, Alexander Toshev, Alex Dimakis, et al.
Dataset Decomposition: Faster LLM Training with Variable Sequence Length Curriculum
Hadi Pour Ansari, Chun-Liang Li, Rick Chang, Pavan Kumar Anasosalu Vasu, Cem Koc, Vaishaal Shankar, Oncel Tuzel
GENOT: Entropic (Gromov) Wasserstein Flow Matching with Applications to Single-Cell Genomics
Dominik Klein, Theo Uscidda, Fabian Theis, Marco Cuturi Cameto
Faster Algorithms for User-Level Private Stochastic Convex Optimization Hilal Asi, Daogao Liu, Andrew Lowy
Grounding of Multimodal Large Language Models in Action Spaces ◊
Andrew Szot, Bogdan Mazoure, Harsh Agrawal, Devon Hjelm, Zsolt Kira, Alexander Toshev
How Far Can Transformers Reason? The Globality Barrier and Inductive Scratchpad
Emmanuel Abbe, Samy Bengio, Aryo Lotfi, Colin Sandon, Omid Saremi
How JEPA Avoids Noisy Features: The Implicit Bias of Deep Linear Self Distillation Networks
Etai Littwin, Omid Saremi, Madhu Advani, Chen Huang, Preetum Nakkiran, Josh Susskind, Vimal Thilak
Instance Optimal Private Density Estimation in the Wasserstein Distance
Vitaly Feldman, Audra McMillan, Satchit Sivakumar, Kunal Talwar
Kaleido Diffusion: Improving Conditional Diffusion Models with Auto-Regressive Latent Modeling
Jiatao Gu, Ying Shen, Shuangfei Zhai, Yizhe Zhang, Navdeep Jaitly, Josh Susskind
Learning Spatially-Aware Language and Audio Embeddings ◊
Bhavika Devnani, Skyler Seto, Zak Aldeneh, Alessandro Toso, Elena Menyaylenko, Barry Theobald, Jonathan Sheaffer, Miguel Sarabia del Castillo
Learning Elastic Costs to Shape Monge Displacements
Michal Klein, Aram Alexandre Pooladian, Pierre Ablin, Eugene Ndiaye, Jonathan Niles Weed, Marco Cuturi
ODGEN: Domain-Specific Object Detection Data Generation with Diffusion Models
JingYuan Zhu, Shiyu Li, Andy Liu, Ping Huang, Jiulong Shan, Huimin Ma, Jian Yuan
PFL-Research: Simulation Framework for Accelerating Research in Private Federated Learning
Filip Granqvist, Congzheng Song, Aine Cahill, Rogier van Dalen, Martin Pelikan, Yi Sheng Chan, Xiaojun Feng, Natarajan Krishnaswami, Vojta Jina, Mona Chitnis
Private and Personalized Frequency Estimation in a Federated Setting
Amrith Setlur, Vitaly Feldman, Kunal Talwar
Private Online Learning via Lazy Algorithms
Hilal Asi, Daogao Liu, Tomer Koren, Kunal Talwar
Private Stochastic Convex Optimization with Heavy Tails
Hilal Asi, Daogao Liu, Kevin Tian
Progressive Entropic Optimal Transport Solvers
Parnian Kassraie, Aram Alexandre Pooladian, Michal Klein, James Thornton, Jonathan Niles-Weed, Marco Cuturi
Strategic Linear Contextual Bandits
Aadi Saha, Thomas Kleine Buening, Christos Dimitrakakis, Haifeng Xu
Transformation-Invariant Learning and Theoretical Guarantees for OOD Generalization ◊
Omar Montasser, Han Shao, Emmanuel Abbe
When is Multicalibration Post-Processing Necessary?
Dutch Hansen, Siddartha Devic, Preetum Nakkiran, Vatsal Sharan
Accepted Workshop Papers
Links to workshop papers with ◊ will be added after the conference, as they become available
AdEMAMix: Leveraging the Surprising Relevance of Very Old Gradients ◊
Matteo Pagliardini, Pierre Ablin, David Grangier
Classifier-Free Guidance is a Predictor-Corrector
Arwen Bradley, Preetum Nakkiran
Computational Bottlenecks of Training Small-Scale Large Language Models
Saleh Ashkboos, Iman Mirzadeh, Keivan Alizadeh, Mohammad Hossein Sekhavat, Moin Nabi, Mehrdad Farajtabar, Fartash Faghri
Device-Directed Speech Detection for Follow-up Conversations Using Large Language Models
Oggi Rudovic, Pranay Dighe, Yi Su, Vineet Garg, Sameer Dharur, Xiaochuan Niu, Ahmed Hussen Abdelaziz, Saurabh Adya, Ahmed Tewfik
Do Compressed LLMs Forget Knowledge? An Experimental Study with Practical Implications
Scott Hoang, Minsik Cho, Thomas Merth, Atlas Wang, Mohammad Rastegari, Devang Naik
Do LLMs Estimate Uncertainty Well in Instruction-Following?
Juyeon Heo, Miao Xiong, Christina Heinze-Deml, Jaya Narain
Do LLMs Internally "Know" When They Follow Instructions?
Juyeon Heo, Christina Heinze-Deml, Shirley Ren, Oussama Elachqar, Udhay Nallasamy, Andy Miller, Jaya Narain
Dueling in the Dark: An Efficient and Optimal O(√T) Mirror Descent Approach for Competing against Adversarial Preferences ◊
Aadi Saha, Yonathan Efroni, Barry Theobald
Duo-LLMs: A Framework for Studying Adaptive Computation in Large Language Models
Keivan Alizadeh Vahid, Iman Mirzadeh, Mohammad Sekhavat, Minsik Cho, Dmitry Belenko, Frank Sun, Hooman Shahrokhi, Moin Nabi, Mehrdad Farajtabar
Efficient and Effective Uncertainty Quantification in LLMs ◊
Miao Xiong, Andrea Santilli, Michael Kirchhof, Adam Golinski, Sinead Williamson
Enhancing JEPAs with Spatial Conditioning: Robust and Efficient Representation Learning
Etai Littwin, Vimal Thilak, Anand Gopalakrishnan
Evaluating Gender Bias Transfer between Pre-trained and Prompt-Adapted Language Models
Niv Sivakumar, Natalie Mackraz, Samira Khorshidi, Krishna Patel, Barry Theobald, Luca Zappella, Nick Apostoloff
Fairness Dynamics During Training ◊
Krishna Patel, Niv Sivakumar, Barry Theobald, Luca Zappella, Nick Apostoloff
Learning Functions on Symmetric Matrices and Point Clouds via Lightweight Invariant Features ◊
Ben Blum-Smith, Teresa Huang, Marco Cuturi, Soledad Villar
Leveraging Periodicity for Robustness with Multi-Modal Mood Pattern Models
Jaya Narain, Jenny Sun, Oussama Elachqar, Haraldur Hallgrimsson, Feng Zhu, Shirley Ren
Memory Retaining Finetuning via Distillation
Zitong Yang, Aonan Zhang, Sam Wiseman, Xiang Kong, Ke Ye, Dong Yin
Momentum Approximation in Asynchronous Private Federated Learning
Tao Yu, Congzheng Song, Jianyu Wang, Mona Chitnis
On a Spurious Interaction Between Uncertainty Scores and Answer Evaluation Metrics in Generative QA Tasks ◊
Andrea Santilli, Miao Xiong, Michael Kirchhof, Pau Rodriguez Lopez, Federico Danieli, Xavier Suau Cuadros, Luca Zappella, Sinead Williamson, Adam Golinski
Ching Fang, Chris Sandino, Behrooz Mahasseni, Juri Minxha, Hadi Pour Ansari, Erdrin Azemi, Ali Moin, Ellen Zippi
SALSA: Soup-Based Alignment Learning for Stronger Adaptation in RLHF ◊
Atoosa Malemir Chegini, Hamid Kazemi, Iman Mirzadeh, Dong Yin, Max Horton, Moin Nabi, Mehrdad Farajtabar, Keivan Alizadeh Vahid
Scaling Smart: Accelerating Large Language Model Pre-Training with Small Model Initialization
Mohammad Samragh Razlighi, Iman Mirzadeh, Keivan Alizadeh Vahid, Fartash Faghri, Minsik Cho, Moin Nabi, Devang Naik, Mehrdad Farajtabar
TiC-LM: A Multi-Year Benchmark for Continual Pretraining of Language Models ◊
Jeffrey Li, Mohammadreza Armandpour, Iman Mirzadeh, Sachin Mehta, Vaishaal Shankar, Raviteja Vemulapalli, Oncel Tuzel, Mehrdad Farajtabar, Hadi Pour Ansari, Fartash Faghri
Towards Time-Series Reasoning with LLMs
Winnie Chow, Lauren Gardiner, Haraldur Hallgrimsson, Maxwell A. Xu, Shirley Ren
Towards Data-Centric RLHF: Simple Metrics for Preference Dataset Comparison
Judy Hanwen Shen, Archit Sharma, Jun Qin
Towards Low-Bit Communication for Tensor Parallel LLM Inference
Harry Dong, Tyler Johnson, Minsik Cho, Emad Soroush
Understanding Compute-Parameter Trade-offs in Sparse Mixture-of-Expert Language Models ◊
Harshay Shah, Samira Abnar, Vimal Thilak, Dan Busbridge, Alaaeldin Mohamed Elnouby Ali, Josh Susskind
Acknowledgements
Samy Bengio is a Board Member.
Kunal Talwar, Marco Cuturi, Pierre Ablin, Samy Bengio, and Sinead Williamson are Senior Area Chairs.
Aadirupa Saha, Byeongjoo Ahn, Natalie Schluter, Navdeep Jaitly, Oncel Tuzel, Pau Rodriguez Lopez, Preetum Nakkiran, Shams Azam, Tatiana Likhomanenko, and Yizhe Zhang are Area Chairs.
Audra McMillan is an Ethics Reviewer.
Arno Blaas, Dapeng Hu, Enrico Fini, Harsh Sharma, Josh Gardner, Louis Béthune, Maartje ter Hoeve, Miguel Sarabia, Mohammad Sekhavat, Niv Sivakumar, Pau Rodriguez Lopez, Ramprasaath Ramasamy Selvaraju, Richard Bai, TT Guo, Vimal Thilak and Yuyang Wang are Conference Reviewers.
Antoine Wehenkel is a co-organizer of the Machine Learning and the Physical Sciences Workshop.
Arno Blaas, Pau Rodriguez Lopez, Rin Metcalf Susa, and Xavier Suau Cuadros are co-organizers of the Workshop on Mechanistic Interventions (MINT).
Marco Cuturi led the Best Paper Award committee for the main track.
Samira Abnar and Vimal Thilak are Workshop Reviewers.
Related readings and updates.
Empirical Methods in Natural Language Processing (EMNLP) 2024
Apple is presenting new research at the Empirical Methods in Natural Language Processing (EMNLP) conference, which takes place in person in Miami, Florida, from November 12 - 16. We are proud to again sponsor the conference, which brings together the scientific and industrial research communities around natural language processing and artificial intelligence. Below is an overview of Apple’s participation at EMNLP 2024.
International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2024
Apple sponsored the International Conference on Acoustics, Speech and Signal Processing (ICASSP), which took place in person from April 14 to 19 in Seoul, South Korea. ICASSP is the IEEE Signal Processing Society's flagship conference on signal processing and its applications.