Machine Learning
Research at Apple

Research highlights

A recent paper from Apple researchers, “The Super Weight in Large Language Models,” reveals that an extremely small subset of parameters in LLMs (in some cases, a single parameter) can exert a disproportionate influence on an LLM’s overall functionality (see Figure 1). This work highlights the critical role of these “super weights” and their corresponding “super activations,”…

Read more

Vision Language Models (VLMs) enable visual understanding alongside textual inputs. They are typically built by passing visual tokens from a pretrained vision encoder to a pretrained Large Language Model (LLM) through a projection layer. By leveraging the rich visual representations of the vision encoder and the world knowledge and reasoning capabilities of the LLM, VLMs can be useful for a wide range of applications, including accessibility…

Read more

Recent publications

AuthorsWenhui Cui**†, Christopher Sandino, Hadi Pouransar, Ran Liu, Juri Minxha, Ellen Zippi, Aman Verma, Anna Sedlackova, Erdrin Azemi, Behrooz Mahasseni
Read paper
AuthorsJiayi Pan*, Xingyao Wang*, Graham Neubig†, Navdeep Jaitly‡, Heng Ji§, Alane Suhr†, Yizhe Zhang‡
Read paper

Explore all publications

Apple is presenting new work at the biennial International Conference on Computer Vision (ICCV), which takes place in person from October 19 to 23, in Honolulu, Hawai’i. The conference alternates each year with the European Conference on Computer Vision (ECCV), and focuses on important topics the field of computer vision.

Read more

Natural language processing (NLP) remains one of the most quickly evolving fields in AI, as new research continues to rapidly advance large language models (LLMs), systems for speech recognition and generation, language agents, and more. This technology is essential to many of today’s AI experiences, including Apple Intelligence and Siri, and fundamental research in NLP will be foundational to future AI.

Read more