News
Entertainment
Science & Technology
Sport
Business & Money
Life
Culture & Art
Hobbies
8 | Follower
Apple Machine Learning Research
23.11.2024
*Equal Contributors A dominant paradigm in large multimodal models is to pair a large language de- coder with a vision encoder. While it is…
This paper was accepted at the Machine Learning and Compression Workshop at NeurIPS 2024. Compressing Large Language Models (LLMs) often…
22.11.2024
This paper was accepted at the Safe Generative AI Workshop (SGAIW) at NeurIPS 2024. Large language models (LLMs) could be valuable personal…
This paper was accepted at the Foundation Model Interventions (MINT) Workshop at NeurIPS 2024. Instruction-following is crucial for building…
We study private stochastic convex optimization (SCO) under user-level differential privacy (DP) constraints. In this setting, there are nnn…
Estimating the density of a distribution from samples is a fundamental problem in statistics. In many practical settings, the Wasserstein…
Learning with identical train and test distributions has been extensively investigated both practically and theoretically. Much remains to…
We study the problem of differentially private stochastic convex optimization (DP-SCO) with heavy-tailed gradients, where we assume a…
We study the problem of private online learning, specifically, online prediction from experts (OPE) and online convex optimization (OCO). We…
21.11.2024
Large language models (LLMs) are commonly trained on datasets consisting of fixed-length token sequences. These datasets are created by…
20.11.2024
This paper was accepted at the Self-Supervised Learning - Theory and Practice (SSLTP) Workshop at NeurIPS 2024. Image-based Joint-Embedding…
This paper considers the learning of logical (Boolean) functions with a focus on the generalization on the unseen (GOTU) setting, a strong…
This paper was accepted at the Efficient Natural Language and Speech Processing (ENLSP) Workshop at NeurIPS 2024. Tensor parallelism…
19.11.2024
This paper was accepted at the Efficient Natural Language and Speech Processing (ENLSP) Workshop at NeurIPS 2024. Large Language Models…
We present Recurrent Drafter (ReDrafter), an advanced speculative decoding approach that achieves state-of-the-art speedup for large…
13.11.2024
This paper was accepted at the Efficient Natural Language and Speech Processing (ENLSP) Workshop at NeurIPS 2024. The pre-training phase of…
07.11.2024
Neural contextual biasing allows speech recognition models to leverage contextually relevant information, leading to improved transcription…
05.11.2024
Apple is presenting new research at the Empirical Methods in Natural Language Processing (EMNLP) conference, which takes place in person in…
This paper was accepted at the Adaptive Foundation Models (AFM) workshop at NeurIPS Workshop 2024. Follow-up conversations with virtual…
Large pretrained vision-language models like CLIP have shown promising generalization capability, but may struggle in specialized domains…
02.11.2024
Many app developers are interested in building on device experiences that integrate increasingly capable large language models (LLMs)…
01.11.2024
This paper was accepted at the Efficient Natural Language and Speech Processing (ENLSP) workshop at NeurIPS Workshop 2024. While large…
31.10.2024
Translating text that contains entity names is a challenging task, as cultural-related references can vary significantly across languages…
30.10.2024
The rapid evolution of Large Language Models (LLMs) and conversational assistants necessitates dynamic, scalable, and configurable…
29.10.2024
Many healthcare applications are inherently multimodal, involving several physiological signals. As sensors for these signals become more…
27.10.2024
Manufacturing quality audits are pivotal for ensuring high product standards in mass production environments. Traditional auditing…
26.10.2024
Recent methods have demonstrated that Large Language Models (LLMs) can solve reasoning tasks better when they are encouraged to solve…
25.10.2024
At Apple, we believe privacy is a fundamental human right. Our work to protect user privacy is informed by a set of privacy principles, and…
24.10.2024
Large Language Models (LLMs) are regularly updated to enhance performance, typically through changes in data or architecture. Within the…
Pretraining robust vision or multimodal foundation models (e.g., CLIP) relies on large-scale datasets that may be noisy, potentially…
The goal of aligning language models to human preferences requires data that reveal these preferences. Ideally, time and money can be spent…
22.10.2024
The growing demand for personalized and private on-device applications highlights the importance of source-free unsupervised domain…
19.10.2024
*Equal Contributors Current multimodal and multitask foundation models like 4M or UnifiedIO show promising results, but in practice their…
17.10.2024
This paper presents Wally, a private search system that supports efficient semantic and keyword search queries against large databases. When…
Time series data are inherently functions of time, yet current transformers often learn time series by modeling them as mere concatenations…
Optimal transport (OT) has profoundly impacted machine learning by providing theoretical and computational tools to realign datasets. In…
While server-side Large Language Models (LLMs) demonstrate proficiency in tool integration and complex reasoning, deploying Small Language…
12.10.2024
Recent advancements in Large Language Models (LLMs) have sparked interest in their formal reasoning capabilities, particularly in…
Hand gesture recognition is becoming a more prevalent mode of human-computer interaction, especially as cameras proliferate across everyday…
10.10.2024
Reinforcement Learning from Human Feedback (RLHF) is an effective approach for aligning language models to human preferences. Central to…