<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://proceedings.mlr.press/v267/feed.xml" rel="self" type="application/atom+xml" /><link href="https://proceedings.mlr.press/v267/" rel="alternate" type="text/html" /><updated>2026-01-05T18:53:52+00:00</updated><id>https://proceedings.mlr.press/v267/feed.xml</id><title type="html">Proceedings of Machine Learning Research</title><subtitle>Proceedings of the 42nd International Conference on Machine Learning
  Held in Vancouver Convention Center, Vancouver, Canada on 13-19 July 2025

Published as Volume 267 by the Proceedings of Machine Learning Research on 06 October 2025.

Volume Edited by:
  Aarti Singh
  Maryam Fazel
  Daniel Hsu
  Simon Lacoste-Julien
  Felix Berkenkamp
  Tegan Maharaj
  Kiri Wagstaff
  Jerry Zhu

Series Editors:
  Neil D. Lawrence
</subtitle><author><name>PMLR</name></author><entry><title type="html">Aggregation of Dependent Expert Distributions in Multimodal Variational Autoencoders</title><link href="https://proceedings.mlr.press/v267/a-mancisidor25a.html" rel="alternate" type="text/html" title="Aggregation of Dependent Expert Distributions in Multimodal Variational Autoencoders" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/a-mancisidor25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/a-mancisidor25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Rogelio&quot;, &quot;family&quot;=&gt;&quot;A. Mancisidor&quot;}, {&quot;given&quot;=&gt;&quot;Robert&quot;, &quot;family&quot;=&gt;&quot;Jenssen&quot;}, {&quot;given&quot;=&gt;&quot;Shujian&quot;, &quot;family&quot;=&gt;&quot;Yu&quot;}, {&quot;given&quot;=&gt;&quot;Michael&quot;, &quot;family&quot;=&gt;&quot;Kampffmeyer&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Lightweight Protocols for Distributed Private Quantile Estimation</title><link href="https://proceedings.mlr.press/v267/aamand25a.html" rel="alternate" type="text/html" title="Lightweight Protocols for Distributed Private Quantile Estimation" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/aamand25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/aamand25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Anders&quot;, &quot;family&quot;=&gt;&quot;Aamand&quot;}, {&quot;given&quot;=&gt;&quot;Fabrizio&quot;, &quot;family&quot;=&gt;&quot;Boninsegna&quot;}, {&quot;given&quot;=&gt;&quot;Abigail&quot;, &quot;family&quot;=&gt;&quot;Gentle&quot;}, {&quot;given&quot;=&gt;&quot;Jacob&quot;, &quot;family&quot;=&gt;&quot;Imola&quot;}, {&quot;given&quot;=&gt;&quot;Rasmus&quot;, &quot;family&quot;=&gt;&quot;Pagh&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Breaking the $n^1.5$ Additive Error Barrier for Private and Efficient Graph Sparsification via Private Expander Decomposition</title><link href="https://proceedings.mlr.press/v267/aamand25b.html" rel="alternate" type="text/html" title="Breaking the $n^1.5$ Additive Error Barrier for Private and Efficient Graph Sparsification via Private Expander Decomposition" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/aamand25b</id><content type="html" xml:base="https://proceedings.mlr.press/v267/aamand25b.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Anders&quot;, &quot;family&quot;=&gt;&quot;Aamand&quot;}, {&quot;given&quot;=&gt;&quot;Justin Y.&quot;, &quot;family&quot;=&gt;&quot;Chen&quot;}, {&quot;given&quot;=&gt;&quot;Mina&quot;, &quot;family&quot;=&gt;&quot;Dalirrooyfard&quot;}, {&quot;given&quot;=&gt;&quot;Slobodan&quot;, &quot;family&quot;=&gt;&quot;Mitrović&quot;}, {&quot;given&quot;=&gt;&quot;Yuriy&quot;, &quot;family&quot;=&gt;&quot;Nevmyvaka&quot;}, {&quot;given&quot;=&gt;&quot;Sandeep&quot;, &quot;family&quot;=&gt;&quot;Silwal&quot;}, {&quot;given&quot;=&gt;&quot;Yinzhan&quot;, &quot;family&quot;=&gt;&quot;Xu&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Improved Approximations for Hard Graph Problems using Predictions</title><link href="https://proceedings.mlr.press/v267/aamand25c.html" rel="alternate" type="text/html" title="Improved Approximations for Hard Graph Problems using Predictions" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/aamand25c</id><content type="html" xml:base="https://proceedings.mlr.press/v267/aamand25c.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Anders&quot;, &quot;family&quot;=&gt;&quot;Aamand&quot;}, {&quot;given&quot;=&gt;&quot;Justin Y.&quot;, &quot;family&quot;=&gt;&quot;Chen&quot;}, {&quot;given&quot;=&gt;&quot;Siddharth&quot;, &quot;family&quot;=&gt;&quot;Gollapudi&quot;}, {&quot;given&quot;=&gt;&quot;Sandeep&quot;, &quot;family&quot;=&gt;&quot;Silwal&quot;}, {&quot;given&quot;=&gt;&quot;Hao&quot;, &quot;family&quot;=&gt;&quot;Wu&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Graph Neural Network Generalization With Gaussian Mixture Model Based Augmentation</title><link href="https://proceedings.mlr.press/v267/abbahaddou25a.html" rel="alternate" type="text/html" title="Graph Neural Network Generalization With Gaussian Mixture Model Based Augmentation" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/abbahaddou25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/abbahaddou25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Yassine&quot;, &quot;family&quot;=&gt;&quot;Abbahaddou&quot;}, {&quot;given&quot;=&gt;&quot;Fragkiskos D.&quot;, &quot;family&quot;=&gt;&quot;Malliaros&quot;}, {&quot;given&quot;=&gt;&quot;Johannes F.&quot;, &quot;family&quot;=&gt;&quot;Lutzeyer&quot;}, {&quot;given&quot;=&gt;&quot;Amine M.&quot;, &quot;family&quot;=&gt;&quot;Aboussalah&quot;}, {&quot;given&quot;=&gt;&quot;Michalis&quot;, &quot;family&quot;=&gt;&quot;Vazirgiannis&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">LMRL Gym: Benchmarks for Multi-Turn Reinforcement Learning with Language Models</title><link href="https://proceedings.mlr.press/v267/abdulhai25a.html" rel="alternate" type="text/html" title="LMRL Gym: Benchmarks for Multi-Turn Reinforcement Learning with Language Models" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/abdulhai25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/abdulhai25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Marwa&quot;, &quot;family&quot;=&gt;&quot;Abdulhai&quot;}, {&quot;given&quot;=&gt;&quot;Isadora&quot;, &quot;family&quot;=&gt;&quot;White&quot;}, {&quot;given&quot;=&gt;&quot;Charlie Victor&quot;, &quot;family&quot;=&gt;&quot;Snell&quot;}, {&quot;given&quot;=&gt;&quot;Charles&quot;, &quot;family&quot;=&gt;&quot;Sun&quot;}, {&quot;given&quot;=&gt;&quot;Joey&quot;, &quot;family&quot;=&gt;&quot;Hong&quot;}, {&quot;given&quot;=&gt;&quot;Yuexiang&quot;, &quot;family&quot;=&gt;&quot;Zhai&quot;}, {&quot;given&quot;=&gt;&quot;Kelvin&quot;, &quot;family&quot;=&gt;&quot;Xu&quot;}, {&quot;given&quot;=&gt;&quot;Sergey&quot;, &quot;family&quot;=&gt;&quot;Levine&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Task Generalization with Autoregressive Compositional Structure: Can Learning from $D$ Tasks Generalize to $D^T$ Tasks?</title><link href="https://proceedings.mlr.press/v267/abedsoltan25a.html" rel="alternate" type="text/html" title="Task Generalization with Autoregressive Compositional Structure: Can Learning from $D$ Tasks Generalize to $D^T$ Tasks?" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/abedsoltan25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/abedsoltan25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Amirhesam&quot;, &quot;family&quot;=&gt;&quot;Abedsoltan&quot;}, {&quot;given&quot;=&gt;&quot;Huaqing&quot;, &quot;family&quot;=&gt;&quot;Zhang&quot;}, {&quot;given&quot;=&gt;&quot;Kaiyue&quot;, &quot;family&quot;=&gt;&quot;Wen&quot;}, {&quot;given&quot;=&gt;&quot;Hongzhou&quot;, &quot;family&quot;=&gt;&quot;Lin&quot;}, {&quot;given&quot;=&gt;&quot;Jingzhao&quot;, &quot;family&quot;=&gt;&quot;Zhang&quot;}, {&quot;given&quot;=&gt;&quot;Mikhail&quot;, &quot;family&quot;=&gt;&quot;Belkin&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Consensus Is All You Get: The Role of Attention in Transformers</title><link href="https://proceedings.mlr.press/v267/abella25a.html" rel="alternate" type="text/html" title="Consensus Is All You Get: The Role of Attention in Transformers" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/abella25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/abella25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Álvaro Rodrı́guez&quot;, &quot;family&quot;=&gt;&quot;Abella&quot;}, {&quot;given&quot;=&gt;&quot;João Pedro&quot;, &quot;family&quot;=&gt;&quot;Silvestre&quot;}, {&quot;given&quot;=&gt;&quot;Paulo&quot;, &quot;family&quot;=&gt;&quot;Tabuada&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Soup-of-Experts: Pretraining Specialist Models via Parameters Averaging</title><link href="https://proceedings.mlr.press/v267/ablin25a.html" rel="alternate" type="text/html" title="Soup-of-Experts: Pretraining Specialist Models via Parameters Averaging" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/ablin25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/ablin25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Pierre&quot;, &quot;family&quot;=&gt;&quot;Ablin&quot;}, {&quot;given&quot;=&gt;&quot;Angelos&quot;, &quot;family&quot;=&gt;&quot;Katharopoulos&quot;}, {&quot;given&quot;=&gt;&quot;Skyler&quot;, &quot;family&quot;=&gt;&quot;Seto&quot;}, {&quot;given&quot;=&gt;&quot;David&quot;, &quot;family&quot;=&gt;&quot;Grangier&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry><entry><title type="html">Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models</title><link href="https://proceedings.mlr.press/v267/abnar25a.html" rel="alternate" type="text/html" title="Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models" /><published>2025-10-06T00:00:00+00:00</published><updated>2025-10-06T00:00:00+00:00</updated><id>https://proceedings.mlr.press/v267/abnar25a</id><content type="html" xml:base="https://proceedings.mlr.press/v267/abnar25a.html"><![CDATA[]]></content><author><name>[{&quot;given&quot;=&gt;&quot;Samira&quot;, &quot;family&quot;=&gt;&quot;Abnar&quot;}, {&quot;given&quot;=&gt;&quot;Harshay&quot;, &quot;family&quot;=&gt;&quot;Shah&quot;}, {&quot;given&quot;=&gt;&quot;Dan&quot;, &quot;family&quot;=&gt;&quot;Busbridge&quot;}, {&quot;given&quot;=&gt;&quot;Alaaeldin&quot;, &quot;family&quot;=&gt;&quot;El-Nouby&quot;}, {&quot;given&quot;=&gt;&quot;Joshua M.&quot;, &quot;family&quot;=&gt;&quot;Susskind&quot;}, {&quot;given&quot;=&gt;&quot;Vimal&quot;, &quot;family&quot;=&gt;&quot;Thilak&quot;}]</name></author><summary type="html"><![CDATA[]]></summary></entry></feed>