{"id":4394,"date":"2025-07-10T07:10:22","date_gmt":"2025-07-10T07:10:22","guid":{"rendered":"https:\/\/techtrendfeed.com\/?p=4394"},"modified":"2025-07-10T07:10:22","modified_gmt":"2025-07-10T07:10:22","slug":"carnegie-mellon-college-at-icml-2025-machine-studying-weblog-mlcmu","status":"publish","type":"post","link":"https:\/\/techtrendfeed.com\/?p=4394","title":{"rendered":"Carnegie Mellon College at ICML 2025 \u2013 Machine Studying Weblog | ML@CMU"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p>CMU researchers are presenting 127 papers on the Forty-Second Worldwide Convention on Machine Studying (ICML 2025), held from July Thirteenth-Nineteenth on the Vancouver Conference Middle. Here&#8217;s a fast overview of the areas our researchers are engaged on:<\/p>\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"1024\" height=\"623\" src=\"https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1024x623.png\" alt=\"\" class=\"wp-image-21219\" srcset=\"https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1024x623.png 1024w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-300x183.png 300w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1536x935.png 1536w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-970x590.png 970w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-320x195.png 320w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-80x49.png 80w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image.png 1962w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-300x183@2x.png 600w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\"\/><\/figure>\n<p>Listed below are our most frequent collaborator establishments:<\/p>\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"1024\" height=\"602\" src=\"https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-1024x602.png\" alt=\"\" class=\"wp-image-21220\" srcset=\"https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-1024x602.png 1024w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-300x176.png 300w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-1536x902.png 1536w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-970x570.png 970w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-320x188.png 320w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-80x47.png 80w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1.png 1954w, https:\/\/blog.ml.cmu.edu\/wp-content\/uploads\/2025\/06\/image-1-300x176@2x.png 600w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\"\/><\/figure>\n<p><meta charset=\"UTF-8\"\/><br \/>\n    <meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0\"\/><\/p>\n<h2 id=\"oral-papers\">Oral Papers<\/h2>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=LCbHsdtvOR\" target=\"_blank\" rel=\"noopener\">Anticipated Variational Inequalities<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Brian Zhang, Ioannis Anagnostides, Emanuel Tewolde, Ratip Emin Berker, Gabriele Farina, Vincent Conitzer, Tuomas Sandholm<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces anticipated variational inequalities (EVIs), a relaxed model of variational inequalities (VIs) the place the objective is to discover a distribution that satisfies the VI situation in expectation. Whereas VIs are typically exhausting to resolve, the authors present that EVIs may be solved effectively, even underneath difficult, non-monotone circumstances, by leveraging concepts from sport principle. EVIs generalize the idea of correlated equilibria and unify varied outcomes throughout easy video games, constrained video games, and settings with non-concave utilities, making them broadly relevant past conventional game-theoretic contexts.\n<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=zf9zwCRKyP\" target=\"_blank\" rel=\"noopener\">Exploring and Mitigating Adversarial Manipulation of Voting-Primarily based Leaderboards<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Yangsibo Huang, Milad Nasr, Anastasios Angelopoulos, Nicholas Carlini, Wei-Lin Chiang, Christopher A. Choquette Choo, Daphne Ippolito, Matthew Jagielski, Katherine Lee, Ken Ziyu Liu, Ion Stoica, Florian Tramer, Chiyuan Zhang<\/p>\n<p class=\"oral-spotlight-space\">This paper reveals that voting-based benchmarks for evaluating LLMs (akin to Chatbot Area) may be susceptible to adversarial manipulation if correct defenses aren\u2019t in place. The authors present that an attacker can determine which mannequin generated a response after which strategically vote to spice up or demote particular fashions, altering the leaderboard with solely round a thousand votes in a simulated surroundings. They collaborate with Chatbot Area\u2019s builders to suggest and implement safety measures akin to reCAPTCHA and login necessities that considerably increase the price of such assaults and improve the platform\u2019s robustness.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=uRAgIVnAO6\" target=\"_blank\" rel=\"noopener\">Excessive-Dimensional Prediction for Sequential Determination Making<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Georgy Noarov, Ramya Ramalingam, Aaron Roth, Stephan Xie<\/p>\n<p class=\"oral-spotlight-space\">This paper presents a brand new algorithmic framework for making dependable, multi-dimensional forecasts in adversarial, nonstationary environments. In contrast to present on-line studying strategies, this strategy gives simultaneous efficiency ensures for a lot of brokers, even after they face totally different goals, act over giant motion areas, or care about particular circumstances (e.g. climate or route alternative). The algorithm ensures low bias throughout many conditional occasions and allows every agent to attain sturdy ensures like diminishing remorse. Functions embody environment friendly options for on-line combinatorial optimization and multicalibration.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=SyQPiZJVWY\" target=\"_blank\" rel=\"noopener\">LLM-SRBench: A New Benchmark for Scientific Equation Discovery with Massive Language Fashions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Parshin Shojaee, Ngoc Hieu Nguyen, Kazem Meidani, Amir Barati Farimani, Khoa Doan, Chandan Reddy<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces LLM-SRBench, a brand new benchmark designed to carefully consider the flexibility of LLMs to find scientific equations (moderately than merely recall them from coaching knowledge). Present assessments typically depend on well-known equations, making it exhausting to inform whether or not fashions are actually reasoning or simply memorizing. LLM-SRBench addresses this by together with 239 difficult issues throughout 4 scientific domains, cut up into two classes: one which disguises acquainted physics equations (LSR-Remodel) and one other that options absolutely artificial, reasoning-driven duties (LSR-Synth). Evaluations present that even one of the best present fashions solely obtain 31.5% accuracy, highlighting the problem of the duty and establishing LLM-SRBench as a priceless software for driving progress in LLM-based scientific discovery.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=kEn7Wt6Yj2\" target=\"_blank\" rel=\"noopener\">On Differential Privateness for Adaptively Fixing Search Issues through Sketching<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Shiyuan Feng, Ying Feng, George Li, Zhao Music, David Woodruff, Lichen Zhang<\/p>\n<p class=\"oral-spotlight-space\">This paper explores  use differential privateness to guard towards data leakage in adaptive search queries, a tougher downside than conventional non-public estimation duties. In contrast to prior work that solely returns numerical summaries (e.g., value), the authors design algorithms that return precise options, like nearest neighbors or regression vectors, even when the inputs or queries change over time. They present how key downside parameters (just like the variety of approximate close to neighbors or situation variety of the info matrix) have an effect on the efficiency of those non-public algorithms. This work has sensible implications for AI programs that depend on non-public database searches or real-time regression, enabling them to offer helpful outcomes whereas safeguarding delicate data from attackers.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=Hi0SyHMmkd\" target=\"_blank\" rel=\"noopener\">Roll the cube &amp; look earlier than you leap: Going past the artistic limits of next-token prediction<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Vaishnavh Nagarajan, Chen Wu, Charles Ding, Aditi Raghunathan<\/p>\n<p class=\"oral-spotlight-space\">This paper proposes a set of straightforward, summary duties designed to probe the artistic limits of as we speak\u2019s language fashions in a managed and measurable approach. These duties mimic real-world open-ended challenges like producing analogies or designing puzzles, the place success requires discovering new connections or establishing novel patterns. The authors present that commonplace next-token prediction tends to be short-sighted and overly reliant on memorization, whereas different approaches like teacherless coaching and diffusion fashions produce extra numerous, unique outputs. Additionally they introduce a method referred to as seed-conditioning, which provides randomness on the enter moderately than the output and may enhance coherence with out sacrificing creativity.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=UeB3Hdrhda\" target=\"_blank\" rel=\"noopener\">Coaching a Usually Curious Agent<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Fahim Tajwar, Yiding Jiang, Abitha Thankaraj, Sumaita Rahman, Zico Kolter, Jeff Schneider, Russ Salakhutdinov<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces Paprika, a fine-tuning methodology that equips language fashions with common decision-making and exploration methods, enabling them to adapt to new duties by means of interplay alone (i.e. with out additional coaching). Paprika trains fashions on artificial environments requiring totally different exploration behaviors, encouraging them to study versatile methods moderately than memorizing options. To enhance effectivity, it makes use of a curriculum learning-based strategy that prioritizes duties with excessive studying worth, taking advantage of restricted interplay knowledge. Fashions educated with Paprika present sturdy switch to utterly new duties, suggesting a promising route for constructing AI brokers that may study to resolve unfamiliar, sequential issues with minimal supervision.<\/p>\n<\/p><\/div>\n<h2 id=\"spotlight-papers\">Highlight Papers<\/h2>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=u6xeKVHS6K\" target=\"_blank\" rel=\"noopener\">GMAIL: Generative Modality Alignment for generated Picture Studying<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Shentong Mo, Sukmin Yun<\/p>\n<p class=\"oral-spotlight-space\">Generative fashions can create lifelike photographs that would assist prepare machine studying fashions, however utilizing them as in the event that they had been actual photographs can result in issues due to variations between the 2. This paper introduces a way referred to as GMAIL that treats actual and generated photographs as separate varieties (or modalities) and aligns them in a shared latent area throughout coaching, moderately than simply mixing them on the pixel stage. The strategy fine-tunes fashions on generated knowledge utilizing a particular loss to bridge the hole, then makes use of these aligned fashions to enhance coaching on duties like picture captioning and retrieval. The outcomes present that GMAIL improves efficiency on a number of vision-language duties and scales nicely as extra generated knowledge is added.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=FKi6yjXwCN\" target=\"_blank\" rel=\"noopener\">LOCATE 3D: Actual-World Object Localization through Self-Supervised Studying in 3D<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Paul McVay, Sergio Arnaud, Ada Martin, Arjun Majumdar, Krishna Murthy Jatavallabhula, Phillip Thomas, Ruslan Partsey, Daniel Dugas, Abha Gejji, Alexander Sax, Vincent-Pierre Berges, Mikael Henaff, Ayush Jain, Ang Cao, Ishita Prasad, Mrinal Kalakrishnan, Michael Rabbat, Nicolas Ballas, Mahmoud Assran, Oleksandr Maksymets, Aravind Rajeswaran, Franziska Meier<\/p>\n<p class=\"oral-spotlight-space\">LOCATE 3D is a mannequin that may discover particular objects in 3D scenes primarily based on pure language descriptions (like \u201cthe small espresso desk between the couch and the lamp\u201d). It achieves state-of-the-art efficiency on commonplace benchmarks and works nicely in real-world settings, like on robots or AR gadgets, through the use of RGB-D sensor knowledge. A key element is 3D-JEPA, a brand new self-supervised studying methodology that makes use of options from 2D imaginative and prescient fashions (like CLIP or DINO) to know 3D level clouds by means of masked prediction duties. The mannequin is educated on a newly launched giant dataset (130K+ examples), serving to it generalize higher throughout totally different environments. <\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=dzwUOiBlQW\" target=\"_blank\" rel=\"noopener\">Masked Autoencoders Are Efficient Tokenizers for Diffusion Fashions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Hao Chen, Yujin Han, Fangyi Chen, Xiang Li, Yidong Wang, Jindong Wang, Ze Wang, Zicheng Liu, Difan Zou, Bhiksha Raj<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces MAETok, a masked autoencoder designed to create a high-quality, semantically significant latent area for diffusion fashions. The authors present that having a well-structured latent area, that means fewer Gaussian modes and extra discriminative options, results in higher picture technology with no need advanced variational autoencoders. MAETok outperforms present strategies on ImageNet utilizing simply 128 tokens, and it\u2019s additionally a lot quicker: 76\u00d7 faster to coach and 31\u00d7 quicker throughout inference. The important thing takeaway is that the construction of the latent area, not variational constraints, is what actually issues for high-quality diffusion-based technology.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=ACzL62Jp4E\" target=\"_blank\" rel=\"noopener\">Place: In-Home Analysis Is Not Sufficient. In direction of Sturdy Third-Occasion Analysis and Flaw Disclosure for Common-Function AI<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Shayne Longpre, Kevin Klyman, Ruth Elisabeth Appel, Sayash Kapoor, Rishi Bommasani, Michelle Sahar, Sean McGregor, Avijit Ghosh, Borhane Blili-Hamelin, Nathan Butters, Alondra Nelson, Amit Elazari, Andrew Sellars, Casey Ellis, Dane Sherrets, Daybreak Music, Harley Geiger, Ilona Cohen, Lauren McIlvenny, Madhulika Srikumar, Mark Jaycox, Markus Anderljung, Nadine Johnson, Nicholas Carlini, Nicolas Miailhe, Nik Marda, Peter Henderson, Rebecca Portnoff, Rebecca Weiss, Victoria Westerhoff, Yacine Jernite, Rumman Chowdhury, Percy Liang, Arvind Narayanan<\/p>\n<p class=\"oral-spotlight-space\">This paper highlights the shortage of strong programs for figuring out and reporting flaws in general-purpose AI (GPAI), particularly in comparison with mature fields like software program safety. The authors suggest three key options: (1) standardized reporting codecs and engagement guidelines to streamline flaw reporting and triaging, (2) formal disclosure applications with authorized protections for researchers (much like bug bounties), and (3) higher infrastructure for distributing flaw stories to related stakeholders. These steps goal to deal with rising dangers like jailbreaks and cross-system vulnerabilities, finally bettering the protection and accountability of GPAI programs.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=beeNgQEfe2\" target=\"_blank\" rel=\"noopener\">Scaling Take a look at-Time Compute With out Verification or RL is Suboptimal<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Amrith Setlur, Nived Rajaraman, Sergey Levine, Aviral Kumar<\/p>\n<p class=\"oral-spotlight-space\">This paper explores  finest scale test-time compute for giant language fashions (LLMs), evaluating two methods: (1) distilling search traces (verifier-free, or VF) and (2) utilizing verifiers or rewards to information studying (verifier-based, or VB). The authors present\u2014each theoretically and thru experiments\u2014that VB strategies considerably outperform VF ones when working with restricted compute or knowledge. They clarify that this efficiency hole grows as fashions and duties get extra advanced, particularly when resolution paths fluctuate in model or high quality. In the end, the paper argues that verification is important for successfully scaling LLM efficiency, particularly for reasoning duties.<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=oa7MYAO6h6\" target=\"_blank\" rel=\"noopener\">ShadowKV: KV Cache in Shadows for Excessive-Throughput Lengthy-Context LLM Inference<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Hanshi Solar, Li-Wen Chang, Wenlei Bao, Dimension Zheng, Ningxin Zheng, Xin Liu, Harry Dong, Yuejie Chi, Beidi Chen<\/p>\n<p class=\"oral-spotlight-space\">As long-context LLMs grow to be extra widespread, their rising reminiscence calls for throughout inference decelerate efficiency, particularly because of the increasing key-value (KV) cache. This paper introduces ShadowKV, a system that considerably improves throughput by compressing the important thing cache utilizing low-rank representations and offloading the worth cache with out main latency prices. It reconstructs solely the mandatory KV pairs throughout decoding to take care of velocity and accuracy. Experiments present ShadowKV helps a lot bigger batch sizes (as much as 6\u00d7) and improves throughput by over 3\u00d7 on commonplace {hardware}, all whereas preserving mannequin high quality throughout a number of LLMs and benchmarks.<\/p>\n<\/p><\/div>\n<h2 id=\"poster-papers\">Poster Papers<\/h2>\n<h3 id=\"accountability,-transparency,-and-interpretability\">Accountability, Transparency, And Interpretability<\/h3>\n<h3 id=\"active-learning-and-interactive-learning\">Lively Studying And Interactive Studying<\/h3>\n<h3 id=\"applications\">Functions<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=W9s817KqYf\" target=\"_blank\" rel=\"noopener\">Home windows Agent Area: Evaluating Multi-Modal OS Brokers at Scale<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Rogerio Bonatti, Dan Zhao, Francesco Bonacci, Dillon Dupont, Sara Abdali, Yinheng Li, Yadong Lu, Justin Wagle, Kazuhito Koishida, Arthur Bucker, Lawrence Jang, Zheng Hui<\/p>\n<\/p><\/div>\n<h3 id=\"causality\">Causality<\/h3>\n<h3 id=\"chemistry,-physics,-and-earth-sciences\">Chemistry, Physics, And Earth Sciences<\/h3>\n<h3 id=\"computer-vision\">Pc Imaginative and prescient<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=w8MCYYAvQD\" target=\"_blank\" rel=\"noopener\">From Hundreds to Billions: 3D Visible Language Grounding through Render-Supervised Distillation from 2D VLMs<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Ang Cao, Sergio Arnaud, Oleksandr Maksymets, Jianing Yang, Ayush Jain, Ada Martin, Vincent-Pierre Berges, Paul McVay, Ruslan Partsey, Aravind Rajeswaran, Franziska Meier, Justin Johnson, Jeong Joon Park, Alexander Sax<\/p>\n<\/p><\/div>\n<h3 id=\"deep-learning\">Deep Studying<\/h3>\n<h3 id=\"discrete-and-combinatorial-optimization\">Discrete And Combinatorial Optimization<\/h3>\n<h3 id=\"domain-adaptation-and-transfer-learning\">Area Adaptation And Switch Studying<\/h3>\n<h3 id=\"evaluation\">Analysis<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=BeoXADC9PW\" target=\"_blank\" rel=\"noopener\">RBench: Graduate-level Multi-disciplinary Benchmarks for LLM &amp; MLLM Advanced Reasoning Analysis<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Meng-Hao Guo, Jiajun Xu, Yi Zhang, Jiaxi Music, Haoyang Peng, Yi-Xuan Deng, Xinzhi Dong, Kiyohiro Nakayama, Zhengyang Geng, Chen Wang, Bolin Ni, Guo-Wei Yang, Yongming Rao, Houwen Peng, Han Hu, Gordon Wetzstein, Shi-min Hu<\/p>\n<\/p><\/div>\n<h3 id=\"everything-else\">Every part Else<\/h3>\n<h3 id=\"fairness\">Equity<\/h3>\n<h3 id=\"foundation-models\">Basis Fashions<\/h3>\n<h3 id=\"game-theory\">Sport Idea<\/h3>\n<h3 id=\"general-machine-learning\">Common Machine Studying<\/h3>\n<h3 id=\"graph-neural-networks\">Graph Neural Networks<\/h3>\n<h3 id=\"graphical-models\">Graphical Fashions<\/h3>\n<h3 id=\"health-\/-medicine\">Well being \/ Medication<\/h3>\n<h3 id=\"language,-speech-and-dialog\">Language, Speech And Dialog<\/h3>\n<h3 id=\"large-language-models\">Massive Language Fashions<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=EGrSMMj37o\" target=\"_blank\" rel=\"noopener\">An Structure Search Framework for Inference-Time Strategies<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Jon Saad-Falcon, Adrian Lafuente, Shlok Natarajan, Nahum Maru, Hristo Todorov, Etash Guha, Estefany Kelly Buchanan, Mayee Chen, Neel Guha, Christopher Re, Azalia Mirhoseini<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=0ERw2196o1\" target=\"_blank\" rel=\"noopener\">Suppose Smarter not Tougher: Adaptive Reasoning with Inference Conscious Optimization<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Zishun Yu, Tengyu Xu, Di Jin, Karthik Abinav Sankararaman, Yun He, Wenxuan Zhou, Zhouhao Zeng, Eryk Helenowski, Chen Zhu, Sinong Wang, Hao Ma, Han Fang<\/p>\n<\/p><\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=jv7bF50spq\" target=\"_blank\" rel=\"noopener\">Unnatural Languages Are Not Bugs however Options for LLMs<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Keyu Duan, Yiran Zhao, Zhili Feng, Jinjie Ni, Tianyu Pang, Qian Liu, Tianle Cai, Longxu Dou, Kenji Kawaguchi, Anirudh Goyal, Zico Kolter, Michael Shieh<\/p>\n<\/p><\/div>\n<h3 id=\"learning-theory\">Studying Idea<\/h3>\n<h3 id=\"multi-agent\">Multi-agent<\/h3>\n<h3 id=\"online-learning-and-bandits\">On-line Studying And Bandits<\/h3>\n<h3 id=\"online-learning,-active-learning-and-bandits\">On-line Studying, Lively Studying And Bandits<\/h3>\n<h3 id=\"optimization\">Optimization<\/h3>\n<h3 id=\"privacy\">Privateness<\/h3>\n<h3 id=\"probabilistic-methods\">Probabilistic Strategies<\/h3>\n<h3 id=\"reinforcement-learning-and-planning\">Reinforcement Studying And Planning<\/h3>\n<h3 id=\"representation-learning\">Illustration Studying<\/h3>\n<h3 id=\"research-priorities,-methodology,-and-evaluation\">Analysis Priorities, Methodology, And Analysis<\/h3>\n<h3 id=\"robotics\">Robotics<\/h3>\n<h3 id=\"safety\">Security<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=WUGrleBcYP\" target=\"_blank\" rel=\"noopener\">SafetyAnalyst: Interpretable, Clear, and Steerable Security Moderation for AI Habits<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Jing-Jing Li, Valentina Pyatkin, Max Kleiman-Weiner, Liwei Jiang, Nouha Dziri, Anne Collins, Jana Schaich Borg, Maarten Sap, Yejin Choi, Sydney Levine<\/p>\n<\/p><\/div>\n<h3 id=\"security\">Safety<\/h3>\n<h3 id=\"sequential-models,-time-series\">Sequential Fashions, Time Collection<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/openreview.net\/forum?id=B6WalMoQJW\" target=\"_blank\" rel=\"noopener\">Enhancing Basis Fashions for Time Collection Forecasting through Wavelet-based Tokenization<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Luca Masserano, Abdul Fatir Ansari, Boran Han, Xiyuan Zhang, Christos Faloutsos, Michael Mahoney, Andrew Wilson, Youngsuk Park, Syama Sundar Yadav Rangapuram, Danielle Maddix, Yuyang Wang<\/p>\n<\/p><\/div>\n<h3 id=\"social-aspects\">Social Elements<\/h3>\n<h3 id=\"structure-learning\">Construction Studying<\/h3>\n<h3 id=\"supervised-learning\">Supervised Studying<\/h3>\n<h3 id=\"theory\">Idea<\/h3>\n<h3 id=\"time-series\">Time Collection<\/h3>\n<\/p><\/div>\n\n","protected":false},"excerpt":{"rendered":"<p>CMU researchers are presenting 127 papers on the Forty-Second Worldwide Convention on Machine Studying (ICML 2025), held from July Thirteenth-Nineteenth on the Vancouver Conference Middle. Here&#8217;s a fast overview of the areas our researchers are engaged on: Listed below are our most frequent collaborator establishments: Oral Papers Anticipated Variational Inequalities Authors: Brian Zhang, Ioannis Anagnostides, [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":4396,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[55],"tags":[110,1838,3922,136,113,1839,442,1840],"class_list":["post-4394","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-machine-learning","tag-blog","tag-carnegie","tag-icml","tag-learning","tag-machine","tag-mellon","tag-mlcmu","tag-university"],"_links":{"self":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/4394","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=4394"}],"version-history":[{"count":1,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/4394\/revisions"}],"predecessor-version":[{"id":4395,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/4394\/revisions\/4395"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/media\/4396"}],"wp:attachment":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=4394"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=4394"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=4394"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69d9690a190636c2e0989534. Config Timestamp: 2026-04-10 21:18:02 UTC, Cached Timestamp: 2026-04-21 09:57:23 UTC -->