{"id":1862,"date":"2025-04-28T04:07:09","date_gmt":"2025-04-28T04:07:09","guid":{"rendered":"https:\/\/techtrendfeed.com\/?p=1862"},"modified":"2025-04-28T04:07:10","modified_gmt":"2025-04-28T04:07:10","slug":"carnegie-mellon-college-at-iclr-2025-machine-studying-weblog-mlcmu","status":"publish","type":"post","link":"https:\/\/techtrendfeed.com\/?p=1862","title":{"rendered":"Carnegie Mellon College at ICLR 2025 \u2013 Machine Studying Weblog | ML@CMU"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p>CMU researchers are presenting 143 papers on the Thirteenth Worldwide Convention on Studying Representations (ICLR 2025), held from April 24 \u2013 28 on the Singapore EXPO. Here&#8217;s a fast overview of the areas our researchers are engaged on: <\/p>\n<p><strong><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXf9ONxtuHRoHZ3kHak_Jr41zzKrKdvHC1vwJiPg5CSe6sc6FmNL8a7Qw6wEuVtxKAET96KA4fgTdlRPe9prshhkVIGunjERA8lAlAi56687XNlHDLNqM4UQK7455nCXryJOsIcr?key=WqEkyGXgVs9Wcabak7GlSnty\" width=\"624\" height=\"383\"\/><\/strong><\/p>\n<p>And listed below are our most frequent collaborator establishments:<\/p>\n<p><strong><img decoding=\"async\" loading=\"lazy\" width=\"624\" height=\"371\" src=\"https:\/\/lh7-rt.googleusercontent.com\/docsz\/AD_4nXdpm6t2zUDw6t5G41V7xiKsQTN_BDJAS20EH6y6hM10gSKFsmzp1naSHpRttr97jeTu_xCA2SxlPQ7ki8hlhaD8NJT4dGtxrlO3RfJ2DjzEvToM8mwJNHZZKQZ80_DSAydkQ1RMbA?key=WqEkyGXgVs9Wcabak7GlSnty\"\/><\/strong><\/p>\n<div class=\"table-of-contents\">\n<h2>Desk of Contents<\/h2>\n<ul>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#oral-papers\">Oral Papers<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#spotlight-papers\">Highlight Papers<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#poster-papers\">Poster Papers<\/a>\n<ul>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#alignment,-fairness,-safety,-privacy,-and-societal-considerations\">Alignment, Equity, Security, Privateness, And Societal Issues<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#applications-to-computer-vision,-audio,-language,-and-other-modalities\">Functions to Pc Imaginative and prescient, Audio, Language, And Different Modalities<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#applications-to-neuroscience-&amp;-cognitive-science\">Functions to Neuroscience &amp; Cognitive Science<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#applications-to-physical-sciences-(physics,-chemistry,-biology,-etc.)\">Functions to Bodily Sciences (Physics, Chemistry, Biology, And many others.)<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#applications-to-robotics,-autonomy,-planning\">Functions to Robotics, Autonomy, Planning<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#causal-reasoning\">Causal Reasoning<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#datasets-and-benchmarks\">Datasets and Benchmarks<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#foundation-or-frontier-models,-including-llms\">Basis or Frontier Fashions, Together with LLMs<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#generative-models\">Generative Fashions<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#infrastructure,-software-libraries,-hardware,-systems,-etc.\">Infrastructure, Software program Libraries, {Hardware}, Methods, and so forth.<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#interpretability-and-explainable-ai\">Interpretability and Explainable AI<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#learning-on-graphs-and-other-geometries-&amp;-topologies\">Studying on Graphs and Different Geometries &amp; Topologies<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#learning-theory\">Studying Principle<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#neurosymbolic-&amp;-hybrid-ai-systems-(physics-informed,-logic-&amp;-formal-reasoning,-etc.)\">Neurosymbolic &amp; Hybrid AI Methods (Physics-Knowledgeable, Logic &amp; Formal Reasoning, and so forth.)<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#optimization\">Optimization<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#other-topics-in-machine-learning-(i.e.,-none-of-the-above)\">Different Subjects in Machine Studying (i.e., not one of the above)<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#probabilistic-methods-(bayesian-methods,-variational-inference,-sampling,-uq,-etc.)\">Probabilistic Strategies (Bayesian Strategies, Variational Inference, Sampling, Uncertainty Quantification, and so forth.)<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#reinforcement-learning\">Reinforcement Studying<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#transfer-learning,-meta-learning,-and-lifelong-learning\">Switch Studying, Meta Studying, and Lifelong Studying<\/a><\/li>\n<li><a rel=\"nofollow\" target=\"_blank\" href=\"#unsupervised,-self-supervised,-semi-supervised,-and-supervised-representation-learning\">Unsupervised, Self-supervised, Semi-supervised, and Supervised Illustration Studying<\/a><\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<\/div>\n<h2 id=\"oral-papers\">Oral Papers<\/h2>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30547\" target=\"_blank\" rel=\"noopener\">Backtracking Improves Era Security<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Yiming Zhang, Jianfeng Chi, Hailey Nguyen, Kartikeya Upasani, Daniel M. Bikel, Jason E Weston, Eric Michael Smith<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces backtracking, a brand new method that enables language fashions to get well from unsafe textual content era through the use of a particular [RESET] token to \u201cundo\u201d problematic outputs. In contrast to conventional security strategies that goal to forestall dangerous responses outright, backtracking trains the mannequin to self-correct mid-generation. The authors show that backtracking considerably improves security with out sacrificing helpfulness, and it additionally supplies robustness in opposition to a number of adversarial assaults.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29245\" target=\"_blank\" rel=\"noopener\">BigCodeBench: Benchmarking Code Era with Various Operate Calls and Complicated Directions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Terry Yue Zhuo, Vu Minh Chien, Jenny Chim, Han Hu, Wenhao Yu, Ratnadira Widyasari, Imam Nur Bani Yusuf, Haolan Zhan, Junda He, Indraneil Paul, Simon Brunner, Chen Gong, James Hoang, Armel Randy Zebaze, Xiaoheng Hong, Wen-ding Li, Jean Kaddour, Ming Xu, Zhihan Zhang, Prateek Yadav, Naman Jain, Alex Gu, Zhoujun Cheng, Jiawei Liu, Qian Liu, Zijian Wang, David Lo, Binyuan Hui, Niklas Muennighoff, Daniel Fried, Xiaoning Du, Hurt De Vries, Leandro Von Werra<\/p>\n<p class=\"oral-spotlight-space\">Current advances in LLMs have enabled process automation by way of Python code, however current benchmarks primarily concentrate on easy, self-contained duties. To evaluate LLMs\u2019 potential to deal with extra sensible challenges requiring numerous and compositional operate use, the authors introduce BigCodeBench\u2014a benchmark masking 1,140 duties throughout 139 libraries and seven domains. Every process contains rigorous testing with excessive department protection, and a variant, BigCodeBench-Instruct, reformulates directions for pure language analysis. Outcomes from testing 60 LLMs reveal important efficiency gaps, highlighting that present fashions wrestle to observe advanced directions and compose operate calls precisely in comparison with human efficiency.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29609\" target=\"_blank\" rel=\"noopener\">Context-Parametric Inversion: Why Instruction Finetuning Could Not Really Enhance Context Reliance<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Sachin Goyal, Christina Baek, J Zico Kolter, Aditi Raghunathan<\/p>\n<p class=\"oral-spotlight-space\">LLMs are anticipated to observe user-provided context, particularly once they include new or conflicting info. Whereas instruction finetuning ought to enhance this potential, the authors uncover a shocking failure mode known as context-parametric inversion: fashions initially rely extra on enter context, however this reliance decreases as finetuning continues\u2014whilst benchmark efficiency improves. By managed experiments and theoretical evaluation, the authors hint the trigger to coaching examples the place context aligns with pretraining data, reinforcing parametric reliance. They recommend mitigation methods and spotlight this as a key problem in instruction tuning.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29314\" target=\"_blank\" rel=\"noopener\">EmbodiedSAM: On-line Section Any 3D Factor in Actual Time<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Xiuwei Xu, Huangxing Chen, Linqing Zhao, Ziwei Wang, Jie Zhou, Jiwen Lu<\/p>\n<p class=\"oral-spotlight-space\">Embodied duties demand fine-grained 3D notion, which is troublesome to realize because of restricted high-quality 3D information. To handle this, the authors suggest a way that leverages the Section Something Mannequin (SAM) for on-line 3D occasion segmentation by remodeling 2D masks into 3D-aware queries. Their strategy allows real-time object matching throughout video frames and environment friendly inference utilizing a similarity matrix. Experiments throughout a number of datasets present that the strategy outperforms offline options and generalizes effectively to new settings with minimal information.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/28487\" target=\"_blank\" rel=\"noopener\">LLM-SR: Scientific Equation Discovery through Programming with Massive Language Fashions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Parshin Shojaee, Kazem Meidani, Shashank Gupta, Amir Barati Farimani, Chandan Okay. Reddy<\/p>\n<p class=\"oral-spotlight-space\">Mathematical equations are remarkably efficient at describing pure phenomena, however discovering them from information is difficult because of huge combinatorial search areas. Current symbolic regression strategies typically overlook area data and depend on restricted representations. To handle this, the authors suggest LLM-SR, a novel strategy that makes use of Massive Language Fashions to generate equation hypotheses knowledgeable by scientific priors and refines them by way of evolutionary search. Evaluated throughout a number of scientific domains, LLM-SR outperforms current strategies, notably in generalization, by effectively exploring the equation area and producing correct, interpretable fashions.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/28441\" target=\"_blank\" rel=\"noopener\">Thoughts the Hole: Analyzing the Self-Enchancment Capabilities of Massive Language Fashions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Yuda Music, Hanlin Zhang, Udaya Ghai, Carson Eisenach, Sham M. Kakade, Dean Foster<\/p>\n<p class=\"oral-spotlight-space\">Self-improvement in Massive Language Fashions entails the mannequin verifying its outputs, filtering information accordingly, and utilizing the refined information for additional studying. Whereas efficient in observe, there was little theoretical grounding for this system. This work presents a complete research of LLM self-improvement, introducing a proper framework centered on the generation-verification hole\u2014a key amount that governs self-improvement. Experiments reveal that this hole scales persistently with pretraining FLOPs throughout duties and mannequin households. The authors additionally discover when and the way iterative self-improvement works and supply insights and techniques to reinforce it.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/28374\" target=\"_blank\" rel=\"noopener\">On the Advantages of Reminiscence for Modeling Time-Dependent PDEs<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Ricardo Buitrago, Tanya Marwah, Albert Gu, Andrej Risteski<\/p>\n<p class=\"oral-spotlight-space\">Information-driven strategies supply an environment friendly various to conventional numerical solvers for PDEs, however most current approaches assume Markovian dynamics, limiting their effectiveness when enter alerts are distorted. Impressed by the Mori-Zwanzig idea, the authors suggest MemNO, a Reminiscence Neural Operator that explicitly incorporates previous states utilizing structured state-space fashions and the Fourier Neural Operator. MemNO demonstrates sturdy efficiency on varied PDE households, particularly on low-resolution inputs, attaining over six instances decrease error than memoryless baselines.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/31125\" target=\"_blank\" rel=\"noopener\">On the Identification of Temporal Causal Illustration with Instantaneous Dependence<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Zijian Li, Yifan Shen, Kaitao Zheng, Ruichu Cai, Xiangchen Music, Mingming Gong, Guangyi Chen, Kun Zhang<\/p>\n<p class=\"oral-spotlight-space\">This work introduces IDOL (Identification framework for Instantaneous Latent dynamics), a way designed to establish latent causal processes in time collection information, even when instantaneous relationships are current. In contrast to current strategies that require interventions or grouping of observations, IDOL imposes a sparse affect constraint, permitting each time-delayed and instantaneous causal relations to be captured. By a temporally variational inference structure and gradient-based sparsity regularization, IDOL successfully estimates latent variables. Experimental outcomes present that IDOL can establish latent causal processes in simulations and real-world human movement forecasting duties, demonstrating its sensible applicability.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/27848\" target=\"_blank\" rel=\"noopener\">Progressive distillation induces an implicit curriculum<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Abhishek Panigrahi, Bingbin Liu, Sadhika Malladi, Andrej Risteski, Surbhi Goel<\/p>\n<p class=\"oral-spotlight-space\">This work explores the idea of progressive distillation, the place a pupil mannequin learns from intermediate checkpoints of a trainer mannequin, somewhat than simply the ultimate mannequin. The authors establish an \u201cimplicit curriculum\u201d that emerges by way of these intermediate checkpoints, which accelerates the scholar\u2019s studying and supplies a pattern complexity profit. Utilizing sparse parity as a sandbox, they show that this curriculum imparts priceless studying steps which might be unavailable from the ultimate trainer mannequin. The research extends this concept to Transformers educated on probabilistic context-free grammars (PCFGs) and real-world datasets, displaying that the trainer progressively teaches the scholar to seize longer contexts. Each theoretical and empirical outcomes spotlight the effectiveness of progressive distillation throughout totally different duties.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/27832\" target=\"_blank\" rel=\"noopener\">Scaling Legal guidelines for Precision<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Tanishq Kumar, Zachary Ankner, Benjamin Frederick Spector, Blake Bordelon, Niklas Muennighoff, Mansheej Paul, Cengiz Pehlevan, Christopher Re, Aditi Raghunathan<\/p>\n<p class=\"oral-spotlight-space\">This work introduces precision-aware scaling legal guidelines that stretch conventional scaling frameworks to account for the consequences of low-precision coaching and inference in language fashions. The authors present that decrease precision successfully reduces a mannequin\u2019s usable parameter depend, enabling predictions of efficiency degradation because of quantization. For inference, they discover that post-training quantization causes growing degradation with extra pretraining information, probably making extra coaching counterproductive. Their unified framework predicts loss throughout various precisions and means that coaching bigger fashions in decrease precision could also be extra compute-efficient. These predictions are validated on over 465 pretraining runs, together with fashions as much as 1.7B parameters.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29372\" target=\"_blank\" rel=\"noopener\">Self-Enchancment in Language Fashions: The Sharpening Mechanism<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Audrey Huang, Adam Block, Dylan J Foster, Dhruv Rohatgi, Cyril Zhang, Max Simchowitz, Jordan T. Ash, Akshay Krishnamurthy<\/p>\n<p class=\"oral-spotlight-space\">This paper presents a theoretical framework for understanding how LLMs can self-improve through the use of themselves as verifiers to refine their very own outputs; a course of the authors name \u201csharpening.\u201d The important thing perception is that LLMs are sometimes higher at judging response high quality than producing high-quality responses outright, so sharpening helps focus chance mass on higher sequences. The paper analyzes two households of self-improvement algorithms: one primarily based on supervised fine-tuning (SFT) and one on reinforcement studying (RLHF). They present that whereas the SFT-based strategy is perfect below sure circumstances, the RLHF-based strategy can outperform it by actively exploring past the mannequin\u2019s current data.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/27799\" target=\"_blank\" rel=\"noopener\">When Choice meets Intervention: Further Complexities in Causal Discovery<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Haoyue Dai, Ignavier Ng, Jianle Solar, Zeyu Tang, Gongxu Luo, Xinshuai Dong, Peter Spirtes, Kun Zhang<\/p>\n<p class=\"oral-spotlight-space\">This work tackles the often-overlooked subject of choice bias in interventional research, the place members are selectively included primarily based on particular standards. Current causal discovery strategies sometimes ignore this bias, resulting in inaccurate conclusions. To handle this, the authors introduce a novel graphical mannequin that distinguishes between the noticed world with interventions and the counterfactual world the place choice happens. They develop a sound algorithm that identifies each causal relationships and choice mechanisms, demonstrating its effectiveness by way of experiments on each artificial and real-world information.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30064\" target=\"_blank\" rel=\"noopener\">miniCTX: Neural Theorem Proving with (Lengthy-)Contexts<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Jiewen Hu, Thomas Zhu, Sean Welleck<\/p>\n<p class=\"oral-spotlight-space\">Actual-world formal theorem proving depends closely on wealthy contextual info, which is commonly absent from conventional benchmarks. To handle this, the authors introduce miniCTX, a benchmark designed to check fashions\u2019 potential to show theorems utilizing beforehand unseen, in depth context from actual Lean tasks and textbooks. In contrast to prior benchmarks, miniCTX contains massive repositories with related definitions, lemmas, and buildings. Baseline experiments present that fashions conditioned on this broader context considerably outperform these relying solely on the native state. The authors additionally present a toolkit to facilitate the enlargement of the benchmark.<\/p>\n<\/div>\n<h2 id=\"spotlight-papers\">Highlight Papers<\/h2>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/28536\" target=\"_blank\" rel=\"noopener\">ADIFF: Explaining audio distinction utilizing pure language<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Soham Deshmukh, Shuo Han, Rita Singh, Bhiksha Raj<\/p>\n<p class=\"oral-spotlight-space\">This paper tackles the novel process of explaining variations between audio recordings, which is essential for purposes like audio forensics, high quality evaluation, and generative audio techniques. The authors introduce two new datasets and suggest a three-tiered clarification framework\u2014starting from concise occasion descriptions to wealthy, emotionally grounded narratives\u2014generated utilizing massive language fashions. They current ADIFF, a brand new methodology that improves on baselines by incorporating audio cross-projection, position-aware captioning, and multi-stage coaching, and present that it considerably outperforms current audio-language fashions each quantitatively and through human analysis.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30792\" target=\"_blank\" rel=\"noopener\">Higher Instruction-Following By Minimal Bayes Threat<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Ian Wu, Patrick Fernandes, Amanda Bertsch, Seungone Kim, Sina Khoshfetrat Pakazad, Graham Neubig<\/p>\n<p class=\"oral-spotlight-space\">This paper explores how LLMs can be utilized as judges to judge and enhance different LLMs. The authors present that utilizing a way known as Minimal Bayes Threat (MBR) decoding\u2014the place an LLM choose selects one of the best output from a set\u2014can considerably enhance mannequin efficiency in comparison with customary decoding strategies. In addition they discover that coaching fashions on these high-quality outputs can result in sturdy beneficial properties even with out counting on MBR at check time, making the fashions quicker and extra environment friendly whereas sustaining or exceeding earlier efficiency.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/31131\" target=\"_blank\" rel=\"noopener\">DeFT: Decoding with Flash Tree-attention for Environment friendly Tree-structured LLM Inference<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Jinwei Yao, Kaiqi Chen, Kexun Zhang, Jiaxuan You, Binhang Yuan, Zeke Wang, Tao Lin<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces DeFT, a brand new algorithm that hurries up how massive language fashions deal with duties involving tree-like buildings with shared textual content prefixes, resembling multi-step reasoning or few-shot prompting. Current strategies waste time and reminiscence by repeatedly accessing the identical information and poorly distributing the workload throughout the GPU. DeFT solves this by neatly grouping and splitting reminiscence utilization to keep away from redundant operations and higher steadiness the work, resulting in as much as 3.6x quicker efficiency on key duties in comparison with present approaches.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/31271\" target=\"_blank\" rel=\"noopener\">Holistically Evaluating the Environmental Influence of Creating Language Fashions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Jacob Morrison, Clara Na, Jared Fernandez, Tim Dettmers, Emma Strubell, Jesse Dodge<\/p>\n<p class=\"oral-spotlight-space\">This paper estimates the complete environmental impression of creating massive language fashions, together with not simply the ultimate coaching runs but in addition mannequin improvement and {hardware} manufacturing\u2014areas sometimes underreported. The authors discovered that coaching a collection of fashions launched 493 metric tons of carbon emissions and used 2.769 million liters of water, even in a extremely environment friendly information heart. Notably, round half of the carbon emissions got here from the event section alone, and energy utilization throughout coaching assorted considerably, elevating issues for vitality grid planning as AI techniques develop.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/32081\" target=\"_blank\" rel=\"noopener\">Language Mannequin Alignment in Multilingual Trolley Issues<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Zhijing Jin, Max Kleiman-weiner, Giorgio Piatti, Sydney Levine, Jiarui Liu, Fernando Gonzalez Adauto, Francesco Ortu, Andr\u00e1s Strausz, Mrinmaya Sachan, Rada Mihalcea, Yejin Choi, Bernhard Sch\u00f6lkopf<\/p>\n<p class=\"oral-spotlight-space\">This paper evaluates how effectively LLMs align with human ethical preferences throughout languages utilizing multilingual trolley issues. The authors introduce MultiTP, a brand new dataset of ethical dilemmas in over 100 languages primarily based on the Ethical Machine experiment, enabling cross-lingual evaluation of LLM decision-making. By assessing 19 fashions throughout six ethical dimensions and analyzing demographic correlations and immediate consistency, they uncover important variation in ethical alignment throughout languages\u2014highlighting moral biases and the necessity for extra inclusive, multilingual approaches to accountable AI improvement.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29611\" target=\"_blank\" rel=\"noopener\">Lean-STaR: Studying to Interleave Pondering and Proving<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Haohan Lin, Zhiqing Solar, Sean Welleck, Yiming Yang<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces Lean-STaR, a framework that improves language model-based theorem proving by incorporating casual \u201cideas\u201d earlier than every proof step. In contrast to conventional approaches that rely solely on formal proof information, Lean-STaR generates artificial thought processes utilizing retrospective proof ways throughout coaching. At inference time, the mannequin generates these ideas to information its subsequent motion, and knowledgeable iteration additional refines its efficiency utilizing the Lean theorem prover. This strategy boosts proof success charges and affords new insights into how structured reasoning improves formal mathematical downside fixing.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30638\" target=\"_blank\" rel=\"noopener\">MagicPIG: LSH Sampling for Environment friendly LLM Era<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Zhuoming Chen, Ranajoy Sadhukhan, Zihao Ye, Yang Zhou, Jianyu Zhang, Niklas Nolte, Yuandong Tian, Matthijs Douze, Leon Bottou, Zhihao Jia, Beidi Chen<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces MagicPIG, a brand new system that hurries up LLM inference by approximating consideration extra effectively. Whereas many strategies assume consideration is sparse and use TopK approximations, the authors present this isn\u2019t at all times correct and may damage efficiency. As an alternative, MagicPIG makes use of a sampling methodology backed by theoretical ensures and accelerates it utilizing Locality Delicate Hashing, offloading computations to the CPU to assist longer inputs and bigger batches with out sacrificing accuracy.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30630\" target=\"_blank\" rel=\"noopener\">Multi-Robotic Movement Planning with Diffusion Fashions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Yorai Shaoul, Itamar Mishani, Shivam Vats, Jiaoyang Li, Maxim Likhachev<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces a way for planning coordinated, collision-free actions for a lot of robots utilizing solely information from particular person robots. The authors mix discovered diffusion fashions with classical planning algorithms to generate real looking, secure multi-robot trajectories. Their strategy, known as Multi-robot Multi-model planning Diffusion, additionally scales to massive environments by stitching collectively a number of diffusion fashions, displaying sturdy leads to simulated logistics situations.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/28960\" target=\"_blank\" rel=\"noopener\">Reinforcement Studying for Management of Non-Markovian Mobile Inhabitants Dynamics<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Josiah C Kratz, Jacob Adamczyk<\/p>\n<p class=\"oral-spotlight-space\">This paper explores how reinforcement studying can be utilized to develop drug dosing methods for controlling cell populations that adapt over time, resembling most cancers cells switching between resistant and prone states. Conventional strategies wrestle when the system\u2019s dynamics are unknown or contain reminiscence of previous environments, making optimum management troublesome. The authors present that deep RL can efficiently study efficient methods even in advanced, memory-based techniques, providing a promising strategy for real-world biomedical purposes.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30649\" target=\"_blank\" rel=\"noopener\">Rewarding Progress: Scaling Automated Course of Verifiers for LLM Reasoning<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Amrith Setlur, Chirag Nagpal, Adam Fisch, Xinyang Geng, Jacob Eisenstein, Rishabh Agarwal, Alekh Agarwal, Jonathan Berant, Aviral Kumar<\/p>\n<p class=\"oral-spotlight-space\">This paper explores how you can enhance massive language fashions\u2019 reasoning by giving suggestions at every step of their pondering course of, somewhat than solely on the remaining reply. The authors introduce a way the place suggestions\u2014known as a course of reward\u2014is predicated on whether or not a step helps make an accurate remaining reply extra seemingly, as judged by a separate mannequin (a \u201cprover\u201d) that may acknowledge progress higher than the mannequin being educated. They present each theoretically and experimentally that this technique makes studying extra environment friendly, resulting in considerably higher and quicker outcomes than conventional outcome-based suggestions strategies.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/27906\" target=\"_blank\" rel=\"noopener\">SVDQuant: Absorbing Outliers by Low-Rank Part for 4-Bit Diffusion Fashions<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Muyang Li, Yujun Lin, Zhekai Zhang, Tianle Cai, Junxian Guo, Xiuyu Li, Enze Xie, Chenlin Meng, Jun-yan Zhu, Music Han<\/p>\n<p class=\"oral-spotlight-space\">This paper introduces SVDQuant, a way for considerably rushing up diffusion fashions by quantizing each weights and activations to 4 bits. Since such aggressive quantization can damage picture high quality, the authors use a intelligent method: they shift problematic \u201coutlier\u201d values right into a separate low-rank part dealt with with increased precision, whereas the remaining is processed with environment friendly low-bit operations. To keep away from slowing issues down because of additional computation, in addition they design a customized inference engine known as Nunchaku, which merges the processing steps to attenuate reminiscence entry. Collectively, these strategies cut back reminiscence utilization and ship over 3x speedups with out sacrificing picture high quality.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30460\" target=\"_blank\" rel=\"noopener\">Stabilizing Reinforcement Studying in Differentiable Multiphysics Simulation<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Eliot Xing, Vernon Luk, Jean Oh<\/p>\n<p class=\"oral-spotlight-space\">This paper tackles the problem of making use of reinforcement studying (RL) to soft-body robotics, the place simulations are often too sluggish for data-hungry RL algorithms. The authors introduce SAPO, a brand new model-based RL algorithm that effectively learns from differentiable simulations utilizing analytic gradients. The authors additionally current Rewarped, a quick, parallel simulation platform that helps each inflexible and deformable supplies, demonstrating that their strategy outperforms current strategies on advanced manipulation and locomotion duties.<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30029\" target=\"_blank\" rel=\"noopener\">Streaming Algorithms For $ell_p$ Flows and $ell_p$ Regression<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Amit Chakrabarti, Jeffrey Jiang, David Woodruff, Taisuke Yasuda<\/p>\n<p class=\"oral-spotlight-space\">This paper investigates how you can clear up underdetermined linear regression issues in a streaming setting, the place the information arrives one column at a time and storing the complete dataset is impractical. The authors develop algorithms that approximate the regression price or output a near-optimal resolution utilizing a lot much less reminiscence than storing your complete dataset\u2014notably related for purposes like computing flows on massive graphs. In addition they set up area decrease bounds, displaying the restrictions of what\u2019s potential, and supply the primary algorithms that obtain nontrivial approximations utilizing sublinear area in varied settings.<\/p>\n<\/div>\n<h2 id=\"poster-papers\">Poster Papers<\/h2>\n<h3 id=\"alignment,-fairness,-safety,-privacy,-and-societal-considerations\">Alignment, Equity, Security, Privateness, And Societal Issues<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"#\" target=\"_blank\" rel=\"noopener\">AgentHarm: Benchmarking Robustness of LLM Brokers on Dangerous Duties<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Maksym Andriushchenko, Alexandra Souly, Mateusz Dziemian, Derek Duenas, Maxwell Lin, Justin Wang, Dan Hendrycks, Andy Zou, J Zico Kolter, Matt Fredrikson, Yarin Gal, Xander Davies<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29856\" target=\"_blank\" rel=\"noopener\">Aligned LLMs Are Not Aligned Browser Brokers<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Priyanshu Kumar, Elaine Lau, Saranya Vijayakumar, Tu Trinh, Elaine T Chang, Vaughn Robinson, Shuyan Zhou, Matt Fredrikson, Sean M. Hendryx, Summer time Yue, Zifan Wang<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/31026\" target=\"_blank\" rel=\"noopener\">Towards Strong Defenses In opposition to LLM Weight Tampering Assaults<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Rishub Tamirisa, Bhrugu Bharathi, Lengthy Phan, Andy Zhou, Alice Gatti, Tarun Suresh, Maxwell Lin, Justin Wang, Rowan Wang, Ron Arel, Andy Zou, Daybreak Music, Bo Li, Dan Hendrycks, Mantas Mazeika<\/p>\n<\/div>\n<h3 id=\"applications-to-computer-vision,-audio,-language,-and-other-modalities\">Functions To Pc Imaginative and prescient, Audio, Language, And Different Modalities<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30598\" target=\"_blank\" rel=\"noopener\">Fugatto 1: Foundational Generative Audio Transformer Opus 1<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Rafael Valle, Rohan Badlani, Zhifeng Kong, Sang-gil Lee, Arushi Goel, Joao Felipe Santos, Aya Aljafari, Sungwon Kim, Shuqi Dai, Siddharth Gururani, Alexander H. Liu, Kevin J. Shih, Ryan Prenger, Wei Ping, Chao-han Huck Yang, Bryan Catanzaro<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29912\" target=\"_blank\" rel=\"noopener\">MetaDesigner: Advancing Creative Typography by way of AI-Pushed, Person-Centric, and Multilingual WordArt Synthesis<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Jun-yan He, Zhi-qi Cheng, Chenyang Li, Jingdong Solar, Qi He, Wangmeng Xiang, Hanyuan Chen, Jin-peng Lan, Xianhui Lin, Kang Zhu, Bin Luo, Yifeng Geng, Xuansong Xie, Alexander G Hauptmann<\/p>\n<\/div>\n<h3 id=\"applications-to-neuroscience-&amp;-cognitive-science\">Functions To Neuroscience &amp; Cognitive Science<\/h3>\n<h3 id=\"applications-to-physical-sciences-(physics,-chemistry,-biology,-etc.)\">Functions To Bodily Sciences (Physics, Chemistry, Biology, And many others.)<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/28747\" target=\"_blank\" rel=\"noopener\">Causal Illustration Studying from Multimodal Organic Observations<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Yuewen Solar, Lingjing Kong, Guangyi Chen, Loka Li, Gongxu Luo, Zijian Li, Yixuan Zhang, Yujia Zheng, Mengyue Yang, Petar Stojanov, Eran Segal, Eric P. Xing, Kun Zhang<\/p>\n<\/div>\n<h3 id=\"applications-to-robotics,-autonomy,-planning\">Functions To Robotics, Autonomy, Planning<\/h3>\n<h3 id=\"causal-reasoning\">Causal Reasoning<\/h3>\n<h3 id=\"datasets-and-benchmarks\">Datasets And Benchmarks<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"#\" target=\"_blank\" rel=\"noopener\">Dynamic-SUPERB Part-2: A Collaboratively Increasing Benchmark for Measuring the Capabilities of Spoken Language Fashions with 180 Duties<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Chien-yu Huang, Wei-chih Chen, Shu-wen Yang, Andy T. Liu, Chen-an Li, Yu-xiang Lin, Wei-cheng Tseng, Anuj Diwan, Yi-jen Shih, Jiatong Shi, William Chen, Xuanjun Chen, Chi-yuan Hsiao, Puyuan Peng, Shih-heng Wang, Chun-yi Kuan, Ke-han Lu, Kai-wei Chang, Chih-kai Yang, Fabian Alejandro Ritter Gutierrez, Huang Kuan-po, Siddhant Arora, You-kuan Lin, Chuang Ming To, Eunjung Yeo, Kalvin Chang, Chung-ming Chien, Kwanghee Choi, Cheng-hsiu Hsieh, Yi-cheng Lin, Chee-en Yu, I-hsiang Chiu, Heitor Guimar\u00e3es, Jionghao Han, Tzu-quan Lin, Tzu-yuan Lin, Homu Chang, Ting-wu Chang, Chun Wei Chen, Shou-jen Chen, Yu-hua Chen, Hsi-chun Cheng, Kunal Dhawan, Jia-lin Fang, Shi-xin Fang, Kuan Yu Fang Chiang, Chi An Fu, Hsien-fu Hsiao, Ching Yu Hsu, Shao-syuan Huang, Lee Chen Wei, Hsi-che Lin, Hsuan-hao Lin, Hsuan-ting Lin, Jian-ren Lin, Ting-chun Liu, Li-chun Lu, Tsung-min Pai, Ankita Pasad, Shih-yun Shan Kuan, Suwon Shon, Yuxun Tang, Yun-shao Tsai, Wei Jui Chiang, Tzu-chieh Wei, Chengxi Wu, Dien-ruei Wu, Chao-han Huck Yang, Chieh-chi Yang, Jia Qi Yip, Shao-xiang Yuan, Haibin Wu, Karen Livescu, David Harwath, Shinji Watanabe, Hung-yi Lee<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29719\" target=\"_blank\" rel=\"noopener\">Scalable Benchmarking and Strong Studying for Noise-Free Ego-Movement and 3D Reconstruction from Noisy Video<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Xiaohao Xu, Tianyi Zhang, Shibo Zhao, Xiang Li, Sibo Wang, Yongqi Chen, Ye Li, Bhiksha Raj, Matthew Johnson-roberson, Sebastian Scherer, Xiaonan Huang<\/p>\n<\/div>\n<h3 id=\"foundation-or-frontier-models,-including-llms\">Basis Or Frontier Fashions, Together with Llms<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29057\" target=\"_blank\" rel=\"noopener\">Variety Empowers Intelligence: Integrating Experience of Software program Engineering Brokers<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Kexun Zhang, Weiran Yao, Zuxin Liu, Yihao Feng, Zhiwei Liu, Rithesh R N, Tian Lan, Lei Li, Renze Lou, Jiacheng Xu, Bo Pang, Yingbo Zhou, Shelby Heinecke, Silvio Savarese, Huan Wang, Caiming Xiong<\/p>\n<\/div>\n<h3 id=\"generative-models\">Generative Fashions<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29672\" target=\"_blank\" rel=\"noopener\">Linear Mixture of Saved Checkpoints Makes Consistency and Diffusion Fashions Higher<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Enshu Liu, Junyi Zhu, Zinan Lin, Xuefei Ning, Shuaiqi Wang, Matthew B. Blaschko, Sergey Yekhanin, Shengen Yan, Guohao Dai, Huazhong Yang, Yu Wang<\/p>\n<\/div>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29729\" target=\"_blank\" rel=\"noopener\">RAG-DDR: Optimizing Retrieval-Augmented Era Utilizing Differentiable Information Rewards<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Xinze Li, Sen Mei, Zhenghao Liu, Yukun Yan, Shuo Wang, Shi Yu, Zheni Zeng, Hao Chen, Ge Yu, Zhiyuan Liu, Maosong Solar, Chenyan Xiong<\/p>\n<\/div>\n<h3 id=\"infrastructure,-software-libraries,-hardware,-systems,-etc.\">Infrastructure, Software program Libraries, {Hardware}, Methods, And many others.<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/29831\" target=\"_blank\" rel=\"noopener\">OpenHands: An Open Platform for AI Software program Builders as Generalist Brokers<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Xingyao Wang, Boxuan Li, Yufan Music, Frank F. Xu, Xiangru Tang, Mingchen Zhuge, Jiayi Pan, Yueqi Music, Bowen Li, Jaskirat Singh, Hoang H. Tran, Fuqiang Li, Ren Ma, Mingzhang Zheng, Invoice Qian, Yanjun Shao, Niklas Muennighoff, Yizhe Zhang, Binyuan Hui, Junyang Lin, Robert Brennan, Hao Peng, Heng Ji, Graham Neubig<\/p>\n<\/div>\n<h3 id=\"interpretability-and-explainable-ai\">Interpretability And Explainable Ai<\/h3>\n<h3 id=\"learning-on-graphs-and-other-geometries-&amp;-topologies\">Studying On Graphs And Different Geometries &amp; Topologies<\/h3>\n<h3 id=\"learning-theory\">Studying Principle<\/h3>\n<h3 id=\"neurosymbolic-&amp;-hybrid-ai-systems-(physics-informed,-logic-&amp;-formal-reasoning,-etc.)\">Neurosymbolic &amp; Hybrid Ai Methods (Physics-informed, Logic &amp; Formal Reasoning, And many others.)<\/h3>\n<h3 id=\"optimization\">Optimization<\/h3>\n<h3 id=\"other-topics-in-machine-learning-(i.e.,-none-of-the-above)\">Different Subjects In Machine Studying (I.e., None Of The Above)<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/28438\" target=\"_blank\" rel=\"noopener\">Zeroth-Order Positive-Tuning of LLMs with Transferable Static Sparsity<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Wentao Guo, Jikai Lengthy, Yimeng Zeng, Zirui Liu, Xinyu Yang, Yide Ran, Jacob R. Gardner, Osbert Bastani, Christopher De Sa, Xiaodong Yu, Beidi Chen, Zhaozhuo Xu<\/p>\n<\/div>\n<h3 id=\"probabilistic-methods-(bayesian-methods,-variational-inference,-sampling,-uq,-etc.)\">Probabilistic Strategies (Bayesian Strategies, Variational Inference, Sampling, Uq, And many others.)<\/h3>\n<h3 id=\"reinforcement-learning\">Reinforcement Studying<\/h3>\n<h3 id=\"transfer-learning,-meta-learning,-and-lifelong-learning\">Switch Studying, Meta Studying, And Lifelong Studying<\/h3>\n<h3 id=\"unsupervised,-self-supervised,-semi-supervised,-and-supervised-representation-learning\">Unsupervised, Self-supervised, Semi-supervised, And Supervised Illustration Studying<\/h3>\n<div class=\"paper\">\n<p class=\"paper-title\"><a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/iclr.cc\/virtual\/2025\/poster\/30157\" target=\"_blank\" rel=\"noopener\">Reminiscence Mosaics<\/a><\/p>\n<p class=\"paper-authors\"><b>Authors:<\/b> Jianyu Zhang, Niklas Nolte, Ranajoy Sadhukhan, Beidi Chen, Leon Bottou<\/p>\n<\/div><\/div>\n\n","protected":false},"excerpt":{"rendered":"<p>CMU researchers are presenting 143 papers on the Thirteenth Worldwide Convention on Studying Representations (ICLR 2025), held from April 24 \u2013 28 on the Singapore EXPO. Here&#8217;s a fast overview of the areas our researchers are engaged on: And listed below are our most frequent collaborator establishments: Desk of Contents Oral Papers Highlight Papers Poster [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":1864,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[55],"tags":[110,1838,1841,136,113,1839,442,1840],"class_list":["post-1862","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-machine-learning","tag-blog","tag-carnegie","tag-iclr","tag-learning","tag-machine","tag-mellon","tag-mlcmu","tag-university"],"_links":{"self":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/1862","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=1862"}],"version-history":[{"count":1,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/1862\/revisions"}],"predecessor-version":[{"id":1863,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/1862\/revisions\/1863"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/media\/1864"}],"wp:attachment":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=1862"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=1862"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=1862"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69d9690a190636c2e0989534. Config Timestamp: 2026-04-10 21:18:02 UTC, Cached Timestamp: 2026-04-19 08:20:17 UTC -->