NeurIPS 2023 compressed

This page contains summaries of all Neurips 2023 accepted papers generated by the compressor, my personal LLM-based project.


What is Flagged in Uncertainty Quantification? Latent Density Models for Uncertainty Categorization

Hao Sun, Boris van Breugel, Jonathan Crabbé, Nabeel Seedat, Mihaela van der Schaar

https://openreview.net/forum?id=zyhxRc9bew

Keywords: Uncertainty Explaination, Uncertainty Quantification, Interpretability

Compressor summary: This paper introduces a framework to classify uncertain examples from machine learning models using the confusion density matrix, which helps compare different uncertainty quantification methods.


Don’t blame Dataset Shift! Shortcut Learning due to Gradients and Cross Entropy

Aahlad Manas Puli, Lily H Zhang, Yoav Wald, Rajesh Ranganath

https://openreview.net/forum?id=zyZkaqNnpa

Keywords: shortcut learning, spurious correlations, perfect stable feature, perception tasks, implicit bias in optimization, improving inductive biases

Compressor summary: The paragraph discusses how default-ERM models tend to use shortcuts in perception tasks due to their preference for maximizing margins, and proposes alternative loss functions called margin control (MARG-CTRL) that encourage uniform-margin solutions and mitigate shortcut learning.


Enhancing Adversarial Contrastive Learning via Adversarial Invariant Regularization

Xilie Xu, Jingfeng Zhang, Feng Liu, Masashi Sugiyama, Mohan Kankanhalli

https://openreview.net/forum?id=zuXyQsXVLF

Keywords: robust pre-training, adversarial contrastive learning

Compressor summary: Adversarial contrastive learning (ACL) enhances standard contrastive learning by using adversarial data and introduces adversarial invariant regularization (AIR) to improve robustness against style factors, as shown by causal reasoning and empirical results.


An Optimization-based Approach To Node Role Discovery in Networks: Approximating Equitable Partitions

Michael Scholkemper, Michael T Schaub

https://openreview.net/forum?id=ztDxO15N7f

Keywords: Role Extraction, Graph Learning, Node Embeddings, Weisfeiler Lehman, Equitable Partition

Compressor summary: The paragraph discusses a new method for identifying structural roles of nodes in complex networks, based on graph-isomorphism tests and the Weisfeiler-Leman algorithm, and presents a benchmark to test the approach.


Generator Identification for Linear SDEs with Additive and Multiplicative Noise

Yuanyuan Wang, Xi Geng, Wei Huang, Biwei Huang, Mingming Gong

https://openreview.net/forum?id=zsOOqjaj2z

Keywords: Linear SDE, Identification, Causal inference

Compressor summary: The paper proposes identifiability conditions for determining the generator of linear SDEs from its solution process distribution, which is useful for causal inference, and provides sufficient and necessary conditions for additive noise and sufficient ones for multiplicative noise, along with geometric interpretations and simulations.


Algorithm Selection for Deep Active Learning with Imbalanced Datasets

Jifan Zhang, Shuai Shao, saurabh verma, Robert D Nowak

https://openreview.net/forum?id=zrUEHZ6s9C

Keywords: Deep Learning, Active Learning

Compressor summary: TAILOR is a meta algorithm that adaptively selects among various active learning strategies to efficiently reduce labeling efforts for deep learning applications using class-balanced rewards.


Discover and Align Taxonomic Context Priors for Open-world Semi-Supervised Learning

Yu Wang, Zhun Zhong, Pengchong Qiao, Xuxin Cheng, Xiawu Zheng, Chang Liu, Nicu Sebe, Rongrong Ji, Jie Chen

https://openreview.net/forum?id=zrLxHYvIFL

Keywords: open-world semi-supervised learning; novel class discovery;

Compressor summary: The paper proposes TIDA, a framework for open-world semi-supervised learning that leverages multi-granularity taxonomic context priors to enhance representation learning and improve pseudo label quality.


Learning Invariant Representations of Graph Neural Networks via Cluster Generalization

Donglin Xia, Xiao Wang, Nian Liu, Chuan Shi

https://openreview.net/forum?id=zrCmeqV3Sz

Keywords: Graph neural networks, network representation learning, deep learning

Compressor summary: The paper proposes CIT, a mechanism that improves GNNs' generalization ability by transferring cluster information and preserving node diversity when the test graph structure differs from the training graph structure.


The expressive power of pooling in Graph Neural Networks

Filippo Maria Bianchi, Veronica Lachi

https://openreview.net/forum?id=zqyVjCjhYD

Keywords: Graph Neural Networks, Graph pooling, Expressive power

Compressor summary: This paper studies how graph pooling affects the expressiveness of Graph Neural Networks (GNNs) and provides a criterion for choosing or designing pooling operators based on their ability to preserve the GNN's message-passing power.


Shared Adversarial Unlearning: Backdoor Mitigation by Unlearning Shared Adversarial Examples

Shaokui Wei, Mingda Zhang, Hongyuan Zha, Baoyuan Wu

https://openreview.net/forum?id=zqOcW3R9rd

Keywords: Backdoor Attack, Trustworthy AI, Backdoor Learning

Compressor summary: The paper proposes a method called Shared Adversarial Unlearning (SAU) to purify backdoored machine learning models using a small clean dataset and adversarial training techniques, achieving state-of-the-art performance.


The Crucial Role of Normalization in Sharpness-Aware Minimization

Yan Dai, Kwangjun Ahn, Suvrit Sra

https://openreview.net/forum?id=zq4vFneRiA

Keywords: Sharpness-Aware Minimization, Normalization, Deep Learning Theory

Compressor summary: The paragraph discusses a paper that explores how normalization, a key component of the SAM optimizer, improves deep neural network performance by stabilizing the algorithm and enabling it to drift along a manifold of minima.


Towards Personalized Federated Learning via Heterogeneous Model Reassembly

Jiaqi Wang, Xingyi Yang, Suhan Cui, Liwei Che, Lingjuan Lyu, Dongkuan Xu, Fenglong Ma

https://openreview.net/forum?id=zpVCITHknd

Keywords: Federated Learning

Compressor summary: The paper introduces pFedHR, a framework for personalized federated learning that handles model heterogeneity by reassembling diverse models using server-side optimization and minimizing the impact of different data distributions.


Active Learning for Semantic Segmentation with Multi-class Label Query

Sehyun Hwang, Sohyun Lee, Hoyoung Kim, Minhyeon Oh, Jungseul Ok, Suha Kwak

https://openreview.net/forum?id=znudaK78u8

Keywords: semantic segmentation; active learning; partial label learning

Compressor summary: The paper presents a new active learning method for semantic segmentation that uses multi-class labeling of local image regions and disambiguates partial labels with two loss functions and pseudo labels, achieving better performance and annotation efficiency.


Time-Reversed Dissipation Induces Duality Between Minimizing Gradient Norm and Function Value

Jaeyeon Kim, Asuman E. Ozdaglar, Chanwoo Park, Ernest K. Ryu

https://openreview.net/forum?id=znY173SCxu

Keywords: Convex Optimization, Acceleration, First-Order methods

Compressor summary: The paper introduces H-duality, a novel concept in convex optimization that reveals symmetries between different first-order optimization methods and leads to new efficient techniques for minimizing function values and gradient magnitudes.


Optimizing over trained GNNs via symmetry breaking

Shiqiang Zhang, Juan S Campos, Christian Wolfgang Feldmann, David Walz, Frederik Sandfort, Miriam Mathea, Calvin Tsay, Ruth Misener

https://openreview.net/forum?id=znW5jNIOED

Keywords: Mixed-integer optimization, Graph neural network, Symmetry-breaking, Molecular design

Compressor summary: The paper proposes methods to optimize machine learning models constrained by trained graph neural networks (GNNs) using symmetry-breaking constraints and applies them to molecular design.


An Alternating Optimization Method for Bilevel Problems under the Polyak-Łojasiewicz Condition

Quan Xiao, Songtao Lu, Tianyi Chen

https://openreview.net/forum?id=zn5ihqknGj

Keywords: Bilevel optimization, nonconvex constrained optimization, convergence analysis

Compressor summary: This paper introduces a new method (GALET) for solving bilevel optimization problems with nonconvex lower-level objectives and shows that it has the same convergence rate as gradient descent.


Scalable Fair Influence Maximization

Xiaobin Rui, Zhixiao Wang, Jiayu Zhao, Lichao Sun, Wei Chen

https://openreview.net/forum?id=zmWNe1V6jg

Keywords: influence maximization, approximation algorithm, social fairness

Compressor summary: The paper proposes an efficient algorithm for maximizing welfare fairness in community structures using a weighted maximum coverage problem approach.


Curriculum Learning With Infant Egocentric Videos

Saber Sheybani, Himanshu Hansaria, Justin Newell Wood, Linda B. Smith, Zoran Tiganj

https://openreview.net/forum?id=zkfyOkBVpz

Keywords: Curriculum learning, Self-supervised learning, Slow changes, Infant development

Compressor summary: Infant visual experiences affect the development of their visual system, and starting with data from younger infants improves learning outcomes for AI models trained on their visual inputs.


Comparing Apples to Oranges: Learning Similarity Functions for Data Produced by Different Distributions

Leonidas Tsepenekas, Ivan Brugere, Freddy Lecue, Daniele Magazzeni

https://openreview.net/forum?id=zjpjsJeVJZ

Keywords: individual fairness; similarity learning; active learning

Compressor summary: The paper proposes an efficient sampling framework to learn similarity functions between different groups using limited expert feedback, and provides theoretical and empirical evidence for its effectiveness.


Temporal Causal Mediation through a Point Process: Direct and Indirect Effects of Healthcare Interventions

Çağlar Hızlı, S. T. John, Anne Tuulikki Juuti, Tuure Tapani Saarinen, Kirsi Hannele Pietiläinen, Pekka Marttinen

https://openreview.net/forum?id=zfHCKDzzC8

Keywords: Machine learning for healthcare, Causal mediation, Gaussian process, Point Process

Compressor summary: The authors propose a new method to analyze the causal effects of an intervention on outcomes and mediators in dynamic systems using temporal point processes.


Interpreting Unsupervised Anomaly Detection in Security via Rule Extraction

Ruoyu Li, Qing Li, Yu Zhang, Dan Zhao, Yong Jiang, Yong Yang

https://openreview.net/forum?id=zfCNwRQ569

Keywords: unsupervised anomaly detection, global explanation, rule extraction

Compressor summary: The paper proposes a post-hoc method to explain black-box unsupervised anomaly detection models using distribution decomposition rules and boundary inference rules, making them more interpretable and trustworthy.


Counting Distinct Elements Under Person-Level Differential Privacy

Thomas Steinke, Alexander Knop

https://openreview.net/forum?id=zdli6OxpWd

Keywords: differential privacy, user-level privacy, person-level privacy, sensitivity

Compressor summary: The authors propose a method to count distinct elements in a dataset with person-level differential privacy, by approximating the problem with a max-flow solution and optimizing the sensitivity bound.


Optimistic Natural Policy Gradient: a Simple Efficient Policy Optimization Framework for Online RL

Qinghua Liu, Gellért Weisz, András György, Chi Jin, Csaba Szepesvari

https://openreview.net/forum?id=zaQ7wV9NOg

Keywords: Theory of reinforcement learning, policy optimization

Compressor summary: This paper introduces Optimistic NPG, a simple and efficient policy optimization framework for online RL with linear MDPs, which improves over existing algorithms in terms of computation and exploration.


Statistical Limits of Adaptive Linear Models: Low-Dimensional Estimation and Inference

Licong Lin, Mufang Ying, Suvrojit Ghosh, Koulik Khamaru, Cun-Hui Zhang

https://openreview.net/forum?id=zXckveawHa

Keywords: Adaptive linear regression, bandit algorithms, high dimensional statistics, statistical inference

Compressor summary: The paper explores how adaptive data collection affects estimation and inference in high-dimensional linear models and proposes a new estimator for single coordinate inference with better performance than OLS.


Universality and Limitations of Prompt Tuning

Yihan Wang, Jatin Chauhan, Wei Wang, Cho-Jui Hsieh

https://openreview.net/forum?id=zWxKYyW9ik

Keywords: prompt-tuning; language model; expressive power

Compressor summary: The paper analyzes the theoretical and practical aspects of prompt tuning for transformer-based models, showing its universality for approximating sequence-to-sequence functions and its limitations for limited-depth models and non-invertible datasets.


Unpaired Multi-Domain Causal Representation Learning

Nils Sturma, Chandler Squires, Mathias Drton, Caroline Uhler

https://openreview.net/forum?id=zW1uVN6Mbv

Keywords: linear structural equation models, causality, representation learning, independent component analysis, structure identifiability, multiple views, graphical model

Compressor summary: The paper proposes a method to identify a shared causal representation of unpaired data from multiple domains using linear models.


$S^3$: Increasing GPU Utilization during Generative Inference for Higher Throughput

Yunho Jin, Chun-Feng Wu, David Brooks, Gu-Yeon Wei

https://openreview.net/forum?id=zUYfbdNl1m

Keywords: Throughput, GPU utilization, Sequence length prediction

Compressor summary: The paper proposes $S^3$, a system that predicts the output sequence length of large language models to increase memory and GPU utilization, and handle mispredictions.


Beta Diffusion

Mingyuan Zhou, Tianqi Chen, Zhendong Wang, Huangjie Zheng

https://openreview.net/forum?id=zTSlm4nmlH

Keywords: Diffusion models, KL-divergence upper bounds, multiplicative transitions, scaled and shifted beta distributions

Compressor summary: Beta diffusion is a new method for generating data within limited ranges using multiplicative transitions based on beta distributions and optimized with KL-divergence upper bounds, which outperform negative ELBOs in experiments with synthetic and natural images.


Sounding Bodies: Modeling 3D Spatial Sound of Humans Using Body Pose and Audio

Xudong XU, Dejan Markovic, Jacob Sandakly, Todd Keebler, Steven Krenn, Alexander Richard

https://openreview.net/forum?id=zQTi3pziFp

Keywords: sound field, spatial audio, virtual humans, human body, body modeling

Compressor summary: The authors present a system that generates realistic 3D spatial audio for human bodies using input from microphones and body pose, and introduce a new multimodal dataset for this task.


Optimized Covariance Design for AB Test on Social Network under Interference

Qianyi Chen, Bo Li, LU DENG, Yong Wang

https://openreview.net/forum?id=zQOYGDc9pu

Keywords: AB test, interference, causal inference, optimization, social network

Compressor summary: The paper proposes a new method for designing randomized network experiments that balances bias and variance by optimizing the covariance matrix of the treatment assignment vector using projected gradient descent.


Multi-scale Diffusion Denoised Smoothing

Jongheon Jeong, Jinwoo Shin

https://openreview.net/forum?id=zQ4yraDiRe

Keywords: adversarial robustness, certified robustness, randomized smoothing, denoised smoothing, diffusion models

Compressor summary: The paper presents a method called multi-scale smoothing that improves the trade-off between robustness and accuracy in denoised smoothing by selectively applying randomized smoothing among multiple noise scales, and proposes diffusion fine-tuning to enhance the performance of diffusion models.


Conformal PID Control for Time Series Prediction

Anastasios Nikolas Angelopoulos, Emmanuel Candes, Ryan Tibshirani

https://openreview.net/forum?id=zPYeYv6YYs

Keywords: conformal prediction, time series, uncertainty quantification, distribution shift

Compressor summary: The paper presents uncertainty quantification algorithms for time series prediction that use conformal prediction and control theory, improve coverage on COVID-19 forecasts, and provide an extendable codebase.


Residual Alignment: Uncovering the Mechanisms of Residual Networks

Jianing Li, Vardan Papyan

https://openreview.net/forum?id=zOCIKYVaF5

Keywords: Deep Learning, Residual Networks, Neural Networks, Generalization, Spectral Analysis

Compressor summary: This paper investigates Residual Alignment in ResNet architectures and shows how it aligns intermediate representations linearly across layers and affects generalization.


Stabilized Neural Differential Equations for Learning Dynamics with Explicit Constraints

Alistair White, Niki Kilbertus, Maximilian Gelbrecht, Niklas Boers

https://openreview.net/forum?id=zO2dAQfvHf

Keywords: neural differential equations, neural ordinary differential equations, constraints, conservation laws, stabilization, dynamical systems, dynamics, scientific machine learning, physics-informed machine learning

Compressor summary: SNDEs are a method to enforce constraints on neural differential equations by adding a stabilization term, which makes them compatible with various models and improves performance.


P-Flow: A Fast and Data-Efficient Zero-Shot TTS through Speech Prompting

Sungwon Kim, Kevin J. Shih, Rohan Badlani, Joao Felipe Santos, Evelina Bakhturina, Mikyas T. Desta, Rafael Valle, Sungroh Yoon, Bryan Catanzaro

https://openreview.net/forum?id=zNA7u7wtIN

Keywords: text-to-speech, zero-shot TTS, flow matching generative model

Compressor summary: P-Flow is a fast and data-efficient zero-shot text-to-speech model that uses speech prompts for speaker adaptation and achieves high quality and pronunciation with less training data and faster sampling speed than existing models.


FAMO: Fast Adaptive Multitask Optimization

Bo Liu, Yihao Feng, Peter Stone, qiang liu

https://openreview.net/forum?id=zMeemcUeXL

Keywords: multitask learning, multitask optimization, conflicting gradients, knowledge transfer

Compressor summary: The paper introduces FAMO, a method that balances task losses in multitask learning using minimal space and time, outperforming existing techniques.


Implicit variance regularization in non-contrastive SSL

Manu Srinath Halvagal, Axel Laborieux, Friedemann Zenke

https://openreview.net/forum?id=zMNUNd9zs1

Keywords: Self-supervised learning, Non-contrastive learning, Learning dynamics

Compressor summary: Non-contrastive self-supervised learning (SSL) methods use different losses to stabilize learning and avoid representational collapse, and a new family of loss functions called IsoLoss can further improve their performance.


New Complexity-Theoretic Frontiers of Tractability for Neural Network Training

Cornelius Brand, Robert Ganian, Mathis Rocton

https://openreview.net/forum?id=zIEaOZ0saA

Keywords: neural network training, computational complexity, ReLU networks, Linear networks

Compressor summary: The article presents new algorithms that improve the efficiency of training neural networks with linear or ReLU activation functions.


Optimal Treatment Regimes for Proximal Causal Learning

Tao Shen, Yifan Cui

https://openreview.net/forum?id=zGdH4tKtOW

Keywords: Optimal treatment regimes, Policy-making, Proximal causal inference, Unmeasured confounding, Value function

Compressor summary: The paragraph discusses a new framework for causal inference using proxy variables, and proposes an optimal individualized treatment regime based on outcome and treatment confounding bridges, with theoretical guarantees and applications shown.


Multi-Swap k-Means++

Lorenzo Beretta, Vincent Cohen-Addad, Silvio Lattanzi, Nikos Parotsidis

https://openreview.net/forum?id=zGRWp7yRqd

Keywords: Clustering, k-means, approximation algorithms

Compressor summary: The paper proposes an improved $k$-means clustering algorithm by combining $k$-means++ sampling with larger local-search neighborhoods, achieving a 9 + ε approximation ratio and practical efficiency.


Automated Classification of Model Errors on ImageNet

Momchil Peychev, Mark Niklas Mueller, Marc Fischer, Martin Vechev

https://openreview.net/forum?id=zEoP4vzFKy

Keywords: ImageNet, evaluation, error classification, error analysis

Compressor summary: The paragraph discusses how automated error classification can help evaluate computer vision models and finds that top-1 accuracy is still useful for measuring model performance despite its limitations.


(Amplified) Banded Matrix Factorization: A unified approach to private training

Christopher A. Choquette-Choo, Arun Ganesh, Ryan McKenna, Hugh Brendan McMahan, J Keith Rush, Abhradeep Guha Thakurta, Zheng Xu

https://openreview.net/forum?id=zEm6hF97Pz

Keywords: Machine Learning, Differential Privacy, Optimization, Private Machine Learning, Federated Learning, Privacy Amplification, Matrix Factorization

Compressor summary: The paper presents a method to improve privacy-utility tradeoffs for machine learning using matrix factorization with banded matrices, which can be applied in both centralized and federated settings.


Leveraging Pre-trained Large Language Models to Construct and Utilize World Models for Model-based Task Planning

Lin Guan, Karthik Valmeekam, Sarath Sreedharan, Subbarao Kambhampati

https://openreview.net/forum?id=zDbsSscmuj

Keywords: LLMs, Planning, Domain Model, LLMs for Planning, LLMs for Heuristic Guidance

Compressor summary: The authors propose a method that uses large language models to construct PDDL domain models and then plan with sound planners, reducing human involvement and improving planning efficiency.


A Novel Framework for Policy Mirror Descent with General Parameterization and Linear Convergence

Carlo Alfano, Rui Yuan, Patrick Rebeschini

https://openreview.net/forum?id=zD6lXmTPPh

Keywords: Theory for Reinforcement Learning, Policy Optimization, Policy Gradient, Mirror Descent.

Compressor summary: The paper introduces a policy optimization framework based on mirror descent that allows general parameterizations and shows linear convergence, sample complexity improvements, and experimental validation.


Quasi-Monte Carlo Graph Random Features

Isaac Reid, Adrian Weller, Krzysztof Marcin Choromanski

https://openreview.net/forum?id=zCFfv49MjE

Keywords: Graph, discrete mathematics, quasi-Monte Carlo, kernel, scalability, Laplacian, clustering, random walks

Compressor summary: The paper introduces a novel method to improve graph random features by inducing negative correlations through antithetic termination, leading to lower-variance estimators and better performance on various tasks.


One-Line-of-Code Data Mollification Improves Optimization of Likelihood-based Generative Models

Ba-Hien Tran, Giulio Franzese, Pietro Michiardi, Maurizio Filippone

https://openreview.net/forum?id=zAXg8dW8ZO

Keywords: Generative Models, Normalizing Flows, Variational Autoencoders

Compressor summary: The paper introduces a simple technique to improve the quality of likelihood-based generative models by using data mollification, which enhances density estimation and low-density region handling without additional computational costs.


Decision Stacks: Flexible Reinforcement Learning via Modular Generative Models

Siyan Zhao, Aditya Grover

https://openreview.net/forum?id=zAQK5r1enm

Keywords: reinforcement learning, generative models, offline RL, sequential decision making

Compressor summary: Decision Stacks is a framework that splits goal-conditioned policy agents into three independent modules for efficient learning and flexible generative decision making in sequential decision making problems.


Towards Foundation Models for Scientific Machine Learning: Characterizing Scaling and Transfer Behavior

Shashank Subramanian, Peter Harrington, Kurt Keutzer, Wahid Bhimji, Dmitriy Morozov, Michael W. Mahoney, Amir Gholami

https://openreview.net/forum?id=zANxvzflMl

Keywords: scientific machine learning, scaling, transfer learning, neural operators, foundation models

Compressor summary: The study shows that pre-training machine learning models can significantly improve performance in scientific applications by adapting them to different tasks with fewer examples and across various physics problems.


CycleNet: Rethinking Cycle Consistency in Text-Guided Diffusion for Image Manipulation

Sihan Xu, Ziqiao Ma, Yidong Huang, Honglak Lee, Joyce Chai

https://openreview.net/forum?id=z9d9DsjAPH

Keywords: Image to image translation, latent diffusion models, conditional diffusion models

Compressor summary: Cyclenet is a method that uses cycle consistency to improve image synthesis with diffusion models, enabling consistent unpaired image-to-image translation and robust generation of high-quality images.


A Regularized Conditional GAN for Posterior Sampling in Image Recovery Problems

Matthew C Bendel, Rizwan Ahmad, Philip Schniter

https://openreview.net/forum?id=z4vKRmq7UO

Keywords: Generative adversarial network, inverse problems, posterior sampling, cGAN, GAN

Compressor summary: The paper proposes a fast method to generate multiple high-quality images from noisy measurements using a conditional Wasserstein GAN with regularization and shows its effectiveness in MRI and inpainting tasks.


Joint Learning of Label and Environment Causal Independence for Graph Out-of-Distribution Generalization

Shurui Gui, Meng Liu, Xiner Li, Youzhi Luo, Shuiwang Ji

https://openreview.net/forum?id=z3HACY5CMa

Keywords: deep learning, graph neural network, out-of-distribution generalization, distribution shift

Compressor summary: The paper introduces LECI, a method to improve graph out-of-distribution generalization by using label and environment information and developing an adversarial training strategy.


Online List Labeling with Predictions

Samuel McCauley, Benjamin Moseley, Aidin Niaparast, Shikha Singh

https://openreview.net/forum?id=z37ki6nqAY

Keywords: Algorithms with Predictions, Data Structures, Learned Indices, Online List Labeling, Resource Allocation, Beyond Worst Case Analysis

Compressor summary: The paper proposes a new list labeling data structure that leverages predictions to improve performance and provides theoretical guarantees for both worst-case and stochastic error models.


Thin and deep Gaussian processes

Daniel Augusto de Souza, Alexander V Nikitin, S. T. John, Magnus Ross, Mauricio A Álvarez, Marc Peter Deisenroth, João Paulo Pordeus Gomes, Diego Mesquita, César Lincoln Mattos

https://openreview.net/forum?id=z2BHMLA8pM

Keywords: Gaussian Processes, Deep Gaussian Processes, non-stationary kernels

Compressor summary: TDGP combines the benefits of deep GPs and shallow GPs by learning interpretable low-dimensional embeddings while maintaining the flexibility to adjust kernel hyperparameters, making it better suited for uncertainty quantification than previous models.


Pre-RMSNorm and Pre-CRMSNorm Transformers: Equivalent and Efficient Pre-LN Transformers

Zixuan Jiang, Jiaqi Gu, Hanqing Zhu, David Z. Pan

https://openreview.net/forum?id=z06npyCwDq

Keywords: Transformer, Normalization, Layer Normalization, RMSNorm, Efficient Machine Learning

Compressor summary: The paragraph discusses different normalization techniques for Transformers and proposes a solution to unify and improve their efficiency without compromising performance.


SparseProp: Efficient Event-Based Simulation and Training of Sparse Recurrent Spiking Neural Networks

Rainer Engelken

https://openreview.net/forum?id=yzZbwQPkmP

Keywords: spiking networks, event-based simulation, sparse networks, backpropagation, algorithm, neuroscience

Compressor summary: SparseProp is a novel event-based algorithm that simulates and trains sparse SNNs with reduced computational cost, enabling efficient and exact simulations of large spiking networks.


What Distributions are Robust to Indiscriminate Poisoning Attacks for Linear Learners?

Fnu Suya, Xiao Zhang, Yuan Tian, David Evans

https://openreview.net/forum?id=yyLFUPNEiT

Keywords: poisoning attacks; adversarial machine learning; machine learning security

Compressor summary: The study explores how linear learners can resist indiscriminate poisoning attacks on some datasets with well-separated and low-variance data distributions.


Revisiting Adversarial Robustness Distillation from the Perspective of Robust Fairness

Xinli Yue, Ningping Mou, Qian Wang, Lingchen Zhao

https://openreview.net/forum?id=ywrPcBEXdC

Keywords: Deep Learning, Knowledge Distillation, Adversarial Training, Fairness

Compressor summary: This paper introduces Fair-ARD, a novel framework to enhance robust fairness in student models by adjusting the weights of difficult classes during adversarial robustness distillation, and shows its effectiveness in improving both overall and class-wise robustness.


Computing Optimal Equilibria and Mechanisms via Learning in Zero-Sum Extensive-Form Games

Brian Hu Zhang, Gabriele Farina, Ioannis Anagnostides, Federico Cacciamani, Stephen Marcus McAleer, Andreas Alexander Haupt, Andrea Celli, Nicola Gatti, Vincent Conitzer, Tuomas Sandholm

https://openreview.net/forum?id=yw1v4RqvPk

Keywords: extensive-form games, deep reinforcement learning, mechanism design, correlated equilibria

Compressor summary: A new method for finding optimal equilibria in extensive-form games is presented, which applies existing zero-sum game learning techniques and achieves state-of-the-art results in tabular games and auction design.


Would I have gotten that reward? Long-term credit assignment by counterfactual contribution analysis

Alexander Meulemans, Simon Schug, Seijin Kobayashi, Nathaniel Daw, Greg Wayne

https://openreview.net/forum?id=yvqqkOn9Pi

Keywords: Reinforcement learning, Long-term credit assignment, contribution analysis, hindsight credit assignment, policy gradient methods

Compressor summary: COCOA is a new model-based credit assignment algorithm that measures the contribution of actions to future rewards by answering counterfactual questions, improving sample efficiency in reinforcement learning.


Canonical normalizing flows for manifold learning

Kyriakos Flouris, Ender Konukoglu

https://openreview.net/forum?id=yubwSWol6K

Keywords: manifold learning flows, normalizing flows, optimization, orthogonalization, sparsity, sparse learning, generative modeling, Riemannian manifold, geometry, metric tensor, orthogonal basis

Compressor summary: The proposed canonical manifold learning flow method generates a sparse and orthogonal low-dimensional representation of the data that improves the efficiency and accuracy of generative modeling techniques compared to other methods.


Epidemic Learning: Boosting Decentralized Learning with Randomized Communication

Martijn De Vos, Sadegh Farhadkhani, Rachid Guerraoui, Anne-marie Kermarrec, Rafael Pires, Rishi Sharma

https://openreview.net/forum?id=ytrhsvGP0r

Keywords: Epidemic, Decentralized Learning, Randomized Communication, Peer sampling

Compressor summary: Epidemic Learning (EL) is a decentralized learning algorithm that uses changing communication topologies to converge faster than conventional approaches, achieving better performance with less communication volume.


Expressivity-Preserving GNN Simulation

Fabian Jogl, Maximilian Thiessen, Thomas Gärtner

https://openreview.net/forum?id=ytTfonl9Wd

Keywords: Graph Neural Networks, GNNs, Graphs, Message Passing, Expressiveness, Graph Transformations, Message Passing Graph Neural Networks

Compressor summary: The paragraph discusses how graph transformations enable message passing operations for simulating graph neural networks (GNNs) without losing expressivity and improving code optimization, while distinguishing between weak and strong simulation methods.


A Unifying Perspective on Multi-Calibration: Game Dynamics for Multi-Objective Learning

Nika Haghtalab, Michael Jordan, Eric Zhao

https://openreview.net/forum?id=ysqlhW0v26

Keywords: multicalibration, multi-objective learning, learning theory, calibration, fairness, games

Compressor summary: The authors present a unified framework for designing and analyzing multi-calibrated predictors by using game dynamics and achieving better guarantees than existing methods.


Emergent Correspondence from Image Diffusion

Luming Tang, Menglin Jia, Qianqian Wang, Cheng Perng Phoo, Bharath Hariharan

https://openreview.net/forum?id=ypOiXjdfnU

Keywords: Correspondence, Diffusion Model

Compressor summary: The paper introduces DIFT, a method to establish correspondences between images using diffusion models without explicit supervision, and shows that it can outperform weakly-supervised methods and off-the-shelf features on various tasks.


CaMP: Causal Multi-policy Planning for Interactive Navigation in Multi-room Scenes

Xiaohan Wang, Yuehu Liu, Xinhang Song, Beibei Wang, Shuqiang Jiang

https://openreview.net/forum?id=yoZTVn0T50

Keywords: Embodied AI, Interactive Navigation, Causal Reinforcement Learning, Hierarchical Reinforcement Learning

Compressor summary: The paper introduces a causal diagram for visual navigation in complex scenes, proposes a multi-policy model to explore counterfactual interactions and reduce exploration, and presents a large-scale dataset for evaluation.


TOA: Task-oriented Active VQA

Xiaoying Xing, Mingfu Liang, Ying Wu

https://openreview.net/forum?id=yoAmURKDJi

Keywords: knowledge-based visual question answering, task-oriented, active image understanding, large language model, visual reasoning, multi-round dialogue

Compressor summary: The authors propose a new method for knowledge-based visual question answering using large language models that actively collects relevant visual evidence to verify their hypotheses, improving performance and interpretability on open-ended datasets.


Covariance-adaptive best arm identification

El Mehdi Saad, Gilles Blanchard, Nicolas Verzelen

https://openreview.net/forum?id=ymHM1qRUeb

Keywords: Multi-armed bandits, Best-arm identification, Adaptive identification

Compressor summary: The paper proposes new algorithms for identifying the best arm in a multi-armed bandit model with dependent arms and covariance, improving on standard methods in applications like clinical trials.


Model-Based Control with Sparse Neural Dynamics

Ziang Liu, Genggeng Zhou, Jeff He, Tobia Marcucci, Li Fei-Fei, Jiajun Wu, Yunzhu Li

https://openreview.net/forum?id=ymBG2xs9Zf

Keywords: model learning, model-based control, neural network sparsification, mixed-integer programming, trajectory optimization

Compressor summary: The paper proposes a new framework for learning and controlling with deep neural networks by sparsifying them and using efficient optimization algorithms, achieving better performance in tasks involving contact dynamics.


Understanding How Consistency Works in Federated Learning via Stage-wise Relaxed Initialization

Yan Sun, Li Shen, Dacheng Tao

https://openreview.net/forum?id=ylPX5D7It7

Keywords: federated learning, local consistency, personalized initialization, excess risk

Compressor summary: The paper proposes FedInit, an efficient federated learning algorithm that uses personalized relaxed initialization to reduce the "client drift" problem and improve performance in distributed training of global models.


Deep Momentum Multi-Marginal Schrödinger Bridge

Tianrong Chen, Guan-Horng Liu, Molei Tao, Evangelos Theodorou

https://openreview.net/forum?id=ykvvv0gc4R

Keywords: Schrödinger Bridge, Trajectory Inference, Optimal Transport

Compressor summary: The paper presents DMSB, a novel computational framework for inferring high-dimensional multi-marginal trajectories from unlabeled samples at coarse time intervals, which outperforms baselines and can reconstruct velocity distributions from position snapshots.


DiffTraj: Generating GPS Trajectory with Diffusion Probabilistic Model

Yuanshao Zhu, Yongchao Ye, Shiyao Zhang, Xiangyu Zhao, James Yu

https://openreview.net/forum?id=ykMdzevPkJ

Keywords: Trajectory Generation, Diffusion Model, Urban Computing, Spatial-temporal Data Mining

Compressor summary: The authors propose a model called DiffTraj that generates realistic and privacy-preserving GPS trajectories by denoising white noise using spatial-temporal features and a deep neural network.


Mind the spikes: Benign overfitting of kernels and neural networks in fixed dimension

Moritz Haas, David Holzmüller, Ulrike von Luxburg, Ingo Steinwart

https://openreview.net/forum?id=yjYwbZBJyl

Keywords: benign overfitting, kernels, neural tangent kernel, consistency, learning theory

Compressor summary: The key to benign overfitting is not high dimension but large derivatives in smooth estimators; this can be achieved by adding small fluctuations to the activation function of wide neural networks.


OBJECT 3DIT: Language-guided 3D-aware Image Editing

Oscar Michel, Anand Bhattad, Eli VanderBilt, Ranjay Krishna, Aniruddha Kembhavi, Tanmay Gupta

https://openreview.net/forum?id=yjWVd8Fhqt

Keywords: computer vision, image editing, generative modeling, diffusion models, 3D

Compressor summary: The authors propose a new task called language-guided 3D-aware editing, where edits are done according to language instructions while preserving the 3D scene, and release a large dataset for this task along with models that can perform well on it.


E2PNet: Event to Point Cloud Registration with Spatio-Temporal Representation Learning

Xiuhong Lin, Changjie Qiu, zhipeng cai, Siqi Shen, Yu Zang, Weiquan Liu, Xuesheng Bian, Matthias Müller, Cheng Wang

https://openreview.net/forum?id=yiehppUCO2

Keywords: event camera, 2D-3D registration, representation learning

Compressor summary: E2PNet is a novel method that registers 2D RGB images to 3D point clouds using event cameras, which improves robustness and enables other vision tasks.


Finite-Time Analysis of Whittle Index based Q-Learning for Restless Multi-Armed Bandits with Neural Network Function Approximation

GUOJUN XIONG, Jian Li

https://openreview.net/forum?id=yhNHpLWJDl

Keywords: Restless bandits, Whittle index policy, Q-learning, Two-timescale stochastic approximation, Neural network function approximation

Compressor summary: Neural-Q-Whittle is a Q-learning algorithm that uses neural networks and Whittle indices to solve intractable restless multi-armed bandits problems with an empirical convergence rate of $\mathcal{O}(1/k^{2/3})$.


Visual Programming for Step-by-Step Text-to-Image Generation and Evaluation

Jaemin Cho, Abhay Zala, Mohit Bansal

https://openreview.net/forum?id=yhBFG9Y85R

Keywords: text-to-image generation; visual programming; text-to-image evaluation; step-by-step generation; interpretability; explainability

Compressor summary: The paragraph introduces two novel visual programming frameworks, VPGen and VPEval, for text-to-image tasks that improve spatial control, handle predefined object classes, and provide interpretable and explainable evaluations with experts in different skills.


FiGURe: Simple and Efficient Unsupervised Node Representations with Filter Augmentations

Chanakya Ekbote, Ajinkya Deshpande, Arun Iyer, SUNDARARAJAN SELLAMANICKAM, Ramakrishna B Bairi

https://openreview.net/forum?id=yh0OkiUk5h

Keywords: Graph Neural Networks, Unsupervised Representation Learning, Graph Filters

Compressor summary: The paper introduces FiGURe, a method to improve node representations using filter-based augmentations and random Fourier feature projections, achieving up to 4.4% better performance on downstream tasks than existing unsupervised models.


Uncertainty Quantification over Graph with Conformalized Graph Neural Networks

Kexin Huang, Ying Jin, Emmanuel Candes, Jure Leskovec

https://openreview.net/forum?id=ygjQCOyNfh

Keywords: Graph Neural Networks, Conformal Prediction, Uncertainty Quantification

Compressor summary: CF-GNN is a graph-based model that provides guaranteed uncertainty estimates by extending conformal prediction to graph data and using a topology-aware output correction model to reduce the prediction set size.


A generative model of the hippocampal formation trained with theta driven local learning rules

Tom George, Kim Stachenfeld, Caswell Barry, Claudia Clopath, Tomoki Fukai

https://openreview.net/forum?id=yft4JlxsRf

Keywords: hippocampus, path integration, local learning, generative models, oscillations, inference, Helmholtz machine, wake-sleep

Compressor summary: The authors propose a biologically plausible model of the hippocampus that uses theta-band oscillations and a wake-sleep algorithm to learn and generate realistic sensory predictions and path integration.


PAC-Bayesian Spectrally-Normalized Bounds for Adversarially Robust Generalization

Jiancong Xiao, Ruoyu Sun, Zhi-Quan Luo

https://openreview.net/forum?id=ydKWoqWZ3t

Keywords: Pac-Bayes, Adversarial Robustness, Generalization

Compressor summary: The paper proposes a new robust generalization bound for deep neural networks that does not require additional assumptions and is tighter than existing bounds, offering insights into the causes of poor robust generalization.


Understanding the detrimental class-level effects of data augmentation

Polina Kirichenko, Mark Ibrahim, Randall Balestriero, Diane Bouchacourt, Shanmukha Ramakrishna Vedantam, Hamed Firooz, Andrew Gordon Wilson

https://openreview.net/forum?id=yageaKlk7S

Keywords: data augmentation, class-dependent bias

Compressor summary: The paper studies how data augmentation affects image classification, especially for classes with ambiguous or fine-grained distinctions, and proposes class-conditional augmentation strategies to improve performance.


Complexity of Derivative-Free Policy Optimization for Structured $\mathcal{H}_\infty$ Control

Xingang Guo, Darioush Keivan, Geir Dullerud, Peter Seiler, Bin Hu

https://openreview.net/forum?id=yaJ4vZPnHX

Keywords: Structured $\mathcal{H}_\infty$ Control, Nonsmooth Optimization, Complexity Analysis

Compressor summary: This paper presents new theoretical results on the difficulty of optimizing policies without derivatives for a class of robust control tasks, and shows how to find approximate solutions using zeroth-order oracles.


Star-Shaped Denoising Diffusion Probabilistic Models

Andrey Okhotin, Dmitry Molchanov, Arkhipkin Sergeevich Vladimir, Grigory Bartosh, Viktor Ohanesian, Aibek Alanov, Dmitry P. Vetrov

https://openreview.net/forum?id=yYUdgbmhh9

Keywords: Generative models, Diffusion, Exponential Family

Compressor summary: The paper introduces Star-Shaped Diffusion Probabilistic Models, which enable generative modeling with non-Gaussian distributions using a simpler and more efficient approach than existing methods.


Diverse Shape Completion via Style Modulated Generative Adversarial Networks

Wesley Khademi, Li Fuxin

https://openreview.net/forum?id=yVMlYSL1Bp

Keywords: multimodal shape completion, point cloud completion, 3d shape generation, generative modeling, generative adversarial networks

Compressor summary: The paper proposes a new network that can generate diverse, plausible 3D shape completions from partial observations using style modulation and multiple discriminators to prevent mode collapse.


PGDiff: Guiding Diffusion Models for Versatile Face Restoration via Partial Guidance

Peiqing Yang, Shangchen Zhou, Qingyi Tao, Chen Change Loy

https://openreview.net/forum?id=yThjbzhIUP

Keywords: Face Restoration, Diffusion

Compressor summary: The paper introduces a new restoration approach called partial guidance that adapts to real-world degradations by modeling desired properties of high-quality images and applying them during the reverse diffusion process, achieving good results across various tasks.


Fast Conditional Mixing of MCMC Algorithms for Non-log-concave Distributions

Xiang Cheng, Bohan Wang, Jingzhao Zhang, Yusong Zhu

https://openreview.net/forum?id=yT0f93CeTw

Keywords: Sampling, MCMC, Conditional Mixing, Non-log-concave Distributions

Compressor summary: The paragraph explains that MCMC algorithms can work efficiently when a subset of the state space has Poincaré-style inequality, even if global mixing is slow, and this can help in various applications like sampling from mixtures of Gaussians or parameter estimation.


Three-Way Trade-Off in Multi-Objective Learning: Optimization, Generalization and Conflict-Avoidance

Lisha Chen, Heshan Devaka Fernando, Yiming Ying, Tianyi Chen

https://openreview.net/forum?id=yPkbdJxQ0o

Keywords: Generalization, algorithm stability, multi-objective optimization, gradient conflict

Compressor summary: The MoDo algorithm studies the stability of dynamic weighting methods for multi-objective learning and reveals how updating along conflict-avoidant directions may hinder optimal generalization performance.


Generalized Information-theoretic Multi-view Clustering

Weitian Huang, Sirui Yang, Hongmin Cai

https://openreview.net/forum?id=yN6NHZOXkg

Keywords: information bottleneck, multi-view clustering, variational autoencoders

Compressor summary: The paper proposes an information-based multi-view clustering model that uses deep neural networks and Stochastic Gradient Variational Bayes for representation learning and clustering, achieving better results than existing methods.


Mechanism Design for Collaborative Normal Mean Estimation

Yiding Chen, Jerry Zhu, Kirthevasan Kandasamy

https://openreview.net/forum?id=yKCLfOOIL7

Keywords: Mechanism design, statistical minimax estimation, federated learning

Compressor summary: The paper proposes a novel mechanism for estimating the mean of a normal distribution collaboratively among strategic agents, using corrupted data sharing and minimax optimal estimators to incentivize truthful reporting and reduce estimation errors.


Towards a Unified Analysis of Kernel-based Methods Under Covariate Shift

Xingdong Feng, Xin HE, Caixing Wang, Chao Wang, Jingnan Zhang

https://openreview.net/forum?id=yIcCkMUCtL

Keywords: kernel methods, covariate shift, reproducing kernel Hilbert space (RKHS)

Compressor summary: The paper proposes a general analysis of nonparametric methods under covariate shift in RKHS, with applications to various learning tasks and theoretical and numerical results.


HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face

Yongliang Shen, Kaitao Song, Xu Tan, Dongsheng Li, Weiming Lu, Yueting Zhuang

https://openreview.net/forum?id=yHdTscY6Ci

Keywords: LLM, ChatGPT, Hugging Face, Autonomous LLM

Compressor summary: HuggingGPT is an LLM-powered agent that uses ChatGPT to connect various AI models in Hugging Face to solve complicated AI tasks across different domains and modalities, moving toward artificial general intelligence.


What Can We Learn from Unlearnable Datasets?

Pedro Sandoval-Segura, Vasu Singla, Jonas Geiping, Micah Goldblum, Tom Goldstein

https://openreview.net/forum?id=yGs9vTRjaE

Keywords: data poisoning, poisons, unlearnable dataset, data protection, imperceptible perturbations, adversarial machine learning

Compressor summary: Unlearnable dataset methods may not protect data privacy as they can enable neural networks to learn useful features, and linear separability of perturbations is not a necessary condition for this.


Augmented Memory Replay-based Continual Learning Approaches for Network Intrusion Detection

Suresh kumar Amalapuram, Sumohana S. Channappayya, Bheemarjuna Tamma

https://openreview.net/forum?id=yGLokEhdh9

Keywords: Continual learning, Class imbalance, scalability, Network intrusion detection, and Cybersecurity

Compressor summary: The paper proposes two methods to improve continual learning-based intrusion detection: one for class imbalance and another for reducing computations, which perform better than baselines on various benchmarks.


Encoding Time-Series Explanations through Self-Supervised Model Behavior Consistency

Owen Queen, Thomas Hartvigsen, Teddy Koker, Huan He, Theodoros Tsiligkaridis, Marinka Zitnik

https://openreview.net/forum?id=yEfmhgwslQ

Keywords: Explainability, Interpretability, Time Series, Explanations, Temporal patterns, Model Understanding, Latent space, Self-supervised learning

Compressor summary: TimeX is a model that trains an interpretable surrogate to mimic a pretrained time series model and provides discrete attribution maps and latent space explanations for interpreting time series predictions.


Convergence of Adam Under Relaxed Assumptions

Haochuan Li, Alexander Rakhlin, Ali Jadbabaie

https://openreview.net/forum?id=yEewbkBNzi

Keywords: Non-convex optimization, Adam, Convergence, Variance reduction

Compressor summary: The paper proves that the Adam algorithm for optimization converges to stationary points under realistic conditions and introduces a variance-reduced version with faster convergence.


Advancing Bayesian Optimization via Learning Correlated Latent Space

Seunghun Lee, Jaewon Chu, Sihyeon Kim, Juyeon Ko, Hyunwoo J. Kim

https://openreview.net/forum?id=yE62KM4qsO

Keywords: Bayesian optimization, smoothness regularization, variational autoencoder

Compressor summary: Correlated latent space Bayesian Optimization (CoBO) is a method that learns correlated latent spaces to reduce the gap between optimization in the latent space and the input space, achieving high performance on discrete data tasks with limited function evaluations.


Closing the gap between the upper bound and lower bound of Adam's iteration complexity

Bohan Wang, Jingwen Fu, Huishuai Zhang, Nanning Zheng, Wei Chen

https://openreview.net/forum?id=yDvb3mlogA

Keywords: Adam, Convergence, Upper Bound, Lower Bound

Compressor summary: This paper derives a new convergence guarantee for Adam optimization with smooth conditions and bounded noise variance, filling a gap in existing literature and providing a tight upper bound.


Credal Marginal MAP

Radu Marinescu, Debarun Bhattacharjya, Junkyu Lee, Fabio Cozman, Alexander G. Gray

https://openreview.net/forum?id=yCBqKTvYe9

Keywords: graphical models, credal networks, probabilistic inference

Compressor summary: Credal Marginal MAP inference is explored in this paper, with new exact and approximation methods proposed and evaluated on various problems.


Siamese Masked Autoencoders

Agrim Gupta, Jiajun Wu, Jia Deng, Li Fei-Fei

https://openreview.net/forum?id=yC3q7vInux

Keywords: Representation Learning, Visual Correspondence, Self-supervised learning, Videos

Compressor summary: Siamese Masked Autoencoders (SiamMAE) is a method for learning visual correspondence from videos by predicting missing patches in one frame based on another frame with masked patches, focusing on object motion and achieving competitive results on various tasks.


Balance, Imbalance, and Rebalance: Understanding Robust Overfitting from a Minimax Game Perspective

Yifei Wang, Liangchen Li, Jiansheng Yang, Zhouchen Lin, Yisen Wang

https://openreview.net/forum?id=yBoVwpGa5E

Keywords: Adversarial Training

Compressor summary: Adversarial Training (AT) has become arguably the state-of-the-art algorithm for extracting robust features. However, researchers recently notice that AT suffers from severe robust overfitting problems, particularly after learning rate (LR) decay. In this paper, we explain this phenomenon by viewing adversarial training as a dynamic minimax game between the model trainer and the attacker. Specifically, we analyze how LR decay breaks the balance between the minimax game by empowering the trainer with a stronger memorization ability, and show such imbalance induces robust overfitting as a result of memorizing non-robust features. We validate this understanding with extensive experiments, and provide a holistic view of robust overfitting from the dynamics of both the two game players. This understanding further inspires us to alleviate robust overfitting by rebalancing the two players by either regularizing the trainer's capacity or improving the attack strength. Experiments show that the proposed ReBalanced Adversarial Training (ReBAT) can attain good robustness and does not suffer from robust overfitting even after very long training. Code is available at https://github.com/PKU-ML/ReBAT.


Error Discovery By Clustering Influence Embeddings

Fulton Wang, Julius Adebayo, Sarah Tan, Diego Garcia-Olano, Narine Kokhlikyan

https://openreview.net/forum?id=yBVLXvJ1sb

Keywords: Debugging, interpretability, influence functions

Compressor summary: The authors introduce InfEmden, a slice discovery method that finds coherent groups of test examples where a model under-performs by using influence embeddings and K-Means clustering.


Operation-Level Early Stopping for Robustifying Differentiable NAS

Shen Jiang, Zipeng Ji, Guanghui Zhu, Chunfeng Yuan, Yihua Huang

https://openreview.net/forum?id=yAOwkf4FyL

Keywords: Differentiable neural architecture search; Image classification; Failure of DARTS

Compressor summary: The paper proposes a method called operation-level early stopping (OLES) to improve Differentiable NAS (DARTS) by addressing the overfitting issue caused by skip connections in neural architecture search.


SNEkhorn: Dimension Reduction with Symmetric Entropic Affinities

Hugues Van Assel, Titouan Vayer, Rémi Flamary, Nicolas Courty

https://openreview.net/forum?id=y9U0IJ2uFr

Keywords: Dimension Reduction, Optimal Transport, Affinities

Compressor summary: The paragraph discusses a novel way of symmetrizing entropic affinities using optimal transport, which improves clustering performance and robustness in dimensionality reduction algorithms like t-SNE.


Private Everlasting Prediction

Moni Naor, Kobbi Nissim, Uri Stemmer, Chao Yan

https://openreview.net/forum?id=y8UAQQHVTX

Keywords: Differential privacy, private learning, private prediction

Compressor summary: The paper proposes private everlasting prediction, a method to protect the privacy of both training data and queries in answering a stream of classification questions, with lower sample complexity than existing methods.


On the Importance of Exploration for Generalization in Reinforcement Learning

Yiding Jiang, J Zico Kolter, Roberta Raileanu

https://openreview.net/forum?id=y5duN2j9s6

Keywords: reinforcement learning, generalization, procgen, crafter

Compressor summary: The paper proposes EDE, a method that encourages exploration in deep RL by using an ensemble of Q-value distributions to handle epistemic uncertainty and improve generalization to new environments.


CSOT: Curriculum and Structure-Aware Optimal Transport for Learning with Noisy Labels

Wanxing Chang, Ye Shi, Jingya Wang

https://openreview.net/forum?id=y50AnAbKp1

Keywords: Learning with Noisy Labels, Optimal Transport, Curriculum Learning

Compressor summary: The paper proposes a novel optimal transport method, CSOT, to address challenges in learning with noisy labels by considering sample distribution structure and assigning reliable labels incrementally.


Learning Causal Models under Independent Changes

Sarah Mameche, David Kaltenpoth, Jilles Vreeken

https://openreview.net/forum?id=y0OlQSZsyp

Keywords: independent mechanisms, causal discovery, information theory, gaussian processes

Compressor summary: The paper proposes a method to discover causal relationships in systems with changing components using Gaussian Process models and independence assumptions.


WITRAN: Water-wave Information Transmission and Recurrent Acceleration Network for Long-range Time Series Forecasting

Yuxin Jia, Youfang Lin, Xinyan Hao, Yan Lin, Shengnan Guo, Huaiyu Wan

https://openreview.net/forum?id=y08bkEtNBK

Keywords: long-range time series forecasting, information transmission, long- and short-term repetitive patterns, global and local correlations

Compressor summary: The WITRAN framework captures semantic information for long-range time series forecasting using a bi-granular information transmission and horizontal vertical gated selective unit, while reducing time and memory complexity compared to previous methods.


Molecule Joint Auto-Encoding: Trajectory Pretraining with 2D and 3D Diffusion

weitao Du, Jiujiu Chen, Xuecang Zhang, Zhi-Ming Ma, Shengchao Liu

https://openreview.net/forum?id=xzmaFfw6oh

Keywords: Molecule Joint Auto-encoding, Molecule Joint Self-supervised Learning, Markov processes, contrastive learning, molecule representation learning

Compressor summary: The proposed method, MoleculeJAE, learns both 2D and 3D molecular information using a pretraining technique, achieving state-of-the-art results on various tasks.


Color Equivariant Convolutional Networks

Attila Lengyel, Ombretta Strafforello, Robert-Jan Bruintjes, Alexander Gielisse, Jan van Gemert

https://openreview.net/forum?id=xz8j3r3oUA

Keywords: color equivariance, equivariance, color robustness, equivariant convolutions

Compressor summary: The paper introduces CEConvs, a novel deep learning building block that allows shape feature sharing across different colors while preserving important color information, improving robustness to color changes and performance on various tasks.


Towards Hybrid-grained Feature Interaction Selection for Deep Sparse Network

Fuyuan Lyu, Xing Tang, Dugang Liu, Chen Ma, Weihong Luo, Liang Chen, xiuqiang He, Xue Liu

https://openreview.net/forum?id=xxfHMqNcum

Keywords: Feature Interaction Search, Deep Sparse Network

Compressor summary: The paper introduces a hybrid-grained feature interaction selection approach for deep sparse networks that targets both feature field and feature value, and shows its effectiveness on three benchmark datasets.


BasisFormer: Attention-based Time Series Forecasting with Learnable and Interpretable Basis

Zelin Ni, Hang Yu, Shizhan Liu, Jianguo Li, Weiyao Lin

https://openreview.net/forum?id=xx3qRKvG0T

Keywords: time series forecasting, basis learning, self-supervised learning

Compressor summary: BasisFormer is a time series forecasting architecture that uses adaptive self-supervised learning, cross-attention, and similarity coefficients to create tailored bases for accurate predictions.


Unsupervised Polychromatic Neural Representation for CT Metal Artifact Reduction

Qing Wu, Lixuan Chen, Ce Wang, Hongjiang Wei, S Kevin Zhou, Jingyi Yu, Yuyao Zhang

https://openreview.net/forum?id=xx3QgKyghS

Keywords: Medical Image, Computed Tomography, Metal Arftiacts, Implicit Neural Representation, Unsupervised Learning

Compressor summary: The paragraph describes a novel polychromatic neural representation (Polyner) method to improve computed tomography (CT) imaging with metallic implants by modeling the metal artifact reduction problem from a nonlinear inverse perspective and achieving better performance than supervised methods.


Personalized Dictionary Learning for Heterogeneous Datasets

Geyu Liang, Naichen Shi, Raed Al Kontar, Salar Fattahi

https://openreview.net/forum?id=xw6Szwu4xz

Keywords: Dictionary Learning, Data Heterogeneity, Personalization

Compressor summary: The paper introduces PerDL, a problem where sparse linear representations are learned from heterogeneous datasets with shared and unique features, and proposes PerMA, an efficient meta-algorithm that can recover the global and local dictionaries for various learning tasks.


TIES-Merging: Resolving Interference When Merging Models

Prateek Yadav, Derek Tam, Leshem Choshen, Colin Raffel, Mohit Bansal

https://openreview.net/forum?id=xtaX3WyCj1

Keywords: Model Merging, Fusing, Collaborative Training, Robust Fine-tuning, Federated Learning

Compressor summary: TIES-Merging is a novel method to combine task-specific models into a multitask model by resolving parameter interference and exploiting sign information, which leads to better performance in various settings.


Faster Discrete Convex Function Minimization with Predictions: The M-Convex Case

Taihei Oki, Shinsaku Sakaue

https://openreview.net/forum?id=xtQ9IGRzIW

Keywords: algorithms with predictions, beyond the worst-case analysis of algorithms, time complexity, combinatorial optimization, discrete convex analysis, submodular functions

Compressor summary: The paper presents a framework to accelerate discrete optimization algorithms, especially laminar convex minimization, by using machine-learned predictions.


Adversarial Robustness in Graph Neural Networks: A Hamiltonian Approach

Kai Zhao, Qiyu Kang, Yang Song, Rui She, Sijie Wang, Wee Peng Tay

https://openreview.net/forum?id=xtADRDRsM2

Keywords: adversarial robustness, graph neural networks

Compressor summary: This paper explores how using conservative Hamiltonian neural flows can improve the adversarial robustness of graph neural networks, which are otherwise vulnerable to perturbations affecting both node features and graph topology.


QuIP: 2-Bit Quantization of Large Language Models With Guarantees

Jerry Chee, Yaohui Cai, Volodymyr Kuleshov, Christopher De Sa

https://openreview.net/forum?id=xrk9g5vcXR

Keywords: Quantization, Large Language Models, Adaptive Rounding, Theoretical Guarantees

Compressor summary: The paper introduces QuIP, a new method for post-training parameter quantization in large language models that uses incoherence processing to improve results and reduce bits per weight needed.


FaceComposer: A Unified Model for Versatile Facial Content Creation

Jiayu Wang, Kang Zhao, Yifeng Ma, Shiwei Zhang, Yingya Zhang, Yujun Shen, Deli Zhao, Jingren Zhou

https://openreview.net/forum?id=xrK3QA9mLo

Keywords: diffusion model; talking face generation; face generation

Compressor summary: FaceComposer is a versatile generative model for creating facial content with text control and animation, based on the latent diffusion framework, and it includes a large-scale multi-modal face database and an easy-to-use interface.


Calibrating “Cheap Signals” in Peer Review without a Prior

Yuxuan Lu, Yuqing Kong

https://openreview.net/forum?id=xr3KAzboHY

Keywords: Peer prediction, Peer review, Calibration

Compressor summary: The paper proposes a one-shot noise calibration process for ranking papers fairly by predicting others' scores, which reduces error probability and outperforms average ratings.


Beyond Unimodal: Generalising Neural Processes for Multimodal Uncertainty Estimation

Myong Chol Jung, He Zhao, Joanna Dipnall, Lan Du

https://openreview.net/forum?id=xq1QvViDdW

Keywords: Uncertainty estimation, multimodality, neural processes

Compressor summary: MNPs are a novel approach for estimating uncertainty in multimodal data that adapts Neural Processes, providing robustness and reliability with faster computation.


StableRep: Synthetic Images from Text-to-Image Models Make Strong Visual Representation Learners

Yonglong Tian, Lijie Fan, Phillip Isola, Huiwen Chang, Dilip Krishnan

https://openreview.net/forum?id=xpjsOQtKqx

Keywords: representation learning, synthetic images, text-to-image models

Compressor summary: The paper explores using text-to-image models to generate synthetic images for learning visual representations, and shows that their method outperforms existing methods on large scale datasets.


Fitting trees to $\ell_1$-hyperbolic distances

Joon-Hyeok Yim, Anna Gilbert

https://openreview.net/forum?id=xo2lbfQE8I

Keywords: tree metric fitting, ultrametric fitting, $\ell_1$-hyperbolicity

Compressor summary: The paper proposes an algorithm to fit trees based on hyperbolicity values, which improves upon previous algorithms and reveals differences between real and synthetic data sets.


Michelangelo: Conditional 3D Shape Generation based on Shape-Image-Text Aligned Latent Representation

Zibo Zhao, Wen Liu, Xin Chen, Xianfang Zeng, Rui Wang, Pei Cheng, BIN FU, Tao Chen, Gang YU, Shenghua Gao

https://openreview.net/forum?id=xmxgMij3LY

Keywords: Conditional 3D Shape Generation, Neural 3D Representation, 3D Reconstruction

Compressor summary: The paper proposes a new method for generating 3D shapes from 2D images or texts by aligning the representations of 3D shapes in different modalities and using two models to encode and decode them.


Text Alignment Is An Efficient Unified Model for Massive NLP Tasks

Yuheng Zha, Yichi Yang, Ruichen Li, Zhiting Hu

https://openreview.net/forum?id=xkkBFePoFn

Keywords: Text Alignment, Efficient Unified Model, NLU Tasks, Factual Consistency Evaluation, QA with Unanswerable Question

Compressor summary: The paper proposes a small and efficient text alignment model that performs well on various NLP tasks, including text entailment, similarity, question answering, and factual consistency, outperforming larger LLMs in some cases.


Asymptotics of Bayesian Uncertainty Estimation in Random Features Regression

Youngsoo Baek, Samuel Berchuck, Sayan Mukherjee

https://openreview.net/forum?id=xgzkuTGBTx

Keywords: asymptotics, random features model, Bayesian inference

Compressor summary: This paper studies how the uncertainty of Bayesian models relates to the performance of their maximum likelihood estimates in high-dimensional settings, finding agreement in certain regimes and similarity in others.


Learning a Neuron by a Shallow ReLU Network: Dynamics and Implicit Bias for Correlated Inputs

Dmitry Chistikov, Matthias Englert, Ranko Lazic

https://openreview.net/forum?id=xgY4QcOiEZ

Keywords: implicit bias, implicit regularization, training dynamics, ReLU networks, gradient flow, theoretical analysis

Compressor summary: The paper investigates how one-hidden layer ReLU networks learn a single neuron task by gradient flow and shows that they implicitly minimize the rank of network parameters while converging to zero loss.


Efficient Bayesian Learning Curve Extrapolation using Prior-Data Fitted Networks

Steven Adriaensen, Herilalaina Rakotoarison, Samuel Müller, Frank Hutter

https://openreview.net/forum?id=xgTV6rmH6n

Keywords: learning curve extrapolation, prior-data fitted networks, transformers, Bayesian inference, uncertainty estimation, model selection

Compressor summary: The authors propose LC-PFN, a fast and accurate Bayesian method for learning curve extrapolation using prior-data fitted neural networks, which can improve model selection efficiency by enabling predictive early stopping.


Physics-Informed Bayesian Optimization of Variational Quantum Circuits

Kim Andrea Nicoli, Christopher J. Anders, Lena Funcke, Tobias Hartung, Karl Jansen, Stefan Kuhn, Klaus Robert Muller, Paolo Stornati, Pan Kessel, Shinichi Nakajima

https://openreview.net/forum?id=xfBeVGJwyL

Keywords: Bayesian optimization, Expected improvement, Quantum computing, Variational Quantum Eigensolvers

Compressor summary: The paper presents a new Bayesian optimization method for VQEs that uses a novel kernel and acquisition function to reduce uncertainty and improve performance.


Efficient Symbolic Policy Learning with Differentiable Symbolic Expression

Jiaming Guo, Rui Zhang, Shaohui Peng, Qi Yi, Xing Hu, Ruizhi Chen, Zidong Du, Xishan Zhang, Ling Li, Qi Guo, Yunji Chen

https://openreview.net/forum?id=xdQpmUPNHC

Keywords: reinforcement learning, context variables, symbolic policy

Compressor summary: The paper proposes ESPL, an efficient gradient-based method to learn simple and interpretable symbolic policies from scratch for sequential decision-making tasks using a differentiable symbolic expression and a path selector.


Towards Data-Agnostic Pruning At Initialization: What Makes a Good Sparse Mask?

Hoang Pham, The-Anh Ta, Shiwei Liu, Lichuan Xiang, Dung D. Le, Hongkai Wen, Long Tran-Thanh

https://openreview.net/forum?id=xdOoCWCYaY

Keywords: Pruning Neural Network, Sparsity, Neural Architecture Search

Compressor summary: The paper proposes a new framework and method for pruning neural networks at initialization that balances effective nodes and paths, improving accuracy and computational efficiency compared to existing methods.


Adversarial Resilience in Sequential Prediction via Abstention

Surbhi Goel, Steve Hanneke, Shay Moran, Abhishek Shetty

https://openreview.net/forum?id=xcGhx9FdxM

Keywords: Sequential prediction, adversarial examples, abstention, out-of-distribution, VC Classes

Compressor summary: The paper proposes a new model for sequential prediction that allows learners to abstain from making predictions on adversarial examples and designs learners with better error guarantees than existing methods.


On Evaluating Adversarial Robustness of Large Vision-Language Models

Yunqing Zhao, Tianyu Pang, Chao Du, Xiao Yang, Chongxuan Li, Ngai-man Cheung, Min Lin

https://openreview.net/forum?id=xbbknN9QFs

Keywords: Large Vision-Language Models, Adversarial Robustness

Compressor summary: The paragraph discusses evaluating the robustness of vision-language models (VLMs) to targeted adversarial examples and their potential security flaws before deployment.


Practical Equivariances via Relational Conditional Neural Processes

Daolang Huang, Manuel Haussmann, Ulpu Remes, S. T. John, Grégoire Clarté, Kevin Sebastian Luck, Samuel Kaski, Luigi Acerbi

https://openreview.net/forum?id=xax5eWeObb

Keywords: neural processes, equivariance, Gaussian processes

Compressor summary: Relational Conditional Neural Processes (RCNPs) are a new model that improves the performance of equivariant neural processes in higher dimensions for various machine learning tasks.


For SALE: State-Action Representation Learning for Deep Reinforcement Learning

Scott Fujimoto, Wei-Di Chang, Edward J. Smith, Shixiang Shane Gu, Doina Precup, David Meger

https://openreview.net/forum?id=xZvGrzRq17

Keywords: Deep reinforcement learning, representation learning

Compressor summary: The paper introduces SALE, a method for learning embeddings from low-level states that improve representation learning in reinforcement learning, and combines it with TD3 to create TD7, which performs better than existing continuous control algorithms on OpenAI gym benchmark tasks.


Learning Curves for Noisy Heterogeneous Feature-Subsampled Ridge Ensembles

Benjamin Samuel Ruben, Cengiz Pehlevan

https://openreview.net/forum?id=xXfDB8kJUs

Keywords: ridge regression, ensembling methods

Compressor summary: Feature bagging reduces prediction variance by combining predictions of many estimators trained on subsets or projections of features, and can mitigate double-descent in noisy least-squares ridge ensembles.


Robust Data Pruning under Label Noise via Maximizing Re-labeling Accuracy

Dongmin Park, Seola Choi, Doyoung Kim, Hwanjun Song, Jae-Gil Lee

https://openreview.net/forum?id=xWCp0uLcpG

Keywords: Data Pruning, Data Subset Selection, Noisy Labels, Relabeling, Self-training

Compressor summary: Prune4Rel is a novel data pruning algorithm that maximizes re-labeling accuracy and generalization performance by finding a subset of training examples with high prediction confidence of their neighbors in the subset.


Fair Graph Distillation

Qizhang Feng, Zhimeng Jiang, Ruiquan Li, Yicheng Wang, Na Zou, Jiang Bian, Xia Hu

https://openreview.net/forum?id=xW0ayZxPWs

Keywords: Graph Distillation, Algorithmic Fairness

Compressor summary: The paper proposes a method to create smaller graphs that are both accurate and fair for graph neural networks, by developing a new metric called coherence and using a bi-level optimization algorithm.


Rethinking Incentives in Recommender Systems: Are Monotone Rewards Always Beneficial?

Fan Yao, Chuanhao Li, Karthik Abinav Sankararaman, Yiming Liao, Yan Zhu, Qifan Wang, Hongning Wang, Haifeng Xu

https://openreview.net/forum?id=xUyBP16Q5J

Keywords: Recommender system, Mechanism design, Potential function, Optimization

Compressor summary: The text discusses how online content recommendation platforms affect media creators' choices and welfare, and proposes a new reward mechanism called Backward Rewarding Mechanisms to improve the platform's design.


Compact Neural Volumetric Video Representations with Dynamic Codebooks

Haoyu Guo, Sida Peng, Yunzhi Yan, Linzhan Mou, Yujun Shen, Hujun Bao, Xiaowei Zhou

https://openreview.net/forum?id=xTgM7XLN9P

Keywords: Computer Vision, 3D Vision, Volumetric Video

Compressor summary: The paper introduces a novel neural representation called dynamic codebook that compresses volumetric videos by merging similar features and compensating with dynamic codes, achieving better quality and storage efficiency.


Act As You Wish: Fine-Grained Control of Motion Diffusion Model with Hierarchical Semantic Graphs

Peng Jin, Yang Wu, Yanbo Fan, Zhongqian Sun, Yang Wei, Li Yuan

https://openreview.net/forum?id=xSEhb2j3TK

Keywords: Text-driven Motion Synthesis, Diffusion Models, Graph networks

Compressor summary: The paper proposes using hierarchical semantic graphs to generate fine-grained human motion from text descriptions, improving over existing methods that may focus too much on action names and lack detail.


Robust Model Reasoning and Fitting via Dual Sparsity Pursuit

Xingyu Jiang, Jiayi Ma

https://openreview.net/forum?id=xRfTcZdQxq

Keywords: Model reasoning; Model fitting; Outliers; Sparse subspace learning; Feature matching

Compressor summary: The paper proposes a unified optimization model for outlier rejection, true model reasoning and parameter estimation, which uses sparse subspace recovery to search for independent bases in an over-embedded data space.


The Tunnel Effect: Building Data Representations in Deep Neural Networks

Wojciech Masarczyk, Mateusz Ostaszewski, Ehsan Imani, Razvan Pascanu, Piotr Miłoś, Tomasz Trzcinski

https://openreview.net/forum?id=xQOHOpe1Fv

Keywords: representation learning, continual learning, training dynamics

Compressor summary: The paper investigates how deep neural networks split into two parts during image classification training, where the initial layers create separable representations and the latter layers, called "the tunnel," compress them without significantly affecting performance.


Stability of Random Forests and Coverage of Random-Forest Prediction Intervals

Yan Wang, Huaiqing Wu, Dan Nettleton

https://openreview.net/forum?id=xPqINp0Eu1

Keywords: Stability, Prediction Intervals, Random Forests

Compressor summary: The paper shows how random forests are stable and can provide good interval predictions using out-of-bag error and weak assumptions.


Learning DAGs from Data with Few Root Causes

Panagiotis Misiakos, Chris Wendler, Markus Püschel

https://openreview.net/forum?id=xPLaXSuSvQ

Keywords: directed acyclic graph, few root causes, structural equation models, linear SEMs, additive noise

Compressor summary: The paper proposes a new algorithm for learning directed acyclic graphs (DAGs) from linear structural equation models (SEMs) with fewer root causes and noise, which improves upon previous methods.


CrossGNN: Confronting Noisy Multivariate Time Series Via Cross Interaction Refinement

Qihe Huang, Lei Shen, Ruixin Zhang, Shouhong Ding, Binwu Wang, Zhengyang Zhou, Yang Wang

https://openreview.net/forum?id=xOzlW2vUYc

Keywords: Time Series Forecasting;

Compressor summary: CrossGNN is a linear complexity GNN model that improves multivariate time series forecasting by refining cross-scale and cross-variable interactions using an adaptive multi-scale identifier and two specialized GNN components.


Proximity-Informed Calibration for Deep Neural Networks

Miao Xiong, Ailin Deng, Pang Wei Koh, Jiaying Wu, Shen Li, Jianqing Xu, Bryan Hooi

https://openreview.net/forum?id=xOJUmwwlJc

Keywords: Calibration, Uncertainty Estimation, Trustworthiness, Fairness, Multicalibration

Compressor summary: The text discusses a common issue in calibration algorithms called proximity bias, which causes models to be more overconfident in low proximity data, and proposes a new algorithm (ProCal) to address it.


RL-based Stateful Neural Adaptive Sampling and Denoising for Real-Time Path Tracing

Antoine Scardigli, Lukas Cavigelli, Lorenz K Muller

https://openreview.net/forum?id=xNyR7DXUzJ

Keywords: computer graphics, rendering, ray tracing, GPU acceleration, RL, spatiotemporal latent space

Compressor summary: The paper proposes a framework that uses reinforcement learning and neural networks to improve Monte-Carlo path tracing, reducing noise and rendering times for realistic image synthesis in real-time applications.


Optimal Algorithms for the Inhomogeneous Spiked Wigner Model

Alexander Pak, Justin Ko, Florent Krzakala

https://openreview.net/forum?id=xNUmTRYtV1

Keywords: Spectral Method, Community detection, Wigner Spike model, Random Matrix, BBP transition, Approximate Message Passing, Spin glasses, Statistical Physics

Compressor summary: The paper proposes an AMP algorithm for a spiked Wigner problem with inhomogeneous noise and compares it with a spectral method, showing good performance and matching information-theoretic results.


Hierarchical Multi-Agent Skill Discovery

Mingyu Yang, Yaodong Yang, Zhenbo Lu, Wengang Zhou, Houqiang Li

https://openreview.net/forum?id=xMgO04HDOS

Keywords: Multi-Agent Reinforcement Learning, Hierarchical Skill Discovery, Probabilistic Graphical Model

Compressor summary: The text describes a hierarchical algorithm called Hierarchical Multi-Agent Skill Discovery (HMASD) that learns team and individual skills for multi-agent tasks without extrinsic rewards, outperforming existing methods.


Knowledge-Augmented Reasoning Distillation for Small Language Models in Knowledge-Intensive Tasks

Minki Kang, Seanie Lee, Jinheon Baek, Kenji Kawaguchi, Sung Ju Hwang

https://openreview.net/forum?id=xJLEQQrFia

Keywords: language model, distillation, reasoning, knowledge augmentation

Compressor summary: KARD is a novel method that enhances small LMs with rationales from large and external sources, achieving better results on knowledge-intensive reasoning tasks.


Robust Bayesian Satisficing

Artun Saday, Y. Cahit Yıldırım, Cem Tekin

https://openreview.net/forum?id=xINPCvgULc

Keywords: robust satisficing, regret minimization, Gaussian processes

Compressor summary: The paper introduces RoBOS, a novel robust Bayesian satisficing algorithm for noisy black-box optimization that performs well under distributional shifts.


Towards Better Dynamic Graph Learning: New Architecture and Unified Library

Le Yu, Leilei Sun, Bowen Du, Weifeng Lv

https://openreview.net/forum?id=xHNzWHbklj

Keywords: dynamic graph learning, Transformer-based architecture, dynamic graph library

Compressor summary: DyGFormer is a new Transformer-based architecture for dynamic graph learning that leverages historical interactions and uses DyGLib, a unified library for rigorous evaluations.


State2Explanation: Concept-Based Explanations to Benefit Agent Learning and User Understanding

Devleena Das, Sonia Chernova, Been Kim

https://openreview.net/forum?id=xGz0wAIJrS

Keywords: Concept-Based Explanations, Reinforcement Learning, Human-AI Interaction

Compressor summary: The paper proposes State2Explanation (S2E), a framework that learns to embed state-action pairs and concept-based explanations, which can improve both an RL agent's learning and users' understanding of the agent's decisions in sequential decision making tasks.


Boosting Spectral Clustering on Incomplete Data via Kernel Correction and Affinity Learning

Fangchen Yu, Runze Zhao, Zhan Shi, Yiwen Lu, Jicong Fan, Yicheng Zeng, Jianfeng Mao, Wenye Li

https://openreview.net/forum?id=xFtuNq23D5

Keywords: Spectral Clustering, Incomplete Data, Kernel Correction, Self-expressive Affinity Learning

Compressor summary: The paper proposes two new methods to improve spectral clustering on incomplete data by enhancing the kernel matrix and learning adaptive affinity matrices.


Kullback-Leibler Maillard Sampling for Multi-armed Bandits with Bounded Rewards

Hao Qin, Kwang-Sung Jun, Chicheng Zhang

https://openreview.net/forum?id=xF89MjFbWp

Keywords: multi-armed bandits, bounded rewards

Compressor summary: The paper analyzes KL-Maillard Sampling, a variant of Maillard Sampling that achieves finite-time gap-dependent regret bounds and asymptotic optimality in bandit problems with Bernoulli rewards.


Dissecting Chain-of-Thought: Compositionality through In-Context Filtering and Learning

Yingcong Li, Kartik Sreenivasan, Angeliki Giannou, Dimitris Papailiopoulos, Samet Oymak

https://openreview.net/forum?id=xEhKwsqxMa

Keywords: chain-of-thought, in-context learning, attention, compositional learning, approximation, length generalization

Compressor summary: The study examines how chain-of-thought (CoT) helps transformers learn compositional functions more efficiently by breaking down the learning process into two phases and reducing sample complexity.


LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical Imaging via Second-order Graph Matching

Duy Minh Ho Nguyen, Hoang Nguyen, Nghiem Tuong Diep, Tan Ngoc Pham, Tri Cao, Binh T. Nguyen, Paul Swoboda, Nhat Ho, Shadi Albarqouni, Pengtao Xie, Daniel Sonntag, Mathias Niepert

https://openreview.net/forum?id=xE7oH5iVGK

Keywords: medical imaging; self-supervised learning; graph matching; large-vision model

Compressor summary: LVM-Med is a family of deep networks trained on large-scale medical datasets that uses a novel self-supervised contrastive learning algorithm to improve performance on various medical tasks.


Probabilistic inverse optimal control for non-linear partially observable systems disentangles perceptual uncertainty and behavioral costs

Dominik Straub, Matthias Schultheis, Heinz Koeppl, Constantin A. Rothkopf

https://openreview.net/forum?id=xDHzQQ4lnC

Keywords: inverse optimal control, probabilistic modeling, motor control, cognitive science

Compressor summary: The paper presents a probabilistic approach to inverse optimal control for partially observable stochastic non-linear systems with unobserved action signals, which can characterize behavior in sequential decision-making tasks under uncertainty and disentangle perceptual factors and behavioral costs.


SODA: Robust Training of Test-Time Data Adaptors

Zige Wang, Yonggang Zhang, Zhen Fang, Long Lan, Wenjing Yang, Bo Han

https://openreview.net/forum?id=xBqjoG0NxM

Keywords: test-time data adaptation, zeroth-order optimization, out-of-distribution generalization

Compressor summary: The paper proposes a method called SODA, which uses reliable labels and preserves data information, to improve test-time data adaptation and adapt deployed models to distribution shifts while maintaining privacy.


Do Not Marginalize Mechanisms, Rather Consolidate!

Moritz Willig, Matej Zečević, Devendra Singh Dhami, Kristian Kersting

https://openreview.net/forum?id=xBhvMu4J03

Keywords: Structural Causal Models, Marginalization, Consolidation, Compression

Compressor summary: Consolidating causal mechanisms helps simplify large-scale structural causal models while maintaining their causality and interventional behavior.


Thrust: Adaptively Propels Large Language Models with External Knowledge

Xinran Zhao, Hongming Zhang, Xiaoman Pan, Wenlin Yao, Dong Yu, Jianshu Chen

https://openreview.net/forum?id=x9FOu3W6iy

Keywords: knowledge-intensive natural language processing, pre-trained language models, instance-level adaptive knowledge usage

Compressor summary: The paper proposes IAPEK, a technique that adaptively retrieves external knowledge for PTLMs using Thrust, a metric that measures models' instance-level knowledgeability, achieving higher cost-efficiency and performance improvement on 26% of the tasks.


Recasting Continual Learning as Sequence Modeling

Soochan Lee, Jaehyeon Son, Gunhee Kim

https://openreview.net/forum?id=x816mCbWpR

Keywords: meta-continual learning, sequence modeling, Transformers, efficient Transformers

Compressor summary: The authors propose a novel approach to continual learning by framing it as a sequence modeling problem and demonstrate its effectiveness using Transformers and their efficient variants on various benchmarks.


Lending Interaction Wings to Recommender Systems with Conversational Agents

Jiarui Jin, Xianyu Chen, Fanghua Ye, Mengyue Yang, Yue Feng, Weinan Zhang, Yong Yu, Jun Wang

https://openreview.net/forum?id=x7q7w07r6Y

Keywords: Conversational Agent, Recommender System, Conversational Recommendation

Compressor summary: CORE is a framework for chatbots to improve online recommendations by minimizing uncertainty through querying attributes or items.


Neural Ideal Large Eddy Simulation: Modeling Turbulence with Neural Stochastic Differential Equations

Anudhyan Boral, Zhong Yi Wan, Leonardo Zepeda-Nunez, James Lottes, Qing Wang, Yi-Fan Chen, John Roberts Anderson, Fei Sha

https://openreview.net/forum?id=x6cOcxRnxG

Keywords: partial differential equations, physics, turbulence, stochastic differential equations, physical simulation, neural differential equations

Compressor summary: The authors propose a neural network approach that combines ideal large eddy simulation and stochastic differential equations to model turbulent flows, achieving better accuracy and stability than traditional methods.


FedNAR: Federated Optimization with Normalized Annealing Regularization

Junbo Li, Ang Li, Chong Tian, Qirong Ho, Eric Xing, Hongyi Wang

https://openreview.net/forum?id=x5fs7TXKDc

Keywords: Federated learning, weight decay, adaptive hyperparameters

Compressor summary: The paper proposes FedNAR, a method to improve federated learning by adjusting weight decay and gradient co-clipping, which enhances convergence speed and model accuracy.


Improving *day-ahead* Solar Irradiance Time Series Forecasting by Leveraging Spatio-Temporal Context

Oussama Boussif, Ghait Boukachab, Dan Assouline, Stefano Massaroli, Tianle Yuan, Loubna Benabbou, Yoshua Bengio

https://openreview.net/forum?id=x5ZruOa4ax

Keywords: Time series forecasting, multi-modal learning, solar irradiance, context-enriched learning

Compressor summary: The paper proposes a deep learning architecture using satellite data to improve solar irradiance forecasting accuracy and uncertainty estimation, as well as introducing a new multi-modal dataset for training and testing.


Stochastic Distributed Optimization under Average Second-order Similarity: Algorithms and Analysis

Dachao Lin, Yuze Han, Haishan Ye, Zhihua Zhang

https://openreview.net/forum?id=x5JCDCvR4b

Keywords: distributed optimization, convex optimization, second-order similarity, client sampling

Compressor summary: The paper proposes two new algorithms for distributed optimization problems with improved communication complexity and shows their effectiveness in smoothness-free and ill-conditioned scenarios.


Online Constrained Meta-Learning: Provable Guarantees for Generalization

Siyuan Xu, Minghui Zhu

https://openreview.net/forum?id=x2xQEszznV

Keywords: meta-learning; generalization

Compressor summary: The paper introduces an online meta-learning framework that learns from constrained tasks and analyzes its optimality gaps and constraint violations, with a practical algorithm and experiments on imitation learning and few-shot classification.


Taming Local Effects in Graph-based Spatiotemporal Forecasting

Andrea Cini, Ivan Marisca, Daniele Zambon, Cesare Alippi

https://openreview.net/forum?id=x2PH6q32LR

Keywords: time series forecasting, spatiotemporal forecasting, graph-based spatiotemporal forecasting, graph neural networks

Compressor summary: The paper explores the trade-off between global and local models in spatiotemporal forecasting using graph neural networks, and proposes a method to incorporate trainable node embeddings for better adaptation to different time series dynamics.


FAST: a Fused and Accurate Shrinkage Tree for Heterogeneous Treatment Effects Estimation

Jia Gu, Caizhi Tang, Han Yan, Qing Cui, Longfei Li, JUN ZHOU

https://openreview.net/forum?id=wzg0BsV8rQ

Keywords: Data fusion, heterogeneous treatment effects estimation, shrinkage estimation, tree-based method

Compressor summary: The paper introduces FAST, a method for estimating treatment effects that uses both trial and observational data to improve accuracy and robustness by combining optimal weighting and tree-based techniques.


Searching for Optimal Per-Coordinate Step-sizes with Multidimensional Backtracking

Frederik Kunstner, Victor S. Portella, Mark Schmidt, Nick Harvey

https://openreview.net/forum?id=wzPcffMZ3b

Keywords: line-search, gradient descent, hypergradient, adaptive methods, smooth, convex, optimization, preconditioning

Compressor summary: The authors introduce multidimensional backtracking, a technique that finds optimal per-coordinate step-sizes for smooth convex problems by using hyper-gradients and cutting-plane methods.


Natural Actor-Critic for Robust Reinforcement Learning with Function Approximation

Ruida Zhou, Tao Liu, Min Cheng, Dileep Kalathil, Panganamala Kumar, Chao Tian

https://openreview.net/forum?id=wxkBdtDbmH

Keywords: robust reinforcement learning, policy-based approach, function approximation, actor-critic

Compressor summary: The authors propose two new uncertainty sets for large-scale robust reinforcement learning and a robust natural actor-critic algorithm that converges to the optimal policy in finite time.


Federated Learning with Bilateral Curation for Partially Class-Disjoint Data

Ziqing Fan, Ruipeng Zhang, Jiangchao Yao, Bo Han, Ya Zhang, Yanfeng Wang

https://openreview.net/forum?id=wwmKVO8bsR

Keywords: federated learning, data heterogeneity, partially class-disjoint data

Compressor summary: FedGELA is a novel approach for federated learning with partially class-disjoint data that uses a simplex ETF classifier to achieve fair and equal discrimination for all classes, improving performance and providing convergence guarantees.


Adapting Fairness Interventions to Missing Values

Raymond Feng, Flavio Calmon, Hao Wang

https://openreview.net/forum?id=wwkQUiaKbo

Keywords: algorithmic fairness, discrimination, missing values, machine learning

Compressor summary: The paper analyzes how missing values in data affect algorithmic fairness, showing that the standard "impute-then-classify" approach can worsen fairness and accuracy, and presents new algorithms that handle missing patterns while preserving their information for better fairness and accuracy.


Subject-driven Text-to-Image Generation via Apprenticeship Learning

Wenhu Chen, Hexiang Hu, YANDONG LI, Nataniel Ruiz, Xuhui Jia, Ming-Wei Chang, William W. Cohen

https://openreview.net/forum?id=wv3bHyQbX7

Keywords: Diffusion Model, Image Generation, Image Editing, In-Context Learning

Compressor summary: SuTI is a fast and high-quality text-to-image generator that learns from millions of online image clusters to produce customized images without fine-tuning for each subject.


Granger Components Analysis: Unsupervised learning of latent temporal dependencies

Jacek Dmochowski

https://openreview.net/forum?id=wqIm0Qsgy0

Keywords: components analysis, unsupervised learning, Granger Causality

Compressor summary: The paper introduces a new unsupervised learning technique for time series data that uses Granger causality to identify driving and driven signals in multivariate data sets.


ContinuAR: Continuous Autoregression For Infinite-Fidelity Fusion

WEI W. XING, Yuxin Wang, Zheng Xing

https://openreview.net/forum?id=wpfsnu5syT

Keywords: Gaussian process, autoregression, multi fidelity, nonparametric Bayesian

Compressor summary: The authors propose ContinuAR, a novel method for multi-fidelity fusion that improves efficiency and performance compared to existing techniques.


BayesDAG: Gradient-Based Posterior Inference for Causal Discovery

Yashas Annadani, Nick Pawlowski, Joel Jennings, Stefan Bauer, Cheng Zhang, Wenbo Gong

https://openreview.net/forum?id=woptnU6fh1

Keywords: Causal Discovery, Structure Learning, Bayesian Inference, Variational Inference, MCMC, Generative Model

Compressor summary: The text introduces a new Bayesian causal discovery method that samples DAGs from the posterior without regularization, using a combination of SG-MCMC and VI, and applies gradient-based MCMC sampling for the first time in this field.


Maximization of Average Precision for Deep Learning with Adversarial Ranking Robustness

Gang Li, Wei Tong, Tianbao Yang

https://openreview.net/forum?id=wm5Ane9VRO

Keywords: Adversarial Average Precision Maximization, Robust Average Precision, Adversarial Ranking Robustness, Adversarial Training

Compressor summary: The paper proposes a new method to optimize both average precision and adversarial robustness, outperforming current methods by a significant margin.


VillanDiffusion: A Unified Backdoor Attack Framework for Diffusion Models

Sheng-Yen Chou, Pin-Yu Chen, Tsung-Yi Ho

https://openreview.net/forum?id=wkIBfnGPTA

Keywords: backdoor, diffusion model, trustworthy

Compressor summary: The paper introduces a backdoor attack framework (VillanDiffusion) for diffusion models, which are generative AI models that can be vulnerable to output manipulation attacks when exposed to malicious patterns at the input.


Zero-shot Visual Relation Detection via Composite Visual Cues from Large Language Models

Lin Li, Jun Xiao, Guikun Chen, Jian Shao, Yueting Zhuang, Long Chen

https://openreview.net/forum?id=wiv21EJ0Vd

Keywords: Visual relation detection, Zero-short learning, Scene graph generation

Compressor summary: RECODE is a novel zero-shot VRD method that uses composite description prompts to improve relation detection by leveraging LLMs and spatial information.


Stein $\Pi$-Importance Sampling

Congye Wang, Wilson Ye Chen, Heishiro Kanagawa, Chris J. Oates

https://openreview.net/forum?id=wiidCRA3at

Keywords: Bayesian, discrepancy, kernel, sampling, Stein's method

Compressor summary: The paper proposes a new method for improving Monte Carlo output using Stein importance sampling and provides a novel variational construction and convergence conditions for it.


Outlier-Robust Wasserstein DRO

Sloan Nietert, Ziv Goldfeld, Soroosh Shafiee

https://openreview.net/forum?id=wg3d2FKAm8

Keywords: distributionally robust optimization, robust statistics, optimal transport, Wasserstein distance

Compressor summary: The paper proposes a new robust optimization approach that handles both geometric and non-geometric data perturbations and shows improved performance on regression and classification tasks.


Consistent Aggregation of Objectives with Diverse Time Preferences Requires Non-Markovian Rewards

Silviu Pitis

https://openreview.net/forum?id=wcdF6jR0Sp

Keywords: Normative Agency Design, Reward Design, Sequential Decision Making, Reinforcement Learning, Intertemporal Fairness, Multi-Objective Decision Making

Compressor summary: The paper shows that optimal AI agents must use non-Markovian rewards when serving multiple objectives with different discount factors, and proposes a practical aggregation scheme for this purpose.


Improved Best-of-Both-Worlds Guarantees for Multi-Armed Bandits: FTRL with General Regularizers and Multiple Optimal Arms

Tiancheng Jin, Junyan Liu, Haipeng Luo

https://openreview.net/forum?id=wbg4JEM5Jp

Keywords: multi-armed bandit, best of both worlds, Follow-the-Regularized-Leader, Tsallis entropy, Shannon entropy, Log-barrier

Compressor summary: The paper studies adaptive multi-armed bandit algorithms and improves the performance of FTRL with various regularizers and a new learning rate schedule, removing the uniqueness assumption and achieving better regret bounds.


High-dimensional Asymptotics of Denoising Autoencoders

Hugo Cui, Lenka Zdeborova

https://openreview.net/forum?id=wbbTqsiKzl

Keywords: statistical physics, replica method, autoencoder, exact asymptotics

Compressor summary: The paper proposes a two-layer non-linear autoencoder with tied weights and a skip connection for denoising high-dimensional data from a Gaussian mixture, and compares it to other methods.


Provable benefits of score matching

Chirag Pabbaraju, Dhruv Rohatgi, Anish Sevekari, Holden Lee, Ankur Moitra, Andrej Risteski

https://openreview.net/forum?id=waXoG35kbb

Keywords: theory, score matching, exponential families, sample complexity, computational hardness

Compressor summary: The paper presents a new exponential family of distributions where score matching is as easy to optimize as maximum likelihood, unlike other cases where score matching is harder or intractable.


Collaboratively Learning Linear Models with Structured Missing Data

Chen Cheng, Gary Cheng, John Duchi

https://openreview.net/forum?id=waDF0oACu2

Keywords: Collaborative Learning, Missing Data, Sensors, Linear Regression

Compressor summary: The paper proposes Collab, a distributed, semi-supervised algorithm for learning least squares estimates for m agents with different features, without communicating the labeled data, achieving near-optimal performance and generalizability.


Optimal Block-wise Asymmetric Graph Construction for Graph-based Semi-supervised Learning

Zixing Song, Yifei Zhang, Irwin King

https://openreview.net/forum?id=wYkfog48Bq

Keywords: Graph-based Semi-supervised Learning, Affinity Graph Construction

Compressor summary: This paper proposes an optimal asymmetric graph structure for label inference in graph-based semi-supervised learning that differentiates the roles of labeled and unlabeled nodes, leading to improved performance on synthetic and real datasets.


Language-driven Scene Synthesis using Multi-conditional Diffusion Model

An Dinh Vuong, Minh Nhat VU, Toan Tien Nguyen, Baoru Huang, Dzung Nguyen, Thieu Vo, Anh Nguyen

https://openreview.net/forum?id=wYKU1C77sa

Keywords: scene synthesis, language-driven, diffusion models, multi-conditional generation, 3D point cloud

Compressor summary: The paper proposes a new language-driven scene synthesis task that combines text, human motion, and objects, and presents a multi-conditional diffusion model to solve it effectively.


What can a Single Attention Layer Learn? A Study Through the Random Features Lens

Hengyu Fu, Tianyu Guo, Yu Bai, Song Mei

https://openreview.net/forum?id=wX8GuzDSJR

Keywords: transformers, attention, deep learning theory, random features

Compressor summary: The paper analyzes how attention layers in Transformers learn and generalize, showing they can efficiently learn certain functions with random features and have advantages over standard neural networks.


Transitivity Recovering Decompositions: Interpretable and Robust Fine-Grained Relationships

Abhra Chaudhuri, Massimiliano Mancini, Zeynep Akata, Anjan Dutta

https://openreview.net/forum?id=wUNPmdE273

Keywords: Interpretability, Robustness, Fine-Grained Representation Learning, Graph Theory, Information Theory

Compressor summary: The paper presents a method called TRD that converts abstract relational representations in image analysis into interpretable graphs and shows its effectiveness and robustness to noisy views.


Improving Diffusion-Based Image Synthesis with Context Prediction

Ling Yang, Jingwei Liu, Shenda Hong, Zhilong Zhang, Zhilin Huang, Zheming Cai, Wentao Zhang, Bin CUI

https://openreview.net/forum?id=wRhLd65bDt

Keywords: Diffusion Model, Image Generation

Compressor summary: ConPreDiff is a novel diffusion model that predicts neighborhood context for better image synthesis and achieves state-of-the-art results in text-to-image generation.


Critical Initialization of Wide and Deep Neural Networks using Partial Jacobians: General Theory and Applications

Darshil Doshi, Tianyu He, Andrey Gromov

https://openreview.net/forum?id=wRJqZRxDEX

Keywords: Criticality, Gaussian Process, Jacobian, LayerNorm, Residual connections, ResNet

Compressor summary: The authors propose a method to diagnose criticality in deep neural networks using partial Jacobians, which helps select optimal initialization and show that LayerNorm and residual connections lead to a critical architecture for any initialization.


Small batch deep reinforcement learning

Johan Samir Obando Ceron, Marc G Bellemare, Pablo Samuel Castro

https://openreview.net/forum?id=wPqEvmwFEh

Keywords: Reinforcement Learning, Deep Reinforcement Learning, Value based, Batch Size

Compressor summary: Reducing the batch size in value-based deep reinforcement learning can improve performance, contrary to the common belief that larger batches are better.


Progressive Ensemble Distillation: Building Ensembles for Efficient Inference

Don Dennis, Abhishek Shetty, Anish Sevekari, Kazuhito Koishida, Virginia Smith

https://openreview.net/forum?id=wNxyDofh74

Keywords: Edge computing, compression, efficient inference, distillation and inference, run-time tradeoff, inference-time tradeoff, on-device, user-side, client-side

Compressor summary: B-DISTIL is a method to create ensembles of smaller student models from a large pretrained teacher model using function composition based aggregation rules, achieving similar performance while allowing flexibility in accuracy vs. inference cost trade-offs.


Leveraging Early-Stage Robustness in Diffusion Models for Efficient and High-Quality Image Synthesis

Yulhwa Kim, Dongwon Jo, Hyesung Jeon, Taesu Kim, Daehyun Ahn, Hyungjun Kim, jae-joon kim

https://openreview.net/forum?id=wNpsGwixjG

Keywords: diffusion models, post-training quantization

Compressor summary: The paper proposes a faster noise estimation method for diffusion models by using low-bit activations in the early reverse diffusion process, while keeping high-bit activations for later stages to maintain image quality.


A Closer Look at the Robustness of Contrastive Language-Image Pre-Training (CLIP)

Weijie Tu, Weijian Deng, Tom Gedeon

https://openreview.net/forum?id=wMNpMe0vp3

Keywords: CLIP

Compressor summary: This paper investigates the safety measures of CLIP models, focusing on their resilience to visual factor variations, uncertainty estimations, and anomaly detection, by testing them on various datasets and conditions.


Calibrating Neural Simulation-Based Inference with Differentiable Coverage Probability

Maciej Falkiewicz, Naoya Takeishi, Imahn Shekhzadeh, Antoine Wehenkel, Arnaud Delaunoy, Gilles Louppe, Alexandros Kalousis

https://openreview.net/forum?id=wLiMhVJ7fx

Keywords: simulation-based inference, inverse problem, bayesian inference, uncertainty quantification, generative modeling

Compressor summary: Bayesian inference uses probability to express uncertainty, but existing methods can be overconfident; a new method adds calibration error to neural models for more accurate uncertainty quantification.


An Efficient and Robust Framework for Approximate Nearest Neighbor Search with Attribute Constraint

Mengzhao Wang, Lingwei Lv, Xiaoliang Xu, Yuxiang Wang, Qiang Yue, Jiongkang Ni

https://openreview.net/forum?id=wLFXTAWa5V

Keywords: approximate nearest neighbor search, attribute filtering, high-dimensional vector, proximity graph

Compressor summary: The paper presents a fast and accurate framework for hybrid query processing using composite indexes based on proximity graphs, improving both approximate nearest neighbor search and attribute filtering.


Learning a 1-layer conditional generative model in total variation

Ajil Jalal, Justin Kang, Ananya Uppal, Kannan Ramchandran, Eric Price

https://openreview.net/forum?id=wImYhdu4VF

Keywords: Generative models, distribution learning, maximum likelihood estimation

Compressor summary: A conditional generative model is a method to sample from a conditional distribution without assumptions on the input distribution, and it can learn deep models using few samples.


A Single-Loop Accelerated Extra-Gradient Difference Algorithm with Improved Complexity Bounds for Constrained Minimax Optimization

Yuanyuan Liu, Fanhua Shang, Weixin An, Junhao Liu, Hongying Liu, Zhouchen Lin

https://openreview.net/forum?id=wIlmx4bHrO

Keywords: Constrained Minimax Optimization; nonconvex- nonconcave

Compressor summary: The paper presents a new algorithm for solving constrained nonconvex-nonconcave problems that improves convergence rates and complexity bounds compared to existing methods.


Online Corrupted User Detection and Regret Minimization

Zhiyong Wang, Jize Xie, Tong Yu, Shuai Li, John C.S. Lui

https://openreview.net/forum?id=wHhPIv5G8Q

Keywords: online learning, online corrupted user detection, clustering of bandits

Compressor summary: The paper proposes an online learning problem and two novel algorithms to learn from user relations and detect corrupted users in web systems.


Restart Sampling for Improving Generative Processes

Yilun Xu, Mingyang Deng, Xiang Cheng, Yonglong Tian, Ziming Liu, Tommi S. Jaakkola

https://openreview.net/forum?id=wFuemocyHZ

Keywords: Generative models, diffusion models, PFGM, sampling

Compressor summary: Restart is a novel sampling algorithm that combines ODE and SDE methods to achieve faster and higher quality results in generative models.


Sharp Calibrated Gaussian Processes

Alexandre Capone, Sandra Hirche, Geoff Pleiss

https://openreview.net/forum?id=wFH5hZAwYz

Keywords: Gaussian Processes, Frequentist Statistics, Kernel Methods, Model Selection and Structure Learning, Regression

Compressor summary: The paragraph describes a new method for improving the uncertainty estimates of Gaussian processes by using a different set of hyperparameters that better satisfy calibration constraints and yield tighter predictive quantiles.


Improving Self-supervised Molecular Representation Learning using Persistent Homology

Yuankai Luo, Lei Shi, Veronika Thost

https://openreview.net/forum?id=wEiUGpcr0M

Keywords: Graph Neural Networks, Molecular Representation Learning, Persistent Homology, Contrastive Learning, Self-supervised Learning

Compressor summary: The paper proposes a self-supervised learning method for molecular representation learning using persistent homology, which offers unique features and improves predictive power, especially in small datasets.


Ambient Diffusion: Learning Clean Distributions from Corrupted Data

Giannis Daras, Kulin Shah, Yuval Dagan, Aravind Gollakota, Alex Dimakis, Adam Klivans

https://openreview.net/forum?id=wBJBLy9kBY

Keywords: corrupted data, generative models, ambient gan, inverse problems, learning from measurements

Compressor summary: The paper proposes a diffusion-based method to learn an unknown distribution using highly-corrupted samples, which can train generative models that are less likely to memorize individual samples and work on standard benchmarks with significant pixel loss or corrupted datasets.


Interpretable Reward Redistribution in Reinforcement Learning: A Causal Approach

Yudi Zhang, Yali Du, Biwei Huang, Ziyan Wang, Jun Wang, Meng Fang, Mykola Pechenizkiy

https://openreview.net/forum?id=w7TyuWhGZP

Keywords: Reinforcement learning, sparse reward, return decomposition, causal modeling

Compressor summary: This paper proposes GRD, a framework that uses causal generative models to redistribute rewards in reinforcement learning for delayed reward scenarios and improves policy optimization by training over the most favorable state subspace.


InfoCD: A Contrastive Chamfer Distance Loss for Point Cloud Completion

Fangzhou Lin, Yun Yue, Ziming Zhang, Songlin Hou, Kazunori Yamada, Vijaya B Kolachalama, Venkatesh Saligrama

https://openreview.net/forum?id=w7LxAZfDfv

Keywords: Contrastive learning; Point cloud completion

Compressor summary: The paper introduces InfoCD, a contrastive Chamfer distance loss that improves point cloud completion and surface similarity estimation by maximizing mutual information between underlying surfaces, achieving state-of-the-art results on benchmark datasets.


Compositional Sculpting of Iterative Generative Processes

Timur Garipov, Sebastiaan De Peuter, Ge Yang, Vikas Garg, Samuel Kaski, Tommi S. Jaakkola

https://openreview.net/forum?id=w79RtqIyoM

Keywords: generative model composition, GFlowNets, diffusion models, classifier guidance, probabilistic methods

Compressor summary: The paper proposes Compositional Sculpting, a method for combining and sampling from iterative generative models like GFlowNets and diffusion models using classifier guidance.


Hyper-HMM: aligning human brains and semantic features in a common latent event space

Caroline Lee, Jane Han, Ma Feilong, Guo Jiahui, James Haxby, Christopher Baldassano

https://openreview.net/forum?id=w6krZiUa7t

Keywords: Brain Imaging, Other Cognitive Science, Other Neuroscience

Compressor summary: The Hyper-HMM is a hybrid model that aligns both temporal and spatial features across brains when processing naturalistic stimuli, allowing for mapping of individual cognitive dynamics and semantic content.


Minimax Optimal Rate for Parameter Estimation in Multivariate Deviated Models

Dat Do, Huy Nguyen, Khai Nguyen, Nhat Ho

https://openreview.net/forum?id=w3ghbKBJg4

Keywords: Mixture Model, Minimax Rate, Maximum Likelihood Estimation

Compressor summary: The paper studies how to estimate parameters in a multivariate model with a mixed density function, using a new condition called \emph{distinguishability} to deal with challenges in convergence rates.


Balanced Training for Sparse GANs

Yite Wang, Jing Wu, Naira Hovakimyan, Ruoyu Sun

https://openreview.net/forum?id=w2F8Fm6Sg3

Keywords: Dynamics sparse training; pruning; neural network pruning; empirical deep learning

Compressor summary: The paper introduces a new metric and method for balancing sparse training in generative adversarial networks (GANs) to improve efficiency without sacrificing performance.


Optimal Learners for Realizable Regression: PAC Learning and Online Learning

Idan Attias, Steve Hanneke, Alkis Kalavasis, Amin Karbasi, Grigoris Velegkas

https://openreview.net/forum?id=w116w62fxH

Keywords: Learning Theory, Regression, PAC Learning, Online Learning

Compressor summary: The authors study the complexity of learning realizable regression in different settings, introduce new dimensions to characterize learnability, and design optimal learners for these settings.


Visual Instruction Tuning

Haotian Liu, Chunyuan Li, Qingyang Wu, Yong Jae Lee

https://openreview.net/forum?id=w0H2xGHlkw

Keywords: visual instruction tuning, instruction tuning, multimodal, LLM, GPT

Compressor summary: The paper introduces LLaVA, a multimodal model that uses language to generate image instructions for GPT-4, improving its performance on various tasks, including Science QA.


GAUCHE: A Library for Gaussian Processes in Chemistry

Ryan-Rhys Griffiths, Leo Klarner, Henry Moss, Aditya Ravuri, Sang T. Truong, Yuanqi Du, Samuel Don Stanton, Gary Tom, Bojana Ranković, Arian Rokkum Jamasb, Aryan Deshwal, Julius Schwartz, Austin Tripp, Gregory Kell, Simon Frieder, Anthony Bourached, Alex James Chan, Jacob Moss, Chengzhi Guo, Johannes P. Dürholt, Saudamini Chaurasia, Ji Won Park, Felix Strieth-Kalthoff, Alpha Lee, Bingqing Cheng, Alan Aspuru-Guzik, Philippe Schwaller, Jian Tang

https://openreview.net/forum?id=vzrA6uqOis

Keywords: Gaussian processes, Bayesian optimization, Chemistry, Molecular Machine Learning, Applications, Software

Compressor summary: GAUCHE is an open-source library that uses Gaussian processes to optimize molecular representations for applications like molecular discovery, chemical reaction optimization, and protein design.


Adaptive whitening with fast gain modulation and slow synaptic plasticity

Lyndon Duong, Eero P Simoncelli, Dmitri Chklovskii, David Lipshutz

https://openreview.net/forum?id=vz7SdRqWGM

Keywords: neuroscience, adaptation, whitening, efficient coding, recurrent neural network, gain modulation, synaptic plasticity, local learning rules

Compressor summary: The paragraph describes a new model that combines synaptic plasticity and gain modulation to rapidly adapt sensory neurons to changing input statistics.


Learning from Rich Semantics and Coarse Locations for Long-tailed Object Detection

Lingchen Meng, Xiyang Dai, Jianwei Yang, Dongdong Chen, Yinpeng Chen, Mengchen Liu, Yi-Ling Chen, Zuxuan Wu, Lu Yuan, Yu-Gang Jiang

https://openreview.net/forum?id=vybQs1Gbuk

Keywords: Long-tail object detection, visual semantics, soft supervision

Compressor summary: RichSem is a simple and effective method for long-tailed object detection that uses semantic branch to learn rich semantics from coarse locations as additional soft supervision without needing accurate bounding boxes.


Optimal Regret Is Achievable with Bounded Approximate Inference Error: An Enhanced Bayesian Upper Confidence Bound Framework

Ziyi Huang, Henry Lam, Amirhossein Meisami, Haofeng Zhang

https://openreview.net/forum?id=vwr4bHHsRT

Keywords: Bayesian bandits, approximate Bayesian inference, Bayesian Upper Confidence Bound, optimal regret order, bounded inference error

Compressor summary: The EBUCB framework improves on existing Bayesian bandit algorithms by achieving optimal regret bounds under constant approximate inference error using two different $\alpha$-divergences, while previous methods only had worst-case linear regret with one $\alpha$-divergence.


InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning

Wenliang Dai, Junnan Li, Dongxu Li, Anthony Tiong, Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, Steven Hoi

https://openreview.net/forum?id=vvoWPYqZJA

Keywords: Vision-Language Models, Instruction Tuning, Zero-shot

Compressor summary: The paper explores vision-language instruction tuning using pretrained BLIP-2 models and introduces a Query Transformer, achieving state-of-the-art results on various tasks.


Train Once, Get a Family: State-Adaptive Balances for Offline-to-Online Reinforcement Learning

Shenzhi Wang, Qisen Yang, Jiawei Gao, Matthieu Gaetan Lin, HAO CHEN, Liwei Wu, Ning Jia, Shiji Song, Gao Huang

https://openreview.net/forum?id=vtoY8qJjTR

Keywords: reinforcement learning, offline-to-online reinforcement learning, offline reinforcement learning, policy improvement, policy constraint

Compressor summary: FamO2O is a framework that adapts improvement and constraint balances for offline-to-online reinforcement learning based on each state, improving performance over existing methods.


The Geometry of Neural Nets' Parameter Spaces Under Reparametrization

Agustinus Kristiadi, Felix Dangel, Philipp Hennig

https://openreview.net/forum?id=vtLNwa6uX0

Keywords: neural network, invariance, equivariance, reparametrization, riemannian geometry, parameter space

Compressor summary: The paper studies how to represent the metric of neural nets under reparametrization using Riemannian geometry, which has implications for flatness analysis and optimization.


Revisiting the Minimalist Approach to Offline Reinforcement Learning

Denis Tarasov, Vladislav Kurenkov, Alexander Nikulin, Sergey Kolesnikov

https://openreview.net/forum?id=vqGWslLeEw

Keywords: Offline Reinforcement Learning

Compressor summary: The authors analyze recent offline reinforcement learning algorithms, propose ReBRAC that integrates several design elements, and show its state-of-the-art performance in various settings and datasets.


Online PCA in Converging Self-consistent Field Equations

Xihan Li, Xiang Chen, Rasul Tutunov, Haitham Bou Ammar, Lei Wang, Jun Wang

https://openreview.net/forum?id=vq11gurmUY

Keywords: Self-consistent Field Equation, Computational Science, Online PCA

Compressor summary: The authors propose a novel approach to solve the SCF equation using online PCA techniques, which improves convergence and applies to electronic structure problems.


TransHP: Image Classification with Hierarchical Prompting

Wenhao Wang, Yifan Sun, Wei Li, Yi Yang

https://openreview.net/forum?id=vpQuCsZXz2

Keywords: hierarchical image classification, hierarchical prompting, vision transformer

Compressor summary: The paper proposes a novel hierarchical image classification method that uses tokenized ancestor-class hints to improve accuracy, efficiency, and explainability of image classification models.


Combinatorial Optimization with Policy Adaptation using Latent Space Search

Felix Chalumeau, Shikha Surana, Clément Bonnet, Nathan Grinsztajn, Arnu Pretorius, Alexandre Laterre, Thomas D Barrett

https://openreview.net/forum?id=vpMBqdt9Hl

Keywords: Reinforcement Learning, Combinatorial Optimization, TSP, CVRP, JSSP

Compressor summary: COMPASS is a novel RL approach that pre-trains diverse policies in a latent space to improve search performance for combinatorial optimization problems.


Conditional score-based diffusion models for Bayesian inference in infinite dimensions

Lorenzo Baldassari, Ali Siahkoohi, Josselin Garnier, Knut Solna, Maarten V. de Hoop

https://openreview.net/forum?id=voG6nEW9BV

Keywords: score-based generative models, diffusion models, inverse problems, bayesian inference, infinite dimensions

Compressor summary: The paper proposes a new method for solving infinite-dimensional Bayesian linear inverse problems using amortized conditional score-based diffusion models, which improves upon previous heuristic approaches by being more theoretically grounded and computationally efficient.


Higher-Order Uncoupled Dynamics Do Not Lead to Nash Equilibrium - Except When They Do

Sarah Asad Toonsi, Jeff S Shamma

https://openreview.net/forum?id=vnTUuecp2v

Keywords: Learning in games, Nash equilibrium, Uncoupled Dynamics

Compressor summary: The paper introduces higher-order gradient play dynamics for multi-agent learning that can lead to Nash Equilibrium or fail depending on the game and dynamics.


On the Overlooked Pitfalls of Weight Decay and How to Mitigate Them: A Gradient-Norm Perspective

Zeke Xie, zhiqiang xu, Jingzhao Zhang, Issei Sato, Masashi Sugiyama

https://openreview.net/forum?id=vnGcubtzR1

Keywords: Weight Decay, Regularization, Optimization, Deep Learning

Compressor summary: The paper proposes a new method called Scheduled Weight Decay (SWD) to reduce large gradient norms caused by weight decay, which can improve convergence and generalization in deep neural networks.


A Novel Approach for Effective Multi-View Clustering with Information-Theoretic Perspective

Chenhang Cui, Yazhou Ren, Jingyu Pu, Jiawei Li, Xiaorong Pu, Tianyi Wu, Yutao Shi, Lifang He

https://openreview.net/forum?id=vlDbqzwczj

Keywords: multi-view learning, clustering

Compressor summary: SUMVC is a new information-theoretic approach for multi-view clustering that improves performance by reducing redundancy and enhancing consistent information across views.


Structured Voronoi Sampling

Afra Amini, Li Du, Ryan Cotterell

https://openreview.net/forum?id=vf77fTbgG3

Keywords: Natural Language Processing, Text Generation, Controlled Generation, MCMC, HMC, Langevin Dynamics

Compressor summary: The paper introduces Structured Voronoi Sampling (SVS), a principled gradient-based technique for sampling from language models that generates fluent and diverse texts while adhering to control targets.


Extremal Domain Translation with Neural Optimal Transport

Milena Gazdieva, Alexander Korotin, Daniil Selikhanovych, Evgeny Burnaev

https://openreview.net/forum?id=vZRiMjo826

Keywords: optimal transport, partial optimal transport, neural networks, domain translation

Compressor summary: The paper introduces extremal transport, a method for unpaired image domain translation based on neural optimal transport and inspired by theoretical limits of possible translations.


ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning

Junguang Jiang, Baixu Chen, Junwei Pan, Ximei Wang, Dapeng Liu, jie jiang, Mingsheng Long

https://openreview.net/forum?id=vZHk1QlBQW

Keywords: Auxiliary-Task Learning, Negative Transfer

Compressor summary: ForkMerge is a novel approach that helps mitigate negative transfer in auxiliary-task learning by periodically forking the model into multiple branches and dynamically merging them based on task weights.


A Unified Framework for Uniform Signal Recovery in Nonlinear Generative Compressed Sensing

Junren Chen, Jonathan Scarlett, Michael Ng, Zhaoqiang Liu

https://openreview.net/forum?id=vUXNNLatFv

Keywords: Compressed sensing, generative models, nonlinearity, uniform recovery

Compressor summary: The paper presents a unified framework for uniform recovery guarantees in nonlinear generative compressed sensing, handling discontinuous or unknown observation models, and achieves similar performance to existing non-uniform results with little additional cost.


Faster Margin Maximization Rates for Generic Optimization Methods

Guanghui Wang, Zihao Hu, Vidya Muthukumar, Jacob Abernethy

https://openreview.net/forum?id=vTug54Uunq

Keywords: Implicit bias, margin maximization, zero-sum game, online learning

Compressor summary: The paper proposes new techniques to improve the implicit bias rates of generic optimization methods like mirror descent and steepest descent in binary classification problems, using a unified framework based on regularized bilinear games.


Going Beyond Linear Mode Connectivity: The Layerwise Linear Feature Connectivity

Zhanpeng Zhou, Yongyi Yang, Xiaojiang Yang, Junchi Yan, Wei Hu

https://openreview.net/forum?id=vORUHrVEnH

Keywords: Linear Mode Connectivity, Permutation Invariance, Optimization Landscape, Science of Deep Learning

Compressor summary: This paper introduces a concept called Layerwise Linear Feature Connectivity (LLFC), which shows how different neural networks can be connected by linear paths in their feature maps, based on the Linear Mode Connectivity phenomenon.


Data Pruning via Moving-one-Sample-out

Haoru Tan, Sitong Wu, Fei Du, Yukang Chen, Zhibin Wang, Fan Wang, XIAOJUAN QI

https://openreview.net/forum?id=vO6ZdPWaHc

Keywords: Data Valuation, Deep Learning, Data Pruning, Coreset Selection.

Compressor summary: The paper presents MoSo, a data-pruning method that uses gradient information to identify and remove uninformative samples from the training set, improving performance at high pruning ratios.


Imitation Learning from Imperfection: Theoretical Justifications and Algorithms

Ziniu Li, Tian Xu, Zeyu Qin, Yang Yu, Zhi-Quan Luo

https://openreview.net/forum?id=vO04AzsB49

Keywords: imitation learning, distribution shift, policy optimization, data selection

Compressor summary: The paper presents a framework for imitation learning with supplementary data that addresses distribution shift issues using importance sampling to improve performance in various tasks.


Suggesting Variable Order for Cylindrical Algebraic Decomposition via Reinforcement Learning

Fuqi Jia, Yuhang Dong, Minghao Liu, Pei Huang, Feifei Ma, Jian Zhang

https://openreview.net/forum?id=vNsdFwjPtL

Keywords: Reinforcement Learning, Graph Neural Network, Cylindrical Algebraic Decomposition.

Compressor summary: The paper introduces two reinforcement learning methods with graph neural networks for suggesting variable order in cylindrical algebraic decomposition, improving efficiency and generalizing well to different datasets.


Exploiting Correlated Auxiliary Feedback in Parameterized Bandits

Arun Verma, Zhongxiang Dai, Yao Shu, Bryan Kian Hsiang Low

https://openreview.net/forum?id=vM5VnNQ4n7

Keywords: Parameterized Bandits, Auxiliary Feedback, Control Variate, Regret Minimization

Compressor summary: The paper proposes a method that uses auxiliary feedback to improve reward estimation, leading to reduced regret in parameterized bandits problems.


Transformer as a hippocampal memory consolidation model based on NMDAR-inspired nonlinearity

Dong-Kyum Kim, Jea Kwon, Meeyoung Cha, C. Justin Lee

https://openreview.net/forum?id=vKpVJxplmB

Keywords: Transformer, NMDA, long-term memory, reference memory, memory consolidation

Compressor summary: The paper proposes a new activation function for deep learning models that mimics the hippocampus's NMDA receptor dynamics, improving working memory and spatial representation in transformers.


Accelerated Quasi-Newton Proximal Extragradient: Faster Rate for Smooth Convex Optimization

Ruichen Jiang, Aryan Mokhtari

https://openreview.net/forum?id=vIGNYQ4Alv

Keywords: convex optimization, quasi-Newton methods, Monteiro-Svaiter acceleration, Nesterov's accelerated gradient, online learning

Compressor summary: The paper presents a new optimization method that beats Nesterov's accelerated gradient in certain regimes and achieves optimal or faster convergence rates based on an online learning approach.


Accelerated Zeroth-order Method for Non-Smooth Stochastic Convex Optimization Problem with Infinite Variance

Nikita Kornilov, Ohad Shamir, Aleksandr Lobanov, Darina Dvinskikh, Alexander Gasnikov, Innokentiy Andreevich Shibaev, Eduard Gorbunov, Samuel Horváth

https://openreview.net/forum?id=vHSQTEIFkp

Keywords: stochastic optimization, gradient-free optimization, zero-order oracle, gradient clipping, infinite variance

Compressor summary: The paper presents an optimal algorithm for non-smooth stochastic convex optimization with infinite noise variance, adapted from a clipped accelerated gradient method.


Generalized Weighted Path Consistency for Mastering Atari Games

Dengwei Zhao, Shikui Tu, Lei Xu

https://openreview.net/forum?id=vHRLS8HhK1

Keywords: Monte Carlo Tree Search, Reinforcement learning, Path consistency.

Compressor summary: The paper proposes GW-PCZero, a generalized version of PCZero that uses a weighting mechanism to reduce variance and improve learning efficiency in reinforcement learning with neural-guided MCTS for real applications with non-zero immediate reward.


Self-supervised video pretraining yields robust and more human-aligned visual representations

Nikhil Parthasarathy, S. M. Ali Eslami, Joao Carreira, Olivier J Henaff

https://openreview.net/forum?id=vF8ukt5l1R

Keywords: self-supervised learning, contrastive, video pretraining, representation learning, visual representation, human alignment, robustness, shape-bias, saliency

Compressor summary: Video pretraining using VITO can learn visually general, robust, and human-like representations from complex transformations in videos.


On Class Distributions Induced by Nearest Neighbor Graphs for Node Classification of Tabular Data

Federico Errica

https://openreview.net/forum?id=vEzcRdiTkP

Keywords: Deep Graph Networks, Graph Neural Networks, Graph Representation Learning, Nearest Neighbors, Node Classification, Tabular Data

Compressor summary: The paper questions the benefits of using nearest neighbor graphs as a way to represent missing graph structures in machine learning problems and suggests exploring other techniques instead.


Neural Algorithmic Reasoning Without Intermediate Supervision

Gleb Rodionov, Liudmila Prokhorenkova

https://openreview.net/forum?id=vBwSACOB3x

Keywords: neural algorithmic reasoning, graph neural networks, self-supervised regularization

Compressor summary: This paper proposes an unsupervised method to learn neural algorithms that can generalize to larger input sizes and achieve state-of-the-art results on classic algorithmic problems like sorting.


An Exploration-by-Optimization Approach to Best of Both Worlds in Linear Bandits

Shinji Ito, Kei Takemura

https://openreview.net/forum?id=vBHKSTgcYQ

Keywords: bandit, linear bandit, best of both worlds, exploration by optimization

Compressor summary: The paper presents a linear bandit algorithm that performs near-optimally in both stochastic and adversarial settings, using exploration by optimization with specific regret bounds depending on dimension, time horizon, and sub-optimality gap.


Reflexion: language agents with verbal reinforcement learning

Noah Shinn, Federico Cassano, Ashwin Gopinath, Karthik R Narasimhan, Shunyu Yao

https://openreview.net/forum?id=vAElhFcKW6

Keywords: language model, reasoning, decision making, programming

Compressor summary: Reflexion is a framework that helps language agents learn from feedback by reflecting on it verbally and storing it in memory for better decision-making in future trials, improving performance across various tasks.


MVDiffusion: Enabling Holistic Multi-view Image Generation with Correspondence-Aware Diffusion

Shitao Tang, Fuyang Zhang, Jiacheng Chen, Peng Wang, Yasutaka Furukawa

https://openreview.net/forum?id=vA0vj1mY77

Keywords: multiview; image generation; generative model; diffusion models

Compressor summary: MVDiffusion is a method for generating multi-view images with high resolution and rich content by concurrently interacting with all views using a correspondence-aware attention mechanism, which includes modules for generation, interpolation, and super-resolution.


Deep Neural Collapse Is Provably Optimal for the Deep Unconstrained Features Model

Peter Súkeník, Marco Mondelli, Christoph H Lampert

https://openreview.net/forum?id=v9yC7sSXf3

Keywords: neural collapse, unconstrained features model, deep learning

Compressor summary: The paper studies deep neural collapse, a phenomenon where neural networks lose structure in earlier layers, and proposes a generalized analytical framework to explain it.


Pitfall of Optimism: Distributional Reinforcement Learning by Randomizing Risk Criterion

Taehyun Cho, Seungyub Han, Heesoo Lee, Kyungjae Lee, Jungwoo Lee

https://openreview.net/forum?id=v8u3EFAyW9

Keywords: distributional reinforcement learning, risk

Compressor summary: The paper proposes a new distributional reinforcement learning method that uses randomized risk criterion for exploration and proves its convergence, optimality, and effectiveness in various environments.


End-To-End Latent Variational Diffusion Models for Inverse Problems in High Energy Physics

Alexander Shmakov, Kevin Greif, Michael James Fenton, Aishik Ghosh, Pierre Baldi, Daniel Whiteson

https://openreview.net/forum?id=v7WWesSiOu

Keywords: Diffusion, Variational, VAE, LDM, Physics, Unfolding

Compressor summary: The paragraph discusses a new deep learning method for correcting detector effects in particle physics measurements at the Large Hadron Collider, which outperforms existing methods in accuracy and efficiency.


Adaptive Algorithms for Relaxed Pareto Set Identification

Cyrille Kone, Emilie Kaufmann, Laura Richert

https://openreview.net/forum?id=v6jIxRRDyD

Keywords: bandit, pure-exploration, pareto front, pareto set

Compressor summary: The paper proposes a new sampling strategy called Adaptive Pareto Exploration that can identify a relevant subset of the Pareto optimal set in multi-objective bandit problems, and demonstrates its effectiveness on vaccination strategies against Covid-19.


Deep Equilibrium Based Neural Operators for Steady-State PDEs

Tanya Marwah, Ashwini Pokle, J Zico Kolter, Zachary Chase Lipton, Jianfeng Lu, Andrej Risteski

https://openreview.net/forum?id=v6YzxwJlQn

Keywords: Deep Equilibrium Models, Partial Differential Equations, Neural Operators

Compressor summary: The authors propose FNO-DEQ, a deep learning architecture that solves steady-state PDEs more efficiently and robustly than existing methods by using weight-tied neural networks and exploiting the fixed point nature of their solutions.


Winner Takes It All: Training Performant RL Populations for Combinatorial Optimization

Nathan Grinsztajn, Daniel Furelos-Blanco, Shikha Surana, Clément Bonnet, Thomas D Barrett

https://openreview.net/forum?id=v6VpqGcGAR

Keywords: Combinatorial Optimization, Reinforcement Learning, TSP, CVRP, JSSP

Compressor summary: The paper introduces Poppy, a training method for populations of reinforcement learning policies that induce unsupervised specialization to solve hard combinatorial optimization problems effectively.


Improving Adversarial Robustness via Information Bottleneck Distillation

Huafeng Kuang, Hong Liu, YONGJIAN WU, Shin'ichi Satoh, Rongrong Ji

https://openreview.net/forum?id=v5Aaxk4sSy

Keywords: Information Bottleneck, Adversarial training, Adversarial robustness, Knowledge distillation

Compressor summary: The study proposes an Information Bottleneck Distillation approach that uses prior knowledge from a pre-trained model to improve the information bottlenecks and robustness of deep neural networks against adversarial attacks.


UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild

Can Qin, Shu Zhang, Ning Yu, Yihao Feng, Xinyi Yang, Yingbo Zhou, Huan Wang, Juan Carlos Niebles, Caiming Xiong, Silvio Savarese, Stefano Ermon, Yun Fu, Ran Xu

https://openreview.net/forum?id=v54eUIayFh

Keywords: Image Generation, Multi-modal, HyperNet

Compressor summary: UniControl is a new generative foundation model that can handle various visual conditions and generate images with precise controls, while still allowing for arbitrary language prompts.


MonoUNI: A Unified Vehicle and Infrastructure-side Monocular 3D Object Detection Network with Sufficient Depth Clues

Jinrang Jia, Zhenjia Li, Yifeng Shi

https://openreview.net/forum?id=v2oGdhbKxi

Keywords: 3D detection, deep learning, autonomous driving

Compressor summary: The paper proposes a unified optimization target called normalized depth for monocular 3D detection of vehicle and infrastructure sides in autonomous driving, and introduces 3D normalized cube depth to improve accuracy.


H2RBox-v2: Incorporating Symmetry for Boosting Horizontal Box Supervised Oriented Object Detection

Yi Yu, Xue Yang, Qingyun Li, Yue Zhou, Feipeng Da, Junchi Yan

https://openreview.net/forum?id=v1VVKaMYbk

Keywords: Oriented object detection, self-supervised learning

Compressor summary: H2RBox-v2 is a new method for oriented object detection that uses reflection symmetry and self-supervision to improve performance and reduce the need for high-quality annotations.


Egocentric Planning for Scalable Embodied Task Achievement

Xiaotian Liu, Hector Palacios, Christian Muise

https://openreview.net/forum?id=v0lkbp66Uw

Keywords: Embodied AI, High-Level Actions, Symbolic Reasoning, Replanning, ALFRED Challenge, Flexible Task Achievement, User-Goal Understanding, Object Types and Actions, Perception Grounding

Compressor summary: Egocentric Planning combines symbolic planning and Object-oriented POMDPs to solve complex domestic tasks with visual perception and natural language processing, achieving a high success rate in ALFRED benchmark and winning the CVPR Embodied AI workshop challenge.


Temporal Continual Learning with Prior Compensation for Human Motion Prediction

Jianwei Tang, Jiangxin Sun, Xiaotong Lin, lifang zhang, Wei-Shi Zheng, Jian-Fang Hu

https://openreview.net/forum?id=v0GzRLvVp3

Keywords: Human Motion Prediction; Temporal Continual Learning; Prior Compensation Factor

Compressor summary: The paper proposes a new framework called Temporal Continual Learning (TCL) for human motion prediction that better preserves prior information and adapts to different models, datasets, and applications.


Online robust non-stationary estimation

Abishek Sankararaman, Murali Balakrishnan

https://openreview.net/forum?id=uzOBDerK1j

Keywords: Estimation, heavy-tails, distribution shifts, regret

Compressor summary: The text describes a new algorithm for estimating time-varying parameters from noisy and corrupted high-dimensional data-streams that is adaptive to drift, robust to heavy-tails and corruptions, requires no distributional knowledge, and can be implemented online.


Dynamic Context Pruning for Efficient and Interpretable Autoregressive Transformers

Sotiris Anagnostidis, Dario Pavllo, Luca Biggio, Lorenzo Noci, Aurelien Lucchi, Thomas Hofmann

https://openreview.net/forum?id=uvdJgFFzby

Keywords: Transformers, Context-pruning, Efficient Transformer

Compressor summary: The study proposes a method that prunes uninformative tokens from context during generation, reducing computational and memory requirements while maintaining model expressiveness and interpretability.


Training Your Image Restoration Network Better with Random Weight Network as Optimization Function

Man Zhou, Naishan Zheng, Yuan Xu, Chun-Le Guo, Chongyi Li

https://openreview.net/forum?id=uv3ge0goPa

Keywords: Image restoration, low-light image enhancement, image de-noising

Compressor summary: The study proposes using random weight networks as constraints for training better image restoration networks, and explores four prototypes with different strategies and variants.


Can semi-supervised learning use all the data effectively? A lower bound perspective

Alexandru Tifrea, Gizem Yüce, Amartya Sanyal, Fanny Yang

https://openreview.net/forum?id=utreNaM1VY

Keywords: semi-supervised learning, statistical lower bound

Compressor summary: The paragraph discusses the limitations of existing theory on semi-supervised learning (SSL) algorithms, which may not always improve upon unsupervised and supervised learning algorithms, but can still perform well in real-world scenarios.


All Points Matter: Entropy-Regularized Distribution Alignment for Weakly-supervised 3D Segmentation

Liyao Tang, Zhe Chen, Shanshan Zhao, Chaoyue Wang, Dacheng Tao

https://openreview.net/forum?id=utQms7PPx5

Keywords: point cloud segmentation, weak supervision

Compressor summary: The paper proposes ERDA, a learning strategy that uses entropy regularization and distribution alignment to improve weakly supervised 3D segmentation tasks with pseudo-labels, achieving state-of-the-art performance with minimal annotations.


Fused Gromov-Wasserstein Graph Mixup for Graph-level Classifications

Xinyu Ma, Xu Chu, Yasha Wang, Yang Lin, Junfeng Zhao, Liantao Ma, Wenwu Zhu

https://openreview.net/forum?id=uqkUguNu40

Keywords: Graph Data Augmentation, Graph Mixup, Fused Gromov Wasserstein

Compressor summary: The paper introduces FGWMixup, a graph data augmentation method that uses optimal transport to find an optimal inter-graph node matching strategy, improving the generalizability and robustness of GNNs.


Optimal approximation using complex-valued neural networks

Paul Geuchen, Felix Voigtlaender

https://openreview.net/forum?id=uotGmrcooz

Keywords: complex-valued neural networks, approximation rates

Compressor summary: The paper analyzes the expressivity and approximation properties of complex-valued neural networks (CVNNs) with various activation functions, showing that their error scales as $m^{-k/(2n)}$ for smooth target functions and is optimal under a continuity assumption.


PriorBand: Practical Hyperparameter Optimization in the Age of Deep Learning

Neeratyoy Mallik, Eddie Bergman, Carl Hvarfner, Danny Stoll, Maciej Janowski, Marius Lindauer, Luigi Nardi, Frank Hutter

https://openreview.net/forum?id=uoiwugtpCH

Keywords: Hyperparameter Optimization, Deep Learning

Compressor summary: PriorBand is an HPO algorithm for DL that combines expert knowledge and cheap proxy tasks to efficiently optimize hyperparameters.


Minimax Forward and Backward Learning of Evolving Tasks with Performance Guarantees

Veronica Alvarez, Santiago Mazuelas, Jose A. Lozano

https://openreview.net/forum?id=uoRiO855Sj

Keywords: Concept drift, Continual learning, Minimax classification, Performance guarantees

Compressor summary: IMRCs use forward and backward learning to improve classification accuracy on evolving tasks with few samples per task.


Sample-efficient Multi-objective Molecular Optimization with GFlowNets

Yiheng Zhu, Jialu Wu, Chaowen Hu, Jiahuan Yan, Chang-Yu Hsieh, Tingjun Hou, Jian Wu

https://openreview.net/forum?id=uoG1fLIK2s

Keywords: drug discovery, multi-objective molecular optimization, Bayesian optimization, generative flow networks

Compressor summary: The paper introduces HN-GFN, a novel algorithm that uses a single hypernetwork to optimize multiple objectives and diversity in molecule design, and a hindsight-like strategy to improve learning and sharing of high-quality molecules.


Energy-Based Sliced Wasserstein Distance

Khai Nguyen, Nhat Ho

https://openreview.net/forum?id=umvV3yvo4N

Keywords: Sliced Wasserstein, Monte Carlo Methods, Point-Cloud, Optimal Transport

Compressor summary: The paper introduces a new way to choose the slicing distribution for the Wasserstein distance, called energy-based sliced Wasserstein (EBSW), which improves on previous approaches in terms of discrimination, cost, and stability.


Enhancing Knowledge Transfer for Task Incremental Learning with Data-free Subnetwork

Qiang Gao, Xiaojun Shan, Yuchen Zhang, Fan Zhou

https://openreview.net/forum?id=uj9PxVTVqq

Keywords: data-free subnetwork, task-incremental learning, knowledge transfer, mask

Compressor summary: DSN is a novel method for neuron-wise incremental learning that transfers knowledge across tasks using masks and data-free replay, avoiding forgetting and privacy issues.


Annotator: A Generic Active Learning Baseline for LiDAR Semantic Segmentation

Binhui Xie, Shuang Li, qingju guo, Chi Harold Liu, Xinjing Cheng

https://openreview.net/forum?id=uiiVSVADDc

Keywords: Active Learning, LiDAR Semantic Segmentation, Domain Adaptation

Compressor summary: Annotator is an active learning method that uses voxel-centric online selection to efficiently annotate LiDAR point clouds for semantic segmentation under various distribution shifts.


Mechanic: A Learning Rate Tuner

Ashok Cutkosky, Aaron Defazio, Harsh Mehta

https://openreview.net/forum?id=uhKtQMn21D

Keywords: optimization, deep learning, online convex optimization

Compressor summary: Mechanic is a technique for automatically adjusting the learning rate in any optimization algorithm, which performs well in various deep learning tasks and can match or surpass manual tuning.


HyenaDNA: Long-Range Genomic Sequence Modeling at Single Nucleotide Resolution

Eric Nguyen, Michael Poli, Marjan Faizi, Armin W Thomas, Michael Wornow, Callum Birch-Sykes, Stefano Massaroli, Aman Patel, Clayton M. Rabideau, Yoshua Bengio, Stefano Ermon, Christopher Re, Stephen Baccus

https://openreview.net/forum?id=ubzNoJjOKj

Keywords: genomics, hyena, foundation models, large language models, transformers

Compressor summary: HyenaDNA is a large language model that can process up to 1 million tokens of human genome data, enabling long-range interactions and single nucleotide resolution for genomic tasks, and outperforms previous models on various benchmarks.


Uncovering Meanings of Embeddings via Partial Orthogonality

Yibo Jiang, Bryon Aragam, Victor Veitch

https://openreview.net/forum?id=ubp5s2tgXq

Keywords: embedding, representation, graphical models, partial orthogonality, Markov boundary

Compressor summary: The paper explores how semantic structure is encoded in vector embeddings using partial orthogonality and introduces embeddings that preserve conditional independence structures.


SnapFusion: Text-to-Image Diffusion Model on Mobile Devices within Two Seconds

Yanyu Li, Huan Wang, Qing Jin, Ju Hu, Pavlo Chemerys, Yun Fu, Yanzhi Wang, Sergey Tulyakov, Jian Ren

https://openreview.net/forum?id=ubgdInLSF9

Keywords: Text-to-Image, Diffusion model, mobile devices, distillation

Compressor summary: The authors introduce a generic approach to make text-to-image diffusion models faster and more accessible on mobile devices, improving efficiency by modifying the network architecture and step distillation, and achieving better results than Stable Diffusion v$1.5$.


Domain Agnostic Fourier Neural Operators

Ning Liu, Siavash Jafarzadeh, Yue Yu

https://openreview.net/forum?id=ubap5FKbJs

Keywords: Operator-Regression Neural Networks, Neural Operators, Data-Driven Physics Modeling, Geometrical and Topological Shape Changes

Compressor summary: The authors propose a novel neural operator architecture, DAFNO, that can learn nonlinear mappings between function spaces on irregular geometries and evolving domains using FFT, achieving state-of-the-art accuracy in material modeling and airfoil simulation, and demonstrating generalizability to unseen crack patterns.


Small Total-Cost Constraints in Contextual Bandits with Knapsacks, with Application to Fairness

Evgenii E Chzhen, Christophe Giraud, Zhen LI, Gilles Stoltz

https://openreview.net/forum?id=uZvG0HLkOB

Keywords: mutli-armed bandits, bandits with knapsacks, primal-dual approaches

Compressor summary: The paper proposes a new algorithm for contextual bandit problems with knapsacks that can handle cost constraints of the order of $\sqrt{T}$, which is more efficient and simpler than existing approaches.


CL-NeRF: Continual Learning of Neural Radiance Fields for Evolving Scene Representation

Xiuzhe Wu, Peng Dai, Weipeng DENG, Handi Chen, Yang Wu, Yan-Pei Cao, Ying Shan, XIAOJUAN QI

https://openreview.net/forum?id=uZjpSBTPik

Keywords: Neural Radiance Field; Continual Learning; Scene Representation

Compressor summary: The paper introduces CL-NeRF, a method for efficiently adapting Neural Radiance Fields to scene changes using continual learning and a new benchmark for evaluating it.


Learning Neural Implicit through Volume Rendering with Attentive Depth Fusion Priors

Pengchong Hu, Zhizhong Han

https://openreview.net/forum?id=uWNqy09dFW

Keywords: 3D Reconstruction, SDF, Neural Rendering, Implicit Representations, SLAM

Compressor summary: The paragraph introduces a novel method for 3D reconstruction from multi-view images using neural implicit representations, volume rendering with an attentive depth fusion prior, and a Truncated Signed Distance Function (TSDF).


Complementary Benefits of Contrastive Learning and Self-Training Under Distribution Shift

Saurabh Garg, Amrith Setlur, Zachary Chase Lipton, Sivaraman Balakrishnan, Virginia Smith, Aditi Raghunathan

https://openreview.net/forum?id=uWGH6jDTVv

Keywords: contrastive learning; self training; distribution shift; semi supervised learning; unsupervised domain adaptation

Compressor summary: The paper investigates how combining self-training and contrastive learning techniques affects unsupervised domain adaptation and semi-supervised learning, showing improved accuracy in the former but not the latter.


Alternating Gradient Descent and Mixture-of-Experts for Integrated Multimodal Perception

Hassan Akbari, Dan Kondratyuk, Yin Cui, Rachel Hornung, Huisheng Wang, Hartwig Adam

https://openreview.net/forum?id=uTlKUAm68H

Keywords: Alternating Gradient Descent, Multimodal, Mixture of Experts, AGD, MoE, Deep Learning, Optimization

Compressor summary: IMP is a multimodal multi-task training method that uses alternating gradient descent and mixture-of-experts to efficiently improve model performance on various tasks, achieving state-of-the-art results in zero-shot video classification.


Self-Supervised Reinforcement Learning that Transfers using Random Features

Boyuan Chen, Chuning Zhu, Pulkit Agrawal, Kaiqing Zhang, Abhishek Gupta

https://openreview.net/forum?id=uRewSnLJAa

Keywords: deep reinforcement learning, self-supervised learning

Compressor summary: The paper proposes a self-supervised RL method that combines model-free and model-based approaches for transferring behaviors across tasks with different rewards in complex environments.


Sampling weights of deep neural networks

Erik Lien Bolager, Iryna Burak, Chinmay Datar, Qing Sun, Felix Dietrich

https://openreview.net/forum?id=uRHpgo6TMR

Keywords: random sampling, neural network parameters, iterative optimization

Compressor summary: The authors propose a method for quickly training fully-connected neural networks without gradient computations by using a data-dependent probability distribution and an efficient sampling algorithm.


A Logic for Expressing Log-Precision Transformers

William Merrill, Ashish Sabharwal

https://openreview.net/forum?id=uR8TtWCIsr

Keywords: transformers, logic, reasoning, circuit complexity, mechanistic interpretability

Compressor summary: The paper investigates whether log-precision transformers, which can attend universally, can be represented by a generalized first-order logic with majority-vote quantifiers.


Data Selection for Language Models via Importance Resampling

Sang Michael Xie, Shibani Santurkar, Tengyu Ma, Percy Liang

https://openreview.net/forum?id=uPSQv0leAu

Keywords: Language models, pretraining, data selection, fine-tuning

Compressor summary: DSIR is a data selection method that uses importance resampling in a reduced feature space to match a target distribution, achieving comparable or better results than expert curation or simple heuristics.


Breadcrumbs to the Goal: Supervised Goal Selection from Human-in-the-Loop Feedback

Marcel Torne Villasevil, Max Balsells I Pamies, Zihan Wang, Samedh Desai, Tao Chen, Pulkit Agrawal, Abhishek Gupta

https://openreview.net/forum?id=uOEeui0rL7

Keywords: Learning from human preferences, self-supervised learning, exploration in reinforcement learning

Compressor summary: HUGE is a technique that uses low-quality human feedback to guide exploration for reinforcement learning without requiring reward specification or exploration bonuses.


Sample Complexity Bounds for Score-Matching: Causal Discovery and Generative Modeling

Zhenyu Zhu, Francesco Locatello, Volkan Cevher

https://openreview.net/forum?id=uNnPWR66b8

Keywords: Causal discovery, Score matching, Score-based generative modeling

Compressor summary: The paper studies how well deep neural networks can estimate a score function for causal discovery using score-matching methods.


Adaptive Linear Estimating Equations

Mufang Ying, Koulik Khamaru, Cun-Hui Zhang

https://openreview.net/forum?id=uNmKBZrRZC

Keywords: bandit algorithm, statistical inference, adaptively collected data, asymptotic normality

Compressor summary: The paper proposes a debiased estimator for sequential data collection that improves statistical inference by achieving both non-asymptotic performance and asymptotic normality.


The Graph Pencil Method: Mapping Subgraph Densities to Stochastic Block Models

Lee M. Gunderson, Gecia Bravo-Hermsdorff, Peter Orbanz

https://openreview.net/forum?id=uN71BdBEG8

Keywords: Stochastic block model, SBM, graphons, matrix pencil method, method of moments

Compressor summary: The paper presents a method to find the parameters of a stochastic block model from a finite set of subgraph densities using a unique translation map that is efficient and easy to compute.


Fair Allocation of Indivisible Chores: Beyond Additive Costs

Bo Li, Fangxiao Wang, Yu Zhou

https://openreview.net/forum?id=uJmsYZiu3E

Keywords: fair allocation of chores, beyond additive cost functions, bin packing, job scheduling

Compressor summary: The paper explores the limitations of achieving exact or good approximations of maximin share fairness in allocating indivisible tasks with subadditive costs and studies two specific settings, bin packing and job scheduling, where constant approximate allocations exist.


Exploring Geometry of Blind Spots in Vision models

Sriram Balasubramanian, Gaurang Sriramanan, Vinu Sankar Sadasivan, Soheil Feizi

https://openreview.net/forum?id=uJ3qNIsDGF

Keywords: Neural networks, Vision models, blind spots, undersensitivity, invariance, level set geometry, input connectivity

Compressor summary: The authors study how vision models like CNNs and Transformers are under-sensitive to input perturbations and propose a method to explore the geometry of "equi-confidence" level sets in these networks, revealing a star-like structure.


FairLISA: Fair User Modeling with Limited Sensitive Attributes Information

Zheng Zhang, Qi Liu, Hao Jiang, Fei Wang, Yan Zhuang, Le Wu, Weibo Gao, Enhong Chen

https://openreview.net/forum?id=uFpjPJMkv6

Keywords: fairness, user modeling

Compressor summary: The paper proposes a new framework called FairLISA to create user models that are fair and accurate even when some sensitive data is missing or unknown.


Focus Your Attention when Few-Shot Classification

Haoqing Wang, Shibo Jie, Zhi-Hong Deng

https://openreview.net/forum?id=uFlE0qgtRO

Keywords: few-shot image classification, fine-tuning, vision transformers

Compressor summary: The paper proposes a method to improve few-shot image classification with pre-trained vision transformers by using attention and gradient information to locate key entities in support images, which helps the model focus on class-related features and generalize better.


Static and Sequential Malicious Attacks in the Context of Selective Forgetting

CHENXU ZHAO, Wei Qian, Zhitao Ying, Mengdi Huai

https://openreview.net/forum?id=uEJfW3OtUm

Keywords: Selective forgetting, static setting, sequential setting, security and robustness

Compressor summary: The paper explores malicious data update attacks on machine unlearning systems and proposes new attack methods and algorithms to exploit them.


Efficient Robust Bayesian Optimization for Arbitrary Uncertain inputs

Lin Yang, Junlong Lyu, Wenlong Lyu, Zhitang Chen

https://openreview.net/forum?id=uDV4lA0gZ6

Keywords: bayesian optimization, robust optimization

Compressor summary: The paper proposes AIRBO, a robust Bayesian Optimization algorithm that models uncertain inputs using Maximum Mean Discrepancy and Nystrom approximation, achieving state-of-the-art performance under various input uncertainties.


(S)GD over Diagonal Linear Networks: Implicit bias, Large Stepsizes and Edge of Stability

Mathieu Even, Scott Pesme, Suriya Gunasekar, Nicolas Flammarion

https://openreview.net/forum?id=uAyElhYKxg

Keywords: SGD, GD, implicit bias, large stepsizes, edge of stability, diagonal linear networks

Compressor summary: The paper studies how randomness and big steps affect gradient descent and stochastic gradient descent in linear networks with two layers, proving their convergence and analysing how these factors influence solution quality in regression problems.


ReDS: Offline RL With Heteroskedastic Datasets via Support Constraints

Anikait Singh, Aviral Kumar, Quan Vuong, Yevgen Chebotar, Sergey Levine

https://openreview.net/forum?id=u8srPlinoj

Keywords: offline RL, support constraints, heteroskedastic data

Compressor summary: Offline reinforcement learning methods struggle to learn from data with non-uniform variability due to distribution constraints; conservative Q-learning (CQL) with reweighting improves performance by allowing the learned policy to choose how closely to follow the behavior policy per state.


Flocks of Stochastic Parrots: Differentially Private Prompt Learning for Large Language Models

Haonan Duan, Adam Dziedzic, Nicolas Papernot, Franziska Boenisch

https://openreview.net/forum?id=u6Xv3FuF8N

Keywords: differential privacy, in-context learning, trustworthy ML

Compressor summary: The authors propose a method to privately learn how to prompt large language models without exposing the data used for prompting, which can maintain high accuracy while protecting user privacy.


Real-World Image Variation by Aligning Diffusion Inversion Chain

Yuechen ZHANG, Jinbo Xing, Eric Lo, Jiaya Jia

https://openreview.net/forum?id=u6Ibs4hTJH

Keywords: image variation, diffusion model, image generation, text-driven image editing

Compressor summary: RIVAL is a novel inference pipeline that generates high-quality image variations by aligning the diffusion process to the source image's inversion chain using cross-image self-attention and step-wise distribution normalization.


MADG: Margin-based Adversarial Learning for Domain Generalization

Aveen Dayal, Vimal K B, Linga Reddy Cenkeramaddi, C Krishna Mohan, Abhinav Kumar, Vineeth N. Balasubramanian

https://openreview.net/forum?id=u6BYyPuD29

Keywords: Domain Generalization, Margin Loss, Adversarial Learning, Domain Adaptation

Compressor summary: The paper proposes a new adversarial learning method for domain generalization called MADG, which uses a margin loss-based discrepancy metric to learn domain-invariant features and achieves better generalization to unseen domains.


Graph Denoising Diffusion for Inverse Protein Folding

Kai Yi, Bingxin Zhou, Yiqing Shen, Pietro Lio, Yu Guang Wang

https://openreview.net/forum?id=u4YXKKG5dX

Keywords: inverse folding, graph neural networks, roto-translation equivariance, diffusion model

Compressor summary: The proposed graph denoising diffusion model generates diverse amino acid sequences for a given protein backbone using physiochemical properties, local environment, and prior knowledge of amino acid replacements, outperforming existing methods in inverse protein folding.


Uncovering motifs of concurrent signaling across multiple neuronal populations

Evren Gokcen, Anna Ivic Jasper, Alison Xu, Adam Kohn, Christian K. Machens, Byron M. Yu

https://openreview.net/forum?id=u39QQh5L8Q

Keywords: neuroscience, multi-population neural recordings, dimensionality reduction, latent variable models, Gaussian processes

Compressor summary: The paragraph describes a new framework for analyzing signals among multiple neuronal populations in different brain networks, which reveals how they communicate and evolve over time.


Robust Data Valuation with Weighted Banzhaf Values

Weida Li, Yaoliang Yu

https://openreview.net/forum?id=u359tNBpxF

Keywords: data valuation, robustness, weighted Banzhaf values

Compressor summary: The paper proposes a new robust data valuation method using weighted Banzhaf values and introduces Kronecker noise to analyze stochasticity.


PlanE: Representation Learning over Planar Graphs

Radoslav Dimitrov, Zeyang Zhao, Ralph Abboud, Ismail Ilkan Ceylan

https://openreview.net/forum?id=u2RJ0I3o3j

Keywords: Graph Representation Learning; Planar Graphs; Graph Property Prediction

Compressor summary: The paragraph introduces PlanE, a framework for learning complete invariants of planar graphs, which are special graph classes with efficient isomorphism testing algorithms, and reports its successful performance on benchmarks.


Knowledge Distillation Performs Partial Variance Reduction

Mher Safaryan, Alexandra Peste, Dan Alistarh

https://openreview.net/forum?id=tzxP9Rx0LV

Keywords: knowledge distillation, stochastic optimization, variance reduction

Compressor summary: The paper explores how knowledge distillation works as a stochastic variance reduction technique for enhancing student models with teacher models, and highlights the importance of careful parametrization for optimal performance.


GEX: A flexible method for approximating influence via Geometric Ensemble

SungYub Kim, Kyungsu Kim, Eunho Yang

https://openreview.net/forum?id=tz4ECtAu8e

Keywords: Influence Function, Geometric Ensemble, Loss Landscape

Compressor summary: The paper proposes a new method to improve Influence Function (IF) approximations for neural networks by addressing limitations and enhancing performance in various tasks.


InstanT: Semi-supervised Learning with Instance-dependent Thresholds

Muyang Li, Runze Wu, Haoyu Liu, Jun Yu, Xun Yang, Bo Han, Tongliang Liu

https://openreview.net/forum?id=txv7TnPvOi

Keywords: semi-supervised learning, pseudo-labeling

Compressor summary: The paper proposes a new SSL method that uses instance-dependent thresholds to select confident unlabeled instances and assign pseudo-labels with probabilistic guarantees.


Fed-FA: Theoretically Modeling Client Data Divergence for Federated Language Backdoor Defense

Zhiyuan Zhang, Deli Chen, Hao Zhou, Fandong Meng, Jie Zhou, Xu Sun

https://openreview.net/forum?id=txPdKZrrZF

Keywords: federated learning, backdoor learning, robust federated aggregation, data divergence

Compressor summary: The paper proposes a method called Federated F-Divergence-Based Aggregation (Fed-FA) to detect and exclude suspicious clients in federated NLP systems by modeling data divergence using f-divergence indicators and synthetic datasets.


DinoSR: Self-Distillation and Online Clustering for Self-supervised Speech Representation Learning

Alexander H. Liu, Heng-Jui Chang, Michael Auli, Wei-Ning Hsu, James R. Glass

https://openreview.net/forum?id=twmHKU3Ds4

Keywords: speech representation learning, self-supervised learning, self-distillation, discrete representation learning

Compressor summary: The paper presents a self-supervised speech representation learning model called DinoSR that combines masked language modeling, self-distillation, and online clustering to achieve state-of-the-art performance in various downstream tasks.


The Bayesian Stability Zoo

Shay Moran, Hilla Schefler, Jonathan Shafer

https://openreview.net/forum?id=tw4QaiiJex

Keywords: Algorithmic stability, Replicability, Differential Privacy, KL Stability, Mutual Information Stability, Global Stability, Perfect Generalization, PAC Learning, Littlestone Dimension, Clique Dimension, PAC Bayes

Compressor summary: The paper explores the connections between different definitions of stability in learning theory, including various types of privacy, generalization, and divergence-based stability notions.


Secure Out-of-Distribution Task Generalization with Energy-Based Models

Shengzhuang Chen, Long-Kai Huang, Jonathan Richard Schwarz, Yilun Du, Ying Wei

https://openreview.net/forum?id=tt7bQnTdRm

Keywords: meta-generalization, out-of-distribution tasks

Compressor summary: The EBML framework uses two neural networks to characterize any meta-training task distribution, enabling detection and adaptation of out-of-distribution tasks for safe and effective meta-learning.


Disentangled Counterfactual Learning for Physical Audiovisual Commonsense Reasoning

Changsheng Lv, Shuai Zhang, Yapeng Tian, Mengshi Qi, Huadong Ma

https://openreview.net/forum?id=trHfuGQyyr

Keywords: Physical Audiovisual;Commonsense Reasoning

Compressor summary: The paper proposes a Disentangled Counterfactual Learning (DCL) method for inferring physical commonsense from audiovisual data, which decouples video factors and models causal relationships using a variational autoencoder and a counterfactual learning module.


NetHack is Hard to Hack

Ulyana Piterbarg, Lerrel Pinto, Rob Fergus

https://openreview.net/forum?id=tp2nEZ5zfP

Keywords: imitation learning, NetHack

Compressor summary: The paper studies why neural policy learning struggles in NetHack and proposes improvements based on a winning symbolic agent's strategy, leading to a better neural agent that still falls behind symbolic and human players.


Derandomized novelty detection with FDR control via conformal e-values

Meshi Bashari, Amir Epstein, Yaniv Romano, Matteo Sesia

https://openreview.net/forum?id=toYvRJ7Zmy

Keywords: Conformal inference, Derandomization, E-values, False discovery rate, Out-of-distribution testing, Testing for outliers, Uncertainty

Compressor summary: The authors propose a method to make conformal inference more stable for novelty detection by using conformal e-values instead of p-values and weighting them with additional data information.


SafeDICE: Offline Safe Imitation Learning with Non-Preferred Demonstrations

Youngsoo Jang, Geon-Hyeong Kim, Jongmin Lee, Sungryull Sohn, Byoungjip Kim, Honglak Lee, Moontae Lee

https://openreview.net/forum?id=toEGuA9Qfn

Keywords: Imitation learning, Preference-based learning, Safe imitation learning

Compressor summary: The paper introduces SafeDICE, an offline imitation learning algorithm that learns a safe policy from non-preferred demonstrations and unlabeled data by estimating stationary distribution corrections.


Dream the Impossible: Outlier Imagination with Diffusion Models

Xuefeng Du, Yiyou Sun, Jerry Zhu, Yixuan Li

https://openreview.net/forum?id=tnRboxQIec

Keywords: Outlier imagination, machine learning

Compressor summary: The paper introduces Dream-OOD, a framework that generates realistic outlier images for machine learning models using in-distribution data and classes, without requiring manual data collection or cleaning.


Add and Thin: Diffusion for Temporal Point Processes

David Lüdke, Marin Biloš, Oleksandr Shchur, Marten Lienen, Stephan Günnemann

https://openreview.net/forum?id=tn9Dldam9L

Keywords: Point Processes, Diffusion, Temporal Data, Generative Model, Forecasting, Density Estimation, Denoising

Compressor summary: The paper introduces ADD-THIN, a probabilistic denoising diffusion model that improves long-term forecasting for event sequences with discrete and continuous components by using temporal point process frameworks.


SPACE: Single-round Participant Amalgamation for Contribution Evaluation in Federated Learning

Yi-Chung Chen, Hsi-Wen Chen, Shun-Guei Wang, Ming-Syan Chen

https://openreview.net/forum?id=tmxjuIFSEc

Keywords: Federated Learning, Contribution Evaluation, Shapley Value, Knowledge Amalgamation

Compressor summary: The paper introduces SPACE, an efficient method for evaluating participant contributions in federated learning using Federated Knowledge Amalgamation and Prototype-based Model Evaluation.


Exponential Lower Bounds for Fictitious Play in Potential Games

Ioannis Panageas, Nikolas Patris, Stratis Skoulakis, Volkan Cevher

https://openreview.net/forum?id=tkenkPYkxj

Keywords: fictitious play, convergence rate, potential games

Compressor summary: This paper studies the convergence rate of Fictitious Play in potential games and shows that it can take exponential time even for two-agent games.


Feature learning via mean-field Langevin dynamics: classifying sparse parities and beyond

Taiji Suzuki, Denny Wu, Kazusato Oko, Atsushi Nitanda

https://openreview.net/forum?id=tj86aGVNb3

Keywords: mean-field regime, feature learning, Neural network optimization, sparse parity function, classification, sample complexity

Compressor summary: The paragraph discusses how mean-field Langevin dynamics (MFLD) can optimize and improve generalization performance of neural networks for binary classification problems by learning features, and provides a new analysis of its sample complexity and convergence rate.


No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models

Jean Kaddour, Oscar Key, Piotr Nawrot, Pasquale Minervini, Matt Kusner

https://openreview.net/forum?id=thbXgJ8gNK

Keywords: language models, transformers, efficient training

Compressor summary: The authors investigate various efficient training algorithms for Transformer-based language models, but find that a fully-decayed learning rate outperforms them all.


GlyphControl: Glyph Conditional Control for Visual Text Generation

Yukang Yang, Dongnan Gui, Yuhui Yuan, Weicong Liang, Haisong Ding, Han Hu, Kai Chen

https://openreview.net/forum?id=thPI8hrA4V

Keywords: Generative Models, Visual Text Generation, Diffusion Models

Compressor summary: The paper introduces a new method called GlyphControl to improve visual text generation by using glyph instructions with an off-the-shelf Stable-Diffusion model and creates a dataset called LAION-Glyph for further research.


Bypassing spike sorting: Density-based decoding using spike localization from dense multielectrode probes

Yizi Zhang, Tianxiao He, Julien Boussard, Charlie Windolf, Olivier Winter, Eric M. Trautmann, Noam Roth, Hailey Barrel, Mark M Churchland, Nick Steinmetz, Erdem Varol, Cole Lincoln Hurwitz, Liam Paninski

https://openreview.net/forum?id=tgQRMrsxht

Keywords: neural decoding, brain-computer interfaces, spike sorting, variational inference, generative models

Compressor summary: The authors propose a new method for decoding neural activity without spike sorting, using mixture of Gaussians to model uncertainty and improve performance in brain-computer interfaces.


SheetCopilot: Bringing Software Productivity to the Next Level through Large Language Models

Hongxin Li, Jingran Su, Yuntao Chen, Qing Li, Zhaoxiang Zhang

https://openreview.net/forum?id=tfyr2zRVoK

Keywords: Large Language Model; Task Planning; Embodied AI; Robotics; Software Automation

Compressor summary: The authors propose SheetCopilot, an agent that uses natural language to control spreadsheets and complete various tasks, and present a dataset and evaluation pipeline to benchmark its performance.


Stable Diffusion is Unstable

Chengbin Du, Yanxi Li, Zhongwei Qiu, Chang Xu

https://openreview.net/forum?id=tesBViWnbx

Keywords: Adversarial Attack, Generative Model, Diffusion Model, Latent Diffusion Model, Conditional Latent Diffusion Model

Compressor summary: Auto-attack on Text-to-image Models (ATM) is a method to make text-to-image models more robust by generating perturbations that can prevent them from blending or losing primary subjects in generated images.


FACE: Evaluating Natural Language Generation with Fourier Analysis of Cross-Entropy

Zuhao Yang, Yingfang Yuan, Yang Xu, SHUO ZHAN, Huajun Bai, Kefan Chen

https://openreview.net/forum?id=tdyLryDebq

Keywords: natural language generation; evaluation metrics; cross-entropy; language model

Compressor summary: The proposed FACE metrics use Fourier Analysis of Cross-Entropy to measure similarity between model-generated and human language, effectively identifying the gap between them and reflecting various factors such as model size and sampling methods.


CQM: Curriculum Reinforcement Learning with a Quantized World Model

Seungjae Lee, Daesol Cho, Jonghae Park, H. Jin Kim

https://openreview.net/forum?id=tcotyjon2a

Keywords: Reinforcement Learning, Curriculum Learning, Goal-conditioned RL

Compressor summary: The authors propose a new method for curriculum RL that automatically defines a semantic goal space from continuous observations and suggests uncertainty and temporal distance-aware goals, improving exploration and performance on goal-reaching tasks.


Tracr: Compiled Transformers as a Laboratory for Interpretability

David Lindner, Janos Kramar, Sebastian Farquhar, Matthew Rahtz, Thomas McGrath, Vladimir Mikulik

https://openreview.net/forum?id=tbbId8u7nP

Keywords: interpretability, transformers, language models, RASP, Tracr, mechanistic interpretability

Compressor summary: The authors introduce a compiler called Tracr that generates structured transformer models from human-readable programs, enabling experiments and interpretability evaluations.


Ignorance is Bliss: Robust Control via Information Gating

Manan Tomar, Riashat Islam, Matthew E. Taylor, Sergey Levine, Philip Bachman

https://openreview.net/forum?id=tW2KSph9o8

Keywords: representation learning, mutual information

Compressor summary: Information gating is a method for learning efficient and robust representations by selectively revealing or hiding information depending on the task at hand.


Language Semantic Graph Guided Data-Efficient Learning

Wenxuan Ma, Shuang Li, Lincan Cai, Jingxuan Kang

https://openreview.net/forum?id=tUyW68cRqr

Keywords: Data-Efficient Learning, Language Semantic Graph

Compressor summary: The paragraph discusses a new approach called Language Semantic Graph (LSG) that uses semantic information from labels to improve data efficiency in machine learning tasks, and demonstrates its effectiveness across different modalities and scenarios.


ForecastPFN: Synthetically-Trained Zero-Shot Forecasting

Samuel Dooley, Gurnoor Singh Khurana, Chirag Mohapatra, Siddartha Venkat Naidu, Colin White

https://openreview.net/forum?id=tScBQRNgjk

Keywords: Forecasting, Zero-shot, Synthetic Data

Compressor summary: ForecastPFN is a new zero-shot forecasting model that uses synthetic data and Bayesian inference to make more accurate and faster predictions than existing methods, even with limited real-life observations.


Human-Guided Complexity-Controlled Abstractions

Andi Peng, Mycal Tucker, Eoin M. Kenny, Noga Zaslavsky, Pulkit Agrawal, Julie Shah

https://openreview.net/forum?id=tSEeRl7ACo

Keywords: human-in-the-loop, representation learning, interpretability

Compressor summary: The authors propose training neural networks to generate a spectrum of discrete representations and control their complexity based on the task, which leads to better generalization and can be informed by human intuition.


Modeling Human Visual Motion Processing with Trainable Motion Energy Sensing and a Self-attention Network

Zitang Sun, Yen-Ju Chen, Yung-Hao Yang, Shin'ya Nishida

https://openreview.net/forum?id=tRKimbAk5D

Keywords: motion perception, optical flow estimation, attention mechanism, psychophysics, In silico neurophysiology, human vision

Compressor summary: The paper proposes a new image-computable model that combines biological and computer vision approaches to simulate human motion perception in complex scenes.


D$^2$CSG: Unsupervised Learning of Compact CSG Trees with Dual Complements and Dropouts

Fenggen Yu, Qimin Chen, Maham Tanveer, Ali Mahdavi Amiri, Hao Zhang

https://openreview.net/forum?id=tQYGjnxPOm

Keywords: 3D reconstruction, constructive solid geometry, unsupervised learning, compact shape assembly

Compressor summary: D$^2$CSG is a neural model that learns to reconstruct 3D CAD shapes using quadric primitives and a dedicated residual branch for complex shape complements.


Non-Stationary Bandits with Auto-Regressive Temporal Dependency

Qinyi Chen, Negin Golrezaei, Djallel Bouneffouf

https://openreview.net/forum?id=tP50lLiZIo

Keywords: non-stationary bandits; autoregressive model; low-regret policy; online learning algorithms

Compressor summary: The paper introduces a new non-stationary MAB framework that considers temporal dynamics in real-world applications like recommendation systems, and proposes an algorithm with two mechanisms for exploration-exploitation and discarding outdated information.


Optimistic Active Exploration of Dynamical Systems

Bhavya Sukhija, Lenart Treven, Cansu Sancaktar, Sebastian Blaes, Stelian Coros, Andreas Krause

https://openreview.net/forum?id=tLrkjK128n

Keywords: Active Exploration, Reinforcement Learning, Dynamical Systems

Compressor summary: OPAX is an algorithm for active exploration in reinforcement learning that uses probabilistic models to quantify uncertainty and maximizes information gain, enabling zero-shot solutions for multiple tasks.


Not All Neuro-Symbolic Concepts Are Created Equal: Analysis and Mitigation of Reasoning Shortcuts

Emanuele Marconato, Stefano Teso, Antonio Vergari, Andrea Passerini

https://openreview.net/forum?id=tLTtqySDFb

Keywords: Neuro-Symbolic Integration, Trustworthy AI, Concept Learning, Learning Shortcuts, Mitigation Strategies

Compressor summary: The paragraph discusses how neuro-symbolic predictive models can suffer from reasoning shortcuts that compromise their advantages, and proposes mitigation strategies to address this issue.


Emergent Communication in Interactive Sketch Question Answering

Zixing Lei, Yiming Zhang, Yuxin Xiong, Siheng Chen

https://openreview.net/forum?id=tLEDsaKuDh

Keywords: Emergent communication, Interactive, Question Answering

Compressor summary: The authors propose a new task (ISQA) and system for vision-based emergent communication, which uses sketches to answer questions about images in multiple rounds, improving accuracy, complexity, and interpretability.


Parallel-mentoring for Offline Model-based Optimization

Can Chen, Christopher Beckham, Zixuan Liu, Xue Liu, Christopher Pal

https://openreview.net/forum?id=tJwyg9Zg9G

Keywords: offline model-based optimization, bi-level optimization

Compressor summary: The authors propose parallel-mentoring, a novel method that uses majority voting and adaptive soft-labeling to improve model-based optimization with multiple proxies, addressing out-of-distribution issues in black-box objective function maximization.


Offline RL with Discrete Proxy Representations for Generalizability in POMDPs

Pengjie Gu, Xinyu Cai, Dong Xing, Xinrun Wang, Mengchen Zhao, Bo An

https://openreview.net/forum?id=tJN664ZNVG

Keywords: Offline RL, POMDP

Compressor summary: ORDER is a probabilistic framework for offline RL that uses discrete state representations and proxy representations to improve robustness against diverse masked observabilities.


Can You Rely on Your Model Evaluation? Improving Model Evaluation with Synthetic Test Data

Boris van Breugel, Nabeel Seedat, Fergus Imrie, Mihaela van der Schaar

https://openreview.net/forum?id=tJ88RBqupo

Keywords: model evaluation, tabular, synthetic data

Compressor summary: The paper proposes 3S Testing, a method that uses generative models to create synthetic test data for small subgroups and simulate distributional shifts, improving the evaluation of machine learning model performance on diverse and underrepresented groups.


Pre-Training Protein Encoder via Siamese Sequence-Structure Diffusion Trajectory Prediction

Zuobai Zhang, Minghao Xu, Aurelie Lozano, Vijil Chenthamarakshan, Payel Das, Jian Tang

https://openreview.net/forum?id=tIzbNQko3c

Keywords: Protein representation learning, diffusion models, self-supervised learning

Compressor summary: The paper proposes a pre-training method for protein encoders using joint sequence-structure diffusion models and enhances it with a technique to capture conformational variations among protein conformers.


Latent Graph Inference with Limited Supervision

Jianglin Lu, Yi Xu, Huan Wang, Yue Bai, Yun Fu

https://openreview.net/forum?id=tGuMwFnRZX

Keywords: Latent Graph Inference, CUR Matrix Decomposition, Graph Neural Networks

Compressor summary: The paper proposes a method to improve latent graph inference by restoring corrupted affinities and replenishing missed supervision for better node representation learning.


Distribution-Free Model-Agnostic Regression Calibration via Nonparametric Methods

Shang Liu, Zhongze Cai, Xiaocheng Li

https://openreview.net/forum?id=tGPx7HdBr4

Keywords: Regression calibration, model recalibration, conditional quantile, nonparametric method

Compressor summary: The paper proposes nonparametric methods for individual calibration of regression models that are efficient, consistent, and provide theoretical guarantees, addressing limitations of existing heuristic methods.


Maximum State Entropy Exploration using Predecessor and Successor Representations

Arnav Kumar Jain, Lucas Lehnert, Irina Rish, Glen Berseth

https://openreview.net/forum?id=tFsxtqGmkn

Keywords: Reinforcement Learning, Maximum state entropy exploration, Non-Markovian exploration, Successor Representation

Compressor summary: The authors propose a new exploration algorithm, $\eta\psi$-Learning, that learns efficient policies by using past experience to predict state visitation entropy and maximize coverage.


Single-Call Stochastic Extragradient Methods for Structured Non-monotone Variational Inequalities: Improved Analysis under Weaker Conditions

Sayantan Choudhury, Eduard Gorbunov, Nicolas Loizou

https://openreview.net/forum?id=tFeaLw9AWn

Keywords: Optimization, Machine Learning, Extragradient Methods, Min-Max Optimization

Compressor summary: The paper studies single-call stochastic extragradient methods for solving large-scale min-max optimization and variational inequalities problems, improves their convergence guarantees, and applies them to two classes of structured non-monotone VIPs using the expected residual condition.


Federated Compositional Deep AUC Maximization

Xinwen Zhang, Yihan Zhang, Tianbao Yang, Richard Souvenir, Hongchang Gao

https://openreview.net/forum?id=tF7W8ai8J3

Keywords: federated learning, compositional optimization, minimax optimization, AUC maximization

Compressor summary: The paragraph discusses a novel federated learning method for imbalanced data that optimizes the AUC score and provides theoretical and empirical evidence of its effectiveness.


PPi: Pretraining Brain Signal Model for Patient-independent Seizure Detection

Zhizhang Yuan, Daoze Zhang, Yang Yang, Junru Chen, Yafeng Li

https://openreview.net/forum?id=tEmFyqjaJh

Keywords: Brain signal, Seizure detection, Pretraining, Domain generalization

Compressor summary: The study proposes a patient-independent seizure detection model using stereoelectroencephalography (SEEG) data that overcomes challenges like domain shift and pattern evolution with novel self-supervised tasks and domain generalization techniques.


Safety Verification of Decision-Tree Policies in Continuous Time

Christian Schilling, Anna Lukina, Emir Demirović, Kim Guldstrand Larsen

https://openreview.net/forum?id=tEKBU5XOTw

Keywords: safety verification, decision tree, reinforcement learning, controller, continuous time

Compressor summary: The paper introduces an algorithm to verify the safety of decision tree controlled systems in continuous time, overcoming the challenges of existing methods.


Doubly Constrained Fair Clustering

John P Dickerson, Seyed A. Esmaeili, Jamie Heather Morgenstern, Claire Jie Zhang

https://openreview.net/forum?id=tECyQO1QOp

Keywords: Fairness, Clustering, Approximation Algorithms

Compressor summary: The paper investigates how two common fairness criteria for clustering, group fairness and diversity in center selection, can be simultaneously satisfied using a constant approximation algorithm, and shows that they are compatible with some but not all distance-based fairness notions.


A Robust and Opponent-Aware League Training Method for StarCraft II

Ruozi Huang, Xipeng Wu, Hongsheng Yu, Zhong Fan, Haobo Fu, QIANG FU, Yang Wei

https://openreview.net/forum?id=tDAu3FPJn9

Keywords: StarCraft II, league training, AlphaStar, opponent-modeling, reinforcement learning

Compressor summary: The paper presents improvements to AlphaStar's league training for StarCraft II, enabling better and more resource-efficient AI agents with advanced exploitation and opponent modeling abilities.


On the Power of SVD in the Stochastic Block Model

Xinyu Mao, Jiapeng Zhang

https://openreview.net/forum?id=tC0r8duG9z

Keywords: Clustering Algorithms, Stochastic Block Model, Spectral Algorithms

Compressor summary: The paper investigates how spectral methods like SVD can improve clustering algorithms and shows that it works well for a specific type of network model called the stochastic block model.


Replicable Reinforcement Learning

ERIC EATON, Marcel Hussing, Michael Kearns, Jessica Sorrell

https://openreview.net/forum?id=tBwRbgsol1

Keywords: Reinforcement Learning, Learning Theory, Replicability, Reproducibility

Compressor summary: The text discusses the development of algorithm frameworks for replicability in various fields and introduces provably replicable algorithms for parallel value iteration and R-Max in episodic reinforcement learning.


Understanding Deep Gradient Leakage via Inversion Influence Functions

Haobo Zhang, Junyuan Hong, Yuyang Deng, Mehrdad Mahdavi, Jiayu Zhou

https://openreview.net/forum?id=tBib2fWr3r

Keywords: Deep Learning, Privacy, Federated Learning, Influence Function

Compressor summary: The paper proposes a method called Inversion Influence Function (I$^2$F) to analyze and understand the privacy leakage caused by Deep Gradient Leakage (DGL) attack in distributed learning settings, and show its effectiveness on various scenarios.


A Bounded Ability Estimation for Computerized Adaptive Testing

Yan Zhuang, Qi Liu, GuanHao Zhao, Zhenya Huang, Weizhe Huang, Zachary Pardos, Enhong Chen, Jinze Wu, Xin Li

https://openreview.net/forum?id=tAwjG5bM7H

Keywords: adaptive learning, computerized adaptive testing, educational measurement, cognitive diagnosis

Compressor summary: The paper proposes a new method (BECAT) for computerized adaptive testing that selects questions to estimate student's ability more accurately and efficiently by matching the gradient of full responses.


Uncertainty Estimation for Safety-critical Scene Segmentation via Fine-grained Reward Maximization

Hongzheng Yang, Cheng Chen, Yueyao Chen, Markus Scheppach, Hon Chi Yip, Qi Dou

https://openreview.net/forum?id=t9Swbo82dB

Keywords: uncertainty estimation, semantic segmentation, medical application

Compressor summary: The paper proposes a novel framework that uses reinforcement learning and reward optimization to improve uncertainty estimation in deep segmentation models for safety-critical applications like medical imaging.


Cheap and Quick: Efficient Vision-Language Instruction Tuning for Large Language Models

Gen Luo, Yiyi Zhou, Tianhe Ren, Shengxin Chen, Xiaoshuai Sun, Rongrong Ji

https://openreview.net/forum?id=t877958UGZ

Keywords: vision-language instruction tuning, multimodal LLM, efficient training

Compressor summary: The paper proposes a cost-effective method called Mixture-of-Modality Adaptation (MMA) to improve the multimodal capabilities of large language models for vision-language tasks, achieving competitive performance and efficiency with low expenditure.


Rewiring Neurons in Non-Stationary Environments

Zhicheng Sun, Yadong MU

https://openreview.net/forum?id=t7ozN4AXd0

Keywords: continual learning, reinforcement learning, brain-inspired learning

Compressor summary: The paragraph describes a novel rewiring approach for the human brain in continual reinforcement learning that promotes adaptivity and exploration, with a focus on stability and plasticity-stability tradeoffs.


Joint Bayesian Inference of Graphical Structure and Parameters with a Single Generative Flow Network

Tristan Deleu, Mizu Nishikawa-Toomey, Jithendaraa Subramanian, Nikolay Malkin, Laurent Charlin, Yoshua Bengio

https://openreview.net/forum?id=t7lnhhi7De

Keywords: bayesian network, bayesian, structure learning, causal discovery, gflownet

Compressor summary: The paper proposes JSP-GFN, a method that uses GFlowNets to approximate the joint posterior over Bayesian Network structure and parameters, achieving good results on various datasets.


Achieving Cross Modal Generalization with Multimodal Unified Representation

Yan Xia, Hai Huang, Jieming Zhu, Zhou Zhao

https://openreview.net/forum?id=t7ZowrDWVw

Keywords: multi-modal, discrete representation, mutual information estimation

Compressor summary: The paper proposes a novel task called Cross Modal Generalization (CMG) that learns a unified discrete representation from paired multimodal data and introduces Uni-Code, which contains two key contributions for fine-grained unified representation of multimodal sequences.


Trans-Dimensional Generative Modeling via Jump Diffusion Models

Andrew Campbell, William Harvey, Christian Dietrich Weilbach, Valentin De Bortoli, Tom Rainforth, Arnaud Doucet

https://openreview.net/forum?id=t6nA7x3GAC

Keywords: diffusion, score-based, score, markov chain, jump diffusion, poisson

Compressor summary: The proposed generative model can handle data of different dimensions by switching between them during the generation process, leading to better performance on tasks like imputation and interpolation.


QuadAttac$K$: A Quadratic Programming Approach to Learning Ordered Top-$K$ Adversarial Attacks

Thomas Paniagua, Ryan Grainger, Tianfu Wu

https://openreview.net/forum?id=t3vPEjgNtj

Keywords: Ordered Top-K Clear-Box Targeted Adversarial Attack, Deep Neural Networks, Quadratic Programming, Robustness

Compressor summary: The paper introduces QuadAttac$K$, a method to learn more aggressive targeted attacks on DNNs by solving quadratic programming in the feature embedding space.


Scalable Membership Inference Attacks via Quantile Regression

Martin Andres Bertran, Shuai Tang, Aaron Roth, Michael Kearns, Jamie Heather Morgenstern, Steven Wu

https://openreview.net/forum?id=t3WCiGjHqd

Keywords: machine learning, privacy, membership inference

Compressor summary: Our proposed attack estimates the distribution of confidence scores on non-training points using quantile regression, and it is more efficient and black-box than shadow model attacks.


Tailoring Self-Attention for Graph via Rooted Subtrees

Siyuan Huang, Yunchong Song, Jiayue Zhou, Zhouhan Lin

https://openreview.net/forum?id=t2hEZadBBk

Keywords: Graph Based Learning

Compressor summary: The paper introduces Subtree Attention, a novel graph attention mechanism that addresses the limitations of local and global attention by allowing computation of attention weights among multi-hop neighbors and has linear time complexity.


Object-Centric Learning for Real-World Videos by Predicting Temporal Feature Similarities

Andrii Zadaianchuk, Maximilian Seitzer, Georg Martius

https://openreview.net/forum?id=t1jLRFvBqm

Keywords: object-centric learning, video, representation learning, self-supervised learning, unsupervised learning

Compressor summary: The paper proposes a new method to learn object-centric representations from videos using pre-trained features and temporal feature similarity loss, achieving state-of-the-art results on synthetic and unconstrained video datasets.


A unified framework for information-theoretic generalization bounds

Yifeng Chu, Maxim Raginsky

https://openreview.net/forum?id=t0fkjO4aZj

Keywords: generalization bounds, information theory, chaining, PAC-Bayes, couplings

Compressor summary: The paper proposes a method to derive generalization bounds for learning algorithms using a probabilistic decorrelation lemma and other techniques.


Riemannian Projection-free Online Learning

Zihao Hu, Guanghui Wang, Jacob Abernethy

https://openreview.net/forum?id=szFqlNRxeS

Keywords: Online learning, Riemannian optimization, projection-free optimization

Compressor summary: The paper proposes projection-free optimization methods for online geodesically convex problems on curved spaces with separation or linear optimization oracles, achieving sub-linear regret guarantees in different settings.


A Neural Collapse Perspective on Feature Evolution in Graph Neural Networks

Vignesh Kothapalli, Tom Tirer, Joan Bruna

https://openreview.net/forum?id=sxao2udWXi

Keywords: Neural collapse, Graph neural networks, Community detection

Compressor summary: The paper investigates how graph topology affects feature evolution in node-wise classification using graph neural networks (GNNs) and compares it to the Neural Collapse phenomenon in instance-wise deep classifiers.


Is RLHF More Difficult than Standard RL? A Theoretical Perspective

Yuanhao Wang, Qinghua Liu, Chi Jin

https://openreview.net/forum?id=sxZLrBqg50

Keywords: reinforcement learning theory, reinforcement learning from human feedback, preference-based reinforcement learning

Compressor summary: The paper shows that existing reward-based RL algorithms can be used to solve preference-based RL with minimal modifications, and applies the theory to various models with preferences.


Meta-in-context learning in large language models

Julian Coda-Forno, Marcel Binz, Zeynep Akata, Matthew Botvinick, Jane X Wang, Eric Schulz

https://openreview.net/forum?id=sx0xpaO0za

Keywords: Large language models, in-context learning, meta-learning, GPT-3

Compressor summary: The paper introduces meta-in-context learning, a method to improve large language models' in-context learning abilities by recursively applying them to different tasks, and shows its effectiveness on various domains and benchmarks.


The noise level in linear regression with dependent data

Ingvar Ziemann, Stephen Tu, George J. Pappas, Nikolai Matni

https://openreview.net/forum?id=swNtr6vGqg

Keywords: Learning Theory, Learning with dependent data, Time-Series

Compressor summary: We find upper limits for random linear regression with dependent data without assuming realizability, and our analysis shows how the error improves as we reduce misspecification.


A Unified, Scalable Framework for Neural Population Decoding

Mehdi Azabou, Vinam Arora, Venkataramana Ganesh, Ximeng Mao, Santosh B Nachimuthu, Michael Jacob Mendelson, Blake Aaron Richards, Matthew G Perich, Guillaume Lajoie, Eva L Dyer

https://openreview.net/forum?id=sw2Y0sirtM

Keywords: neural population, brain decoder, transformer, tokenization, sequence-to-sequence, electrophysiology, brain-computer interfaces

Compressor summary: The paper introduces a method to model neural activity across diverse recordings using tokenization, cross-attention, and PerceiverIO, enabling few-shot performance with minimal labels in new sessions.


CEIL: Generalized Contextual Imitation Learning

Jinxin Liu, Li He, Yachen Kang, Zifeng Zhuang, Donglin Wang, Huazhe Xu

https://openreview.net/forum?id=suzMI2P1rT

Keywords: imitation learning, reinforcement learning, offline imitation learning

Compressor summary: The paper introduces ContExtual Imitation Learning (CEIL), a versatile algorithm for imitation learning that learns from expert behaviors using hindsight embeddings, and demonstrates its effectiveness on various settings and benchmarks.


Riemannian SAM: Sharpness-Aware Minimization on Riemannian Manifolds

Jihun Yun, Eunho Yang

https://openreview.net/forum?id=strvrjSi3C

Keywords: optimization, riemannian, manifolds, sharpness-aware

Compressor summary: The paper introduces Riemannian SAM, a novel optimization algorithm for training geometric deep learning models, and shows its benefits on knowledge graph completion and machine translation tasks.


The Simplicity Bias in Multi-Task RNNs: Shared Attractors, Reuse of Dynamics, and Geometric Representation

Elia Turner, Omri Barak

https://openreview.net/forum?id=stDm3S0CV7

Keywords: Computational Neural Models; Recurrent Neural Networks; Multiple Tasks; Geometry;Dynamical Systems;Attractors;Neuroscience

Compressor summary: The authors study how Recurrent Neural Networks (RNNs) handle multiple tasks and find that they tend to reuse existing dynamics and opt for simple solutions, which they call the "simplicity bias".


Stable and low-precision training for large-scale vision-language models

Mitchell Wortsman, Tim Dettmers, Luke Zettlemoyer, Ari S. Morcos, Ali Farhadi, Ludwig Schmidt

https://openreview.net/forum?id=sqqASmpA2R

Keywords: CLIP, int8, stability

Compressor summary: The authors propose SwitchBack, a fast linear layer for int8 training, and recommend a hybrid of AdamW and Adafactor to prevent loss spikes in large language-vision models.


HASSOD: Hierarchical Adaptive Self-Supervised Object Detection

Shengcao Cao, Dhiraj Joshi, Liangyan Gui, Yu-Xiong Wang

https://openreview.net/forum?id=sqkGJjIRfG

Keywords: self-supervised learning, object detection

Compressor summary: HASSOD is a novel self-supervised object detection method that learns to detect objects and understand their compositions without human supervision, achieving improved performance and interpretability over existing methods.


Sparsity-Preserving Differentially Private Training of Large Embedding Models

Badih Ghazi, Yangsibo Huang, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Amer Sinha, Chiyuan Zhang

https://openreview.net/forum?id=sqTcCXkG4P

Keywords: Differential Privacy, Recommendation Systems, Embedding Models, Efficient Machine Learning

Compressor summary: The paper introduces DP-FEST and DP-AdaFEST, two new privacy-preserving algorithms for training large embedding models that reduce gradient size and maintain accuracy.


What’s Left? Concept Grounding with Logic-Enhanced Foundation Models

Joy Hsu, Jiayuan Mao, Joshua B. Tenenbaum, Jiajun Wu

https://openreview.net/forum?id=sq4o3tjWaj

Keywords: concept learning, visual reasoning, large language models, neuro-symbolic learning

Compressor summary: The paper proposes LEFT, a framework that learns to ground and reason with concepts across domains using a logic-based program executor and an LLM interpreter, outperforming previous methods in reasoning tasks.


Belief Projection-Based Reinforcement Learning for Environments with Delayed Feedback

Jangwon Kim, Hangyeol Kim, Jiwook Kang, Jongchan Baek, Soohee Han

https://openreview.net/forum?id=sq0m11cUMV

Keywords: time-delay system, reinforcement learning

Compressor summary: The paragraph describes a new actor-critic algorithm called BPQL that solves delayed feedback problems in reinforcement learning by reducing the state space size and outperforming traditional methods in continuous control tasks.


Unsupervised Semantic Correspondence Using Stable Diffusion

Eric Hedlin, Gopal Sharma, Shweta Mahajan, Hossam Isack, Abhishek Kar, Andrea Tagliasacchi, Kwang Moo Yi

https://openreview.net/forum?id=sovxUzPzLN

Keywords: Semantic Correspondence, Stable Diffusion, Optimization-based Inference

Compressor summary: The authors propose a method to find semantic correspondences between images using diffusion models' semantic knowledge without any training, achieving results comparable to strongly supervised methods.


Dynamically Masked Discriminator for GANs

Wentian Zhang, Haozhe Liu, Bing Li, Jinheng Xie, Yawen Huang, Yuexiang Li, Yefeng Zheng, Bernard Ghanem

https://openreview.net/forum?id=sodl2c3aTM

Keywords: Generative model, Generative Adversarial Network

Compressor summary: The paper proposes a new GAN method that helps the discriminator adapt faster to changes in generated data, improving the quality of generated results.


AV-NeRF: Learning Neural Fields for Real-World Audio-Visual Scene Synthesis

Susan Liang, Chao Huang, Yapeng Tian, Anurag Kumar, Chenliang Xu

https://openreview.net/forum?id=snY3FOnlQi

Keywords: Scene synthesis, audio-visual, NeRF

Compressor summary: The authors propose a novel approach to synthesize realistic audio-visual scenes using NeRF, integrating knowledge of sound propagation and source-centric acoustic fields, and present a new dataset and demo videos.


Beyond MLE: Convex Learning for Text Generation

Chenze Shao, Zhengrui Ma, Min Zhang, Yang Feng

https://openreview.net/forum?id=sla7V80uWA

Keywords: Maximum likelihood estimation, Convex function, Text generation

Compressor summary: The paper proposes convex functions as an alternative to maximum likelihood estimation (MLE) for training text generation models, which can improve performance and efficiency on various tasks.


Survival Instinct in Offline Reinforcement Learning

Anqi Li, Dipendra Misra, Andrey Kolobov, Ching-An Cheng

https://openreview.net/forum?id=shePL2nbwl

Keywords: Offline RL, safe RL

Compressor summary: Offline reinforcement learning can produce good and safe policies even with wrong reward labels due to a survival instinct from pessimism and implicit biases in data collection.


FD-Align: Feature Discrimination Alignment for Fine-tuning Pre-Trained Models in Few-Shot Learning

Kun Song, Huimin Ma, Bochao Zou, Huishuai Zhang, Weiran Huang

https://openreview.net/forum?id=shXnfALjuH

Keywords: few-shot learning, CLIP, fine-tuning

Compressor summary: The paper introduces FD-Align, a method that improves few-shot learning by preserving spurious feature consistency and generalizing better to ID and OOD tasks.


Data Market Design through Deep Learning

Sai Srivatsa Ravindranath, Yanchen Jiang, David C. Parkes

https://openreview.net/forum?id=sgCrNMOuXp

Keywords: Data Markets, Information Design, Differentiable Economics, Economics, Deep Learning, Mechanism Design, Algorithmic Game Theory

Compressor summary: The paper proposes a deep learning approach to design revenue-optimal data markets by learning signaling schemes and handling obedience and incentive constraints.


On Sample-Efficient Offline Reinforcement Learning: Data Diversity, Posterior Sampling and Beyond

Thanh Nguyen-Tang, Raman Arora

https://openreview.net/forum?id=sdlh4gVOj8

Keywords: reinforcement learning, offline reinforcement learning

Compressor summary: The paper investigates what makes offline reinforcement learning sample-efficient and unifies three classes of algorithms using a new concept of data diversity.


Retrieval-Augmented Multiple Instance Learning

Yufei CUI, Ziquan Liu, Yixin CHEN, Yuchen Lu, Xinyue Yu, Xue Liu, Tei-Wei Kuo, Miguel R. D. Rodrigues, Chun Jason Xue, Antoni B. Chan

https://openreview.net/forum?id=scaKiAtbI3

Keywords: Multiple Instance Learning, Whole Slide Imaging, Nearest Neighbor Retrieval

Compressor summary: The paper introduces RAM-MIL, a framework that uses Optimal Transport to improve MIL performance on out-of-domain data in medical diagnosis tasks based on whole slide images.


VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and Dataset

Sihan Chen, Handong Li, Qunbo Wang, Zijia Zhao, Mingzhen Sun, Xinxin Zhu, Jing Liu

https://openreview.net/forum?id=scYa9DYUAy

Keywords: Cross-Modality Foundation Model, Cross-Modality Pretraining Dataset

Compressor summary: This paper introduces a large dataset (VAST-27M) for multi-modal video captioning and a model (VAST) that can process vision, audio, and subtitle modalities for various tasks.


Unleashing the Full Potential of Product Quantization for Large-Scale Image Retrieval

Yu Liang, Shiliang Zhang, Kenli Li, Xiaoyu Wang

https://openreview.net/forum?id=scG0cwftEe

Keywords: Deep Hash, Image Retrieval, Product Quantization

Compressor summary: The paragraph introduces a novel deep hashing framework using product quantization that addresses the limitations of current methods on large-scale datasets and achieves better performance.


Quantizable Transformers: Removing Outliers by Helping Attention Heads Do Nothing

Yelysei Bondarenko, Markus Nagel, Tijmen Blankevoort

https://openreview.net/forum?id=sbusw6LD41

Keywords: transformers, LLM, softmax, attention, outliers, quantization, post-training quantization

Compressor summary: The paper proposes two modifications to the attention mechanism in transformer models to reduce outliers and enable efficient quantization without compromising performance.


DreamSparse: Escaping from Plato’s Cave with 2D Diffusion Model Given Sparse Views

Paul Yoo, Jiaxian Guo, Yutaka Matsuo, Shixiang Shane Gu

https://openreview.net/forum?id=sZNBYvunEr

Keywords: Novel View Synthesis, Diffusion Model

Compressor summary: DreamSparse is a framework that uses a pre-trained diffusion model to generate high-quality novel view images from sparse views by incorporating 3D geometry information as a prior.


DiffPack: A Torsional Diffusion Model for Autoregressive Protein Side-Chain Packing

Yangtian Zhang, Zuobai Zhang, Bozitao Zhong, Sanchit Misra, Jian Tang

https://openreview.net/forum?id=sXMQPKbLXf

Keywords: protein side-chain packing, diffusion models, autoregressive models, geometric deep learning

Compressor summary: The authors propose DiffPack, a torsional diffusion model that predicts protein side-chain packing by learning joint distributions of side-chain angles and improves accuracy while reducing model size.


DISCOVER: Making Vision Networks Interpretable via Competition and Dissection

Konstantinos P. Panousis, Sotirios Chatzis

https://openreview.net/forum?id=sWNOvNXGLP

Keywords: Interpretability, Explainability, Network Dissection, Competitive Networks, Sparsity, Multimodal Models

Compressor summary: The authors present a method to generate textual descriptions of individual neuron functions in deep vision networks, enabling easier interpretation and exploration of their decision processes.


Ordering-based Conditions for Global Convergence of Policy Gradient Methods

Jincheng Mei, Bo Dai, Alekh Agarwal, Mohammad Ghavamzadeh, Csaba Szepesvari, Dale Schuurmans

https://openreview.net/forum?id=sW8yGZ4uVJ

Keywords: reinforcement learning, policy gradient, policy optimization, function approximation, global convergence

Compressor summary: The paper shows that policy gradient methods can converge globally for finite-arm bandits with linear function approximation, depending on properties between the policy update and the representation, without relying heavily on approximation error as a key quantity.


Weakly-Supervised Audio-Visual Segmentation

Shentong Mo, Bhiksha Raj

https://openreview.net/forum?id=sUqG96QqZM

Keywords: audio-visual learning, visual sound localization, audio-visual segmentation

Compressor summary: The paragraph describes a novel framework called WS-AVS that simplifies the supervision for audio-visual segmentation by using instance-level annotations and multi-scale contrastive learning, and shows its effectiveness on various scenarios.


D-Separation for Causal Self-Explanation

Wei Liu, Jun Wang, Haozhao Wang, Ruixuan Li, Zhiying Deng, YuanKai Zhang, Yang Qiu

https://openreview.net/forum?id=sUFGPYS25Q

Keywords: interpretability, causal inference, rationalization, self-explaining

Compressor summary: The Minimum Conditional Dependence (MCD) criterion is a novel way to uncover the causal rationale of NLP models by minimizing the dependence between non-selected input parts and the target label, which improves interpretability and performance compared to previous MMI-based methods.


Let the Flows Tell: Solving Graph Combinatorial Problems with GFlowNets

Dinghuai Zhang, Hanjun Dai, Nikolay Malkin, Aaron Courville, Yoshua Bengio, Ling Pan

https://openreview.net/forum?id=sTjW3JHs2V

Keywords: graph; combinatorial optimization; sampling; gflownets

Compressor summary: The paper presents a method to train conditional GFlowNets on MDPs for solving combinatorial optimization problems by sampling from the solution space and demonstrates their effectiveness in finding high-quality solutions.


Understanding and Addressing the Pitfalls of Bisimulation-based Representations in Offline Reinforcement Learning

Hongyu Zang, Xin Li, Leiji Zhang, Yang Liu, Baigui Sun, Riashat Islam, Remi Tachet des Combes, Romain Laroche

https://openreview.net/forum?id=sQyRQjun46

Keywords: Bisimulation metrics, Reinforcement Learning, Representation Learning, Offline RL

Compressor summary: The paper analyzes why bisimulation methods work well in online RL but not in offline RL, and proposes using the expectile operator and reward scaling to improve performance on two benchmarks.


Cocktail: Mixing Multi-Modality Control for Text-Conditional Image Generation

Minghui Hu, Jianbin Zheng, Daqing Liu, Chuanxia Zheng, Chaoyue Wang, Dacheng Tao, Tat-Jen Cham

https://openreview.net/forum?id=sQBGVw5qH9

Keywords: Multi-modality, Image Generation, Diffusion

Compressor summary: Cocktail is a pipeline that mixes different modalities into one embedding and uses a hyper-network called gControlNet to align and infuse control signals from various sources into a pre-trained diffusion model, resulting in high-quality image synthesis with diverse contents.


A Measure-Theoretic Axiomatisation of Causality

Junhyung Park, Simon Buchholz, Bernhard Schölkopf, Krikamol Muandet

https://openreview.net/forum?id=sPLTQSf6GI

Keywords: Causality, probability theory, causal models

Compressor summary: The text introduces a new concept of causal space that combines probability theory with causality to overcome some limitations of current approaches.


Contrast Everything: A Hierarchical Contrastive Framework for Medical Time-Series

Yihe Wang, Yu Han, Haishuai Wang, Xiang Zhang

https://openreview.net/forum?id=sOQBHlCmzp

Keywords: Deep Learning, Contrastive Learning, Self-supervised Learning, Time Series, Healthcare

Compressor summary: COMET is a hierarchical framework that leverages data consistencies at multiple levels in medical time series to improve contrastive representation learning and outperforms existing methods on diverse datasets.


Projection-Free Online Convex Optimization via Efficient Newton Iterations

Khashayar Gatmiry, Zakaria Mhammedi

https://openreview.net/forum?id=sOOg1xJADA

Keywords: Online Learning, Online convex optimization, projection-free, Newton method

Compressor summary: The paper proposes a new online convex optimization algorithm that uses self-concordant barriers to avoid projections and achieve low regret without computing matrix inverses often.


Error Bounds for Learning with Vector-Valued Random Features

Samuel Lanthaler, Nicholas H. Nelsen

https://openreview.net/forum?id=sLr1sohnmo

Keywords: random features, random feature model, operator learning, vector-valued

Compressor summary: The paper analyzes vector-valued random features ridge regression in an infinite-dimensional setting, providing strong consistency and minimax optimal convergence rates without relying on random matrix theory.


On skip connections and normalisation layers in deep optimisation

Lachlan Ewen MacDonald, Jack Valmadre, Hemanth Saratchandran, Simon Lucey

https://openreview.net/forum?id=sLhXMkI0kx

Keywords: optimisation, optimization, skip, connection, normalisation, normalization, deep, learning, polyak, lojasiewicz, lipschitz

Compressor summary: The paper presents a theory for gradient optimisation of deep neural networks that explains the roles of normalisation layers and skip connections, and shows how they enable training to global optima and acceleration of training.


ELDEN: Exploration via Local Dependencies

Zizhao Wang, Jiaheng Hu, Peter Stone, Roberto Martín-Martín

https://openreview.net/forum?id=sL4pJBXkxu

Keywords: reinforcement learning; intrinsic motivation; exploration

Compressor summary: ELDEN is a novel intrinsic reward function that encourages reinforcement learning agents to explore environments with complex dependencies by exploiting uncertainty in the learned dynamics.


Cross-links Matter for Link Prediction: Rethinking the Debiased GNN from a Data Perspective

Zihan Luo, Hong Huang, Jianxun Lian, Xiran Song, Xing Xie, Hai Jin

https://openreview.net/forum?id=sJDkwMVqb9

Keywords: Cross-links, Debias, Graph Neural Networks, Link Prediction

Compressor summary: The paper proposes a twin-structure framework to reduce data bias between internal-links and cross-links in GNN-based link prediction, enhancing both bias mitigation and utility.


VOCE: Variational Optimization with Conservative Estimation for Offline Safe Reinforcement Learning

Jiayi Guan, Guang Chen, Jiaming Ji, Long Yang, Ao Zhou, Zhijun Li, changjun jiang

https://openreview.net/forum?id=sIU3WujeSl

Keywords: Offline safe reinforcement learning, Pessimistic conservative estimation, Variational optimization, Reinforcement Learning

Compressor summary: VOCE is a new offline safe RL algorithm that uses probabilistic inference and pessimistic estimation to learn policies that satisfy safety constraints while optimizing rewards.


Q-DM: An Efficient Low-bit Quantized Diffusion Model

Yanjing Li, Sheng Xu, Xianbin Cao, Xiao Sun, Baochang Zhang

https://openreview.net/forum?id=sFGkL5BsPi

Keywords: network quantization, diffusion model, image synthesize

Compressor summary: The paper proposes TaQ and NeM methods to improve low-bit quantized diffusion models, reducing computation and memory costs while maintaining high-quality data generation.


SutraNets: Sub-series Autoregressive Networks for Long-Sequence, Probabilistic Forecasting

Shane Bergsma, Tim Zeyl, Lei Guo

https://openreview.net/forum?id=sC4RbbVKbu

Keywords: time series, probabilistic forecasting, autoregressive generative models, neural networks

Compressor summary: SutraNets are a new method for predicting long time series by factoring likelihood into conditional probabilities and generating multivariate outputs to reduce error accumulation.


Doubly Robust Augmented Transfer for Meta-Reinforcement Learning

Yuankun Jiang, Nuowen Kan, Chenglin Li, Wenrui Dai, Junni Zou, Hongkai Xiong

https://openreview.net/forum?id=sABYNWKcwK

Keywords: Meta-reinforcement learning, doubly robust (DR), sample transfer

Compressor summary: The paper proposes a new Meta-RL method called DRaT that addresses dynamics and reward function variations across tasks by using a doubly robust estimator to transfer informative samples from other tasks.


RH-BrainFS: Regional Heterogeneous Multimodal Brain Networks Fusion Strategy

Hongting Ye, Yalu Zheng, Yueying Li, Ke Zhang, Youyong Kong, Yonggui Yuan

https://openreview.net/forum?id=s97ezbqoDZ

Keywords: Multimodal, Neuroscience, Subgraph, Transformer

Compressor summary: The authors present RH-BrainFS, a novel multimodal brain network fusion strategy that addresses regional heterogeneity between structural and functional connectivity, and shows its effectiveness in various neuroscience tasks.


AND: Adversarial Neural Degradation for Learning Blind Image Super-Resolution

Fangzhou Luo, Xiaolin Wu, Yanhui Guo

https://openreview.net/forum?id=s8QsYV1VZ2

Keywords: Blind Image Super-Resolution

Compressor summary: The paper proposes an adversarial neural degradation (AND) model that improves image super-resolution by generating a wide range of complex degradation effects without explicit supervision, resulting in better generalization and performance on real-world images.


Soft-Unification in Deep Probabilistic Logic

Jaron Maene, Luc De Raedt

https://openreview.net/forum?id=s86M8naPSv

Keywords: neuro-symbolic AI, probabilistic logic, embeddings

Compressor summary: The paragraph discusses soft-unification, a technique for combining logic and neural concepts in AI, and proposes a new framework called DeepSoftLog that improves upon previous systems by introducing probabilistic semantics and achieving better performance on neuro-symbolic benchmarks.


Don’t Stop Pretraining? Make Prompt-based Fine-tuning Powerful Learner

Zhengxiang Shi, Aldo Lipani

https://openreview.net/forum?id=s7xWeJQACI

Keywords: Continued Pre-training, Prompt-based Fine-tuning, Language Models

Compressor summary: The study proposes Prompt-based Continued Pre-training (PCP), which improves prompt-based fine-tuning in natural language processing tasks by combining instruction tuning with conventional continued pre-training using unsupervised objectives.


On Certified Generalization in Structured Prediction

Bastian Boll, Christoph Schnoerr

https://openreview.net/forum?id=s1jQ91yFAb

Keywords: Structured Prediction, PAC-Bayes, Concentration Inequalities, Statistical Learning Theory, Knothe-Rosenblatt Rearrangement

Compressor summary: The paper proposes a new way to measure how well structured prediction models generalize using generative models and Wasserstein dependency matrices.


Focused Transformer: Contrastive Training for Context Scaling

Szymon Tworkowski, Konrad Staniszewski, Mikołaj Pacek, Yuhuai Wu, Henryk Michalewski, Piotr Miłoś

https://openreview.net/forum?id=s1FjXzJ0jy

Keywords: Transformers, Language Models, Natural Language Processing

Compressor summary: The Focused Transformer (FoT) technique addresses the distraction issue in large language models by enhancing the structure of the (key, value) space and extending the effective context length through contrastive learning.


GALOPA: Graph Transport Learning with Optimal Plan Alignment

Yejiang Wang, Yuhai Zhao, Daniel Zhengkui Wang, Ling Li

https://openreview.net/forum?id=rzlqOVExUA

Keywords: graph neural network; self-supervised learning; optimal transport;

Compressor summary: The paper proposes a self-supervised graph learning method that preserves both structural and matching information between graphs using optimal transport plans, and shows its advantages over existing methods in different scenarios.


Private Federated Frequency Estimation: Adapting to the Hardness of the Instance

Jingfeng Wu, Wennan Zhu, Peter Kairouz, Vladimir Braverman

https://openreview.net/forum?id=rzDBoh1tBh

Keywords: sketch, federated analytics, privacy

Compressor summary: The paper proposes an improved sketch algorithm for federated frequency estimation with multiple communication rounds, and a two-phase approach that adapts the sketch size based on problem complexity and ensures differential privacy.


EgoEnv: Human-centric environment representations from egocentric video

Tushar Nagarajan, Santhosh Kumar Ramakrishnan, Ruta Desai, James Hillis, Kristen Grauman

https://openreview.net/forum?id=rybsHQ4DXy

Keywords: egocentric video, 3D environment, sim2real, sim-to-real, episodic memory

Compressor summary: The paper proposes an approach to understand human activities in their environment using egocentric video and 3D simulation, and shows its effectiveness on real-world videos.


Matrix Compression via Randomized Low Rank and Low Precision Factorization

Rajarshi Saha, Varun Srivastava, Mert Pilanci

https://openreview.net/forum?id=rxsCTtkqA9

Keywords: Matrix compression, Randomized low rank factorization, Randomized SVD, Sketching, Quantized embeddings, Random matrices

Compressor summary: The paper proposes an algorithm to decompose large matrices into low rank factors and quantize them for efficient storage and processing.


Reconstructing the Mind's Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors

Paul Steven Scotti, Atmadeep Banerjee, Jimmie Goode, Stepan Shabalin, Alex Nguyen, Cohen Ethan, Aidan James Dempster, Nathalie Verlinde, Elad Yundler, David Weisberg, Kenneth Norman, Tanishq Mathew Abraham

https://openreview.net/forum?id=rwrblCYb2A

Keywords: fMRI, computational neuroscience, mind reading, diffusion models

Compressor summary: MindEye is a novel method to reconstruct and retrieve images from brain activity using fMRI and specialized submodules for retrieval and reconstruction.


Break It Down: Evidence for Structural Compositionality in Neural Networks

Michael A. Lepori, Thomas Serre, Ellie Pavlick

https://openreview.net/forum?id=rwbzMiuFQl

Keywords: Deep Learning, Compositionality, Cognitive Science

Compressor summary: The authors investigate whether neural networks use modular subnetworks or template matching to solve complex tasks, and find evidence that they can learn structural compositionality without relying on symbolic mechanisms.


DäRF: Boosting Radiance Fields from Sparse Input Views with Monocular Depth Adaptation

Jiuhn Song, Seonghoon Park, Honggyu An, Seokju Cho, Min-Seop Kwak, Sungjin Cho, Seungryong Kim

https://openreview.net/forum?id=rsrfEIdawr

Keywords: Neural Radiance Fields, 3D Reconstruction, Few-shot NeRF, Monocular Priors

Compressor summary: The DäRF framework combines NeRF and monocular depth estimation to improve novel view synthesis and 3D geometry reconstruction with a small number of real-world images.


PointGPT: Auto-regressively Generative Pre-training from Point Clouds

Guangyan Chen, Meiling Wang, Yi Yang, Kai Yu, Li Yuan, Yufeng Yue

https://openreview.net/forum?id=rqE0fEQDqs

Keywords: Generative Pre-training Transformer; GPT; Auto-regressively Generative Pre-training; Self-supervised Learning; Point clouds

Compressor summary: PointGPT extends the GPT language model to point clouds, using a novel auto-regressive generation task and a dual masking strategy, achieving state-of-the-art results on various downstream tasks.


On the Role of Noise in the Sample Complexity of Learning Recurrent Neural Networks: Exponential Gaps for Long Sequences

Alireza Fathollah Pour, Hassan Ashtiani

https://openreview.net/forum?id=rpuEARqB54

Keywords: PAC Learning, Recurrent Neural Networks, Noise, Sample Complexity

Compressor summary: The paper studies how adding noise to multi-layered sigmoid networks affects their sample complexity for classifying sequences, and shows an exponential gap between noisy and non-noisy networks.


ConRad: Image Constrained Radiance Fields for 3D Generation from a Single Image

Senthil Purushwalkam, Nikhil Naik

https://openreview.net/forum?id=roGYQvarnC

Keywords: 3D, generation, diffusion, viewpoint

Compressor summary: The paper introduces ConRad, a novel method that uses pretrained image generation models to reconstruct 3D objects from a single RGB image, capturing the appearance and preserving details while producing realistic 3D reconstructions.


AdaPlanner: Adaptive Planning from Feedback with Language Models

Haotian Sun, Yuchen Zhuang, Lingkai Kong, Bo Dai, Chao Zhang

https://openreview.net/forum?id=rnKgbKmelt

Keywords: Large language models, decision making, closed-loop planning

Compressor summary: AdaPlanner is a closed-loop approach that allows large language models to adaptively refine their plans in response to environmental feedback and outperforms state-of-the-art methods with fewer samples.


Managing Temporal Resolution in Continuous Value Estimation: A Fundamental Trade-off

Zichen Zhang, Johannes Kirschner, Junxi Zhang, Francesco Zanini, Alex Ayoub, Masood Dehghan, Dale Schuurmans

https://openreview.net/forum?id=rmQgQCZWiP

Keywords: Reinforcement Learning, Policy Evaluation, Temporal Discretization, Continuous Time, LQR

Compressor summary: The paper analyzes how time discretization affects Monte-Carlo policy evaluation for LQR systems and finds a trade-off between approximation and statistical error, leading to an optimal choice of temporal resolution for data efficiency.


False Discovery Proportion control for aggregated Knockoffs

Alexandre Blain, Bertrand Thirion, Olivier Grisel, Pierre Neuvial

https://openreview.net/forum?id=rlPUJ60bwM

Keywords: Knockoffs, Derandomization of Knockoffs, False Discoveries Proportion control, Controlled variable selection, Statistical inference, High-dimensional inference

Compressor summary: KOPI is a new method that improves Knockoff-based inference by controlling the actual proportion of false discoveries and using a novel aggregation technique.


Transformed Low-Rank Parameterization Can Help Robust Generalization for Tensor Neural Networks

Andong Wang, Chao Li, Mingyuan Bai, Zhong Jin, Guoxu Zhou, Qibin Zhao

https://openreview.net/forum?id=rih3hsSWx8

Keywords: Tensor SVD; Tensor Neural Networks; Transformed Low-rankness; Adversarial Generalization; Implicit Bias.

Compressor summary: The paper analyzes the theoretical generalization error of neural networks with t-product layers (t-NNs) and shows that transformed low-rank parameterization can improve their adversarial robustness.


DreamHuman: Animatable 3D Avatars from Text

Nikos Kolotouros, Thiemo Alldieck, Andrei Zanfir, Eduard Gabriel Bazavan, Mihai Fieraru, Cristian Sminchisescu

https://openreview.net/forum?id=rheCTpRrxI

Keywords: text to 3d; 3d avatars

Compressor summary: DreamHuman is a method that creates realistic 3D human avatars from text descriptions by combining different models and optimization techniques, resulting in diverse and high-quality animatable models.


Counterfactual Evaluation of Peer-Review Assignment Policies

Martin Saveski, Steven Jecmen, Nihar B Shah, Johan Ugander

https://openreview.net/forum?id=rhIfzCZoXG

Keywords: peer review, causal inference, counterfactual policy evaluation

Compressor summary: This paper proposes and applies novel methods to evaluate how changes in peer-review assignment algorithms affect review quality using randomness and data from two computer science venues.


Policy Optimization for Continuous Reinforcement Learning

Hanyang Zhao, Wenpin Tang, David Yao

https://openreview.net/forum?id=rfcak9EV99

Keywords: exploratory stochastic control, occupation time, performance difference, policy optimization

Compressor summary: The authors develop a new concept called occupation time for continuous reinforcement learning and show how it can improve performance and optimization methods in this setting.


Parallel Spiking Neurons with High Efficiency and Ability to Learn Long-term Dependencies

Wei Fang, Zhaofei Yu, Zhaokun Zhou, Ding Chen, Yanqi Chen, Zhengyu Ma, Timothée Masquelier, Yonghong Tian

https://openreview.net/forum?id=rfTFJvTkr2

Keywords: Spiking Neural Network, SNN, deep learning, spiking neuron, neuromorphic computing

Compressor summary: The authors propose Parallel Spiking Neurons, which improve simulation speed and accuracy in spiking neural networks by reformulating neuronal dynamics without reset and enabling parallel processing of inputs.


Why think step by step? Reasoning emerges from the locality of experience

Ben Prystawski, Michael Y. Li, Noah Goodman

https://openreview.net/forum?id=rcXXNFVlEn

Keywords: chain-of-thought; language models; reasoning

Compressor summary: The text discusses how chain-of-thought reasoning helps language models make better inferences when trained on locally structured data with strong variable dependencies, reducing bias and improving efficiency.


Adaptive Test-Time Personalization for Federated Learning

Wenxuan Bao, Tianxin Wei, Haohan Wang, Jingrui He

https://openreview.net/forum?id=rbw9xCU6Ci

Keywords: federated learning, personalized federated learning, test-time adaptation

Compressor summary: The paper introduces ATP, a novel test-time personalized federated learning algorithm that adapts global models in an unsupervised way without labeled data from multiple clients with different distributions, achieving strong generalization and outperforming existing methods.


Swarm Reinforcement Learning for Adaptive Mesh Refinement

Niklas Freymuth, Philipp Dahlinger, Tobias Daniel Würth, Simon Reisch, Luise Kärger, Gerhard Neumann

https://openreview.net/forum?id=rZqRu8e4uc

Keywords: Adaptive Mesh Refinement, Finite Element Method, Swarm Reinforcement Learning, Graph Neural Networks

Compressor summary: The paper proposes ASMR, a novel Adaptive Swarm Markov Decision Process for adaptive mesh refinement in engineering simulations, which achieves significant speedup and quality improvements compared to existing methods.


On the impact of activation and normalization in obtaining isometric embeddings at initialization

Amir Joudaki, Hadi Daneshmand, Francis Bach

https://openreview.net/forum?id=rY4sA9qYKy

Keywords: dynamical isometry, Lyapunov analysis, random neural networks

Compressor summary: In this paper, we explore the structure of the penultimate Gram matrix in deep neural networks, which contains the pairwise inner products of outputs corresponding to a batch of inputs. In several architectures it has been observed that this Gram matrix becomes degenerate with depth at initialization, which dramatically slows training. Normalization layers, such as batch or layer normalization, play a pivotal role in preventing the rank collapse issue. Despite promising advances, the existing theoretical results do not extend to layer normalization, which is widely used in transformers, and can not quantitatively characterize the role of non-linear activations. To bridge this gap, we prove that layer normalization, in conjunction with activation layers, biases the Gram matrix of a multilayer perceptron towards the identity matrix at an exponential rate with depth at initialization. We quantify this rate using the Hermite expansion of the activation function.


Wasserstein Gradient Flows for Optimizing Gaussian Mixture Policies

Hanna Ziesche, Leonel Rozo

https://openreview.net/forum?id=rW4mNcDxpS

Keywords: Policy optimization, robot learning, reinforcement learning, Gaussian mixture models, optimal transport, robotics

Compressor summary: The paper presents a novel policy optimization method for robots based on optimal transport and Riemannian optimization of GMMs, which enhances performance in different robot tasks.


PRED: Pre-training via Semantic Rendering on LiDAR Point Clouds

Hao Yang, Haiyang Wang, Di Dai, Liwei Wang

https://openreview.net/forum?id=rUldfB4SPT

Keywords: Pre-Train, Autonomous Driving, LiDAR, 3D Object Detection

Compressor summary: The paper proposes a new method, PRED, that uses images and neural rendering to help train point cloud encoders for 3D perception tasks in autonomous driving, addressing the challenges of incompleteness and occlusions.


Locality Sensitive Hashing in Fourier Frequency Domain For Soft Set Containment Search

Indradyumna Roy, Rishi Agarwal, Soumen Chakrabarti, Anirban Dasgupta, Abir De

https://openreview.net/forum?id=rUf0GV5CuU

Keywords: Locality sensitive hashing, Fourier transform, Order embeddings

Compressor summary: The paper proposes FourierHashNet, a data-sensitive and trainable index for fast retrieval of relevant documents using hinge distance transformed into dominance similarity in the frequency domain.


Homotopy-based training of NeuralODEs for accurate dynamics discovery

Joon-Hyuk Ko, Hankyul Koh, Nojun Park, Wonho Jhe

https://openreview.net/forum?id=rUFckPrzXR

Keywords: neural ordinary differential equations, synchronization, homotopy optimization, loss landscape, dynamical systems

Compressor summary: This paper introduces a new training method for NeuralODEs that improves their performance and speed by using synchronization and homotopy optimization without changing the model architecture.


Efficient Learning of Linear Graph Neural Networks via Node Subsampling

Seiyun Shin, Ilan Shomorony, Han Zhao

https://openreview.net/forum?id=rQI3FOzo1f

Keywords: Graph neural networks, Random sampling, Regression

Compressor summary: Graph Neural Networks (GNNs) aim to learn from large graphs, but their computational cost can be high; this paper proposes an efficient training method for a specific type of GNN by subsampling nodes and using leverage score sampling.


Continuous-time Analysis of Anchor Acceleration

Jaewook J. Suh, Jisun Park, Ernest K. Ryu

https://openreview.net/forum?id=rN99gLCBe4

Keywords: acceleration, convex optimization, continuous-time analysis, monotone operator, monotone inclusion, minimax optimization, fixed-point problem, anchor acceleration

Compressor summary: This paper analyzes the continuous-time models of anchor acceleration, a minimization technique with an unclear mechanism, and proposes an adaptive method based on the findings.


Preconditioning Matters: Fast Global Convergence of Non-convex Matrix Factorization via Scaled Gradient Descent

Xixi Jia, Hailin Wang, Jiangjun Peng, Xiangchu Feng, Deyu Meng

https://openreview.net/forum?id=rLpLjCBW4J

Keywords: Non-convex optimization, matrix factorization, low rank, scaled gradient descent

Compressor summary: The paper proposes two preconditioned gradient descent methods for low-rank matrix factorization that converge faster and do not require small learning rates or initializations.


ARTIC3D: Learning Robust Articulated 3D Shapes from Noisy Web Image Collections

Chun-Han Yao, Amit Raj, Wei-Chih Hung, Michael Rubinstein, Yuanzhen Li, Ming-Hsuan Yang, Varun Jampani

https://openreview.net/forum?id=rJc5Lsn5QU

Keywords: 3D articulated shape, animal body estimation, diffusion for 3D

Compressor summary: ARTIC3D is a self-supervised framework that reconstructs 3D shapes from sparse images using skeleton-based surface representation and 2D diffusion priors, achieving high-quality results even with noisy and occluded inputs.


Adversarial Counterfactual Environment Model Learning

Xiong-Hui Chen, Yang Yu, Zhengmao Zhu, ZhiHua Yu, Chen Zhenjun, Chenghe Wang, Yinan Wu, Rong-Jun Qin, Hongqiu Wu, Ruijin Ding, Huang Fangsheng

https://openreview.net/forum?id=rHAX0LRwk8

Keywords: environment model learning, offline reinforcement learning, off-policy evaluation, individual treatment effects estimation, causal inference, adversarial learning

Compressor summary: The paper proposes a new model-learning objective called AWRM to address the selection bias of behavior policies in environment dynamics models and shows that it improves counterfactual prediction and downstream tasks using the GALILEO algorithm.


Formalizing locality for normative synaptic plasticity models

Colin Bredenberg, Ezekiel Williams, Cristina Savin, Blake Aaron Richards, Guillaume Lajoie

https://openreview.net/forum?id=rGN3X9jnEg

Keywords: synaptic plasticity, computational neuroscience

Compressor summary: The authors propose formal definitions of locality for machine learning algorithms that aim to be biologically plausible, in order to make clear what quantities cannot be included in a learning rule and test their predictions against them.


FLuID: Mitigating Stragglers in Federated Learning using Invariant Dropout

Irene Wang, Prashant J. Nair, Divya Mahajan

https://openreview.net/forum?id=rG1M3kOVba

Keywords: Federated Learning

Compressor summary: FLuID is a method to balance the training load in Federated Learning by creating sub-models for low-performance devices without sacrificing accuracy.


Learning better with Dale’s Law: A Spectral Perspective

Pingsheng Li, Jonathan Cornford, Arna Ghosh, Blake Aaron Richards

https://openreview.net/forum?id=rDiMgZulwi

Keywords: Dale's Law, RNNs, brain-inspired neural networks, DANNs, computational neuroscience, spectral properties, inhibition

Compressor summary: The authors extend Dale's ANNs to RNNs, showing that good performance is possible while respecting Dale's Law, which states that neurons must be excitatory or inhibitory. They also find that the spectral properties of the recurrent weight matrix at initialisation are more impactful on network performance than sign constraints.


Learning Invariant Molecular Representation in Latent Discrete Space

Xiang Zhuang, Qiang Zhang, Keyan Ding, Yatao Bian, Xiao Wang, Jingsong Lv, Hongyang Chen, Huajun Chen

https://openreview.net/forum?id=r9fzp8eyhZ

Keywords: molecular representation learning, out-of-distribution

Compressor summary: The authors propose a new molecular representation learning framework that improves generalization and robustness against distribution shifts, using a novel strategy of first encoding and then separating features in the latent space.


Learning to Group Auxiliary Datasets for Molecule

Tinglin Huang, Ziniu Hu, Zhitao Ying

https://openreview.net/forum?id=r9eZH6WNm2

Keywords: molecule, routing mechanism, meta gradient

Compressor summary: MolGroup is a method that uses graph structure similarity and task similarity to select the best auxiliary datasets for small molecule machine learning models, improving their performance.


Grounding Neural Inference with Satisfiability Modulo Theories

Zifan Wang, Saranya Vijayakumar, Kaiji Lu, Vijay Ganesh, Somesh Jha, Matt Fredrikson

https://openreview.net/forum?id=r8snfquzs3

Keywords: Satisfiability Modulo Theories, Solver Layer, Combinatorial Problem, MAXSAT, SAT

Compressor summary: The paper proposes SMTLayer, a technique that integrates SMT solvers into DNNs to encode domain knowledge as mathematical formulas, enabling more efficient, robust, and interpretable learning.


TexQ: Zero-shot Network Quantization with Texture Feature Distribution Calibration

Xinrui Chen, Yizhi Wang, Renao Yan, Yiqing Liu, Tian Guan, Yonghong He

https://openreview.net/forum?id=r8LYNleLf9

Keywords: Zero-shot quantization, Texture feature calibration, Post-training quantization, low bit width, Neural network compression

Compressor summary: The paper proposes TexQ, a novel zero-shot quantization method for neural networks that synthesizes calibration images and uses mixup knowledge distillation to generate diverse samples, improving performance in ultra-low bit width quantization.


Learning Adaptive Tensorial Density Fields for Clean Cryo-ET Reconstruction

YUANHAO WANG, Ramzi Idoughi, Wolfgang Heidrich

https://openreview.net/forum?id=r7g9nFsulw

Keywords: Neural density fields, Coordinate-based representations, Quadtree structure, Cryo-electron microscope

Compressor summary: The authors propose a new learning-based method for improving 3D structure reconstruction from cryo-ET data, which addresses challenges such as missing data and high noise levels by using an adaptive tensor representation and a novel loss function.


Meta-Learning Adversarial Bandit Algorithms

Mikhail Khodak, Ilya Osadchiy, Keegan Harris, Nina Balcan, Kfir Yehuda Levy, Ron Meir, Steven Wu

https://openreview.net/forum?id=r6xGZ0XL2g

Keywords: online learning, multi-armed bandits, meta-learning, multi-task learning, bandit linear optimization

Compressor summary: The authors propose meta-algorithms that improve performance across multiple similar tasks using bandit feedback, by tuning hyperparameters of inner learners in online-within-online settings for MAB and BLO.


Global Update Tracking: A Decentralized Learning Algorithm for Heterogeneous Data

Sai Aparna Aketi, Abolfazl Hashemi, Kaushik Roy

https://openreview.net/forum?id=qyixBZl8Ph

Keywords: Federated Learning, Decentralized Learning, Non-IID Data, Heterogeneous data distribution, Peer-to-peer connectivity

Compressor summary: The paper proposes Global Update Tracking (GUT), a decentralized learning algorithm that reduces the impact of heterogeneous data distribution and achieves state-of-the-art performance on various Computer Vision datasets.


Landscape Surrogate: Learning Decision Losses for Mathematical Optimization Under Partial Information

Arman Zharmagambetov, Brandon Amos, Aaron M Ferber, Taoan Huang, Bistra Dilkina, Yuandong Tian

https://openreview.net/forum?id=qyEm4tF2p1

Keywords: learning surrogates, predict+optimize framework, combinatorial nonlinear optimization, argmin differentiation

Compressor summary: The paper proposes a learnable landscape surrogate to improve learning-integrated optimization by addressing challenges such as problem uncertainty and sparse gradients.


Diffusion Schrödinger Bridge Matching

Yuyang Shi, Valentin De Bortoli, Andrew Campbell, Arnaud Doucet

https://openreview.net/forum?id=qy07OHsJT5

Keywords: diffusion Schrödinger bridge, bridge matching, optimal transport

Compressor summary: The paragraph discusses a new method, Iterative Markovian Fitting (IMF), for solving Schrödinger bridges (SBs) problems and a novel algorithm, Diffusion Schrödinger Bridge Matching (DSBM), which significantly improves SB numerics and recovers various recent transport methods.


Reinforcement Learning with Simple Sequence Priors

Tankred Saanum, Noemi Elteto, Peter Dayan, Marcel Binz, Eric Schulz

https://openreview.net/forum?id=qxF8Pge6vM

Keywords: Deep Reinforcement Learning, Compression, Sequence learning, Information bottleneck, Mutual information

Compressor summary: The paper proposes an RL algorithm that learns simple action sequences, either learned by autoregressive models or compressed by data algorithms, and shows it outperforms model-free approaches in continuous control tasks.


Multimodal Deep Learning Model Unveils Behavioral Dynamics of V1 Activity in Freely Moving Mice

Aiwen Xu, Yuchen Hou, Cris M. Niell, Michael Beyeler

https://openreview.net/forum?id=qv5UZJTNda

Keywords: neuroscience, cognitive science, multimodal learning, representation learning, network architecture, computational biology, visual perception

Compressor summary: The authors introduce a multimodal recurrent neural network that can predict visual cortex activity in free-moving mice by integrating visual input, behavior, and temporal dynamics, revealing new insights into cortical function.


On the Robustness of Mechanism Design under Total Variation Distance

Anuran Makur, Marios Mertzanidis, Alexandros Psomas, Athina Terzoglou

https://openreview.net/forum?id=qumBHr77ht

Keywords: mechanism design, revenue maximization, correlated distributions, total variation distance

Compressor summary: The paper studies how to design truthful mechanisms for agents with unknown and correlated valuation functions, showing that some existing mechanisms are robust to variations in the distribution, and providing various applications and extensions.


Performance Scaling via Optimal Transport: Enabling Data Selection from Partially Revealed Sources

Feiyang Kang, Hoang Anh Just, Anit Kumar Sahu, Ruoxi Jia

https://openreview.net/forum?id=quMBEd27x9

Keywords: data-centric AI, data acquisition, data valuation, performance prediction, data markets, optimal transport, scaling laws

Compressor summary: The paper introduces , a framework that predicts model performance and supports data selection decisions using partial samples of prospective data sources, improving over existing methods in accuracy and computation costs.


TabMT: Generating tabular data with masked transformers

Manbir S Gulati, Paul F Roysdon

https://openreview.net/forum?id=qs4swxtIAQ

Keywords: Tabular Data, Deep Learning, Generative Modeling, Transformers, Masked Transformers, Synthetic data

Compressor summary: TabMT is a new Masked Transformer model that generates synthetic tabular data with strong performance, handling missing data and various data types, and providing good privacy tradeoffs.


PHOTOSWAP: Personalized Subject Swapping in Images

Jing Gu, Yilin Wang, Nanxuan Zhao, Tsu-Jui Fu, Wei Xiong, Qing Liu, Zhifei Zhang, HE Zhang, Jianming Zhang, HyunJoon Jung, Xin Eric Wang

https://openreview.net/forum?id=qqcIM8NiiB

Keywords: image editing, diffusion model, text to image generation

Compressor summary: The text introduces Photoswap, a novel approach for personalized subject swapping in images without training, which preserves the pose and coherence of the image while swapping subjects.


Lower Bounds on Adaptive Sensing for Matrix Recovery

Praneeth Kacham, David Woodruff

https://openreview.net/forum?id=qptO6YDZEP

Keywords: Compressed Sensing, Matrix Recovery, Low rank approximation

Compressor summary: The authors study the lower bounds on adaptive algorithms for recovering low rank matrices using linear measurements, showing that any such algorithm must perform a certain number of rounds to achieve a satisfactory approximation.


Aligning Synthetic Medical Images with Clinical Knowledge using Human Feedback

Shenghuan Sun, Gregory Goldgof, Atul Butte, Ahmed Alaa

https://openreview.net/forum?id=qlnlamFQEa

Keywords: Synthetic clinical data, Machine learning for healthcare

Compressor summary: The paper introduces a pathologist-in-the-loop framework that uses human feedback to improve the quality and plausibility of synthetic medical images generated by conditional diffusion models, addressing challenges in assessing clinical sensibility and incorporating domain knowledge.


Bayesian nonparametric (non-)renewal processes for analyzing neural spike train variability

David Liu, Máté Lengyel

https://openreview.net/forum?id=qlJoo2y3gY

Keywords: Gaussian processes, renewal processes, point processes, neural data analysis, Bayesian machine learning, non-stationary time series

Compressor summary: The authors propose a Bayesian approach to model neural spiking activity that captures instantaneous variability in response to covariates, and apply it to two datasets of animal navigation, showing improved predictive power and richer patterns of modulation.


Stability and Generalization of the Decentralized Stochastic Gradient Descent Ascent Algorithm

Miaoxi Zhu, Li Shen, Bo Du, Dacheng Tao

https://openreview.net/forum?id=ql6LVyi2Dg

Keywords: decentralized algorithm, minimax problem, algorithmic stability, generalization analysis

Compressor summary: The paper studies how decentralized algorithms like D-SGDA can generalize well despite their structure, by analyzing their primal-dual generalization bound using algorithmic stability.


Rank-1 Matrix Completion with Gradient Descent and Small Random Initialization

Daesung Kim, Hye Won Chung

https://openreview.net/forum?id=qjqJL2lfkH

Keywords: Matrix completion, gradient descent, random initialization

Compressor summary: The paper analyzes gradient descent algorithm's convergence behavior on rank-1 symmetric matrix completion problem with small random initialization, and shows that it converges to the ground truth when more samples are available.


High-Fidelity Audio Compression with Improved RVQGAN

Rithesh Kumar, Prem Seetharaman, Alejandro Luebs, Ishaan Kumar, Kundan Kumar

https://openreview.net/forum?id=qjnl1QUnFA

Keywords: audio generation, audio compression, GAN, audio, speech

Compressor summary: The authors introduce a universal neural audio compression algorithm that achieves 90x compression of various audio domains with high fidelity, outperforming existing methods and providing open-source code and weights.


Transformers learn through gradual rank increase

Emmanuel Abbe, Samy Bengio, Enric Boix-Adserà, Etai Littwin, Joshua M. Susskind

https://openreview.net/forum?id=qieeNlO3C7

Keywords: transformers, low-rank bias, incremental learning

Compressor summary: The text describes how transformers learn incrementally, leading to a growing difference between trained and initial weights, and provides theoretical and empirical evidence for this behavior.


Diffusion Self-Guidance for Controllable Image Generation

Dave Epstein, Allan Jabri, Ben Poole, Alexei A Efros, Aleksander Holynski

https://openreview.net/forum?id=qgv56R2YJ7

Keywords: generative models, image editing, diffusion, guidance

Compressor summary: Self-guidance is a technique that uses internal representations of diffusion models to control various aspects of image generation and manipulation without additional training or models.


Convolutional Visual Prompt for Robust Visual Perception

Yun-Yun Tsai, Chengzhi Mao, Junfeng Yang

https://openreview.net/forum?id=qgmrC8jhCo

Keywords: self-supervised learning, representation learning, visual prompts, domain generalization, input adaptation

Compressor summary: Convolutional visual prompts (CVP) are a lightweight method for adapting vision models in test-time settings without labels, improving their robustness against out-of-distribution samples.


Affinity-Aware Graph Networks

Ameya Velingker, Ali Kemal Sinop, Ira Ktena, Petar Veličković, Sreenivas Gollapudi

https://openreview.net/forum?id=qgiG7WZohZ

Keywords: graph neural networks, message passing, effective resistance, hitting time

Compressor summary: The paper explores using random walk measures as features in graph neural networks to improve performance on various node and graph property prediction tasks with low computational complexity.


ANTN: Bridging Autoregressive Neural Networks and Tensor Networks for Quantum Many-Body Simulation

Zhuo Chen, Laker Newhouse, Eddie Chen, Di Luo, Marin Soljacic

https://openreview.net/forum?id=qdsDy0zbn4

Keywords: Autoregressive neural network, tensor network, quantum many-body physics, variational Monte Carlo

Compressor summary: The Autoregressive Neural TensorNet (ANTN) is a novel architecture that combines tensor networks and autoregressive neural networks to improve quantum many-body physics simulation, outperforming existing methods.


Cross-Domain Policy Adaptation via Value-Guided Data Filtering

Kang Xu, Chenjia Bai, Xiaoteng Ma, Dong Wang, Bin Zhao, Zhen Wang, Xuelong Li, Wei Li

https://openreview.net/forum?id=qdM260dXsa

Keywords: Reinforcement Learning; Domain Adaptation; Online Dynamics Adaptation

Compressor summary: The paper proposes a new reinforcement learning method that adapts policies across domains with different dynamics by using value consistency to selectively share data, and shows its superior performance.


Parsel🐍: Algorithmic Reasoning with Language Models by Composing Decompositions

Eric Zelikman, Qian Huang, Gabriel Poesia, Noah Goodman, Nick Haber

https://openreview.net/forum?id=qd9qcbVAwQ

Keywords: reasoning, language models, code synthesis, decomposition

Compressor summary: Parsel is a framework that helps large language models generate complex algorithms with code by breaking down tasks into hierarchical natural language function descriptions and searching over combinations of possible function implementations using tests.


Multi-Head Adapter Routing for Cross-Task Generalization

Lucas Caccia, Edoardo Ponti, Zhan Su, Matheus Pereira, Nicolas Le Roux, Alessandro Sordoni

https://openreview.net/forum?id=qcQhBli5Ho

Keywords: Parameter Efficient Finetuning, Multitask Learning, Transfer Learning, Natural Language Processing

Compressor summary: The paper explores how adapter routing affects PEFT and proposes $\texttt{MHR}$, a new routing method that improves performance by combining subsets of adapter parameters, and $\texttt{MHR}$-$\mu$, a variant that discards routing for single-adapter fine-tuning and zero-shot transfer.


One Risk to Rule Them All: A Risk-Sensitive Perspective on Model-Based Offline Reinforcement Learning

Marc Rigter, Bruno Lacerda, Nick Hawes

https://openreview.net/forum?id=qZjl2TKvUY

Keywords: offline reinforcement learning, model-based reinforcement learning, risk, uncertainty

Compressor summary: The paper proposes a model-based offline reinforcement learning algorithm that combines risk-aversion to aleatoric and epistemic uncertainty to address distributional shift and safety in safety-critical domains.


Multi-Objective Intrinsic Reward Learning for Conversational Recommender Systems

Zhendong Chu, Nan Wang, Hongning Wang

https://openreview.net/forum?id=qYAp31KwU2

Keywords: Conversational Recommendation, Reinforcement Learning, Meta Learning

Compressor summary: The paper proposes a new method to improve Conversational Recommender Systems by learning user preferences and designing task-specific rewards from user interactions, leading to better recommendations with fewer turns.


Are GATs Out of Balance?

Nimrah Mustafa, Aleksandar Bojchevski, Rebekka Burkholz

https://openreview.net/forum?id=qY7UqLoora

Keywords: graph attention networks, gradient flow, conservation law

Compressor summary: The paper explores the optimization and learning dynamics of Graph Attention Networks (GATs), derives a conservation law for their gradient flow dynamics, proposes an initialization scheme to improve trainability and speedup, and serves as a foundation for future studies on positive homogeneous models with attention.


CoPriv: Network/Protocol Co-Optimization for Communication-Efficient Private Inference

Wenxuan Zeng, Meng Li, Haichuan Yang, Wen-jie Lu, Runsheng Wang, Ru Huang

https://openreview.net/forum?id=qVeDwgYsho

Keywords: Private Inference, Network/Protocol Co-Optimization, Winograd Convolution, Structural Re-parameterization

Compressor summary: CoPriv is a framework that optimizes both the DNN architecture and the 2PC inference protocol to significantly reduce communication overhead while maintaining accuracy for secure privacy protection in DNN inference.


LambdaBeam: Neural Program Search with Higher-Order Functions and Lambdas

Kensen Shi, Hanjun Dai, Wen-Ding Li, Kevin Ellis, Charles Sutton

https://openreview.net/forum?id=qVMPXrX4FR

Keywords: Program Synthesis, Programming By Example, Lambdas, Functional Programming

Compressor summary: LambdaBeam is a search algorithm that uses neural networks to synthesize lambda functions for longer and more general programs, overcoming the limitations of prior approaches in handling iterative loops and higher-order functions.


Multi-resolution Spectral Coherence for Graph Generation with Score-based Diffusion

Hyuna Cho, Minjae Jeong, Sooyeon Jeon, Sungsoo Ahn, Won Hwa Kim

https://openreview.net/forum?id=qUlpDjYnsp

Keywords: graph wavelet transform, multi-scale wavelet filtering, graph generation, diffusion model

Compressor summary: The paper introduces a new method called Wavelet Graph Diffusion Model (Wave-GD) that generates realistic graphs with accurate node and edge frequency characteristics by modeling their joint distribution in the spectral space.


Alleviating the Semantic Gap for Generalized fMRI-to-Image Reconstruction

Tao Fang, Qian Zheng, Gang Pan

https://openreview.net/forum?id=qSS9izTOpo

Keywords: fMRI, image reconstruction, brain decoding

Compressor summary: The paper proposes a method (GESS) to improve fMRI-to-image reconstruction by using CLIP features to bridge the semantic gap between training and testing data, as well as leveraging structural information and uncertainty estimation.


The Double-Edged Sword of Implicit Bias: Generalization vs. Robustness in ReLU Networks

Spencer Frei, Gal Vardi, Peter Bartlett, Nathan Srebro

https://openreview.net/forum?id=qSCziWQBPD

Keywords: adversarial robustness, neural networks, implicit bias, generalization

Compressor summary: The paper investigates how the biased gradient flow in two-layer ReLU networks affects generalization and adversarial robustness, showing that it leads to solutions that are good at fitting data but vulnerable to small perturbations.


Provably Efficient Offline Goal-Conditioned Reinforcement Learning with General Function Approximation and Single-Policy Concentrability

Hanlin Zhu, Amy Zhang

https://openreview.net/forum?id=qS9aHF8bXz

Keywords: offline goal-conditioned RL, provably efficient algorithm, single-policy concentrability, general function approximation

Compressor summary: The paper provides a rigorous theoretical analysis of an existing offline goal-conditioned reinforcement learning algorithm, showing its efficiency, stability, and effectiveness in various real-world environments.


Complex Query Answering on Eventuality Knowledge Graph with Implicit Logical Constraints

Jiaxin Bai, Xin Liu, Weiqi Wang, Chen Luo, Yangqiu Song

https://openreview.net/forum?id=qQnO1HLQHe

Keywords: Knowledge Graph, Complex Query Answering, Eventuality Graph

Compressor summary: The paper proposes a new framework, CEQA, that uses implicit logical constraints and theorem provers to answer complex queries on eventuality-centric knowledge graphs, improving neural query encoders' performance.


Statistical Insights into HSIC in High Dimensions

Tao Zhang, Yaowu Zhang, Tingyou Zhou

https://openreview.net/forum?id=qPyvuFT0U9

Keywords: High dimensionality; Independence test; Kernel method; Nonlinear dependency.

Compressor summary: The paper studies how a popular measure of dependence, HSIC, behaves when the dimensions of the variables it measures grow at different rates and provides conditions for its performance in high-dimensional settings.


Monitor-Guided Decoding of Code LMs with Static Analysis of Repository Context

Lakshya Agrawal, Aditya Kanade, Navin Goyal, Shuvendu K Lahiri, Sriram Rajamani

https://openreview.net/forum?id=qPUbKxKvXq

Keywords: Language models, code generation, correctness, program analysis

Compressor summary: Monitor-guided decoding (MGD) helps language models of code achieve better compilation rates and agreement with ground truth by using static analysis to provide context, especially when dealing with global information.


Read and Reap the Rewards: Learning to Play Atari with the Help of Instruction Manuals

Yue Wu, Yewen Fan, Paul Pu Liang, Amos Azaria, Yuanzhi Li, Tom Mitchell

https://openreview.net/forum?id=qP0Drg2HuH

Keywords: Games, Instruction Manual, Atari Games, Large Language Models, Language Models, Zero-shot, In-context prompting

Compressor summary: The Read and Reward framework uses instruction manuals to enhance the performance and efficiency of reinforcement learning agents on Atari games.


TriRE: A Multi-Mechanism Learning Paradigm for Continual Knowledge Retention and Promotion

Preetha Vijayan, Prashant Shivaram Bhat, Bahram Zonooz, Elahe Arani

https://openreview.net/forum?id=qL3zPoWJda

Keywords: Continual Learning, Catastrophic Forgetting, Experience Replay, Lifelong Learning, Bio-Inspired, Active Forgetting, Scalable Neurogenesis

Compressor summary: The paragraph introduces TriRE, a novel continual learning paradigm that combines multiple neurophysiological processes to reduce catastrophic forgetting and promote knowledge transfer in deep neural networks.


VaRT: Variational Regression Trees

Sebastian Salazar

https://openreview.net/forum?id=qJRlz3SucN

Keywords: Probabilistic Machine Learning, Variational Inference, Bayesian Inference, Bayesian Nonparametrics

Compressor summary: The paper presents a new Bayesian model for decision trees that uses variational inference and shows its effectiveness in regression and causal inference problems, with a PyTorch implementation available.


Is Heterogeneity Notorious? Taming Heterogeneity to Handle Test-Time Shift in Federated Learning

Yue Tan, Chen Chen, Weiming Zhuang, Xin Dong, Lingjuan Lyu, Guodong Long

https://openreview.net/forum?id=qJJmu4qsLO

Keywords: Federated Learning, Test-Time Shift, Contrastive Learning

Compressor summary: FedICON is a framework that uses contrastive learning to capture and adapt to heterogeneous data in federated learning, improving performance under test-time shifts.


Sketching Algorithms for Sparse Dictionary Learning: PTAS and Turnstile Streaming

Gregory Dexter, Petros Drineas, David Woodruff, Taisuke Yasuda

https://openreview.net/forum?id=qHzEFxtheD

Keywords: dictionary learning, k means clustering, sketching, ptas, streaming

Compressor summary: The paper develops new techniques to extend sketching-based approaches to sparse dictionary learning and Euclidean k-means clustering problems, and obtains new upper and lower bounds for these problems in various settings.


Don't be so Monotone: Relaxing Stochastic Line Search in Over-Parameterized Models

Leonardo Galli, Holger Rauhut, Mark Schmidt

https://openreview.net/forum?id=qHrZszJSXj

Keywords: line search, nonmonotone, stochastic gradient descent, over-parametrized models, Polyak step size, optimization

Compressor summary: Nonmonotone line search methods can improve the speed and generalization of SGD/Adam by allowing larger step sizes without compromising convergence rates, as demonstrated by the PoNoS method and a new resetting technique.


Towards Revealing the Mystery behind Chain of Thought: A Theoretical Perspective

Guhao Feng, Bohang Zhang, Yuntian Gu, Haotian Ye, Di He, Liwei Wang

https://openreview.net/forum?id=qHrADgAdYu

Keywords: Chain-of-Thought Prompting, Large Language Models, Theory, Circuit Complexity, Dynamic Programming

Compressor summary: The paper investigates how Chain-of-Thought prompting enhances Large Language Models' performance in solving mathematical and decision-making problems, using circuit complexity theory and experiments.


Gradient Descent with Linearly Correlated Noise: Theory and Applications to Differential Privacy

Anastasia Koloskova, Ryan McKenna, Zachary Charles, J Keith Rush, Hugh Brendan McMahan

https://openreview.net/forum?id=qCglMj6A4z

Keywords: optimization, machine learning, differential privacy

Compressor summary: The paper analyzes how gradient descent works with linearly correlated noise, which arises in some privacy-preserving optimization methods, and proposes improved matrix factorizations for better performance.


VLATTACK: Multimodal Adversarial Attacks on Vision-Language Tasks via Pre-trained Models

Ziyi Yin, Muchao Ye, Tianrong Zhang, Tianyu Du, Jinguo Zhu, Han Liu, Jinghui Chen, Ting Wang, Fenglong Ma

https://openreview.net/forum?id=qBAED3u1XZ

Keywords: vision-language, adversarial attacks, pre-trained model, fine-tuned model

Compressor summary: This paper proposes VLATTACK, a method to craft image and text perturbations that can attack black-box fine-tuned Vision-Language models on different tasks by using block-wise similarity attacks for images and existing methods for texts.


Complex-valued Neurons Can Learn More but Slower than Real-valued Neurons via Gradient Descent

Jin-Hui Wu, Shao-Qun Zhang, Yuan Jiang, Zhi-Hua Zhou

https://openreview.net/forum?id=qA0uHmaVKk

Keywords: Complex-valued Neural Networks; Learning Neurons; Real-valued Neural Networks; Convergence Rate

Compressor summary: Complex-valued neural networks can efficiently learn functions of real-valued neurons and complex-valued neurons faster than real-valued neurons can learn each other or be learned by complex-valued neurons.


Trust Region-Based Safe Distributional Reinforcement Learning for Multiple Constraints

Dohyeong Kim, Kyungjae Lee, Songhwai Oh

https://openreview.net/forum?id=q9WMXjUxxT

Keywords: Reinforcement learning, Safety, Multiple Constraints, Distributional Critic

Compressor summary: The paper proposes a trust region-based safe reinforcement learning algorithm for robotic tasks with multiple constraints, and shows its effectiveness in experiments.


Deep Contract Design via Discontinuous Networks

Tonghan Wang, Paul Duetting, Dmitry Ivanov, Inbal Talgam-Cohen, David C. Parkes

https://openreview.net/forum?id=q8mH2d6uw2

Keywords: Automated contract design, discontinuous neural networks

Compressor summary: The paper proposes using deep learning and a new type of network (DeLU) to automate the design of optimal contracts, considering both agent's incentives and principal's utility.


Learning from Active Human Involvement through Proxy Value Propagation

Zhenghao Peng, Wenjie Mo, Chenda Duan, Quanyi Li, Bolei Zhou

https://openreview.net/forum?id=q8SukwaEBy

Keywords: Machine Learning, Human-in-the-loop Reinforcement Learning, Safety, Sample Efficiency, Reward-free

Compressor summary: The proposed method uses a proxy value function to express human intents and label state-action pairs for policy optimization, allowing the AI agent to learn from active human involvement in various control tasks.


Triple Eagle: Simple, Fast and Practical Budget-Feasible Mechanisms

Kai Han, You Wu, He Huang, Shuang Cui

https://openreview.net/forum?id=q6bVqOgGxP

Keywords: mechanism design, budget-feasible, truthful

Compressor summary: The paper introduces TripleEagle, a new algorithmic framework for designing Budget-Feasible Mechanisms with better approximation ratios, linear complexities, and strategyproofness for submodular valuation functions.


Predict, Refine, Synthesize: Self-Guiding Diffusion Models for Probabilistic Time Series Forecasting

Marcel Kollovieh, Abdul Fatir Ansari, Michael Bohlke-Schneider, Jasper Zschiegner, Hao Wang, Bernie Wang

https://openreview.net/forum?id=q6X038vKgU

Keywords: diffusion models, time series forecasting, generative modeling, deep learning

Compressor summary: The authors propose a task-agnostic diffusion model for time series applications and show its effectiveness in forecasting, refinement, and synthetic data generation.


Diff-Foley: Synchronized Video-to-Audio Synthesis with Latent Diffusion Models

Simian Luo, Chuanhao Yan, Chenxu Hu, Hang Zhao

https://openreview.net/forum?id=q5FAZAIooz

Keywords: Video-to-Audio Generation; Contrastive Audio-Visual Pretraining; Latent Diffusion Model;

Compressor summary: Diff-Foley is a new method for generating high-quality audio from silent videos with improved synchronization and relevance, using a latent diffusion model and contrastive pretraining.


Discriminative Feature Attributions: Bridging Post Hoc Explainability and Inherent Interpretability

Usha Bhalla, Suraj Srinivas, Himabindu Lakkaraju

https://openreview.net/forum?id=q4HlFS7B7Y

Keywords: Machine Learning Explainability, Machine Learning Interpretability

Compressor summary: The paragraph discusses a method called Distractor Erasure Tuning (DiET), which improves the faithfulness of post hoc explanations for machine learning models by making them robust to distractor erasure in inputs.


Efficient Data Subset Selection to Generalize Training Across Models: Transductive and Inductive Networks

Eeshaan Jain, Tushar Nandy, Gaurav Aggarwal, Ashish V. Tendulkar, Rishabh K Iyer, Abir De

https://openreview.net/forum?id=q3fCWoC9l0

Keywords: Data Subset Selection, Efficient Learning

Compressor summary: The paragraph describes $\texttt{SubSelNet}$, a non-adaptive framework for subset selection in efficient learning that uses an attention-based neural gadget and can be either transductive or inductive, achieving better results than other methods on various real datasets.


Brain Dissection: fMRI-trained Networks Reveal Spatial Selectivity in the Processing of Natural Images

Gabriel Herbert Sarch, Michael J. Tarr, Katerina Fragkiadaki, Leila Wehbe

https://openreview.net/forum?id=q3fA5tTod3

Keywords: Computational Neuroscience, Deep Neural Networks, Visual Neuroscience, Visual Streams, Scene Perception, Brain Imaging

Compressor summary: The paper trains neural networks to predict brain responses to images and uses "network dissection" to explore how different regions of the visual cortex interpret various features of natural scenes.


Learning to Search Feasible and Infeasible Regions of Routing Problems with Flexible Neural k-Opt

Yining Ma, Zhiguang Cao, Yeow Meng Chee

https://openreview.net/forum?id=q1JukwH2yP

Keywords: learning to optimize, vehicle routing problem, combinatorial optimization

Compressor summary: The paper introduces NeuOpt, a learning-to-search solver for routing problems that learns flexible k-opt exchanges and uses GIRE, D2A, and reward shaping to explore both feasible and infeasible regions effectively, achieving superior performance on TSP and CVRP compared to existing methods.


Learning Linear Causal Representations from Interventions under General Nonlinear Mixing

Simon Buchholz, Goutham Rajendran, Elan Rosenfeld, Bryon Aragam, Bernhard Schölkopf, Pradeep Kumar Ravikumar

https://openreview.net/forum?id=q131tA7HCT

Keywords: Causal Representation Learning, Interventional data, Gaussian Structural Causal models

Compressor summary: The authors study how to learn causal representations from unknown interventions using Gaussian latent distributions and a general mixing function, and they prove identifiability results without access to intervention targets.


SPQR: Controlling Q-ensemble Independence with Spiked Random Model for Reinforcement Learning

Dohyeok Lee, Seungyub Han, Taehyun Cho, Jungwoo Lee

https://openreview.net/forum?id=q0sdoFIfNg

Keywords: Deep Reinforcement Learning, Ensemble Q-learning

Compressor summary: The paper proposes a new method, spiked Wishart Q-ensemble independence regularization (SPQR), to improve deep reinforcement learning by ensuring diversity of multiple Q-functions using random matrix theory.


On the Consistency of Maximum Likelihood Estimation of Probabilistic Principal Component Analysis

Arghya Datta, Sayak Chakrabarty

https://openreview.net/forum?id=q0RfX96un8

Keywords: maximum likelihood estimate, non-identifiability, Redner approach, quotient topological spaces, consistency

Compressor summary: The text describes PPCA, a widely used statistical tool with applications in various fields, and proposes a novel approach to resolve the identifiability issue of its maximum likelihood solution using quotient topological spaces.


StateMask: Explaining Deep Reinforcement Learning through State Mask

Zelei Cheng, Xian Wu, Jiahao Yu, Wenhai Sun, Wenbo Guo, Xinyu Xing

https://openreview.net/forum?id=pzc6LnUxYN

Keywords: deep reinforcement learning, interpretation, explanation

Compressor summary: StateMask is a novel method to identify critical states for an agent's final reward by learning a mask net that temporarily blinds the agent without affecting its performance.


Unified Enhancement of Privacy Bounds for Mixture Mechanisms via $f$-Differential Privacy

Chendi Wang, Buxin Su, Jiayuan Ye, Reza Shokri, Weijie J Su

https://openreview.net/forum?id=pw5hEuEroL

Keywords: Differential privacy, $f$-DP, mixture mechanisms, shuffling, differentially private gradient descent

Compressor summary: The paper proposes an improved analysis of differential privacy for shuffling models and one-iteration DP-GD using $f$-DP, with applications to random initialization and mixture mechanisms.


Flow-Attention-based Spatio-Temporal Aggregation Network for 3D Mask Detection

Yuxin Cao, Yian Li, Yumeng Zhu, Derui Wang, Minhui Xue

https://openreview.net/forum?id=pvSKVt3EsM

Keywords: 3D mask detection, spatio-temporal aggregation, optical flow, deep learning

Compressor summary: FASTEN is a novel 3D mask detection framework that uses flow attention and spatio-temporal aggregation to quickly and accurately detect spoofing attacks in face recognition systems with low computational overhead and only five frames of input.


Most Neural Networks Are Almost Learnable

Amit Daniely, Nathan Srebro, Gal Vardi

https://openreview.net/forum?id=pvPujuvjQd

Keywords: learning neural networks, computational complexity, random networks

Compressor summary: The paper proposes a polynomial-time approximation scheme (PTAS) for learning random constant-depth neural networks with an additive error of epsilon under certain conditions on the network size and activation functions.


GraphPatcher: Mitigating Degree Bias for Graph Neural Networks via Test-time Augmentation

Mingxuan Ju, Tong Zhao, Wenhao Yu, Neil Shah, Yanfang Ye

https://openreview.net/forum?id=puupdGOWUp

Keywords: Graph neural network, Test-time Augmentation

Compressor summary: The paragraph discusses a test-time augmentation framework, GraphPatcher, that improves the performance of graph neural networks on low-degree nodes without compromising their ability to handle high-degree nodes.


Self-Predictive Universal AI

Elliot Catt, Jordi Grau-Moya, Marcus Hutter, Matthew Aitchison, Tim Genewein, Gregoire Deletang, Li Kevin Wenliang, Joel Veness

https://openreview.net/forum?id=psXVkKO9No

Keywords: General Reinforcement Learning, Reinforcement Learning, Self-Modeling, Bayes-optimality, Policy Distillation, Uncertainty, Universal AI

Compressor summary: Self-AIXI is a universal agent that uses learning to obtain good policies by self-predicting its own action data, converging to AIXI and having maximal intelligence and self-optimization properties.


Label Poisoning is All You Need

Rishi Dev Jha, Jonathan Hayase, Sewoong Oh

https://openreview.net/forum?id=prftZp6mDH

Keywords: security, backdoor attack

Compressor summary: The paper introduces FLIP, a novel label-only backdoor attack method that can successfully manipulate image predictions with corrupted training labels, while maintaining high clean test accuracy.


Distributional Pareto-Optimal Multi-Objective Reinforcement Learning

Xin-Qiang Cai, Pushi Zhang, Li Zhao, Jiang Bian, Masashi Sugiyama, Ashley Juan Llorens

https://openreview.net/forum?id=prIwYTU9PV

Keywords: Multi-Objective Reinforcement Learning

Compressor summary: DPMORL is a novel approach that learns policies balancing multiple objectives while considering return uncertainty and satisfying diverse distributional preferences.


ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation

Zhengyi Wang, Cheng Lu, Yikai Wang, Fan Bao, Chongxuan Li, Hang Su, Jun Zhu

https://openreview.net/forum?id=ppJuFSOAnM

Keywords: diffusion models, text to 3D

Compressor summary: VSD is a particle-based variational framework that improves text-to-3D generation by modeling the 3D parameter as a random variable and addressing over-saturation, over-smoothing, and low-diversity issues.


Wasserstein Quantum Monte Carlo: A Novel Approach for Solving the Quantum Many-Body Schrödinger Equation

Kirill Neklyudov, Jannes Nys, Luca Thiede, Juan Felipe Carrasquilla Alvarez, qiang liu, Max Welling, Alireza Makhzani

https://openreview.net/forum?id=pjSzKhSrfs

Keywords: Quantum Monte Carlo, Schrödinger equation, Wasserstein Fisher-Rao gradient flow

Compressor summary: The paper introduces Wasserstein Quantum Monte Carlo, a new method for solving quantum many-body problems that uses transportation-based gradients and converges faster than traditional methods.


Kernelized Reinforcement Learning with Order Optimal Regret Bounds

Sattar Vakili, Julia Olkhovskaya

https://openreview.net/forum?id=pirH9ycaNg

Keywords: Reinforcement Learning, Kernel ridge regression, Gaussian processes, LSVI

Compressor summary: The paper proposes $\pi$-KRVI, an optimistic modification of least-squares value iteration for reinforcement learning with nonlinear function approximation using kernel ridge regression, and proves its sublinear regret bound under a general setting.


Learning Probabilistic Symmetrization for Architecture Agnostic Equivariance

Jinwoo Kim, Dat Tien Nguyen, Ayhan Suleymanzade, Hyeokjun An, Seunghoon Hong

https://openreview.net/forum?id=phnN1eu5AX

Keywords: equivariant machine learning, transformers, graphs, general-purpose architectures

Compressor summary: The authors propose a framework that uses a non-equivariant backbone and an equivariant network to learn functions with group symmetries, achieving competitive results against tailored equivariant architectures.


FedFed: Feature Distillation against Data Heterogeneity in Federated Learning

Zhiqin Yang, Yonggang Zhang, Yu Zheng, Xinmei Tian, Hao Peng, Tongliang Liu, Bo Han

https://openreview.net/forum?id=phnGilhPH8

Keywords: Federated Learning, Data Heterogeneity

Compressor summary: The paper proposes a method called FedFed that shares some features of data to improve federated learning while preserving privacy.


Recurrent Hypernetworks are Surprisingly Strong in Meta-RL

Jacob Beck, Risto Vuorio, Zheng Xiong, Shimon Whiteson

https://openreview.net/forum?id=pefAAzu8an

Keywords: meta-RL, RL, reinforcement learning, memory, rnn, recurrent, hypernetwork, few-shot

Compressor summary: The paper investigates how using hypernetworks can improve the performance of simple recurrent networks in meta-RL tasks.


Reducing Blackwell and Average Optimality to Discounted MDPs via the Blackwell Discount Factor

Julien Grand-Clément, Marek Petrik

https://openreview.net/forum?id=pcuC65JWAa

Keywords: Markov Decision Process, Blackwell optimality, average optimality, robust optimization

Compressor summary: The paper introduces a discount factor for Markov Decision Processes called the Blackwell discount factor and shows how it can be used to find optimal policies without strict assumptions on the MDP structure.


Initialization-Dependent Sample Complexity of Linear Predictors and Neural Networks

Roey Magen, Ohad Shamir

https://openreview.net/forum?id=pcpjtYNJCH

Keywords: sample complexity; learning theory; neural networks; linear predictors

Compressor summary: The text discusses new findings on how many examples are needed to learn vector-valued linear predictors and neural networks, showing different sample complexity behavior compared to scalar-valued linear predictors.


Binarized Spectral Compressive Imaging

Yuanhao Cai, Yuxin Zheng, Jing Lin, Xin Yuan, Yulun Zhang, Haoqian Wang

https://openreview.net/forum?id=pcKwgdVAlq

Keywords: Applications, Computer Vision, Low-level Vision, Image Restoration, Snapshot Compressive Imaging, Hyperspectral Image Reconstruction

Compressor summary: The paper proposes a novel method, BiSRNet, for efficient HSI restoration from compressed measurements in SCI systems using binarized convolutional modules that outperform existing binarization algorithms.


Learning Generalizable Agents via Saliency-guided Features Decorrelation

Sili Huang, Yanchao Sun, Jifeng Hu, Siyuan Guo, Hechang Chen, Yi Chang, Lichao Sun, Bo Yang

https://openreview.net/forum?id=pb1OwZNgr2

Keywords: reinforcement learning, generalization

Compressor summary: SGFD is a method to improve generalization in visual RL by decorrelating features using saliency maps and random Fourier functions.


Free-Bloom: Zero-Shot Text-to-Video Generator with LLM Director and LDM Animator

Hanzhuo Huang, Yufan Feng, Cheng Shi, Lan Xu, Jingyi Yu, Sibei Yang

https://openreview.net/forum?id=paa2OU5jN8

Keywords: Text-to-Video, Zero-Shot Generation, Large Language Model, Latent Diffusion Models

Compressor summary: Text-to-video generation using large language models and latent diffusion models to create vivid and high-quality videos without video data or training.


Kernel-Based Tests for Likelihood-Free Hypothesis Testing

Patrik Robert Gerber, Tianze Jiang, Yury Polyanskiy, Rui Sun

https://openreview.net/forum?id=paTESG8iSE

Keywords: Kernel methods, Maximum mean discrepancy, Likelihood-free inference, Hypothesis testing, Minimax statistics

Compressor summary: The paper studies the trade-off between labeling more inputs and using less training data in two applications: detecting the Higgs boson and identifying planted images in CIFAR-10 dataset.


Enhancing Robot Program Synthesis Through Environmental Context

Tianyi Chen, Qidi Wang, Zhen Dong, Liwei Shen, Xin Peng

https://openreview.net/forum?id=pZ2Ww45GkL

Keywords: program synthesis, partial envrionment, robotic programming, domain-specific language

Compressor summary: The paper presents a framework for synthesizing programs for robots in partially observed environments by learning an embedding space and using graph structure to rectify errors.


The Best of Both Worlds in Network Population Games: Reaching Consensus and Convergence to Equilibrium

Shuyue Hu, Harold Soh, Georgios Piliouras

https://openreview.net/forum?id=pXtVyj4R33

Keywords: Multi-Agent Learning, Consensus Formation, Smooth Fictitious Play, Network Game, Population Game

Compressor summary: The paper explores how consensus and equilibrium are related in multi-agent systems with multiple interacting sub-populations and shows that smooth fictitious play can achieve both goals.


Principled Weight Initialisation for Input-Convex Neural Networks

Pieter-Jan Hoedt, Günter Klambauer

https://openreview.net/forum?id=pWZ97hUQtQ

Keywords: initialization, signal propagation, input-convex networks

Compressor summary: ICNNs are networks with guaranteed convexity that require a new weight initialization method for better learning and generalization, which can also be trained without skip-connections and applied to drug discovery tasks.


RETVec: Resilient and Efficient Text Vectorizer

Elie Bursztein, Marina Zhang, Owen Skipper Vallis, Xinyu Jia, Alexey Kurakin

https://openreview.net/forum?id=pVlC0reMKq

Keywords: language modeling, text embedding, adversarial text attack, text vectorization

Compressor summary: RETVec is a robust, multilingual text vectorizer with a character encoding and an optional embedding model that can handle typos and adversarial attacks.


CorresNeRF: Image Correspondence Priors for Neural Radiance Fields

Yixing Lao, Xiaogang Xu, zhipeng cai, Xihui Liu, Hengshuang Zhao

https://openreview.net/forum?id=pTCZWSDltG

Keywords: Neural Radiance Fields, 3D Reconstruction, Few View

Compressor summary: CorresNeRF uses image correspondence priors to supervise NeRF training, improving performance under sparse-view settings for novel view synthesis and surface reconstruction tasks.


Parameter-efficient Tuning of Large-scale Multimodal Foundation Model

Haixin Wang, Xinlong Yang, Jianlong Chang, Dian Jin, Jinan Sun, Shikun Zhang, Xiao Luo, Qi Tian

https://openreview.net/forum?id=pT8DIhsJCw

Keywords: parameter-efficient transfer learning; multi-modal learning; prompt learning

Compressor summary: The paper introduces AURORA, a framework for multimodal transfer learning that uses mode approximation and Informative Context Enhancement to achieve high performance with few parameters.


Detecting Any Human-Object Interaction Relationship: Universal HOI Detector with Spatial Prompt Learning on Foundation Models

Yichao Cao, Qingfei Tang, Xiu Su, Song Chen, Shan You, Xiaobo Lu, Chang Xu

https://openreview.net/forum?id=pQvAL40Cdj

Keywords: Human-object interaction, Commonsense Knowledge, Foundation Models

Compressor summary: UniHOI is a method for recognizing complex human-object interactions in open world settings using vision-language foundation models and large language models to guide decoding, interpret interactions, and handle various input types.


Leveraging Vision-Centric Multi-Modal Expertise for 3D Object Detection

Linyan Huang, Zhiqi Li, Chonghao Sima, Wenhai Wang, Jingdong Wang, Yu Qiao, Hongyang Li

https://openreview.net/forum?id=pQF9kbM8Ea

Keywords: camera-only detection, multi-modal distillation, multi-view object detection

Compressor summary: The paper proposes VCD, a framework that improves camera-only 3D object detection by using a multi-modal expert with LiDAR input and trajectory-based distillation supervision, achieving state-of-the-art results on nuScenes.


ESSEN: Improving Evolution State Estimation for Temporal Networks using Von Neumann Entropy

Qiyao Huang, Yingyue Zhang, Zhihong Zhang, Edwin Hancock

https://openreview.net/forum?id=pO7d6iFdnc

Keywords: Temporal Network, Graph Neural Network, Von Neumann Entropy

Compressor summary: ESSEN is a novel framework that uses von Neumann entropy and thermodynamic temperature to measure temporal network evolution and improve link prediction performance.


Statistical Knowledge Assessment for Large Language Models

Qingxiu Dong, Jingjing Xu, Lingpeng Kong, Zhifang Sui, Lei Li

https://openreview.net/forum?id=pNtG6NAmx0

Keywords: Large Language Models, Knowledge Assessment, Evaluation

Compressor summary: The paper introduces KaRR, a statistical approach that quantifies how well large language models (LLMs) can generate factually correct answers for given facts, based on the ratio of LLM text to random chances. The paper tests 20 LLMs and finds strong agreement with human judgments.


PRODIGY: Enabling In-context Learning Over Graphs

Qian Huang, Hongyu Ren, Peng Chen, Gregor Kržmanc, Daniel Zeng, Percy Liang, Jure Leskovec

https://openreview.net/forum?id=pLwYhNNnoR

Keywords: Graph Neural Network, in-context learning, pretraining

Compressor summary: PRODIGY is a novel framework that enables in-context learning over graphs using a prompt graph representation and a graph neural network architecture, achieving strong performance on citation networks and knowledge graphs tasks.


Kernelized Cumulants: Beyond Kernel Mean Embeddings

Patric Bonnier, Harald Oberhauser, Zoltán Szabó

https://openreview.net/forum?id=pLsPFxqn7J

Keywords: kernel, cumulant, mean embedding, Hilbert-Schmidt independence criterion, maximum mean discrepancy

Compressor summary: The paper introduces kernelized cumulants for reproducing kernel Hilbert spaces, which are computationally efficient and provide new all-purpose statistics for data analysis.


An active learning framework for multi-group mean estimation

Abdellah Aznag, Rachel Cummings, Adam N. Elmachtoub

https://openreview.net/forum?id=pLcSrn8NpJ

Keywords: Active learning, mean estimation, bandit feedback, data acquisition

Compressor summary: The paper proposes an active learning algorithm called Variance-UCB to learn the means of multiple unknown groups by minimizing the $p$-norm of variances, with optimal regret bounds for finite and infinite $p$.


Non-Smooth Weakly-Convex Finite-sum Coupled Compositional Optimization

Quanqi Hu, Dixian Zhu, Tianbao Yang

https://openreview.net/forum?id=pLOWV1UGF6

Keywords: non-smooth optimization, weakly-convex optimization, compositional optimization, AUC maximization

Compressor summary: The paper studies non-smooth weakly-convex finite-sum coupled compositional optimization problems and proposes algorithms for solving them in machine learning applications.


PID-Inspired Inductive Biases for Deep Reinforcement Learning in Partially Observable Control Tasks

Ian Char, Jeff Schneider

https://openreview.net/forum?id=pKnhUWqZTJ

Keywords: Reinforcement Learning, Control, POMDP

Compressor summary: The paragraph discusses using PID features for history encoding in deep reinforcement learning to improve robustness and performance on various control tasks.


Adversarial Examples Exist in Two-Layer ReLU Networks for Low Dimensional Linear Subspaces

Odelia Melamed, Gilad Yehudai, Gal Vardi

https://openreview.net/forum?id=pJbEXBBN88

Keywords: Adversarial Examples, Robustness, Neural Networks, Classification

Compressor summary: This paper studies how gradient methods affect the robustness of two-layer neural networks trained on low dimensional linear subspaces and suggests ways to improve their resistance to adversarial examples in certain directions.


Optimal Exploration for Model-Based RL in Nonlinear Systems

Andrew Wagenmaker, Guanya Shi, Kevin Jamieson

https://openreview.net/forum?id=pJQu0zpKCS

Keywords: reinforcement learning, control theory, system identification, experiment design, active learning

Compressor summary: The paper proposes an algorithm for exploring nonlinear dynamical systems to learn relevant parameters for controller optimization more efficiently.


What Makes Good Examples for Visual In-Context Learning?

Yuanhan Zhang, Kaiyang Zhou, Ziwei Liu

https://openreview.net/forum?id=pIXTMrBe7f

Keywords: computer vision, visual in-context learning, prompt learning

Compressor summary: This paper proposes prompt retrieval frameworks for large vision models to improve their performance in in-context learning by selecting appropriate input-output pairs without updating any internal model parameters.


Causal Effect Identification in Uncertain Causal Networks

Sina Akbari, Fateme Jamshidi, Ehsan Mokhtarian, Matthew James Vowels, Jalal Etesami, Negar Kiyavash

https://openreview.net/forum?id=pH4Fv7C3yC

Keywords: Causal effect, identifiability, causal DAGs, probabilistic graphs

Compressor summary: The paper studies how to find the most plausible probabilistic causal graph that can identify a specific causal effect, using algorithms that solve an NP-hard optimization problem called edge ID.


FlatMatch: Bridging Labeled Data and Unlabeled Data with Cross-Sharpness for Semi-Supervised Learning

Zhuo Huang, Li Shen, Jun Yu, Bo Han, Tongliang Liu

https://openreview.net/forum?id=pE3yaP0Eqg

Keywords: Semi-Supervised Learning

Compressor summary: FlatMatch is a new SSL method that balances learning on labeled and unlabeled data by penalizing the prediction difference between worst-case and original models, improving generalization performance.


Latent Space Translation via Semantic Alignment

Valentino Maiorca, Luca Moschella, Antonio Norelli, Marco Fumero, Francesco Locatello, Emanuele Rodolà

https://openreview.net/forum?id=pBa70rGHlr

Keywords: latent space translation, relative representation, Procrustes analysis, zero-shot, stitching, latent communication, representation learning, manifold alignment, multimodal

Compressor summary: The paper proposes a method to translate and stitch representations learned from different neural networks, enabling better understanding of their intrinsic similarity and improving performance in various tasks and domains.


Demystifying the Optimal Performance of Multi-Class Classification

Minoh Jeong, Martina Cardone, Alex Dytso

https://openreview.net/forum?id=p9k5MS0JAL

Keywords: Bayes error, estimation, classification, minimum error probability

Compressor summary: The authors propose a new method to estimate the Bayes error rate for multi-class classification problems and analyze its properties and performance on synthetic and real data.


From Trainable Negative Depth to Edge Heterophily in Graphs

Yuchen Yan, Yuzhong Chen, Huiyuan Chen, Minghua Xu, Mahashweta Das, Hao Yang, Hanghang Tong

https://openreview.net/forum?id=p8lowHbuv8

Keywords: graph convolutional network

Compressor summary: The paper proposes a new GCN model that can adjust its depth continuously to handle both homophilic and heterophilic graph topologies, and shows improved performance on node classification tasks.


Modality-Independent Teachers Meet Weakly-Supervised Audio-Visual Event Parser

Yung-Hsuan Lai, Yen-Chun Chen, Yu-Chiang Frank Wang

https://openreview.net/forum?id=p8gTWkFIvx

Keywords: Audio-Visual Video Parsing, Audio-Visual Learning

Compressor summary: The paper introduces VALOR, a method that uses pre-trained models as teachers to help learn audio and visual labels from weak labels in an unaligned setting, improving performance on the Look, Listen, and Parse dataset and Audio-Visual Event Localization tasks.


DIN-SQL: Decomposed In-Context Learning of Text-to-SQL with Self-Correction

Mohammadreza Pourreza, Davood Rafiei

https://openreview.net/forum?id=p53QDxSIc5

Keywords: In-Context Learning, Text-to-SQL, Task Decomposition, Spider Challenge, Natural Language Interfaces to Databases

Compressor summary: The paper proposes decomposing the text-to-SQL task into smaller sub-tasks and using LLMs to solve them, leading to improved performance and setting new benchmarks.


Riemannian stochastic optimization methods avoid strict saddle points

Ya-Ping Hsieh, Mohammad Reza Karimi Jaghargh, Andreas Krause, Panayotis Mertikopoulos

https://openreview.net/forum?id=p4SjKPchJy

Keywords: Riemannian optimization, saddle points, stochastic approximation

Compressor summary: The paper investigates if retraction-based stochastic Riemannian optimization algorithms can avoid saddle points with certain assumptions, ensuring the convergence to a local minimizer.


How does GPT-2 compute greater-than?: Interpreting mathematical abilities in a pre-trained language model

Michael Hanna, Ollie Liu, Alexandre Variengien

https://openreview.net/forum?id=p4PckNQR8k

Keywords: interpretability, language models, NLP

Compressor summary: The paper examines how pre-trained language models like GPT-2 small use basic math skills and identifies the specific components and mechanisms involved in solving a simple mathematical task.


Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision

Zhiqing Sun, Yikang Shen, Qinhong Zhou, Hongxin Zhang, Zhenfang Chen, David Daniel Cox, Yiming Yang, Chuang Gan

https://openreview.net/forum?id=p40XRfBX96

Keywords: AI Alignment, Large Language Models, In Context Learning, Neural Symbolics

Compressor summary: The proposed SELF-ALIGN approach enables AI assistant agents to align themselves with human intentions using minimal human supervision by combining LLMs and principle-driven reasoning.


FlowPG: Action-constrained Policy Gradient with Normalizing Flows

Janaka Chathuranga Brahmanage, Jiajing Ling, Akshat Kumar

https://openreview.net/forum?id=p1gzxzJ4Y5

Keywords: action-constrained reinforcement learning, decision making

Compressor summary: ACRL uses a normalizing flow model to learn an invertible mapping between feasible actions and latent variables, enabling faster and more accurate action sampling and constraint satisfaction in DDPG-based RL problems.


Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training

Yefan Zhou, Tianyu Pang, Keqin Liu, charles h martin, Michael W. Mahoney, Yaoqing Yang

https://openreview.net/forum?id=oyV9FslE3j

Keywords: Heavy-tail self-regularization, learning rate schedule

Compressor summary: The paper introduces TempBalance, a layer-wise learning rate method for neural network training, which uses Heavy-Tailed Self-Regularization Theory to balance the temperature across layers and improve performance on various datasets.


Language-based Action Concept Spaces Improve Video Self-Supervised Learning

Kanchana Ranasinghe, Michael S Ryoo

https://openreview.net/forum?id=oyFyOPZUCs

Keywords: self-supervised learning for videos, zero-shot action recognition

Compressor summary: The paper proposes a method to adapt image CLIP models to the video domain using language tied self-supervised learning, which improves action recognition performance on benchmarks.


MIMONets: Multiple-Input-Multiple-Output Neural Networks Exploiting Computation in Superposition

Nicolas Menet, Michael Hersche, Geethan Karunaratne, Luca Benini, Abu Sebastian, Abbas Rahimi

https://openreview.net/forum?id=ox7aynitoW

Keywords: Computation in superposition, Vector-symbolic architectures, Convolutional neural networks, Transformers

Compressor summary: The authors propose MIMONets, which can process multiple inputs at once using variable binding and unbinding. This enables computation in superposition, improving inference efficiency and performance-accuracy trade-offs for CNN and Transformer models. They also provide theoretical bounds on the interference between superposition channels.


On the Properties of Kullback-Leibler Divergence Between Multivariate Gaussian Distributions

Yufeng Zhang, Jialu Pan, Kenli Li, Wanwei Liu, Zhenbang Chen, Xinwang Liu, J Wang

https://openreview.net/forum?id=ouLe91yibj

Keywords: Kullback-Leibler divergence, statistical divergence, multivariate Gaussian distribution, mathematical optimization, Lambert $W$ function, machine learning, flow-based model, reinforcement learning

Compressor summary: The paper studies properties of KL divergence between multivariate Gaussian distributions, deriving bounds for different scenarios and showing that it follows a relaxed triangle inequality.


Linear Time Algorithms for k-means with Multi-Swap Local Search

Junyu Huang, Qilong Feng, Ziyun Huang, Jinhui Xu, Jianxin Wang

https://openreview.net/forum?id=oss2jXD1Zs

Keywords: Approximation Algorithms, k-means Clustering, Local Search

Compressor summary: The paper proposes a multi-swap local search algorithm for the $k$-means problem with linear running time and improved approximation ratio, and also introduces a sampling-based method and a recombination mechanism to enhance the performance.


Model and Feature Diversity for Bayesian Neural Networks in Mutual Learning

Cuong Pham, Cuong C. Nguyen, Trung Le, Dinh Phung, Gustavo Carneiro, Thanh-Toan Do

https://openreview.net/forum?id=os2BdbiGwX

Keywords: Bayesian Neural Networks, Deep Mutual Learning

Compressor summary: The paper proposes a deep mutual learning approach that increases diversity in parameter and feature distributions of Bayesian Neural Networks, leading to significant performance improvements in classification accuracy, negative log-likelihood, and expected calibration error.


Bypassing the Simulator: Near-Optimal Adversarial Linear Contextual Bandits

Haolin Liu, Chen-Yu Wei, Julian Zimmert

https://openreview.net/forum?id=orh4e0AO9R

Keywords: adversarial linear contextual bandits, log-determinant barrier

Compressor summary: The paper proposes a new method for solving the adversarial linear contextual bandit problem that achieves lower regret and maintains computational efficiency in the presence of stochastic arm availability and additive misspecification errors.


Sample-Conditioned Hypothesis Stability Sharpens Information-Theoretic Generalization Bounds

Ziqiao Wang, Yongyi Mao

https://openreview.net/forum?id=oqDSDKLd3S

Keywords: generalization, information-theoretic bounds, stability

Compressor summary: The authors propose a new way to measure generalization in machine learning using a matrix and a stability concept that improve existing bounds, especially for stochastic convex optimization problems.


Explore In-Context Learning for 3D Point Cloud Understanding

Zhongbin Fang, Xiangtai Li, Xia Li, Joachim M. Buhmann, Chen Change Loy, Mengyuan Liu

https://openreview.net/forum?id=ooXpTZYwXa

Keywords: In-context learning, Point cloud, Prompt tuning

Compressor summary: The paper introduces Point-In-Context, a novel framework for in-context learning in 3D point clouds, which addresses technical challenges and outperforms individual models.


Multi-Fidelity Multi-Armed Bandits Revisited

Xuchuang Wang, Qingyun Wu, Wei Chen, John C.S. Lui

https://openreview.net/forum?id=oi45JlpSOT

Keywords: Multi-fidelity, multi-armed bandits

Compressor summary: The paper investigates the multi-fidelity multi-armed bandit problem, where each arm has different costs and observation accuracies, and proposes algorithms for best arm identification and regret minimization.


Successor-Predecessor Intrinsic Exploration

Changmin Yu, Neil Burgess, Maneesh Sahani, Samuel Gershman

https://openreview.net/forum?id=ohKbQp0jIY

Keywords: Exploration, reinforcement learning

Compressor summary: The paper introduces SPIE, an exploration algorithm for reinforcement learning that combines prospective and retrospective information to generate more efficient and ethologically plausible behavior in environments with sparse rewards and bottleneck states.


Disentangling Cognitive Diagnosis with Limited Exercise Labels

Xiangzhi Chen, Le Wu, Fei Liu, Lei Chen, Kun Zhang, Richang Hong, Meng Wang

https://openreview.net/forum?id=ogPBujRhiN

Keywords: Intelligent Education System, Cognitive Diagnosis, Disentangled Representation Learning, Interpretability

Compressor summary: The paper proposes a new method called Disentanglement based Cognitive Diagnosis (DCD) for measuring students' proficiency in specific knowledge concepts with limited exercise labels, using response records and two novel modules.


Learning Efficient Coding of Natural Images with Maximum Manifold Capacity Representations

Thomas Edward Yerxa, Yilun Kuang, Eero P Simoncelli, SueYeon Chung

https://openreview.net/forum?id=og9V7NgOrQ

Keywords: computational neuroscience, theoretical neuroscience, efficient coding, representation geometry, neural manifolds, self-supervised learning, statistical physics of learning

Compressor summary: The efficient coding hypothesis suggests sensory systems adapt to their inputs for maximal information capture, but measuring this is difficult; researchers have developed Maximum Manifold Capacity Representations (MMCR) which use a novel efficiency metric and are competitive with self-supervised learning methods.


Online (Multinomial) Logistic Bandit: Improved Regret and Constant Computation Cost

Yu-Jie Zhang, Masashi Sugiyama

https://openreview.net/forum?id=ofa1U5BJVJ

Keywords: Logistic Bandit, Generalized Linear Bandit, Regret Bound, Computation Cost

Compressor summary: The paper proposes a logistic bandit algorithm that is statistically efficient, computationally fast, and adaptable to complex decision-making problems, improving on existing binary and multinomial methods.


Demystifying Structural Disparity in Graph Neural Networks: Can One Size Fit All?

Haitao Mao, Zhikai Chen, Wei Jin, Haoyu Han, Yao Ma, Tong Zhao, Neil Shah, Jiliang Tang

https://openreview.net/forum?id=oef30oScVB

Keywords: Graph Neural Network

Compressor summary: The study shows that Graph Neural Networks perform well on nodes with similar structures in homophilic graphs and heterophilic nodes in heterophilic graphs, but struggle with nodes having different structures in the opposite graph type.


FedGCN: Convergence-Communication Tradeoffs in Federated Training of Graph Convolutional Networks

Yuhang Yao, Weizhao Jin, Srivatsan Ravi, Carlee Joe-Wong

https://openreview.net/forum?id=ody3RBUuJS

Keywords: Federated Graph Learning;

Compressor summary: FedGCN is a novel algorithm that trains GCN models on graphs distributed among clients using federated learning, achieving fast convergence, high accuracy, and low communication.


SQ Lower Bounds for Learning Mixtures of Linear Classifiers

Ilias Diakonikolas, Daniel Kane, Yuxin Sun

https://openreview.net/forum?id=obCNIzeSrg

Keywords: mixtures models, linear classifier, Statistical Query model, spherical designs

Compressor summary: The paper studies how to learn mixtures of linear classifiers with Gaussian features and shows that current algorithms are nearly optimal using a new construction of spherical designs.


FGPrompt: Fine-grained Goal Prompting for Image-goal Navigation

Xinyu Sun, Peihao Chen, Jugang Fan, Jian Chen, Thomas H. Li, Mingkui Tan

https://openreview.net/forum?id=oaJEB5Qcia

Keywords: Visual Navigation, Image-Goal Navigation, Embodied AI

Compressor summary: The paper proposes a Fine-grained Goal Prompting method to improve image-goal navigation for autonomous systems by using high-resolution feature maps in the goal image as prompts.


TFLEX: Temporal Feature-Logic Embedding Framework for Complex Reasoning over Temporal Knowledge Graph

Xueyuan Lin, Haihong E, Chengjin Xu, Gengxian Zhou, Haoran Luo, Tianyi Hu, Fenglong Su, Ningyuan Li, Mingzhi Sun

https://openreview.net/forum?id=oaGdsgB18L

Keywords: Temporal Knowledge Graph Reasoning, Temporal Knowledge Graph Embedding, Temporal Knowledge Graph, Temporal Logic, Knowledge Graph Reasoning, Knowledge Graph Embedding, Knowledge Graph, Machine Learning

Compressor summary: TFLEX is a novel framework for multi-hop logical reasoning over temporal knowledge graphs that uses fuzzy logic to handle complex queries involving entities, timestamps, and set operations.


Follow-ups Also Matter: Improving Contextual Bandits via Post-serving Contexts

Chaoqi Wang, Ziyu Ye, Zhe Feng, Ashwinkumar Badanidiyuru, Haifeng Xu

https://openreview.net/forum?id=oaCDiKoJ2w

Keywords: linear stochastic bandits, online learning, partial information, contextual bandits

Compressor summary: The paragraph discusses a novel contextual bandit problem with post-serving contexts, where additional valuable information about user rewards can be observed after arm selection, and presents poLinUCB, a new algorithm that achieves tight regret under standard assumptions using a robustified version of the Elliptical Potential Lemma.


Learning Cuts via Enumeration Oracles

Daniel Thuerck, Boro Sofranac, Marc Pfetsch, Sebastian Pokutta

https://openreview.net/forum?id=oU4QHdcIWW

Keywords: Integer Programming, Cutting Planes, Optimization

Compressor summary: The paper presents a novel method for learning facets of polyhedra by embedding an enumeration oracle in a Frank-Wolfe algorithm to generate strong cutting planes.


Softmax Output Approximation for Activation Memory-Efficient Training of Attention-based Networks

Changhyeon Lee, Seulki Lee

https://openreview.net/forum?id=oScaeIibRx

Keywords: Memory efficient, Activation saving memory, NLP, Transformer

Compressor summary: The paper proposes a method to reduce memory usage in attention-based networks by approximating the softmax output during training and shows that it improves performance with less memory.


Neural Harmonics: Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning

Marina Munkhoeva, Ivan Oseledets

https://openreview.net/forum?id=oSYjkJKHZx

Keywords: unsupervised learning, self-supervised learning, representation learning, matrix completion

Compressor summary: The paper explores how self-supervised learning methods use a Laplace operator and low-rank matrix completion to learn representations without labels, and analyzes their convergence and performance.


Analyzing the Sample Complexity of Self-Supervised Image Reconstruction Methods

Tobit Klug, Dogukan Atik, Reinhard Heckel

https://openreview.net/forum?id=oRn953uhFq

Keywords: image reconstruction, denoising, accelerated MRI, self-supervised, sample complexity

Compressor summary: The paper compares the performance and sample complexity of supervised and self-supervised methods for image reconstruction tasks, finding that self-supervised methods require more training data but eventually achieve similar results.


Activity Grammars for Temporal Action Segmentation

Dayoung Gong, Joonseok Lee, Deunsol Jung, Suha Kwak, Minsu Cho

https://openreview.net/forum?id=oOXZ5JEjPb

Keywords: neuro-symbolic approach, Temporal action segmentation, grammar

Compressor summary: The paper introduces KARI, a grammar induction algorithm for temporal action segmentation, and BEP, a parser that transforms frame-level probabilities into actions according to the induced grammar.


Neural Fields with Hard Constraints of Arbitrary Differential Order

Fangcheng Zhong, Kyle Thomas Fogarty, Param Hanji, Tianhao Walter Wu, Alejandro Sztrajman, Andrew Everett Spielberg, Andrea Tagliasacchi, Petra Bosilj, Cengiz Oztireli

https://openreview.net/forum?id=oO1IreC6Sd

Keywords: neural fields, constrained optimization

Compressor summary: The authors present Constrained Neural Fields (CNF), a method for imposing hard constraints on neural networks during optimization, inspired by meshless interpolation and spectral collocation techniques in scientific computing.


Particle-based Variational Inference with Generalized Wasserstein Gradient Flow

Ziheng Cheng, Shiyue Zhang, Longlin Yu, Cheng Zhang

https://openreview.net/forum?id=oNuam8eFz2

Keywords: Particle-based VI, generalized Wasserstein gradient flow

Compressor summary: The paper introduces GWG, a novel ParVI framework that improves on ParVIs by using a generalized Wasserstein gradient flow of the KL divergence, which allows for more flexible kernel design and stronger convergence guarantees.


Unbiased constrained sampling with Self-Concordant Barrier Hamiltonian Monte Carlo

Maxence Noble, Valentin De Bortoli, Alain Durmus

https://openreview.net/forum?id=oMm1dfo3tK

Keywords: Hamiltonian Monte Carlo, Riemannian manifold, self-concordant barrier, constrained sampling

Compressor summary: The paper introduces Barrier Hamiltonian Monte Carlo, a method for sampling from constrained Gibbs distributions using Hamiltonian dynamics and a new filter step, which overcomes bias issues in existing Riemannian HMC methods.


When Demonstrations meet Generative World Models: A Maximum Likelihood Framework for Offline Inverse Reinforcement Learning

Siliang Zeng, Chenliang Li, Alfredo Garcia, Mingyi Hong

https://openreview.net/forum?id=oML3v2cFg2

Keywords: Inverse Reinforcement Learning, Model-based Offline Inverse Reinforcement Learning

Compressor summary: The paragraph introduces a bi-level optimization approach to improve offline inverse reinforcement learning by incorporating uncertainty in the expert's model of the world.


Worst-case Performance of Popular Approximate Nearest Neighbor Search Implementations: Guarantees and Limitations

Piotr Indyk, Haike Xu

https://openreview.net/forum?id=oKqaWlEfjY

Keywords: Nearest neighbor search; graph-based algorithms; worst-case analysis

Compressor summary: The paper analyzes the theoretical guarantees of graph-based approximate nearest neighbor search algorithms, finding that some variants have slow preprocessing but good performance, while others have fast preprocessing but high query time.


UniT: A Unified Look at Certified Robust Training against Text Adversarial Perturbation

Muchao Ye, Ziyi Yin, Tianrong Zhang, Tianyu Du, Jinghui Chen, Ting Wang, Fenglong Ma

https://openreview.net/forum?id=oGxE2Nvlda

Keywords: certified robust training, text adversarial defense

Compressor summary: The paragraph describes a new framework called UniT that unifies and improves existing certified robust training pipelines for text classification by working in the word embedding space and using decoupled regularization loss to enhance the base model's robustness.


Generate What You Prefer: Reshaping Sequential Recommendation via Guided Diffusion

Zhengyi Yang, Jiancan Wu, Zhicai Wang, Xiang Wang, Yancheng Yuan, Xiangnan He

https://openreview.net/forum?id=oFpBnt6bgC

Keywords: Sequential Recommendation, Recommendation System, Generative Model, Diffusion Model

Compressor summary: The text describes a new sequential recommendation method, DreamRec, that generates an oracle item based on the user's historical interactions, rather than relying on negative sampling and classification.


Gradient Informed Proximal Policy Optimization

Sanghyun Son, Laura Yu Zheng, Ryan Sullivan, Yi-Ling Qiao, Ming Lin

https://openreview.net/forum?id=oFaLc6fHSt

Keywords: Reinforcement Learning, Analytic Gradient-Based Policy Learning, Proximal Policy Optimization, Differentiable Programming

Compressor summary: The paper introduces a new policy learning method that combines analytical gradients with PPO, using an α-policy to adjust the influence of these gradients and improve performance in different environments.


Directional diffusion models for graph representation learning

Run Yang, Yuling Yang, Fan Zhou, Qiang Sun

https://openreview.net/forum?id=oDtyJt5JLk

Keywords: diffusion models, graph representation learning, unsupervised learning

Compressor summary: The paper proposes directional diffusion models for unsupervised graph representation learning, which use data-dependent, anisotropic, and directional noises to better capture graph structures and outperform existing methods on 12 datasets.


Unified Off-Policy Learning to Rank: a Reinforcement Learning Perspective

Zeyu Zhang, Yi Su, Hui Yuan, Yiran Wu, Rishab Balasubramanian, Qingyun Wu, Huazheng Wang, Mengdi Wang

https://openreview.net/forum?id=oDcWnfZyZW

Keywords: learning to rank, off-policy learning, reinforcement learning, click model

Compressor summary: The paper proposes a method (CUOLR) for off-policy Learning to Rank that adapts to various click models using offline reinforcement learning without complex debiasing techniques or prior knowledge.


Distributional Policy Evaluation: a Maximum Entropy approach to Representation Learning

Riccardo Zamboni, Alberto Maria Metelli, Marcello Restelli

https://openreview.net/forum?id=o91in9tDEs

Keywords: reinforcement learning, distributional reinforcement learning, maximum entropy estimation, representation learning

Compressor summary: The paper proposes a new Max-Ent framework for evaluating policies in distributional RL, which can consider the state representation complexity and guide state space learning using progressive factorization.


Recovering Unbalanced Communities in the Stochastic Block Model with Application to Clustering with a Faulty Oracle

Chandra Sekhar Mukherjee, Pan Peng, Jiapeng Zhang

https://openreview.net/forum?id=o7W0Zet6p3

Keywords: SBM, Unbalanced SBM, Spectral algorithms, Small cluster barrier

Compressor summary: The paper proposes an SVD-based algorithm to detect communities of varying sizes in the stochastic block model and improves upon previous results, also providing an efficient clustering algorithm with sublinear query complexity.


Regression with Cost-based Rejection

Xin Cheng, Yuzhou Cao, Haobo Wang, Hongxin Wei, Bo An, Lei Feng

https://openreview.net/forum?id=o7HckkxOZH

Keywords: regression, rejection costs, surrogate loss

Compressor summary: The paper proposes a regression problem with cost-based rejection, where models reject examples based on their variance and mean squared error, and shows that this approach can improve prediction performance.


Label-Retrieval-Augmented Diffusion Models for Learning from Noisy Labels

Jian Chen, Ruiyi Zhang, Tong Yu, Rohan Sharma, zhiqiang xu, Tong Sun, Changyou Chen

https://openreview.net/forum?id=o778eWSr1S

Keywords: diffusion model, label noise, retrieval augmented learning

Compressor summary: The paper proposes a new diffusion model that uses generative uncertainty and neighbor consistency to learn from noisy labels and achieve state-of-the-art results on real-world datasets.


Recursion in Recursion: Two-Level Nested Recursion for Length Generalization with Scalability

Jishnu Ray Chowdhury, Cornelia Caragea

https://openreview.net/forum?id=o6yTKfdnbA

Keywords: Recursive Neural Networks, Long Range Arena, RvNN, Long Range Sequence Modeling, Length Generalization, LRA, Structured Encoding, Inductive Bias, Hierarchical Model, Recursive Models

Compressor summary: RIR combines a balanced tree model with a Beam Tree RvNN inner recursion to achieve length-generalization on ListOps and scalability on long sequence tasks like LRA, outperforming Transformers.


Arbitrarily Scalable Environment Generators via Neural Cellular Automata

Yulun Zhang, Matthew Christopher Fontaine, Varun Bhatt, Stefanos Nikolaidis, Jiaoyang Li

https://openreview.net/forum?id=o6Dnt1uEyZ

Keywords: Multi-robot systems, quality diversity, automatic environment generation, neural cellular automata

Compressor summary: The paper proposes optimizing Neural Cellular Automata (NCA) environment generators via Quality Diversity (QD) algorithms to create arbitrarily large environments for multi-robot systems, improving their scalability and throughput.


Certifiably Robust Graph Contrastive Learning

Minhua Lin, Teng Xiao, Enyan Dai, Xiang Zhang, Suhang Wang

https://openreview.net/forum?id=o50nH0sV9x

Keywords: Certifiable Robustness, Graph Contrastive Learning

Compressor summary: The paper proposes a unified criterion and a novel technique called RES to certify and enhance the robustness of Graph Contrastive Learning models against adversarial attacks, which can be proven to be effective in downstream tasks.


Causal de Finetti: On the Identification of Invariant Causal Structure in Exchangeable Data

Siyuan Guo, Viktor Tóth, Bernhard Schölkopf, Ferenc Huszár

https://openreview.net/forum?id=o4RtDFMSNL

Keywords: Independent Causal Mechanism, Causal Discovery, Exchangeable, Bayesian Statistics

Compressor summary: The text discusses how exchangeable data can help constraint-based causal discovery methods identify more accurate causal structures than i.i.d. data by using independent causal mechanism generative processes.


Identifiability Guarantees for Causal Disentanglement from Soft Interventions

Jiaqi Zhang, Kristjan Greenewald, Chandler Squires, Akash Srivastava, Karthikeyan Shanmugam, Caroline Uhler

https://openreview.net/forum?id=o16sYKHk3S

Keywords: Causality, Identifiability, Disentanglement

Compressor summary: The paper proposes a method for causal disentanglement using latent variables, which can identify the causal model from unpaired observational and interventional data even when some causal variables are unobserved, and demonstrates its application to predicting combinatorial perturbation effects in genomics.


Active Observing in Continuous-time Control

Samuel Holt, Alihan Hüyük, Mihaela van der Schaar

https://openreview.net/forum?id=o0ggjFD24U

Keywords: Sensing, Model-Based Reinforcement Learning

Compressor summary: The paper introduces a new continuous-time control problem with costly observations, where irregular observation policies can improve expected utility, and presents a simple initial method to solve it.


Individual Arbitrariness and Group Fairness

Carol Xuan Long, Hsiang Hsu, Wael Alghamdi, Flavio Calmon

https://openreview.net/forum?id=nzkWhoXUpv

Keywords: predictive multiplicity, fairness in machine learning, Rashomon effect

Compressor summary: The paper discusses how group fairness and accuracy optimizations can increase predictive multiplicity, a phenomenon where similar models produce different outputs for individual samples, and proposes an ensemble algorithm to improve consistency in individual-level decision-making.


Variational Gaussian Processes with Decoupled Conditionals

Xinran Zhu, Kaiwen Wu, Natalie Maus, Jacob R. Gardner, David Bindel

https://openreview.net/forum?id=nwK8UkK3uB

Keywords: Gaussian processes, variational inference, variational Gaussian processes, Bayesian optimization

Compressor summary: The paper proposes a new way to improve variational Gaussian processes by making the predictive mean and covariance more flexible, leading to better performance on regression tasks and Bayesian optimization.


State-Action Similarity-Based Representations for Off-Policy Evaluation

Brahma S Pavse, Josiah P. Hanna

https://openreview.net/forum?id=nvX3MiQM0G

Keywords: reinforcement learning, off-policy evaluation, off-policy RL, representation learning, behavioral similarity metrics

Compressor summary: The paper proposes a new method to improve off-policy evaluation in reinforcement learning by using a learned encoder to transform the dataset before applying fitted q-evaluation, and introduces an OPE-specific state-action similarity metric to learn the encoder.


A Scale-Invariant Sorting Criterion to Find a Causal Order in Additive Noise Models

Alexander Gilbert Reisach, Myriam Tami, Christof Seiler, Antoine Chambaz, Sebastian Weichwald

https://openreview.net/forum?id=nrbR2F29vU

Keywords: Causal Discovery, Directed Acyclic Graph, Varsortability, Additive Noise Model, Structural Causal Model, Simulation, Benchmark

Compressor summary: The paper studies how different parameter choices in additive noise models affect causal discovery and introduces a new method called $R^2$-SortnRegress that leverages high $R^2$ values to find causal orders.


Leave No Stone Unturned: Mine Extra Knowledge for Imbalanced Facial Expression Recognition

Yuhang Zhang, Yaqi Li, lixiong Qin, Xuannan Liu, Weihong Deng

https://openreview.net/forum?id=nrQif5tH7O

Keywords: Facial expression recognition, imbalanced learning

Compressor summary: The paper proposes a novel method for facial expression recognition that extracts extra knowledge from both major and minor class samples using re-balanced attention maps and smooth labels to handle the imbalance problem.


CLIP4HOI: Towards Adapting CLIP for Practical Zero-Shot HOI Detection

Yunyao Mao, Jiajun Deng, Wengang Zhou, Li Li, Yao Fang, Houqiang Li

https://openreview.net/forum?id=nqIIWnwe73

Keywords: human-object interaction detection, zero-shot learning, CLIP model adaptatiion

Compressor summary: CLIP4HOI is a novel framework for zero-shot human-object interaction detection that uses CLIP to independently identify humans and objects and adapts the model into a fine-grained classifier for proposal discrimination, improving transferability and performance on rare and unseen categories.


Truncated Affinity Maximization: One-class Homophily Modeling for Graph Anomaly Detection

Hezhe Qiao, Guansong Pang

https://openreview.net/forum?id=nq4OhifyEe

Keywords: Anomaly Detection, Graph Neural Network, Graph Anomaly Detection, One-Class Homophily, Local Node Affinity

Compressor summary: The paper introduces a novel anomaly scoring method for graph anomaly detection, based on local node affinity, and a tailored representation learning approach called Truncated Affinity Maximization (TAM) that leverages the one-class homophily phenomenon.


Marginal Density Ratio for Off-Policy Evaluation in Contextual Bandits

Muhammad Faaiz Taufiq, Arnaud Doucet, Rob Cornish, Jean-Francois Ton

https://openreview.net/forum?id=noyleECBam

Keywords: contextual bandits, variance reduction, off-policy evaluation

Compressor summary: The Marginal Ratio (MR) estimator is a new off-policy evaluation method for contextual bandits that reduces variance compared to existing methods like Inverse Probability Weighting and Doubly Robust, and performs well in causal inference settings.


Joint Feature and Differentiable $ k $-NN Graph Learning using Dirichlet Energy

Lei Xu, Lei Chen, Rong Wang, Feiping Nie, Xuelong Li

https://openreview.net/forum?id=noMktb4ait

Keywords: Feature Selection, Differential k-NN Graph, Dirichlet Energy

Compressor summary: The paper proposes a deep feature selection method that uses differentiable k-NN graph learning based on Dirichlet Energy to identify important features and learn dynamic graphs, and applies Optimal Transport theory to address non-differentiability issues.


Practical Sharpness-Aware Minimization Cannot Converge All the Way to Optima

Dongkuk Si, Chulhee Yun

https://openreview.net/forum?id=nijJN0LHqM

Keywords: Sharpness-Aware Minimization, convex optimization

Compressor summary: The paper studies how varying the parameters in an optimizer called SAM affects its convergence properties on smooth and non-smooth functions.


Beyond Pretrained Features: Noisy Image Modeling Provides Adversarial Defense

Zunzhi You, Daochang Liu, Bohyung Han, Chang Xu

https://openreview.net/forum?id=niHkj9ixUZ

Keywords: self-supervised learning, adversarial robustness

Compressor summary: The paper proposes De^3, a method that uses noisy image modeling to improve adversarial robustness of deep neural networks by exploiting the pretrained decoder for denoising and sampling the noise scale hyperparameter from random distributions.


Post-processing Private Synthetic Data for Improving Utility on Selected Measures

Hao Wang, Shivchander Sudalairaj, John Henning, Kristjan Greenewald, Akash Srivastava

https://openreview.net/forum?id=neu9JlNweE

Keywords: differential privacy, synthetic data

Compressor summary: The paper introduces a post-processing technique to improve the utility of synthetic data for specific end-user requirements while maintaining privacy and quality, by resampling and optimizing weights using efficient algorithms.


Symbolic Discovery of Optimization Algorithms

Xiangning Chen, Chen Liang, Da Huang, Esteban Real, Kaiyuan Wang, Hieu Pham, Xuanyi Dong, Thang Luong, Cho-Jui Hsieh, Yifeng Lu, Quoc V Le

https://openreview.net/forum?id=ne6zeqLFCZ

Keywords: AutoML

Compressor summary: The authors propose a method to discover optimization algorithms for deep neural network training and introduce Lion, an efficient and effective optimizer that performs well on various tasks.


Learning Curves for Deep Structured Gaussian Feature Models

Jacob A Zavatone-Veth, Cengiz Pehlevan

https://openreview.net/forum?id=nbG6zfJtIe

Keywords: random feature models, generalization, deep networks, ridge regression

Compressor summary: The study examines how weight anisotropy and structured Gaussian features impact generalization in deep learning models using the replica trick from statistical physics.


Generating Behaviorally Diverse Policies with Latent Diffusion Models

Shashank Hegde, Sumeet Batra, K.R. Zentner, Gaurav S. Sukhatme

https://openreview.net/forum?id=nafgeYknRT

Keywords: Latent Diffusion, Quality Diversity, Reinforcement Learning, Graph Neural Networks

Compressor summary: The paper proposes using diffusion models to compress a collection of diverse policies into one generative model that can select and sequence behaviors with language.


Uncertainty Quantification via Neural Posterior Principal Components

Elias Nehme, Omer Yair, Tomer Michaeli

https://openreview.net/forum?id=nZ0jnXizyR

Keywords: Uncertainty Quantification, Inverse Problems, Probabilistic Modelling, Principal Components Analysis, Deep Learning

Compressor summary: This paper proposes a fast method to predict the principal components of the posterior distribution for image restoration models, enabling better uncertainty quantification and visualization.


Regret Matching+: (In)Stability and Fast Convergence in Games

Gabriele Farina, Julien Grand-Clément, Christian Kroer, Chung-Wei Lee, Haipeng Luo

https://openreview.net/forum?id=nYgs0qZJ97

Keywords: Regret Matching, Predictive algorithms, Extensive-Form Games

Compressor summary: The paper investigates the instability of regret matching algorithms, proposes fixes, and demonstrates their effectiveness in various game scenarios using theory and experiments.


Mitigating Source Bias for Fairer Weak Supervision

Changho Shin, Sonia Cromp, Dyah Adila, Frederic Sala

https://openreview.net/forum?id=nXPqMyWUnx

Keywords: Weak supervision, fairness

Compressor summary: Weak supervision can produce biased pseudolabels, and a fairness-based technique can improve accuracy and reduce bias in weak supervision without tradeoffs.


Aggregating Capacity in FL through Successive Layer Training for Computationally-Constrained Devices

Kilian Pfeiffer, Ramin Khalili, Joerg Henkel

https://openreview.net/forum?id=nXNsqB4Yr1

Keywords: Federated Learning, Memory, Resource Constraints

Compressor summary: The paper proposes a new method for federated learning on edge devices that reduces resource requirements and improves model accuracy by enabling successive freezing and training of model parameters.


FedGame: A Game-Theoretic Defense against Backdoor Attacks in Federated Learning

Jinyuan Jia, Zhuowen Yuan, Dinuka Sahabandu, Luyao Niu, Arezoo Rajabi, Bhaskar Ramasubramanian, Bo Li, Radha Poovendran

https://openreview.net/forum?id=nX0zYBGEka

Keywords: backdoor defense, federated learning, game theory

Compressor summary: Federated learning faces backdoor attacks from dynamic attackers, and the paper proposes FedGame, a minimax game-based interactive defense mechanism that enhances robustness against such attacks.


RegBN: Batch Normalization of Multimodal Data with Regularization

MORTEZA GHAHREMANI, Christian Wachinger

https://openreview.net/forum?id=nUbdkXqC8R

Keywords: Multimodal Data, Multimodality, Batch Normalization, Heterogeneous data, Regularization, Confounder, Confounding Effect Removal, Data Dependency

Compressor summary: The paper introduces RegBN, a novel approach for multimodal Batch Normalization with REGularization, which uses the Frobenius norm as a regularizer term to address confounders and dependencies among different data sources, simplifying training and inference, and improving performance across diverse modalities and architectures.


Robust Matrix Sensing in the Semi-Random Model

Xing Gao, Yu Cheng

https://openreview.net/forum?id=nSr2epejn2

Keywords: Matrix sensing, Optimization, Low-rank matrix recovery, Semi-random, Adversarial input, Robustness

Compressor summary: The paper proposes a new non-convex optimization algorithm for semi-random matrix sensing that avoids bad local optima by reweighting the input matrices based on the current solution and takes weighted gradient steps.


Shape Non-rigid Kinematics (SNK): A Zero-Shot Method for Non-Rigid Shape Matching via Unsupervised Functional Map Regularized Reconstruction

Souhaib Attaiki, Maks Ovsjanikov

https://openreview.net/forum?id=nSgMh5v5Ne

Keywords: shape matching

Compressor summary: Shape Non-rigid Kinematics (SNK) is a new method for matching shapes that requires no training or ground truth data, and uses an encoder-decoder architecture to deform the source shape to match the target shape.


Differentiable Clustering with Perturbed Spanning Forests

Lawrence Stewart, Francis Bach, Felipe Llinares-López, Quentin Berthet

https://openreview.net/forum?id=nRfcVBsF9n

Keywords: Structured learning, Clustering, Differentiable, weakly supervised, semi-supervised, representation learning

Compressor summary: A new method for clustering data that can be trained end-to-end with efficient gradients and works well in noisy and complex settings is presented, along with a custom loss function for learning from partial clustering data.


Interpretability at Scale: Identifying Causal Mechanisms in Alpaca

Zhengxuan Wu, Atticus Geiger, Thomas Icard, Christopher Potts, Noah Goodman

https://openreview.net/forum?id=nRfClnMhVX

Keywords: Mechanistic Interpretability

Compressor summary: The paper introduces Boundless DAS, an improved method to find interpretable causal structures in large language models, and applies it to the Alpaca model, revealing its use of two boolean variables for numerical reasoning.


Multiclass Boosting: Simple and Intuitive Weak Learning Criteria

Nataly Brukhim, Amit Daniely, Yishay Mansour, Shay Moran

https://openreview.net/forum?id=nQ84YY9Iut

Keywords: Boosting, Multiclass classification, PAC Learning, List PAC Learning

Compressor summary: The paper presents a new boosting method for multiclass classification that generalizes weak learnability, has low sample complexity, and simplifies theoretical applications in list learning.


Eliminating Domain Bias for Federated Learning in Representation Space

Jianqing Zhang, Yang Hua, Jian Cao, Hao Wang, Tao Song, Zhengui XUE, Ruhui Ma, Haibing Guan

https://openreview.net/forum?id=nO5i1XdUS0

Keywords: Federated Learning, Personalized Federated Learning, Representation, Knowledge Transfer

Compressor summary: The paper introduces a framework called Domain Bias Eliminator (DBE) to address representation degeneration in federated learning and improve generalization and personalization abilities.


Memory Efficient Optimizers with 4-bit States

Bingrui Li, Jianfei Chen, Jun Zhu

https://openreview.net/forum?id=nN8TnHB5nw

Keywords: memory efficiency, optimizer, Adam, quantization

Compressor summary: The paper proposes 4-bit quantization of optimizer states for neural networks, achieving memory efficiency without sacrificing accuracy on various tasks.


GPT-ST: Generative Pre-Training of Spatio-Temporal Graph Neural Networks

Zhonghang Li, Lianghao Xia, Yong Xu, Chao Huang

https://openreview.net/forum?id=nMH5cUaSj8

Keywords: Spatial Temporal Prediction, Deep Neural Networks, Pre-training Model

Compressor summary: The authors propose a spatio-temporal pre-training framework that enhances the performance of downstream models for traffic management and travel planning by learning customized representations and adaptive mask strategies.


Provably Efficient Algorithm for Nonstationary Low-Rank MDPs

Yuan Cheng, Jing Yang, Yingbin Liang

https://openreview.net/forum?id=nMB41QjLDY

Keywords: Reinforcement Learning, Nonstationary Environment, Representation Learning, Policy Optimization, Statistical Complexity

Compressor summary: The paper proposes two algorithms, PORTAL and Ada-PORTAL, for reinforcement learning in nonstationary low-rank Markov Decision Processes, with theoretical guarantees on their sample efficiency and suboptimality gap.


A Framework for Fast and Stable Representations of Multiparameter Persistent Homology Decompositions

David Loiseaux, Mathieu Carrière, Andrew Blumberg

https://openreview.net/forum?id=nKCUDd9GYu

Keywords: Topological Data Analysis, Multiparameter Persistent Homology, Kernel Methods, Convergence Rate, Statistical Learning

Compressor summary: The article proposes a new representation framework for multiparameter persistent homology in topological data analysis, which is stable, efficient, and informative, and shows its effectiveness on geometric and point cloud data.


Evaluating Robustness and Uncertainty of Graph Models Under Structural Distributional Shifts

Gleb Bazhenov, Denis Kuznedelev, Andrey Malinin, Artem Babenko, Liudmila Prokhorenkova

https://openreview.net/forum?id=nJFJcgjnGo

Keywords: graph, distributional shift, structural shift, uncertainty, robustness, graph neural networks

Compressor summary: The authors propose a method to create diverse distributional shifts in graph learning based on node structure, evaluate its difficulty, and show that simple models often perform better than complex ones.


PromptRestorer: A Prompting Image Restoration Method with Degradation Perception

Cong Wang, Jinshan Pan, Wei Wang, Jiangxin Dong, Mengzhu Wang, Yakun Ju, Junyang Chen

https://openreview.net/forum?id=nIaNgaQvsV

Keywords: Degradation Vanishing, Prompting Learning, Image Restoration

Compressor summary: PromptRestorer is a novel image restoration model that uses degradation features to guide the restoration process and achieves state-of-the-art results on 4 tasks.


$H$-Consistency Bounds: Characterization and Extensions

Anqi Mao, Mehryar Mohri, Yutao Zhong

https://openreview.net/forum?id=nI7EmXq2PL

Keywords: consistency, H-consistency, characterization, learning theory

Compressor summary: The paper introduces general characterizations and extensions of $H$-consistency bounds for surrogate losses in multi-class classification, covering constrained and comp-sum losses, and providing tighter bounds than previous studies.


What Truly Matters in Trajectory Prediction for Autonomous Driving?

Tran Phong, Haoran Wu, Cunjun Yu, Panpan Cai, Sifa Zheng, David Hsu

https://openreview.net/forum?id=nG35q8pNL9

Keywords: trajectory prediction; autonomous driving

Compressor summary: The paper studies the dynamics gap between trajectory prediction accuracy on fixed datasets and real-world driving performance, and emphasizes the trade-off between computational efficiency and prediction accuracy for autonomous driving systems.


Regret-Optimal Model-Free Reinforcement Learning for Discounted MDPs with Short Burn-In Time

Xiang Ji, Gen Li

https://openreview.net/forum?id=nFsbQHFmj2

Keywords: reinforcement learning theory, regret minimization, minimax optimality

Compressor summary: The paper presents a new model-free reinforcement learning algorithm that achieves optimal performance with low memory and computational cost, and fast sample efficiency.


Label Correction of Crowdsourced Noisy Annotations with an Instance-Dependent Noise Transition Model

Hui Guo, Boyu Wang, Grace Yi

https://openreview.net/forum?id=nFEQNYsjQO

Keywords: Noisy Label, Instance-Dependent Transition Matrix, Label Correction, Crowdsourcing

Compressor summary: The paper proposes a new label correction algorithm using a Bayesian network to model instance-dependent noise transitions in crowdsourced annotations, with theoretical guarantees and experimental validation.


Contrastive Training of Complex-Valued Autoencoders for Object Discovery

Aleksandar Stanić, Anand Gopalakrishnan, Kazuki Irie, Jürgen Schmidhuber

https://openreview.net/forum?id=nF6X3u0FaA

Keywords: object-centric learning, complex-valued networks, unsupervised learning, temporal correlation hypothesis

Compressor summary: The authors propose a new method for synchrony-based object-centric models that improves their ability to learn from multi-object color datasets and store more than three objects.


Private Distribution Learning with Public Data: The View from Sample Compression

Shai Ben-David, Alex Bie, Clement Louis Canonne, Gautam Kamath, Vikrant Singhal

https://openreview.net/forum?id=nDIrJmKPd5

Keywords: differential privacy, distribution learning, gaussians, mixture of gaussians, compression schemes, robust compression schemes, privacy

Compressor summary: The paper studies how to privately learn a distribution from public and private data, and shows that this depends on compressing the data and learning from lists of examples.


FouriDown: Factoring Down-Sampling into Shuffling and Superposing

Qi Zhu, Man Zhou, Jie Huang, Naishan Zheng, Hongzhi Gao, Chongyi Li, Yuan Xu, Feng Zhao

https://openreview.net/forum?id=nCwStXFDQu

Keywords: Image restoration, Down-Sampling, Fourier transform

Compressor summary: The authors propose a new down-sampling method called FouriDown, which uses a learnable and context-adaptive Fourier function to improve image restoration tasks like de-blurring and low-light enhancement.


The Equivalence of Dynamic and Strategic Stability under Regularized Learning in Games

Victor Boone, Panayotis Mertikopoulos

https://openreview.net/forum?id=nCLdsEzZBV

Keywords: Regularized learning, dynamic stability, strategic stability, Nash equilibrium

Compressor summary: The paper studies how no-regret learning in multiplayer games leads to setwise rationality properties and closedness under better replies, and explores the convergence rate of different regularization methods.


Predicting Global Label Relationship Matrix for Graph Neural Networks under Heterophily

Langzhang Liang, Xiangjing Hu, Zenglin Xu, Zixing Song, Irwin King

https://openreview.net/forum?id=nBFMCyEi0j

Keywords: graph neural networks, heterophily problem, global label relationship matrix

Compressor summary: LRGNN is a generic GNN that works on both homophilous and heterophilous graphs by predicting the low-rank label relationship matrix using robust low-rank approximation, which has two advantages for graph modeling.


Training Transitive and Commutative Multimodal Transformers with LoReTTa

Manuel Tran, Yashin Dicente Cid, Amal Lahiani, Fabian J Theis, Tingying Peng, Eldad Klaiman

https://openreview.net/forum?id=nArzDm353Y

Keywords: generative pre-training, causal modeling, masked modeling, commutative modeling, transitive modeling, multimodal learning

Compressor summary: LoReTTa is a self-supervised framework that links modalities with causal modeling and masked modeling to handle multimodal datasets with missing modalities.


Deconstructing Data Reconstruction: Multiclass, Weight Decay and General Losses

Gon Buzaglo, Niv Haim, Gilad Yehudai, Gal Vardi, Yakir Oz, Yaniv Nikankin, michal Irani

https://openreview.net/forum?id=nA9Fh3HFHJ

Keywords: memorization, data reconstruction, implicit bias

Compressor summary: This paper extends a previous study on reconstructing training data from neural networks and explores factors affecting this process, such as weight decay and neuron count.


Environment-Aware Dynamic Graph Learning for Out-of-Distribution Generalization

Haonan Yuan, Qingyun Sun, Xingcheng Fu, Ziwei Zhang, Cheng Ji, Hao Peng, Jianxin Li

https://openreview.net/forum?id=n8JWIzYPRz

Keywords: dynamic graph learning, out-of-distribution generalization, invariant learning, link prediction

Compressor summary: The EAGLE framework models and exploits spatio-temporal environments for out-of-distribution generalization in dynamic graph neural networks, achieving superior performance against existing methods under distribution shifts.


Clifford Group Equivariant Neural Networks

David Ruhe, Johannes Brandstetter, Patrick Forré

https://openreview.net/forum?id=n84bzMrGUD

Keywords: Clifford algebras, geometric deep dearning, Clifford group equivariance, E(n)-equivariant neural networks, O(n)-equivariant neural networks

Compressor summary: The paragraph introduces a new method for building equivariant neural networks using the Clifford group, which has desirable properties and shows state-of-the-art performance on various tasks.


Learning with Explanation Constraints

Rattana Pukdee, Dylan Sam, J Zico Kolter, Nina Balcan, Pradeep Kumar Ravikumar

https://openreview.net/forum?id=n6ztJ3Lrdj

Keywords: Interpretable ML, Semi-supervised learning, Learning theory

Compressor summary: The paper proposes a learning framework that uses explanation constraints from prior knowledge to improve the performance and interpretability of deep learning models, especially when using gradient information as explanations.


Robust Distributed Learning: Tight Error Bounds and Breakdown Point under Data Heterogeneity

Youssef Allouah, Rachid Guerraoui, Nirupam Gupta, Rafael Pinot, Geovani Rizk

https://openreview.net/forum?id=n3fPDW87is

Keywords: Optimization, Byzantine resilience, Distributed machine learning, federated learning

Compressor summary: The paper proposes a more realistic heterogeneity model for robust distributed learning algorithms, showing lower breakdown points and better matching with empirical observations.


DRAUC: An Instance-wise Distributionally Robust AUC Optimization Framework

Siran Dai, Qianqian Xu, Zhiyong Yang, Xiaochun Cao, Qingming Huang

https://openreview.net/forum?id=n3ZVdny7OH

Keywords: Robust Learning AUC

Compressor summary: The paper proposes a method to optimize AUC in long-tailed classification scenarios using distributionally robust optimization and a surrogate loss, while addressing label bias.


Fast Optimal Transport through Sliced Generalized Wasserstein Geodesics

Guillaume Mahey, Laetitia Chapel, Gilles Gasso, Clément Bonet, Nicolas Courty

https://openreview.net/forum?id=n3XuYdvhNW

Keywords: Optimal Transport, Wasserstein distance, Generalized Geodesics, Sliced Wasserstein

Compressor summary: The paper introduces a new proxy for Wasserstein distance called min-SWGG, which is based on one-dimensional projections and has benefits in various applications such as gradient flows, shape matching, and image colorization.


2Direction: Theoretically Faster Distributed Training with Bidirectional Communication Compression

Alexander Tyurin, Peter Richtárik

https://openreview.net/forum?id=n18MhTsSGb

Keywords: convex optimization, accelerated method, communication compression, bidirectional compression, distributed optimization

Compressor summary: The 2Direction method speeds up distributed convex optimization by using bidirectional compressed communication and a new error-feedback mechanism, outperforming previous methods in both communication complexity and acceleration.


Optimality in Mean Estimation: Beyond Worst-Case, Beyond Sub-Gaussian, and Beyond $1+\alpha$ Moments

Trung Dang, Jasper C.H. Lee, Maoyuan Song, Paul Valiant

https://openreview.net/forum?id=mvSDs51eqQ

Keywords: mean estimation, instance optimality

Compressor summary: The paper explores if mean estimation algorithms can benefit from knowing specific features of the input distribution and finds that it's possible in some cases, but generally not, while introducing a new framework for analyzing algorithm optimality.


Thinker: Learning to Plan and Act

Stephen Chung, Ivan Anokhin, David Krueger

https://openreview.net/forum?id=mumEBl0arj

Keywords: Reinforcement learning, model-based reinforcement learning, planning, Monte Carlo Tree Search, Markov Decision Process

Compressor summary: The Thinker algorithm is a novel reinforcement learning approach that allows agents to interact with and utilize a learned world model for autonomous planning, achieving state-of-the-art performance in Sokoban and competitive results in Atari 2600.


Reliable Off-Policy Learning for Dosage Combinations

Jonas Schweisthal, Dennis Frauen, Valentyn Melnychuk, Stefan Feuerriegel

https://openreview.net/forum?id=muVKSb8gi5

Keywords: off-policy learning, causal inference, reliable machine learning, medicine, dosaging, normalizing flows

Compressor summary: The paper proposes a novel method for personalized medicine that estimates individualized dose-response functions and optimal dosage combinations using neural networks, conditional normalizing flows, and gradient-based learning.


Efficient Sampling of Stochastic Differential Equations with Positive Semi-Definite Models

Anant Raj, Umut Simsekli, Alessandro Rudi

https://openreview.net/forum?id=mookk2nLO9

Keywords: Kernel Methods, Sampling, Fokker-Planck Equation, Fractional Fokker-Planck Equation, Stochastic Differential Equations, Partial Differential Equations

Compressor summary: The paper proposes an efficient method to sample from a stochastic differential equation using a recent probabilistic model and shows that it works well for smooth solutions.


Transient Neural Radiance Fields for Lidar View Synthesis and 3D Reconstruction

Anagh Malik, Parsa Mirdehghan, Sotiris Nousias, Kyros Kutulakos, David B. Lindell

https://openreview.net/forum?id=mmmd2vp0n0

Keywords: neural radiance fields, 3D reconstruction, single-photon lidar, computational imaging

Compressor summary: The paper proposes a novel method to render transient NeRFs using single-photon lidar data and captures light transport phenomena at picosecond timescales, showing improved geometry and appearance compared to point cloud-based supervision.


Optimizing Solution-Samplers for Combinatorial Problems: The Landscape of Policy-Gradient Method

Constantine Caramanis, Dimitris Fotakis, Alkis Kalavasis, Vasilis Kontonis, Christos Tzamos

https://openreview.net/forum?id=mmTy1iyU5G

Keywords: Policy Gradient, Combinatorial Optimization, Gradient Descent

Compressor summary: The authors develop a theoretical framework to analyze effective deep neural network and reinforcement learning methods for solving combinatorial problems, showing they can find near-optimal solutions with a limited number of parameters and without getting stuck in sub-optimal states. They also propose a novel regularization process for gradient descent that improves performance.


A Causal Framework for Decomposing Spurious Variations

Drago Plecko, Elias Bareinboim

https://openreview.net/forum?id=mm9svgvwvk

Keywords: Causal Inference, Confounding, Fair and Explainable AI

Compressor summary: The authors develop formal tools for decomposing spurious associations in causal mechanisms using Markovian and Semi-Markovian models, providing applications in various fields and demonstrating their approach on a real dataset.


Language Quantized AutoEncoders: Towards Unsupervised Text-Image Alignment

Hao Liu, Wilson Yan, Pieter Abbeel

https://openreview.net/forum?id=mlxRLIy7kc

Keywords: Large Language Model, VQVAE, Vector Quantization, Multimodal

Compressor summary: The paper proposes LQAE, a method to align images and texts without supervision, using pretrained language models, which improves few-shot learning in vision tasks like image classification and VQA.


Unleashing the Power of Randomization in Auditing Differentially Private ML

Krishna Pillutla, Galen Andrew, Peter Kairouz, Hugh Brendan McMahan, Alina Oprea, Sewoong Oh

https://openreview.net/forum?id=mlbes5TAAg

Keywords: Differential privacy auditing, multiple canaries, randomization, lifting, adaptive confidence intervals

Compressor summary: The authors propose a method to audit differentially private machine learning by adding multiple randomized examples (canaries) and use lifted differential privacy, statistical tests, and confidence intervals to improve sample complexity.


Robust Mean Estimation Without Moments for Symmetric Distributions

Gleb Novikov, David Steurer, Stefan Tiegel

https://openreview.net/forum?id=mkve1raJUc

Keywords: Robust Mean Estimation, Unbounded First Moment, Symmetric Distributions (Spherical, Elliptical, Product), Filtering Algorithm, Huber Loss

Compressor summary: The paper proposes efficient and robust algorithms for estimating the mean of symmetric distributions, including product Cauchy and elliptical distributions, without moment assumptions or strong distributional assumptions.


Memory-Constrained Algorithms for Convex Optimization

Moise Blanchard, Junhui Zhang, Patrick Jaillet

https://openreview.net/forum?id=mkKQr56xdB

Keywords: Convex optimization, feasibility problem, first-order methods, memory constraints, cutting planes, oracle complexity

Compressor summary: The paper introduces recursive cutting-plane algorithms for solving feasibility problems and convex optimization with constrained memory, achieving sub-polynomial complexity and optimal memory trade-offs in some regimes.


Beyond Average Return in Markov Decision Processes

Alexandre Marthe, Aurélien Garivier, Claire Vernade

https://openreview.net/forum?id=mgNu8nDFwa

Keywords: Markov Decision Process, Dynamic Programming, statistical functionnals, Distributionnal Reinforcement Learning, Policy Evaluation, Planning

Compressor summary: The paragraph discusses how Dynamic Programming can handle certain operations efficiently in Markov Decision Processes, but only for specific statistics classes, and explores the use of Distributional Reinforcement Learning to evaluate functionals approximately.


Large Language Models Are Zero-Shot Time Series Forecasters

Nate Gruver, Marc Anton Finzi, Shikai Qiu, Andrew Gordon Wilson

https://openreview.net/forum?id=md68e8iZK1

Keywords: large language models, time series, probabilistic forecasting

Compressor summary: The authors propose using large language models to predict future values in time series by encoding them as text and show their effectiveness in handling various aspects of time series data.


Neural Lighting Simulation for Urban Scenes

Ava Pun, Gary Sun, Jingkang Wang, Yun Chen, Ze Yang, Sivabalan Manivasagam, Wei-Chiu Ma, Raquel Urtasun

https://openreview.net/forum?id=mcx8IGneYw

Keywords: Scene Relighting, Lighting Estimation, Camera Simulation, Self-Driving, Lighting Simulation, Scene Editing

Compressor summary: LightSim is a neural lighting camera simulation system that creates realistic and diverse images under different illumination conditions for training image-based robot perception models.


SEENN: Towards Temporal Spiking Early Exit Neural Networks

Yuhang Li, Tamar Geller, Youngeun Kim, Priyadarshini Panda

https://openreview.net/forum?id=mbaN0Y0QTw

Keywords: Spiking Neural Networks, ANN-SNN Conversion, Conditional Computing

Compressor summary: This paper introduces Spiking Early-Exit Neural Networks (SEENNs), which adjust the number of timesteps in spiking neural networks based on input samples, achieving better accuracy and efficiency tradeoffs.


Bridging Discrete and Backpropagation: Straight-Through and Beyond

Liyuan Liu, Chengyu Dong, Xiaodong Liu, Bin Yu, Jianfeng Gao

https://openreview.net/forum?id=mayAyPrhJI

Keywords: discrete random variables, back-propagation, straight through

Compressor summary: ReinMax is a novel method that approximates gradients for discrete latent variables in deep learning using Heun's method, achieving second-order accuracy with negligible computation overheads and outperforming existing approaches.


Towards the Difficulty for a Deep Neural Network to Learn Concepts of Different Complexities

Dongrui Liu, Huiqi Deng, Xu Cheng, Qihan Ren, Kangrui Wang, Quanshi Zhang

https://openreview.net/forum?id=mZ3hnyL9bS

Keywords: representation complexity, deep learning

Compressor summary: The paper explores why deep neural networks find it harder to learn complex concepts involving many input variables, and identifies the specific factor that increases learning difficulty.


Class-Conditional Conformal Prediction with Many Classes

Tiffany Ding, Anastasios Nikolas Angelopoulos, Stephen Bates, Michael Jordan, Ryan Tibshirani

https://openreview.net/forum?id=mYz6ApeU4J

Keywords: conformal prediction, uncertainty quantification, class imbalance

Compressor summary: Clustered conformal prediction is a method that groups similar classes together and improves the probability of correctly predicting their labels when there is limited labeled data per class.


Reining Generalization in Offline Reinforcement Learning via Representation Distinction

Yi Ma, Hongyao Tang, Dong Li, Zhaopeng Meng

https://openreview.net/forum?id=mVywRIDNIl

Keywords: Offline Reinforcement Learning

Compressor summary: The paper proposes Representation Distinction (RD), a technique that improves offline reinforcement learning by differentiating between in-sample and out-of-distribution data representations to reduce overgeneralization.


Hierarchical Gaussian Mixture based Task Generative Model for Robust Meta-Learning

Yizhou Zhang, Jingchao Ni, Wei Cheng, Zhengzhang Chen, Liang Tong, Haifeng Chen, Yan Liu

https://openreview.net/forum?id=mVTyeQIiE4

Keywords: Few-Shot Learning, Meta Learning, Task Representation

Compressor summary: The paper proposes a meta-learning framework that models the density of task instances using a Hierarchical Gaussian Mixture based Task Generative Model (HTGM), which helps adapt to new tasks from different or unseen distributions.


InfoPrompt: Information-Theoretic Soft Prompt Tuning for Natural Language Understanding

Junda Wu, Tong Yu, Rui Wang, Zhao Song, Ruiyi Zhang, Handong Zhao, Chaochao Lu, Shuai Li, Ricardo Henao

https://openreview.net/forum?id=mSNfjOcDUv

Keywords: soft prompt tuning

Compressor summary: The paragraph describes a new method called InfoPrompt that improves soft prompt tuning by maximizing the mutual information between prompts and other model parameters, leading to faster convergence and better performance.


The Rise of AI Language Pathologists: Exploring Two-level Prompt Learning for Few-shot Weakly-supervised Whole Slide Image Classification

Linhao Qu, xiaoyuan Luo, Kexue Fu, Manning Wang, Zhijian Song

https://openreview.net/forum?id=mSDfBXr8Py

Keywords: multiple instance learning, whole slide image classification, prompt learning, vision-language model, few-shot learning

Compressor summary: The paper proposes a novel few-shot weakly supervised learning method for pathology Whole Slide Image (WSI) classification using prompt learning and GPT-4 to handle challenges posed by the weak bag labels within the Multiple Instance Learning framework.


Scaling Open-Vocabulary Object Detection

Matthias Minderer, Alexey A. Gritsenko, Neil Houlsby

https://openreview.net/forum?id=mQPNcBWjGc

Keywords: object detection, open-vocabulary object detection, vision transformers, vision-language models, scaling, self-training

Compressor summary: OWLv2 is a model that uses self-training with pseudo-annotations to scale up object detection data, achieving significant improvements in performance on rare classes.


Slimmed Asymmetrical Contrastive Learning and Cross Distillation for Lightweight Model Training

Jian Meng, Li Yang, Kyungmin Lee, Jinwoo Shin, Deliang Fan, Jae-sun Seo

https://openreview.net/forum?id=mOVEJletyD

Keywords: Contrastive Learning, Self-supervised Learning, Energy-efficient contrastive learning

Compressor summary: SACL-XD is a new self-supervised contrastive learning scheme that improves the performance of lightweight models and reduces energy consumption in AI applications.


(Provable) Adversarial Robustness for Group Equivariant Tasks: Graphs, Point Clouds, Molecules, and More

Jan Schuchardt, Yan Scholten, Stephan Günnemann

https://openreview.net/forum?id=mLe63bAYc7

Keywords: Adversarial robustness, Geometric machine learning, Equivariances, Robustness Certification, Graph neural networks

Compressor summary: The paragraph discusses a new approach to adversarial robustness that considers task equivariance in real-world tasks like molecular property prediction, and proposes methods to achieve provable robustness for various models and architectures.


Efficient Testable Learning of Halfspaces with Adversarial Label Noise

Ilias Diakonikolas, Daniel Kane, Vasilis Kontonis, Sihan Liu, Nikos Zarifis

https://openreview.net/forum?id=mIm0hsUUt1

Keywords: Machine Learning

Compressor summary: The paper presents a fast algorithm for learning halfspaces under the Gaussian distribution with adversarial label noise, using iterative soft localization and testers to ensure data quality.


Theoretically Guaranteed Bidirectional Data Rectification for Robust Sequential Recommendation

yatong sun, Bin Wang, Zhu Sun, Xiaochun Yang, Yan Wang

https://openreview.net/forum?id=mHsxsrLl0y

Keywords: recommender systems, sequential recommendation

Compressor summary: The paper proposes BirDRec, a framework that rectifies unreliable data in sequential recommender systems by adjusting both input and target items, and reduces its complexity with sampling and self-ensemble methods.


Provably Fast Convergence of Independent Natural Policy Gradient for Markov Potential Games

Youbang Sun, Tao Liu, Ruida Zhou, Panganamala Kumar, Shahin Shahrampour

https://openreview.net/forum?id=mA7nTGXjD3

Keywords: Multi Agent Reinforcement Learning, Markov Potential Games, Natural Policy Gradient, Nash Equilibrium

Compressor summary: The paper proposes an efficient algorithm for multi-agent reinforcement learning that finds approximate Nash Equilibria using policy gradients, and demonstrates its effectiveness on two examples.


Learning Motion Refinement for Unsupervised Face Animation

Jiale Tao, Shuhang Gu, Wen Li, Lixin Duan

https://openreview.net/forum?id=m9uHv1Pxq7

Keywords: Face animation, Motion refinement, Structure correlation

Compressor summary: The paper presents a new unsupervised face animation approach that learns both coarse and fine facial motions using a local affine motion model and a novel motion refinement module based on dense correlation between source and driving images.


Characterizing Graph Datasets for Node Classification: Homophily-Heterophily Dichotomy and Beyond

Oleg Platonov, Denis Kuznedelev, Artem Babenko, Liudmila Prokhorenkova

https://openreview.net/forum?id=m7PIJWOdlY

Keywords: graph characteristics, homophily, heterophily, label informativeness, constant baseline, GNN

Compressor summary: The paper discusses the drawbacks of common homophily measures in graph neural networks and proposes new measures called adjusted homophily and label informativeness.


Equivariant Adaptation of Large Pretrained Models

Arnab Kumar Mondal, Siba Smarak Panigrahi, Sékou-Oumar Kaba, Sai Rajeswar, Siamak Ravanbakhsh

https://openreview.net/forum?id=m6dRQJw280

Keywords: deep learning, large pretrained models, symmetry, equivariance, group theory, computer vision, point clouds, foundation models

Compressor summary: Equivariant networks can be made from large pretrained models by using a canonicalization network that transforms input to a canonical form, improving robustness to deterministic data transformations.


Better with Less: A Data-Active Perspective on Pre-Training Graph Neural Networks

Jiarong Xu, Renhong Huang, XIN JIANG, Yuxuan Cao, Carl Yang, Chunping Wang, Yang Yang

https://openreview.net/forum?id=m2WR1yJ8N9

Keywords: graph neural networks, pre-training

Compressor summary: The paper proposes a graph pre-training framework (APT) that uses fewer, carefully selected data points to improve downstream tasks, by choosing representative and instructive data based on graph properties and predictive uncertainty.


Learning List-Level Domain-Invariant Representations for Ranking

Ruicheng Xian, Honglei Zhuang, Zhen Qin, Hamed Zamani, Jing Lu, Ji Ma, Kai Hui, Han Zhao, Xuanhui Wang, Michael Bendersky

https://openreview.net/forum?id=m21rQusNgb

Keywords: learning to rank, domain adaptation, text ranking

Compressor summary: This paper introduces list-level alignment, a new invariant representation learning method for ranking problems that leverages the list structure of data and provides theoretical and empirical improvements over existing methods.


Efficient Hyper-parameter Optimization with Cubic Regularization

Zhenqian Shen, Hansi Yang, Yong Li, James Kwok, quanming yao

https://openreview.net/forum?id=m11TbsaQQI

Keywords: hyper-parameter optimization, cubic regularization

Compressor summary: The paper proposes a new hyper-parameter optimization method that uses cubic regularization and stochastic relaxation to avoid local optima and work without hyper-gradients.


Learn to Categorize or Categorize to Learn? Self-Coding for Generalized Category Discovery

Sarah Rastegar, Hazel Doughty, Cees G. M. Snoek

https://openreview.net/forum?id=m0vfXMrLwF

Keywords: Generalized category discovery, Open world learning, Open-set recognition

Compressor summary: The paper proposes a self-supervised method for discovering unknown categories at test time by viewing them as optimal solutions to well-defined problems and using minimal length codes to represent the hierarchy in real-world data.


StyleTTS 2: Towards Human-Level Text-to-Speech through Style Diffusion and Adversarial Training with Large Speech Language Models

Yinghao Aaron Li, Cong Han, Vinay S Raghavan, Gavin Mischler, Nima Mesgarani

https://openreview.net/forum?id=m0RbqrUM26

Keywords: Speech Processing, Text-to-Speech, Diffusion Model, Large Language Model, Self-Supervised Speech Model, WavLM

Compressor summary: StyleTTS 2 is a text-to-speech model that uses style diffusion, adversarial training, and large speech language models to generate high-quality, natural-sounding speech with different styles.


DiffComplete: Diffusion-based Generative 3D Shape Completion

Ruihang Chu, Enze Xie, Shentong Mo, Zhenguo Li, Matthias Nießner, Chi-Wing Fu, Jiaya Jia

https://openreview.net/forum?id=lzqaQRsITh

Keywords: 3d shape completion, conditional generation, diffusion models

Compressor summary: The authors present a new diffusion-based method called DiffComplete that balances realism, multi-modality, and high fidelity in 3D shape completion by using hierarchical feature aggregation and occupancy-aware fusion.


Spontaneous symmetry breaking in generative diffusion models

Gabriel Raya, Luca Ambrogioni

https://openreview.net/forum?id=lxGFGMMSVl

Keywords: generative models;diffusion models;score-based generative models; symmetry-breaking

Compressor summary: The paper presents a new understanding of diffusion models' dynamics, leading to improved performance and diversity in generating high-dimensional data.


CARE: Modeling Interacting Dynamics Under Temporal Environmental Variation

Xiao Luo, Haixin Wang, Zijie Huang, Huiyu Jiang, Abhijeet Sadashiv Gangan, Song Jiang, Yizhou Sun

https://openreview.net/forum?id=lwg3ohkFRv

Keywords: Dynamical System, Distribution Shift, Neural ODE, Graph Neural Network

Compressor summary: The paper proposes a probabilistic model called Context-attended Graph ODE (CARE) for time-varying interacting dynamical systems and shows its effectiveness on four datasets.


H-InDex: Visual Reinforcement Learning with Hand-Informed Representations for Dexterous Manipulation

Yanjie Ze, Yuyao Liu, Ruizhe Shi, Jiaxin Qin, Zhecheng Yuan, Jiashun Wang, Huazhe Xu

https://openreview.net/forum?id=lvvaNwnP6M

Keywords: Visual Reinforcement Learning, Representation Learning, Dexterous Manipulation

Compressor summary: The paper proposes a framework called H-InDex that uses human hand poses to improve robotic dexterous manipulation tasks using reinforcement learning.


K-Nearest-Neighbor Local Sampling Based Conditional Independence Testing

Shuai Li, Yingjie Zhang, Hongtu Zhu, Christina Dan Wang, Hai Shu, Ziqi Chen, Zhuoran Sun, Yanfeng Yang

https://openreview.net/forum?id=luyXPdkNSN

Keywords: Conditional Independence testing, causal inference, conditional mutual information, k-nearest neighbor, conditional randomization test, conditional permutation test

Compressor summary: The article introduces a novel conditional independence testing method that is efficient, powerful, and robust, without assumptions about distributions or dependencies.


SPA: A Graph Spectral Alignment Perspective for Domain Adaptation

Zhiqing Xiao, Haobo Wang, Ying Jin, Lei Feng, Gang Chen, Fei Huang, Junbo Zhao

https://openreview.net/forum?id=lpx9LZPVtZ

Keywords: Domain Adaptation, Self-training, Graph Spectra

Compressor summary: The paper proposes a new graph-based method for unsupervised domain adaptation, called SPA, which improves the performance of machine learning models in different target domains by aligning their structures and enhancing discriminability.


ProtoDiff: Learning to Learn Prototypical Networks by Task-Guided Diffusion

Yingjun Du, Zehao Xiao, Shengcai Liao, Cees G. M. Snoek

https://openreview.net/forum?id=lp9GR2t3hn

Keywords: Meta-learning, few-shot learning, diffusion model, prototype

Compressor summary: ProtoDiff is a novel framework that uses task-guided diffusion to generate overfitted prototypes for few-shot learning challenges, improving classification performance.


A Computation and Communication Efficient Method for Distributed Nonconvex Problems in the Partial Participation Setting

Alexander Tyurin, Peter Richtárik

https://openreview.net/forum?id=loxinzXlCx

Keywords: Nonconvex Optimization, Partial Participation, Variance Reduction, Compressed Communication, Distributed Optimization

Compressor summary: The new method optimizes distributed optimization and federated learning by combining variance reduction, partial participation, and compressed communication with optimal oracle and communication complexities.


Robust Learning for Smoothed Online Convex Optimization with Feedback Delay

Pengfei Li, Jianyi Yang, Adam Wierman, Shaolei Ren

https://openreview.net/forum?id=loixpHDZKj

Keywords: Online optimization, competitive algorithm, switching cost

Compressor summary: The paper proposes an online optimization method called Robustness-Constrained Learning that uses machine learning predictions to improve performance while accounting for multi-step costs and feedback delay.


Sample Complexity for Quadratic Bandits: Hessian Dependent Bounds and Optimal Algorithms

Qian Yu, Yining Wang, Baihe Huang, Qi Lei, Jason D. Lee

https://openreview.net/forum?id=lnTpBUge5G

Keywords: optimization, quadratic bandits, sample complexity, optimality

Compressor summary: The paper studies how to efficiently optimize quadratic functions using stochastic zeroth-order methods and introduces the concept of energy allocation to analyze the information-theoretic limitations and proposes a Hessian-independent algorithm that works well for all cases.


Flexible Attention-Based Multi-Policy Fusion for Efficient Deep Reinforcement Learning

Zih-Yun Chiu, Yi-Lin Tuan, William Yang Wang, Michael C. Yip

https://openreview.net/forum?id=lmXNcKhj4c

Keywords: Reinforcement Learning, Deep Reinforcement Learning, Sample Efficiency, Generalizability, Multi-Policy Decision Making, Multi-Policy Continuous Control

Compressor summary: The paper introduces Knowledge-Grounded RL (KGRL), a reinforcement learning paradigm that fuses multiple knowledge policies for efficient and flexible learning, and proposes Knowledge-Inclusive Attention Network (KIAN) as a new actor architecture.


A General Framework for Robust G-Invariance in G-Equivariant Networks

Sophia Sanborn, Nina Miolane

https://openreview.net/forum?id=llP6lmMiXE

Keywords: equivariance, group-equivariant cnns, invariance, pooling, convolutional neural networks

Compressor summary: The $G$-TC layer is a robust group-invariance method for $G$-CNNs that uses triple correlation theory to preserve signal structure and improve classification accuracy, working for both commutative and non-commutative groups.


Single-Pass Pivot Algorithm for Correlation Clustering. Keep it simple!

Konstantin Makarychev, Sayak Chakrabarty

https://openreview.net/forum?id=lkEiOZlmPm

Keywords: correlation clustering, Pivot algorithm, streaming

Compressor summary: The paper proposes a simplified and efficient algorithm for correlation clustering with improved memory usage compared to previous approaches.


Do SSL Models Have Déjà Vu? A Case of Unintended Memorization in Self-supervised Learning

Casey Meehan, Florian Bordes, Pascal Vincent, Kamalika Chaudhuri, Chuan Guo

https://openreview.net/forum?id=lkBygTc0SI

Keywords: self-supervised learning, privacy, data reconstruction, memorization

Compressor summary: SSL models can accidentally remember specific parts of images they've seen before, which could expose private information and be a security risk.


A Theoretical Analysis of the Test Error of Finite-Rank Kernel Ridge Regression

Tin Sum Cheng, Aurelien Lucchi, Anastasis Kratsios, Ivan Dokmanić, David Belius

https://openreview.net/forum?id=lk6KDG6qI7

Keywords: Kernel, regression, bias-variance, generalization

Compressor summary: Existing statistical learning guarantees for general kernel regressors often yield loose bounds when used with finite-rank kernels. Yet, finite-rank kernels naturally appear in a number of machine learning problems, e.g. when fine-tuning a pre-trained deep neural network's last layer to adapt it to a novel task when performing transfer learning. We address this gap for finite-rank kernel ridge regression (KRR) by deriving sharp non-asymptotic upper and lower bounds for the KRR test error of any finite-rank KRR. Our bounds are tighter than previously derived bounds on finite-rank KRR and, unlike comparable results, they also remain valid for any regularization parameters.


Nominality Score Conditioned Time Series Anomaly Detection by Point/Sequential Reconstruction

Chih-Yu Lai, Fan-Keng Sun, Zhengqi Gao, Jeffrey Lang, Duane S Boning

https://openreview.net/forum?id=ljgM3vNqfQ

Keywords: time series, anomaly detection, point anomalies, contextual anomalies, nominality score, induced anomaly score

Compressor summary: The paper presents a framework for unsupervised time series anomaly detection using point-based and sequence-based models, and shows it beats most existing methods.


Transformers as Statisticians: Provable In-Context Learning with In-Context Algorithm Selection

Yu Bai, Fan Chen, Huan Wang, Caiming Xiong, Song Mei

https://openreview.net/forum?id=liMSqUuVg9

Keywords: in-context learning, transformers, deep learning theory, learning theory

Compressor summary: The paragraph discusses how neural sequence models based on the transformer architecture can perform in-context learning and adaptively select different algorithms or tasks without explicit prompting, using a comprehensive statistical theory and experimental evidence.


Toward Re-Identifying Any Animal

Bingliang Jiao, Lingqiao Liu, Liying Gao, Ruiqi Wu, Guosheng Lin, PENG WANG, Yanning Zhang

https://openreview.net/forum?id=leS8668NJm

Keywords: Re-identification, Category-generalizable

Compressor summary: The paper introduces a new task called Re-identify Any Animal in the Wild (ReID-AW) and proposes UniReID, a universal re-identification model that can handle any unseen wildlife category using dynamic prompts and semantic knowledge from GPT-4.


A Tale of Two Features: Stable Diffusion Complements DINO for Zero-Shot Semantic Correspondence

Junyi Zhang, Charles Herrmann, Junhwa Hur, Luisa Polania Cabrera, Varun Jampani, Deqing Sun, Ming-Hsuan Yang

https://openreview.net/forum?id=lds9D17HRd

Keywords: Semantic Correspondence, Diffusion Models, Vision Transformer, Representation

Compressor summary: The text describes how diffusion models can be used for image processing tasks and explores their properties, performance, and potential applications.


A Single 2D Pose with Context is Worth Hundreds for 3D Human Pose Estimation

Qitao Zhao, Ce Zheng, Mengyuan Liu, Chen Chen

https://openreview.net/forum?id=lclQ2RvWYu

Keywords: Human Pose Estimation; 2D-to-3D Lifting; Context-Aware

Compressor summary: The proposed method improves 3D human pose estimation by using spatial context from 2D pose detectors without relying on temporal clues or additional computation.


Lift Yourself Up: Retrieval-augmented Text Generation with Self-Memory

Xin Cheng, Di Luo, Xiuying Chen, Lemao Liu, Dongyan Zhao, Rui Yan

https://openreview.net/forum?id=lYNSvp51a7

Keywords: natural language processing, retrieval-augmented text generation, self memory

Compressor summary: The paper proposes a novel framework called selfmem that uses its own output as memory to improve text generation tasks like translation, summarization, and dialogue generation.


DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining

Sang Michael Xie, Hieu Pham, Xuanyi Dong, Nan Du, Hanxiao Liu, Yifeng Lu, Percy Liang, Quoc V Le, Tengyu Ma, Adams Wei Yu

https://openreview.net/forum?id=lXuByUeHhd

Keywords: language models, pretraining, domain reweighting, data curation

Compressor summary: DoReMi is a method that uses a small proxy model to optimize mixture proportions of pretraining data domains and improve language model performance without knowing downstream tasks.


Entropy-based Training Methods for Scalable Neural Implicit Samplers

Weijian Luo, Boya Zhang, Zhihua Zhang

https://openreview.net/forum?id=lXOoR4KYcJ

Keywords: implicit sampler, learning to sample, generative models

Compressor summary: The paper introduces a neural implicit sampler that efficiently generates samples from un-normalized target distributions using two novel training methods, and shows its effectiveness, efficiency, and scalability on three sampling benchmarks.


Not All Out-of-Distribution Data Are Harmful to Open-Set Active Learning

Yang Yang, Yuxuan Zhang, XIN SONG, Yi Xu

https://openreview.net/forum?id=lV3LIGlc1w

Keywords: Out-of-Distribution, Active Learning

Compressor summary: Progressive Active Learning (PAL) is a sampling scheme that selects valuable out-of-distribution instances and balances pseudo-ID and pseudo-OOD instances to improve both ID classifier and OOD detector performance in open-set active learning.


Unconstrained Dynamic Regret via Sparse Coding

Zhiyu Zhang, Ashok Cutkosky, Ioannis Paschalidis

https://openreview.net/forum?id=lT9n36RH1w

Keywords: Dynamic online learning, parameter-free online learning, time series forecasting, wavelet

Compressor summary: The paper proposes adaptive regret bounds for online convex optimization under nonstationary comparator sequences by using sparse coding and energy-sparse complexity measures.


Rewarded soups: towards Pareto-optimal alignment by interpolating weights fine-tuned on diverse rewards

Alexandre Rame, Guillaume Couairon, Corentin Dancette, Jean-Baptiste Gaya, Mustafa Shukor, Laure Soulier, Matthieu Cord

https://openreview.net/forum?id=lSbbC2VyCu

Keywords: Deep learning, Foundation models, Fine-tuning, Reward optimization, Linear mode connectivity, Weight averaging, Model soups, Robustness, Generalization, Alignment, Multi objective learning.

Compressor summary: The paper introduces rewarded soup, a multi-policy approach that aims for Pareto-optimal generalization across different rewards and preferences, by interpolating specialized networks' weights linearly.


Masked Space-Time Hash Encoding for Efficient Dynamic Scene Reconstruction

Feng Wang, Zilong Chen, Guokang Wang, Yafei Song, Huaping Liu

https://openreview.net/forum?id=lSLYXuLqRQ

Keywords: NeRF, Dynamic Scenes

Compressor summary: MSTH is a novel method that efficiently reconstructs dynamic 3D scenes from videos by using a learnable mask to guide a weighted combination of 3D and 4D hash encodings, reducing redundancy and achieving better results than previous methods with less training time and memory storage.


Extensible Prompts for Language Models on Zero-shot Language Style Customization

Tao Ge, Jing Hu, Li Dong, Shaoguang Mao, Yan Xia, Xun Wang, Si-Qing Chen, Furu Wei

https://openreview.net/forum?id=lRxpVfDMzz

Keywords: large language model, prompt, imaginary words, OOD robustness, natural language, zero-shot

Compressor summary: X-Prompt uses imaginary words that can be understood by large language models to help them comprehend complex concepts and enable better interactions with humans.


Social Motion Prediction with Cognitive Hierarchies

Wentao Zhu, Jason Qin, Yuke Lou, Hang Ye, Xiaoxuan Ma, Hai Ci, Yizhou Wang

https://openreview.net/forum?id=lRu0dN7BY6

Keywords: multi-person motion prediction

Compressor summary: The study introduces a new benchmark, formulation, and framework for predicting 3D multi-person motions in team sports using behavioral cloning, generative adversarial imitation learning, and cognitive hierarchy.


Functional-Group-Based Diffusion for Pocket-Specific Molecule Generation and Elaboration

Haitao Lin, Yufei Huang, Odin Zhang, Yunfan Liu, Lirong Wu, Siyuan Li, Zhiyuan Chen, Stan Z. Li

https://openreview.net/forum?id=lRG11M91dx

Keywords: sturcture-based drug design; molecule generation; diffusion model

Compressor summary: \textsc{D3FG} is a functional-group-based diffusion model that generates realistic and interacting molecules for protein pocket-specific drug design using graph neural networks.


Energy-Based Cross Attention for Bayesian Context Update in Text-to-Image Diffusion Models

Geon Yeong Park, Jeongsol Kim, Beomsu Kim, Sang Wan Lee, Jong Chul Ye

https://openreview.net/forum?id=lOCHMGO6ow

Keywords: Diffusion model, Energy-based model, Text-to-image generation

Compressor summary: The paper proposes a novel energy-based model framework that adapts context control for text-to-image diffusion models, improving semantic alignment and enabling zero-shot compositional generation for various image generation tasks.


Hypothesis Selection with Memory Constraints

Maryam Aliakbarpour, Mark Bun, Adam Smith

https://openreview.net/forum?id=lM1UnEssuX

Keywords: Hypothesis selection, memory constrained algorithms, density estimation, limited space

Compressor summary: The paper studies how to select the best distribution from a set of candidates using limited memory and queries, achieving nearly optimal tradeoffs between memory usage and sample complexity.


On the Interplay between Social Welfare and Tractability of Equilibria

Ioannis Anagnostides, Tuomas Sandholm

https://openreview.net/forum?id=lM0xyViO90

Keywords: learning in games, optimistic gradient descent, Nash equilibrium, price of anarchy, smooth games, social welfare

Compressor summary: The paper connects efficiency, smoothness, and no-regret learning algorithms to study equilibrium computation in different game types and improve welfare bounds.


PDP: Parameter-free Differentiable Pruning is All You Need

Minsik Cho, Saurabh Adya, Devang Naik

https://openreview.net/forum?id=lLztVBaBVU

Keywords: pruning, cnn, transformers

Compressor summary: PDP is an efficient and effective train-time pruning scheme for DNNs that works on various tasks, architectures, and pruning constraints, achieving state-of-the-art results in model size, accuracy, and training cost.


Unlimiformer: Long-Range Transformers with Unlimited Length Input

Amanda Bertsch, Uri Alon, Graham Neubig, Matthew R. Gormley

https://openreview.net/forum?id=lJWUJWLCJo

Keywords: retrieval augmentation, summarization, long-context, generation, long-input, encoder-decoder, transformers, language models, natural language generation, natural language processing, deep learning, neural networks

Compressor summary: The paper introduces Unlimiformer, a method that allows transformers to handle long inputs by using a kNN index for attention instead of attending to every token.


Gold-YOLO: Efficient Object Detector via Gather-and-Distribute Mechanism

Chengcheng Wang, Wei He, Ying Nie, Jianyuan Guo, Chuanjian Liu, Yunhe Wang, Kai Han

https://openreview.net/forum?id=lJDoPAjkCV

Keywords: YOLO, object detection, computer vision

Compressor summary: The paper introduces Gold-YOLO, a new real-time object detection model that uses convolution and self-attention to improve feature fusion and achieves state-of-the-art performance while maintaining low latency.


The CLIP Model is Secretly an Image-to-Prompt Converter

Yuxuan Ding, Chunna Tian, Haoxuan Ding, Lingqiao Liu

https://openreview.net/forum?id=lHa7gFbmvS

Keywords: Diffusion Model, CLIP model, Image Variation, Customized Generation

Compressor summary: The paper shows how to use CLIP to convert images into text prompts for text-to-image generation, improving this ability with some training data or online steps, and enabling better image-text interactions for tasks like variation and editing.


AutoGO: Automated Computation Graph Optimization for Neural Network Evolution

Mohammad Salameh, Keith G Mills, Negar Hassanpour, Fred X. Han, Shuting Zhang, Wei Lu, SHANGLING JUI, CHUNHUA ZHOU, Fengyu Sun, Di Niu

https://openreview.net/forum?id=lDI3ZuyzM9

Keywords: Neural Architecture Search, Optimization Framework, Performance Prediction

Compressor summary: AutoGO is a framework that optimizes neural networks by evolving their low-level computation graphs, improving performance and hardware compatibility for various computer vision tasks.


An information-theoretic quantification of the content of communication between brain regions

Marco Celotto, Jan Bím, Alejandro Tlaie, Vito De Feo, Alessandro Toso, Stefan M Lemke, Daniel Chicharro, Hamed Nili, Malte Bieler, Ileana Livia Hanganu-Opatz, Tobias H. Donner, Andrea Brovelli, Stefano Panzeri

https://openreview.net/forum?id=lD8xaUWw24

Keywords: Information transmission; Brain data analysis; Sensory processing; Partial information decomposition

Compressor summary: The authors develop a new method called Feature-specific Information Transfer (FIT) that can measure how much information about specific features is transmitted between brain regions, beyond what traditional methods can reveal.


Team-PSRO for Learning Approximate TMECor in Large Team Games via Cooperative Reinforcement Learning

Stephen Marcus McAleer, Gabriele Farina, Gaoyue Zhou, Mingzhi Wang, Yaodong Yang, Tuomas Sandholm

https://openreview.net/forum?id=lCThtrJxoH

Keywords: PSRO, team games, TMECor, populations, equilibrium, game theory, RL

Compressor summary: The paper introduces two algorithms for multi-player team games that balance game-theoretic guarantees with scalability and outperform existing methods on some experiments.


Adversarial Learning for Feature Shift Detection and Correction

Míriam Barrabés, Daniel Mas Montserrat, Margarita Geleta, Xavier Giró-i-Nieto, Alexander G Ioannidis

https://openreview.net/forum?id=lBhRTO2uWf

Keywords: feature shift detection, distribution shift, shift, data-centric AI

Compressor summary: The paper proposes a method to detect and fix feature shifts in data using adversarial learning and simple iterative heuristics, outperforming existing approaches.


Normalization Layers Are All That Sharpness-Aware Minimization Needs

Maximilian Mueller, Tiffany Joyce Vlaar, David Rolnick, Matthias Hein

https://openreview.net/forum?id=lArwl3y9x6

Keywords: sharpness-aware minimization, flatness, generalization, normalization layers

Compressor summary: The paper explores how perturbing only the normalization parameters in the adversarial step of Sharpness-aware minimization (SAM) can improve generalization performance compared to perturbing all parameters or using alternative sparse perturbation methods.


DrugCLIP: Contrasive Protein-Molecule Representation Learning for Virtual Screening

Bowen Gao, Bo Qiang, Haichuan Tan, Yinjun Jia, Minsi Ren, Minsi Lu, Jingjing Liu, Wei-Ying Ma, Yanyan Lan

https://openreview.net/forum?id=lAbCgNcxm7

Keywords: Application, Drug Discovery, Representation Learning, Dataset Augmentation

Compressor summary: The paper proposes a new contrastive learning method called DrugCLIP for faster and better virtual screening of potential drugs using protein pocket and molecule representations without explicit binding-affinity scores.


Unsupervised Learning for Solving the Travelling Salesman Problem

Yimeng Min, Yiwei Bai, Carla P Gomes

https://openreview.net/forum?id=lAEc7aIW20

Keywords: Combinatorial Optimization, Graph Neural Network, Travelling Salesman Problem

Compressor summary: UTSP is an unsupervised learning framework that uses a graph neural network and local search to solve the TSP, achieving better results than existing data-driven heuristics while being more efficient in terms of parameters and training samples.


Globally solving the Gromov-Wasserstein problem for point clouds in low dimensional Euclidean spaces

Martin Ryner, Jan Kronqvist, Johan Karlsson

https://openreview.net/forum?id=l9MbuqzlZt

Keywords: Gromov-Wasserstein problem, QAP, Global optimization

Compressor summary: The paper proposes an efficient method for solving the Gromov-Wasserstein problem, which finds the best assignment between two sets of points preserving distances, using low-dimensional optimization.


Visual Instruction Inversion: Image Editing via Image Prompting

Thao Nguyen, Yuheng Li, Utkarsh Ojha, Yong Jae Lee

https://openreview.net/forum?id=l9BsCh8ikK

Keywords: image editing, diffusion models, visual prompting

Compressor summary: The paper proposes a method for image editing using visual prompts and text-to-image diffusion models to generate editing directions from examples.


Generative Category-level Object Pose Estimation via Diffusion Models

Jiyao Zhang, Mingdong Wu, Hao Dong

https://openreview.net/forum?id=l6ypbj6Nv5

Keywords: Category-Level Object Pose Estimation, Diffusion Model

Compressor summary: The authors propose a novel solution for category-level object pose estimation using conditional generative modeling with score-based diffusion models, achieving state-of-the-art results on the REAL275 dataset and generalizing well to new categories.


Practical Contextual Bandits with Feedback Graphs

Mengxiao Zhang, Yuheng Zhang, Olga Vrousgou, Haipeng Luo, Paul Mineiro

https://openreview.net/forum?id=l6pYRbuHpO

Keywords: Online learning with feedback graphs, Contextual Bandits, Practical algorithms

Compressor summary: The paper proposes a method for contextual bandits with feedback graphs that simplifies learning by reducing it to regression and achieving minimal performance loss.


Fine-Grained Visual Prompting

Lingfeng Yang, Yueze Wang, Xiang Li, Xinlong Wang, Jian Yang

https://openreview.net/forum?id=l6R4Go3noz

Keywords: visual prompting, zero-shot, visual language model, referring expression comprehension

Compressor summary: The authors propose a fine-grained visual prompting method that uses pixel-level masks to improve zero-shot instance-level recognition of objects in images, achieving significant improvements over existing methods on several benchmarks.


Relative Entropic Optimal Transport: a (Prior-aware) Matching Perspective to (Unbalanced) Classification

Liangliang Shi, Haoyu Zhen, Gu Zhang, Junchi Yan

https://openreview.net/forum?id=l61Kp1zBwC

Keywords: Optimal Transport; Unbalanced Classification

Compressor summary: The paper proposes a new optimal transport method for classification, which connects optimal transport and machine learning concepts like barycentric projection and transfer learning.


FOCAL: Contrastive Learning for Multimodal Time-Series Sensing Signals in Factorized Orthogonal Latent Space

Shengzhong Liu, Tomoyoshi Kimura, Dongxin Liu, Ruijie Wang, Jinyang Li, Suhas Diggavi, Mani Srivastava, Tarek Abdelzaher

https://openreview.net/forum?id=l4CZCKXoSn

Keywords: Multimodal Time Series; Contrastive Learning; Factorized Latent Space

Compressor summary: FOCAL is a novel contrastive learning framework for multimodal time series that considers modality-specific features and temporal information locality, improving performance in downstream tasks.


BIRD: Generalizable Backdoor Detection and Removal for Deep Reinforcement Learning

Xuan Chen, Wenbo Guo, Guanhong Tao, Xiangyu Zhang, Dawn Song

https://openreview.net/forum?id=l3yxZS3QdT

Keywords: Backdoor Defense, Deep Reinforcement Learning

Compressor summary: BIRD is a technique that detects and removes backdoors from pretrained deep reinforcement learning policies without needing information about the attack or access to its training process.


Scan and Snap: Understanding Training Dynamics and Token Composition in 1-layer Transformer

Yuandong Tian, Yiping Wang, Beidi Chen, Simon Shaolei Du

https://openreview.net/forum?id=l3HUgVHqGQ

Keywords: transformer, training dynamics, theoretical analysis, self-attention, interpretability, neural network understanding

Compressor summary: The paper analyzes how self-attention in a 1-layer transformer works during training for next token prediction, showing that it acts as a discriminative scanning algorithm with a scan and snap dynamics.


Feature Likelihood Score: Evaluating the Generalization of Generative Models Using Samples

Marco Jiralerspong, Joey Bose, Ian Gemp, Chongli Qin, Yoram Bachrach, Gauthier Gidel

https://openreview.net/forum?id=l2VKZkolT7

Keywords: Generative model, FID, Evaluation, Precision, Recall, Likelihood

Compressor summary: FLS is a new metric that assesses the novelty, fidelity, and diversity of generated data by deep generative models, addressing limitations of existing methods.


Look Beneath the Surface: Exploiting Fundamental Symmetry for Sample-Efficient Offline RL

Peng Cheng, Xianyuan Zhan, Zhihao Wu, Wenjia Zhang, Youfang Lin, Shou cheng Song, Han Wang, Li Jiang

https://openreview.net/forum?id=kyXMU3H7RB

Keywords: sample efficiency; offline reinforcement learning; fundamental symmetry

Compressor summary: The paper proposes a new offline reinforcement learning algorithm (TSRL) that leverages time-reversal symmetry of system dynamics to improve performance on small datasets and enhance data efficiency and generalizability.


Minimum-Risk Recalibration of Classifiers

Zeyu Sun, Dogyoon Song, Alfred Hero

https://openreview.net/forum?id=kvXcHfBghm

Keywords: probability calibration, optimal number of bins, label shift adaptation

Compressor summary: The paper introduces minimum-risk recalibration using MSE decomposition, analyzes UMB recalibration, and proposes a two-stage approach for label shift adaptation.


PRIOR: Personalized Prior for Reactivating the Information Overlooked in Federated Learning.

Mingjia Shi, Yuhao Zhou, Kai Wang, Huaizheng Zhang, Shudong Huang, Qing Ye, Jiancheng Lv

https://openreview.net/forum?id=kuxu4lCRr5

Keywords: Federated Learning, Personalized Federated Learning, Expectation Maximization, Relaxed Mirror Descent

Compressor summary: The paper proposes a new method, pFedBreD, for personalized federated learning that injects prior knowledge into the global model to improve performance on various datasets.


Computing Optimal Nash Equilibria in Multiplayer Games

Youzhi Zhang, Bo An, Venkatramanan Siva Subrahmanian

https://openreview.net/forum?id=kupNhxLc6k

Keywords: Algorithmic game theory, Optimal Nash equilibrium

Compressor summary: The paper proposes a new algorithm, CRM, to efficiently compute a Nash Equilibrium in multiplayer games by using correlation plans to reduce the solution space and improve the speed.


DDCoT: Duty-Distinct Chain-of-Thought Prompting for Multimodal Reasoning in Language Models

Ge Zheng, Bin Yang, Jiajin Tang, Hong-Yu Zhou, Sibei Yang

https://openreview.net/forum?id=ktYjrgOENR

Keywords: Chain-of-Thought Reasoning, Multimodal Science Question Answering, Vision and Langauge

Compressor summary: The text discusses challenges in applying chain of thought reasoning to multimodal contexts and proposes a new method called DDCoT that improves reasoning abilities, generalizability, and explainability of AI models.


Multi-task learning with summary statistics

Parker Knight, Rui Duan

https://openreview.net/forum?id=ktTSji9ZIs

Keywords: multi-task learning, genetic risk prediction, summary statistics

Compressor summary: The paper proposes a multi-task learning framework using summary statistics and adaptive parameter selection for applications where data-sharing is restricted, such as healthcare, and analyzes its performance theoretically and empirically.


Towards Last-layer Retraining for Group Robustness with Fewer Annotations

Tyler LaBonte, Vidya Muthukumar, Abhishek Kumar

https://openreview.net/forum?id=kshC3NOP6h

Keywords: spurious correlations, group robustness, last-layer retraining, distribution shift

Compressor summary: The paper proposes a method called SELF that improves group robustness of neural networks by retuning the last layer using disagreements or misclassifications, without needing group annotations.


SpatialRank: Urban Event Ranking with NDCG Optimization on Spatiotemporal Data

BANG AN, Xun Zhou, Yongjian Zhong, Tianbao Yang

https://openreview.net/forum?id=ks7Mf5lzSx

Keywords: urban event, NDCG optimization, ranking, traffic accident, crime, spatiotemporal data

Compressor summary: SpatialRank is a novel method for urban event ranking that uses adaptive graph convolution layers and a hybrid NDCG loss to predict the top-k most risky locations based on spatiotemporal dependencies and neighboring spatial proximity.


Easy Learning from Label Proportions

Robert Istvan Busa-Fekete, Heejin Choi, Travis Dick, Claudio Gentile, Andres Munoz medina

https://openreview.net/forum?id=kqBUgrkm1c

Keywords: learning with partial information, unbiased loss, classification, proportion matching

Compressor summary: EASYLLP is a debiasing technique for weakly supervised classification that uses aggregate labels and works with various loss functions, improving instance level performance on Learning from Label Proportions.


Understanding Multi-phase Optimization Dynamics and Rich Nonlinear Behaviors of ReLU Networks

Mingze Wang, Chao Ma

https://openreview.net/forum?id=konBXvt2iS

Keywords: non-convex optimization, training dynamics, neural network

Compressor summary: The paper analyzes the training process of a two-layer ReLU network using Gradient Flow and identifies four phases and specific nonlinear behaviors that occur during learning.


Accountability in Offline Reinforcement Learning: Explaining Decisions with a Corpus of Examples

Hao Sun, Alihan Hüyük, Daniel Jarrett, Mihaela van der Schaar

https://openreview.net/forum?id=kmbG9iBRIb

Keywords: Accountability, Reinforcement Learning, Batched Control, Accountable Decision-Making, Offline RL, Interpretability in RL

Compressor summary: The paper presents AOC, a controller that uses an offline dataset as the Decision Corpus and selects a subset of examples for accountable control in low-data scenarios, evaluating it in simulated and real healthcare settings.


Offline Minimax Soft-Q-learning Under Realizability and Partial Coverage

Masatoshi Uehara, Nathan Kallus, Jason D. Lee, Wen Sun

https://openreview.net/forum?id=kjkLJ7NJJZ

Keywords: Reinforcement learning theory, PAC RL, Offline Reinforcement learning

Compressor summary: The paper proposes value-based algorithms for offline reinforcement learning that can achieve PAC guarantees even with partial coverage of data, using novel minimax loss functions derived from Lagrange functions in nonlinear convex optimization problems.


Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation

David Brandfonbrener, Ofir Nachum, Joan Bruna

https://openreview.net/forum?id=kjMGHTo8Cs

Keywords: representation learning, imitation learning

Compressor summary: This paper explores how large datasets with multitask demonstrations can be pretrained using inverse dynamics modeling to learn low dimensional representations for imitation learning in unknown environments.


On permutation symmetries in Bayesian neural network posteriors: a variational perspective

Simone Rossi, Ankit Singh, Thomas Hannagan

https://openreview.net/forum?id=kj33zJ9Vue

Keywords: Bayesian deep learning, approximate inference, permutation symmetries

Compressor summary: The authors extend the formalism of marginalized loss barrier and solution interpolation to Bayesian neural networks, propose an algorithm to search for linearly connected solutions, and experiment with different architectures and datasets.


End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes

Alexandre Max Maraval, Matthieu Zimmer, Antoine Grosnit, Haitham Bou Ammar

https://openreview.net/forum?id=kfWzpZvEUh

Keywords: meta-learning, bayesian optimisation, neural process, transformer, end-to-end, reinforcement learning

Compressor summary: The paper presents a differentiable meta-BO framework that uses neural processes and transformers to learn acquisition functions with RL and an auxiliary task to handle sparse rewards.


TextDiffuser: Diffusion Models as Text Painters

Jingye Chen, Yupan Huang, Tengchao Lv, Lei Cui, Qifeng Chen, Furu Wei

https://openreview.net/forum?id=ke3RgcDmfO

Keywords: Diffusion Model; Text Rendering

Compressor summary: TextDiffuser is a new method for generating images with readable and coherent text that works by first generating keyword layouts with a Transformer model and then using diffusion models to create images based on the layout and text prompts. It also introduces a large-scale dataset and evaluation tool for text rendering quality.


Prompt Pre-Training with Twenty-Thousand Classes for Open-Vocabulary Visual Recognition

Shuhuai Ren, Aston Zhang, Yi Zhu, Shuai Zhang, Shuai Zheng, Mu Li, Alex Smola, Xu Sun

https://openreview.net/forum?id=kdFR6IUEW6

Keywords: Prompt Pre-Training, CLIP, Open-Vocabulary Visual Recognition

Compressor summary: POMP is a prompt pre-training method for vision-language models that improves visual recognition tasks with zero-shot learning and achieves state-of-the-art performances on 21 datasets.


DESSERT: An Efficient Algorithm for Vector Set Search with Vector Set Queries

Joshua Engels, Benjamin Coleman, Vihan Lakshman, Anshumali Shrivastava

https://openreview.net/forum?id=kXfrlWXLwH

Keywords: Embedding Based Retrieval, Passage Ranking, Locality Sensitive Hashing, Randomized Algorithms

Compressor summary: The paper introduces DESSERT, a fast approximate search algorithm for vector set queries, which can be integrated into semantic search models like ColBERT to improve their speed and performance.


DaTaSeg: Taming a Universal Multi-Dataset Multi-Task Segmentation Model

Xiuye Gu, Yin Cui, Jonathan Huang, Abdullah Rashwan, Xuan Yang, Xingyi Zhou, Golnaz Ghiasi, Weicheng Kuo, Huizhong Chen, Liang-Chieh Chen, David A Ross

https://openreview.net/forum?id=kXOXrVnwbb

Keywords: universal segmentation, multi-task segmentation, multi-dataset segmentation, panoptic segmentation, semantic segmentation, instance segmentation, weakly-supervised segmentation

Compressor summary: The paper proposes DaTaSeg, a universal multi-dataset multi-task segmentation model that leverages weak-supervision, text embeddings, and merge operations to improve performance across different tasks and datasets, including open-vocabulary segmentation.


Towards Efficient Pre-Trained Language Model via Feature Correlation Distillation

Kun Huang, Xin Guo, Meng Wang

https://openreview.net/forum?id=kVfHQV668B

Keywords: Knowledge Distillation; Pre-Trained Language Model

Compressor summary: FCD is a new method for compressing PLMs that uses output feature relationships to transfer knowledge more effectively and achieves better results than previous KD approaches.


Bayesian Active Causal Discovery with Multi-Fidelity Experiments

Zeyu Zhang, Chaozhuo Li, Xu Chen, Xing Xie

https://openreview.net/forum?id=kS8rIH43Zc

Keywords: Causal Discovery, Active Learning, Multi-fidelity

Compressor summary: The paper presents a probabilistic model for multi-fidelity active causal discovery, with a mutual-information based acquisition function and a cascading model to capture correlations between different fidelity oracles, and extends it to batch intervention scenarios.


A Fractional Graph Laplacian Approach to Oversmoothing

Sohir Maskey, Raffaele Paolino, Aras Bacho, Gitta Kutyniok

https://openreview.net/forum?id=kS7ED7eE74

Keywords: Graph Neural Networks, Graph Neural ODE, Fractional Laplacian, Oversmoothing

Compressor summary: The paper introduces fractional graph Laplacian neural ODEs for directed graphs to capture long-range dependencies and mitigate oversmoothing in graph neural networks.


Inferring Hybrid Neural Fluid Fields from Videos

Hong-Xing Yu, Yang Zheng, Yuan Gao, Yitong Deng, Bo Zhu, Jiajun Wu

https://openreview.net/forum?id=kRdaTkaBwC

Keywords: neural scene representations, fluid dynamics, flow reconstruction, physics-based learning

Compressor summary: The paper presents HyFluid, a neural method to jointly estimate fluid density and velocity from sparse multiview videos, using physics-based losses and a hybrid neural velocity representation that handles visual ambiguities and turbulence in fluid flows.


NuTrea: Neural Tree Search for Context-guided Multi-hop KGQA

Hyeong Kyu Choi, Seunghun Lee, Jaewon Chu, Hyunwoo J. Kim

https://openreview.net/forum?id=kR5ycmBclj

Keywords: Knowledge Graph Question Answering, Knowledge Graph, Graph Neural Networks

Compressor summary: NuTrea is a tree search-based GNN model that improves multi-hop KGQA by incorporating broader KG context and handling ambiguous nodes.


Subclass-Dominant Label Noise: A Counterexample for the Success of Early Stopping

Yingbin Bai, Zhongyi Han, Erkun Yang, Jun Yu, Bo Han, Dadong Wang, Tongliang Liu

https://openreview.net/forum?id=kR21XsZeAr

Keywords: learning with noisy labels, weakly supervised learning

Compressor summary: The paper studies a type of label noise called subclass-dominant label noise (SDN) and proposes a method called NoiseCluster that uses long-trained representations to identify and correct it, achieving better results than existing methods on synthetic and real data.


Online Ad Allocation with Predictions

Fabian Christian Spaeh, Alina Ene

https://openreview.net/forum?id=kPfd3pcwHV

Keywords: Learning Augmented Algorithms, Display Ads, Generalized Assignment Problem

Compressor summary: The paper presents a learning-augmented algorithm for online packing problems that uses machine-learned predictions to improve performance beyond worst-case algorithms, and shows its effectiveness on synthetic and real data.


Abide by the law and follow the flow: conservation laws for gradient flows

Sibylle Marcotte, Rémi Gribonval, Gabriel Peyré

https://openreview.net/forum?id=kMueEV8Eyy

Keywords: Implicit bias, conservation laws, gradient flow, linear neural network, matrix factorization

Compressor summary: The article explores how to identify and count conservation laws, which describe quantities conserved during gradient flows of a given model, and how these laws relate to optimization initialization properties in over-parameterized models.


Point Cloud Completion with Pretrained Text-to-Image Diffusion Models

Yoni Kasten, Ohad Rahamim, Gal Chechik

https://openreview.net/forum?id=kMmAYbT0VL

Keywords: Point Cloud, Text, 3D

Compressor summary: SDS-Complete uses a text-to-image diffusion model to complete incomplete point clouds of various objects without needing extensive 3D data, improving performance on out-of-distribution objects.


On the choice of Perception Loss Function for Learned Video Compression

Sadaf Salehkalaibar, Truong Buu Phan, Jun Chen, Wei Yu, Ashish J Khisti

https://openreview.net/forum?id=kLIieSS2P3

Keywords: Video Compression, Information Theory, Neural Compression

Compressor summary: The choice of perception loss function affects video compression quality, especially at low bit rates, but encoded representations can be near universal and work for either choice of function.


ConDaFormer: Disassembled Transformer with Local Structure Enhancement for 3D Point Cloud Understanding

Lunhao Duan, Shanshan Zhao, Nan Xue, Mingming Gong, Gui-Song Xia, Dacheng Tao

https://openreview.net/forum?id=kKXJkiniOx

Keywords: Point Cloud, Transformer, 3D Segmentation, 3D object detection

Compressor summary: The paper proposes ConDaFormer, a new transformer block for 3D point cloud understanding that reduces costs and models local geometry prior by disassembling cubic windows into three orthogonal 2D planes and using depth-wise convolution to enhance local structure.


On Learning Necessary and Sufficient Causal Graphs

Hengrui Cai, Yixin Wang, Michael Jordan, Rui Song

https://openreview.net/forum?id=kKFDMtpeDW

Keywords: Causal structural learning, Necessity and sufficiency, Natural causal effects, Probabilities of causation, Variable selection

Compressor summary: The paper proposes a method to learn a subset of causally relevant variables from a complex graph using probabilities of causation, which can improve causal estimation in various fields.


When Do Graph Neural Networks Help with Node Classification? Investigating the Homophily Principle on Node Distinguishability

Sitao Luan, Chenqing Hua, Minkai Xu, Qincheng Lu, Jiaqi Zhu, Xiao-Wen Chang, Jie Fu, Jure Leskovec, Doina Precup

https://openreview.net/forum?id=kJmYu3Ti2z

Keywords: Graph Neural Networks, Homophily, Heterophily, Low-pass filter, High-pass filter, Node Distinguishability, Metrics

Compressor summary: The paper studies how the similarity of node neighborhood patterns (intra-class Node Distinguishability) and differences between classes (inter-class ND) affect Graph Neural Networks' performance, proposing new metrics to measure them and a hypothesis-testing based performance metric that goes beyond homophily.


On the Identifiability of Sparse ICA without Assuming Non-Gaussianity

Ignavier Ng, Yujia Zheng, Xinshuai Dong, Kun Zhang

https://openreview.net/forum?id=kJIibP5bq2

Keywords: independent component analysis, second-order statistics, sparsity

Compressor summary: The paper presents a new method for identifying Gaussian sources using independent component analysis that does not require non-Gaussian assumptions or restrictive connective structures, and provides experimental validation.


Limits, approximation and size transferability for GNNs on sparse graphs via graphops

Thien Le, Stefanie Jegelka

https://openreview.net/forum?id=kDQwossJuI

Keywords: graph neural networks, convolution, graph limits, size transferability

Compressor summary: The paper investigates how well graph neural networks (GNNs) can generalize to different graphs, especially sparse ones, using a theoretical approach based on graphops.


Can Pre-Trained Text-to-Image Models Generate Visual Goals for Reinforcement Learning?

Jialu Gao, Kaizhe Hu, Guowei Xu, Huazhe Xu

https://openreview.net/forum?id=kChEBODIx9

Keywords: Visual Reinforcement Learning, Large Generative Models, Image Editing, Robotics

Compressor summary: LfVoid is a method that uses text-to-image models and image editing to guide robots based on natural language instructions, even without in-domain training or true goal observations.


Coherent Soft Imitation Learning

Joe Watson, Sandy Huang, Nicolas Heess

https://openreview.net/forum?id=kCCD8d2aEu

Keywords: imitation learning, inverse reinforcement learning, behavioral cloning, learning from demonstration

Compressor summary: The authors propose a hybrid imitation learning method that combines behavioral cloning and reinforcement learning to learn from expert demonstrations and achieve complex tasks with minimal hyperparameter tuning.


SAME: Uncovering GNN Black Box with Structure-aware Shapley-based Multipiece Explanations

Ziyuan Ye, Rihan Huang, Qilin Wu, Quanying Liu

https://openreview.net/forum?id=kBBsj9KRgh

Keywords: GNN explainability, Shapley value, Monte Carlo tree search, structure awareness, multi-grained explanation

Compressor summary: SAME is a new method for explaining graph neural networks that uses Monte Carlo tree search to explore structure-aware substructures and improves explanation fidelity on various benchmarks.


Discovering General Reinforcement Learning Algorithms with Adversarial Environment Design

Matthew Thomas Jackson, Minqi Jiang, Jack Parker-Holder, Risto Vuorio, Chris Lu, Gregory Farquhar, Shimon Whiteson, Jakob Nicolaus Foerster

https://openreview.net/forum?id=kAU6Cdq1gV

Keywords: Reinforcement Learning, Meta-Learning, Meta-RL, Meta-Optimization, Policy Meta-Optimization, Environment Design, Unsupervised Environment Design, Auto-Curricula

Compressor summary: This paper explores how meta-learning update rules for deep reinforcement learning can improve generalization and proposes GROOVE, a novel method that generates curricula to maximize the regret of a meta-learned optimizer.


Open Compound Domain Adaptation with Object Style Compensation for Semantic Segmentation

Tingliang Feng, Hao Shi, Xueyang Liu, Wei Feng, Liang Wan, Yanlin Zhou, Di Lin

https://openreview.net/forum?id=k9zSU3pdi4

Keywords: Object Style Compensation, Open Compound Domain Adaptation, Semantic Segmentation

Compressor summary: The paper introduces Object Style Compensation, a method that adapts object styles across domains for semantic image segmentation by using an Object-Level Discrepancy Memory.


PDF: Point Diffusion Implicit Function for Large-scale Scene Neural Representation

Yuhan Ding, Fukun Yin, Jiayuan Fan, Hui Li, Xin Chen, Wen Liu, Chongshan Lu, Gang YU, Tao Chen

https://openreview.net/forum?id=k8U8ZijXHh

Keywords: implicit neural representation; diffusion; point cloud; volume rendering

Compressor summary: The authors propose a new method called Point Diffusion implicit Function (PDF) that uses point cloud super-resolution and background modeling to represent and synthesize large-scale outdoor scenes for novel view synthesis, achieving better results than existing methods.


L2T-DLN: Learning to Teach with Dynamic Loss Network

Zhaoyang Hai, Liyuan Pan, Xiabi Liu, Zhengzheng Liu, Mirna Yunita

https://openreview.net/forum?id=k6yNi6DEqK

Keywords: Learning to teach, dynamic loss function, optimization

Compressor summary: The paper proposes a teacher model with memory units and a Dynamic Loss Network to adjust loss functions adaptively based on both the states of the student model and the loss function, leading to improved deep learning performance on various tasks.


Streaming Algorithms and Lower Bounds for Estimating Correlation Clustering Cost

Sepehr Assadi, Vihan Shah, Chen Wang

https://openreview.net/forum?id=k4ZCORSFEd

Keywords: Correlation Clustering, Graph Streaming Algorithms, Large-scale Clustering, Graph Learning

Compressor summary: This paper studies streaming correlation clustering with very limited memory, proposing two novel algorithms that estimate the optimal cost up to a constant factor and some additional error.


L-C2ST: Local Diagnostics for Posterior Approximations in Simulation-Based Inference

Julia Linhart, Alexandre Gramfort, Pedro L. C. Rodrigues

https://openreview.net/forum?id=k2UVKezeWn

Keywords: machine learning, calibration, simulation-based inference, neuroscience, normalizing flows, classifier two-sample tests

Compressor summary: The paper introduces a new method, $\ell$-C2ST, for evaluating deep generative models' approximations of complex posterior distributions at specific observations, improving interpretability and performance over existing approaches.


Lookaround Optimizer: $k$ steps around, 1 step average

Jiangtao Zhang, Shunyu Liu, Jie Song, Tongtian Zhu, Zhengqi Xu, Mingli Song

https://openreview.net/forum?id=k1Xy5zCNOJ

Keywords: Deep Learning, Computer Vision, Mode Connectivity, Weight Average

Compressor summary: Lookaround is a new optimizer that improves diversity and generalization in deep networks by training multiple networks on transformed data and averaging them, leading to flatter minima.


Metropolis Sampling for Constrained Diffusion Models

Nic Fishman, Leo Klarner, Emile Mathieu, Michael John Hutchinson, Valentin De Bortoli

https://openreview.net/forum?id=jzseUq55eP

Keywords: diffusion model, generative modelling, manifold, constraints, proteins, robotics

Compressor summary: The paper proposes a simple noising scheme based on Metropolis sampling for diffusion models with domain-informed constraints, improving efficiency and performance over existing methods.


Generalized test utilities for long-tail performance in extreme multi-label classification

Erik Schultheis, Marek Wydmuch, Wojciech Kotlowski, Rohit Babbar, Krzysztof Dembczynski

https://openreview.net/forum?id=jze2r6RDFz

Keywords: extreme multi-label classification, long-tail labels performance, complex performance measures

Compressor summary: The paper proposes a new metric for extreme multi-label classification that captures the importance of rare labels and optimizes expected test utility using efficient prediction rules.


Ess-InfoGAIL: Semi-supervised Imitation Learning from Imbalanced Demonstrations

Huiqiao Fu, Kaiqiang Tang, Yuanyang Lu, Yiming Qi, Guizhou Deng, Flood Sung, Chunlin Chen

https://openreview.net/forum?id=jxhUNLoi4m

Keywords: Generative adversarial imitation learning, semi-supervised learning, multi-modal behaviors, imbalanced data

Compressor summary: Imitation learning can reproduce expert behaviors without explicit rewards, but real-world demos are challenging; a new semi-supervised method learns disentangled behavior representations from imbalanced data using generative adversarial networks and information maximization.


Auxiliary Losses for Learning Generalizable Concept-based Models

Ivaxi Sheth, Samira Ebrahimi Kahou

https://openreview.net/forum?id=jvYXln6Gzn

Keywords: Interpretability, concept bottleneck models, explainability

Compressor summary: The paper introduces coop-CBM, a transparent neural network model that preserves concept representations and improves performance under different data distributions.


Improving Language Plasticity via Pretraining with Active Forgetting

Yihong Chen, Kelly Marchisio, Roberta Raileanu, David Ifeoluwa Adelani, Pontus Stenetorp, Sebastian Riedel, Mikel Artetxe

https://openreview.net/forum?id=jvEbQBxd8X

Keywords: plasticity, continual learning, meta-learning, embeddings, cross-lingual transfer, forgetting

Compressor summary: The paper proposes using active forgetting during pretraining to create PLMs that can adapt quickly and effectively to new languages, especially when there is little data available.


Deep Reinforcement Learning with Plasticity Injection

Evgenii Nikishin, Junhyuk Oh, Georg Ostrovski, Clare Lyle, Razvan Pascanu, Will Dabney, Andre Barreto

https://openreview.net/forum?id=jucDLW6G9l

Keywords: deep reinforcement learning, continual learning, loss of plasticity

Compressor summary: The paper proposes plasticity injection, a method to increase neural network plasticity in deep reinforcement learning without changing parameters or predictions, and demonstrates its usefulness as a diagnostic and improvement tool for RL agents.


Simple and Controllable Music Generation

Jade Copet, Felix Kreuk, Itai Gat, Tal Remez, David Kant, Gabriel Synnaeve, Yossi Adi, Alexandre Défossez

https://openreview.net/forum?id=jtiQ26sCJi

Keywords: Music generation, Generative AI, Transformer, Language Models

Compressor summary: MusicGen is a single transformer model that generates high-quality music based on text or melodic features, outperforming previous methods.


Fine-grained Expressivity of Graph Neural Networks

Jan Böker, Ron Levie, Ningyuan Teresa Huang, Soledad Villar, Christopher Morris

https://openreview.net/forum?id=jt10uWlEbc

Keywords: graphons, universal approximation, weisfeiler-leman, graph metric, tree homomorphisms, tree distance, optimal transport, GNNs

Compressor summary: The paper presents a continuous extension of the Weisfeiler--Leman test for analyzing the expressive power of message-passing graph neural networks on graphons, providing a theoretical and empirical framework to understand their limitations and performance.


Domain Re-Modulation for Few-Shot Generative Domain Adaptation

Yi Wu, Ziqiang Li, Chaoyue Wang, Heliang Zheng, Shanshan Zhao, Bin Li, Dacheng Tao

https://openreview.net/forum?id=jown9RvYn7

Keywords: StyleGAN, Few-Shot Generative Domain Adaptation

Compressor summary: The study proposes a new generator structure called Domain Re-Modulation (DoRM) for few-shot Generative Domain Adaptation, which incorporates memory and domain association to achieve high quality, diversity, and consistency across domains.


Implicit Differentiable Outlier Detection Enable Robust Deep Multimodal Analysis

Zhu Wang, Sourav Medya, Sathya N. Ravi

https://openreview.net/forum?id=jooPcatnVF

Keywords: Implicit layer, Out-of-distribution detection, multimodal learning

Compressor summary: The authors propose a method to combine deep neural networks and semantic knowledge for vision and language tasks, using an implicit out-of-distribution detector to filter irrelevant information.


Learning Reliable Logical Rules with SATNet

Zhaoyu Li, Jinpei Guo, Yuhe Jiang, Xujie Si

https://openreview.net/forum?id=jnIBiP2di1

Keywords: Logical Reasoning, Rule Learning, Interpretation, SATNet

Compressor summary: The paper proposes a new framework that generates interpretable and verifiable logical rules for advanced AI systems using differentiable learning and MaxSAT solver.


D-CIPHER: Discovery of Closed-form Partial Differential Equations

Krzysztof Kacprzyk, Zhaozhi Qian, Mihaela van der Schaar

https://openreview.net/forum?id=jnCPN1vpSR

Keywords: differential equations, symbolic regression

Compressor summary: D-CIPHER is a new method for finding differential equations from data that works with noisy and infrequent observations and discovers more phenomena than existing approaches.


Energy-based learning algorithms for analog computing: a comparative study

Benjamin Scellier, Maxence Ernoult, Jack Kendall, Suhas Kumar

https://openreview.net/forum?id=jl5a3t78Uh

Keywords: energy-based learning algorithm, contrastive learning, equilibrium propagation, coupled learning, convolutional Hopfield network

Compressor summary: The paragraph compares seven energy-based learning algorithms on deep convolutional Hopfield networks for five vision tasks and finds that negative perturbations perform better than positive ones.


Statistical Guarantees for Variational Autoencoders using PAC-Bayesian Theory

Sokhna Diarra Mbacke, Florence Clerc, Pascal Germain

https://openreview.net/forum?id=jkPDRHff3s

Keywords: Variational Autoencoders, PAC-Bayes, Statistical Learning Theory

Compressor summary: The authors use PAC-Bayesian theory to provide statistical guarantees for Variational Autoencoders (VAEs) and bounds on their performance metrics.


Adaptive Online Replanning with Diffusion Models

Siyuan Zhou, Yilun Du, Shun Zhang, Mengdi Xu, Yikang Shen, Wei Xiao, Dit-Yan Yeung, Chuang Gan

https://openreview.net/forum?id=jhs8F63xI6

Keywords: Decision making, Robotics, Planning-based

Compressor summary: The paper proposes a method to improve diffusion planning by deciding when to replan and adjusting trajectories, leading to better performance in robotic control tasks.


Finite-Time Analysis of Single-Timescale Actor-Critic

Xuyang Chen, Lin Zhao

https://openreview.net/forum?id=jh3UNSQK0l

Keywords: Finite-time analysis, single-timescale actor-critic

Compressor summary: The paper analyzes the convergence of online single-timescale actor-critic with linear function approximation on continuous state space, achieving an $\epsilon$-approximate stationary point with improved sample complexity.


Debiasing Scores and Prompts of 2D Diffusion for View-consistent Text-to-3D Generation

Susung Hong, Donghoon Ahn, Seungryong Kim

https://openreview.net/forum?id=jgIrJeHHlz

Keywords: Text-to-3D, Diffusion Models

Compressor summary: The paper proposes two methods to improve view consistency in text-to-3D generation by debiasing scores from 2D diffusion models and aligning user prompts with object views.


From ViT Features to Training-free Video Object Segmentation via Streaming-data Mixture Models

Roy Uziel, Or Dinari, Oren Freifeld

https://openreview.net/forum?id=jfsjKBDB1z

Keywords: Unsupervised, video segmentation, clustering

Compressor summary: The paper presents a training-free video object segmentation method that uses pre-trained deep features and classical clustering, achieving state-of-the-art results with low memory and fast inference.


MKOR: Momentum-Enabled Kronecker-Factor-Based Optimizer Using Rank-1 Updates

Mohammad Mozaffari, Sikan Li, Zhao Zhang, Maryam Mehri Dehnavi

https://openreview.net/forum?id=jcnvDO96N5

Keywords: machine learning, deep learning, optimizers, distributed training;second-order optimization;

Compressor summary: MKOR is a new optimizer that improves training speed and convergence of DNNs by reducing complexity and increasing second-order updates frequency.


Interaction Measures, Partition Lattices and Kernel Tests for High-Order Interactions

Zhaolu Liu, Robert Peach, Pedro A. M. Mediano, Mauricio Barahona

https://openreview.net/forum?id=jcRB6xHdJ2

Keywords: High-order interactions; Lattice theory; Kernel tests

Compressor summary: The text introduces a method to measure and test interactions between groups of more than two variables in complex systems using kernel-based tests and lattice theory connections.


Generator Born from Classifier

Runpeng Yu, Xinchao Wang

https://openreview.net/forum?id=jcJVgIFY2r

Keywords: Generative Model

Compressor summary: The paper presents a new method to reconstruct an image generator using a pre-trained classifier, by training the generator to satisfy convergence conditions based on maximum-margin bias theory.


Design from Policies: Conservative Test-Time Adaptation for Offline Policy Optimization

Jinxin Liu, Hongyin Zhang, Zifeng Zhuang, Yachen Kang, Donglin Wang, Bin Wang

https://openreview.net/forum?id=jZYf1GxH1V

Keywords: offline reinforcement learning, test-time adaptation

Compressor summary: The paper proposes a non-iterative bi-level offline RL method called DROP that learns an MBO score model and a behavior embedding for safe policy extraction during testing.


Moral Responsibility for AI Systems

Sander Beckers

https://openreview.net/forum?id=jYIknUIgkd

Keywords: responsibility, causation, causal models

Compressor summary: The paper defines moral responsibility for AI systems using causal models, compares its approach to existing ones, and extends the concept to a degree of responsibility.


Beyond Deep Ensembles: A Large-Scale Evaluation of Bayesian Deep Learning under Distribution Shift

Florian Seligmann, Philipp Becker, Michael Volpp, Gerhard Neumann

https://openreview.net/forum?id=jX49iKr6vb

Keywords: bayesian deep learning, distribution shift, calibration

Compressor summary: The paper surveys recent BDL methods on realistic tasks, focusing on how they generalize and calibrate under distribution shift. It explores different aspects of BDL, such as signed error, fine-tuning, and ensembles, and compares various algorithms in terms of accuracy and calibration.


RAPHAEL: Text-to-Image Generation via Large Mixture of Diffusion Paths

Zeyue Xue, Guanglu Song, Qiushan Guo, Boxiao Liu, Zhuofan Zong, Yu Liu, Ping Luo

https://openreview.net/forum?id=jUdZCcoOu3

Keywords: Diffusion Model, Text-to-Image Generation

Compressor summary: RAPHAEL is a text-conditional image diffusion model that generates highly artistic images by stacking tens of mixture-of-experts layers, outperforming other models in image quality and style switching, with a zero-shot FID score of 6.61 on the COCO dataset.


SPRING: Studying Papers and Reasoning to play Games

Yue Wu, So Yeon Min, Shrimai Prabhumoye, Yonatan Bisk, Ruslan Salakhutdinov, Amos Azaria, Tom Mitchell, Yuanzhi Li

https://openreview.net/forum?id=jU9qiRMDtR

Keywords: Games, Instruction Manual, Crafter, Open-world games, Large Language Models, Language Models, Zero-shot, In-context prompting

Compressor summary: SPRING is a novel approach that uses a large language model to reason and play open-world survival games like Crafter, outperforming state-of-the-art RL baselines without any training.


Spuriosity Rankings: Sorting Data to Measure and Mitigate Biases

Mazda Moayeri, Wenxiao Wang, Sahil Singla, Soheil Feizi

https://openreview.net/forum?id=jSuhnO9QJv

Keywords: spurious correlations, interpretability, bias, distributional robustness

Compressor summary: The paper proposes a simple and effective way to reduce model biases due to spurious features by ranking images based on how much they rely on such cues, using deep neural features of an interpretable network. The method works on any data and model without major modifications, and shows that class-wise biases are strongly related across models.


When are ensembles really effective?

Ryan Theisen, Hyunsuk Kim, Yaoqing Yang, Liam Hodgkinson, Michael W. Mahoney

https://openreview.net/forum?id=jS4DUGOtBD

Keywords: Ensembling, theory, deep learning

Compressor summary: The paper investigates when ensembling improves classification performance, finding that it depends on the disagreement rate and the average error rate of the models involved, with non-interpolating models benefiting more from ensembling than interpolating ones.


Learning Better with Less: Effective Augmentation for Sample-Efficient Visual Reinforcement Learning

Guozheng Ma, Linrui Zhang, Haoyu Wang, Lu Li, Zilin Wang, Zhen Wang, Li Shen, Xueqian Wang, Dacheng Tao

https://openreview.net/forum?id=jRL6ErxMVB

Keywords: Data Augmentation, Visual Reinforcement Learning, Sample Efficiency

Compressor summary: This work investigates the attributes of data augmentation (DA) in visual reinforcement learning, introducing new DA operations and fusion schemes to improve sample efficiency.


Is Learning in Games Good for the Learners?

William Brown, Jon Schneider, Kiran Vodrahalli

https://openreview.net/forum?id=jR2FkqW6GB

Keywords: learning in games, correlated equilibria, Stackelberg equilibria, swap regret, dynamic regret

Compressor summary: This text discusses tradeoffs between reward and regret in repeated games with two agents using generalized equilibrium, and shows that different algorithms can lead to different outcomes depending on the opponent's algorithm choice.


IEBins: Iterative Elastic Bins for Monocular Depth Estimation

Shuwei Shao, Zhongcai Pei, Xingming Wu, Zhong Liu, Weihai Chen, Zhengguo Li

https://openreview.net/forum?id=jOuxQGRVoQ

Keywords: Monocular depth estimation, Iterative refinement, Deep learning

Compressor summary: The paper introduces a novel iterative elastic bins method for monocular depth estimation, which progressively optimizes the search range and utilizes temporal context modeling with GRU-based architecture to achieve state-of-the-art results.


Fast Partitioned Learned Bloom Filter

Atsuki Sato, Yusuke Matsui

https://openreview.net/forum?id=jL2eJxPK88

Keywords: optimization, data structures, algorithms, theory, learned algorithms

Compressor summary: This paper presents two methods to construct a learned Bloom filter faster and with similar memory efficiency, while one method can achieve the same data structure as the original.


CluB: Cluster Meets BEV for LiDAR-Based 3D Object Detection

Yingjie Wang, Jiajun Deng, Yuenan Hou, Yao Li, Yu Zhang, Jianmin Ji, Wanli Ouyang, Yanyong Zhang

https://openreview.net/forum?id=jIhX7SpfCz

Keywords: 3D object detection ; Point clouds

Compressor summary: The paper proposes CluB, a 3D object detection framework that combines BEV-based and cluster-based detectors by enriching object representation at feature and query levels.


Gradient Flossing: Improving Gradient Descent through Dynamic Control of Jacobians

Rainer Engelken

https://openreview.net/forum?id=jEQRoJzDx8

Keywords: exploding/vanishing gradients, Lyapunov exponents, Lyapunov spectrum, chaos, RNN, condition number, Jacobian

Compressor summary: Gradient flossing is a new method to stabilize recurrent neural networks by regularizing Lyapunov exponents during learning, leading to improved convergence speed and success rate for tasks with long time horizons.


The Surprising Effectiveness of Diffusion Models for Optical Flow and Monocular Depth Estimation

Saurabh Saxena, Charles Herrmann, Junhwa Hur, Abhishek Kar, Mohammad Norouzi, Deqing Sun, David J. Fleet

https://openreview.net/forum?id=jDIlzSU8wJ

Keywords: Monocular depth, optical flow, diffusion, depth, flow

Compressor summary: Denoising diffusion probabilistic models can estimate optical flow and monocular depth effectively without specialized architectures, providing uncertainty and multimodality information.


Learning Repeatable Speech Embeddings Using An Intra-class Correlation Regularizer

Jianwei Zhang, Suren Jayasuriya, Visar Berisha

https://openreview.net/forum?id=jCPRG3FuHV

Keywords: repeatability, embeddings, metric learning, intra-class correlation, intra-class variance

Compressor summary: The authors propose a method to measure and improve the repeatability of embeddings for machine learning tasks using intra-class correlation coefficients (ICC) and show its effectiveness on speech tasks.


Hierarchical Adaptive Value Estimation for Multi-modal Visual Reinforcement Learning

Yangru Huang, Peixi Peng, Yifan Zhao, Haoran Xu, Mengyue Geng, Yonghong Tian

https://openreview.net/forum?id=jB4wsc1DQW

Keywords: vision-based reinforcement learning, multi-modal, event camera

Compressor summary: The paper proposes a local value estimation framework for vision-based reinforcement learning that adapts to different modalities and improves performance in autonomous driving.


Jailbroken: How Does LLM Safety Training Fail?

Alexander Wei, Nika Haghtalab, Jacob Steinhardt

https://openreview.net/forum?id=jA235JGM09

Keywords: red teaming, safety, RLHF, large language models

Compressor summary: The text discusses how large language models trained for safety still face adversarial misuse, investigates why this happens, and evaluates new attacks on state-of-the-art models like GPT-4 and Claude v1.3.


On Masked Pre-training and the Marginal Likelihood

Pablo Moreno-Muñoz, Pol G. Recasens, Søren Hauberg

https://openreview.net/forum?id=j9wGUcS30B

Keywords: Marginal likelihood, masked pre-training, Bayesian inference

Compressor summary: The paper explains how masked pre-training works by maximizing a model's generalization measure and explores its potential for Bayesian models.


Learning from Both Structural and Textual Knowledge for Inductive Knowledge Graph Completion

Kunxun Qi, Jianfeng Du, Hai Wan

https://openreview.net/forum?id=j7x9wW3tCf

Keywords: Knowledge graph completion; Neural approximate rule learning; Neural rule-based system

Compressor summary: The paper proposes a two-stage framework for knowledge graph completion using both structural and textual knowledge, which improves performance over existing rule-based systems.


DynPoint: Dynamic Neural Point For View Synthesis

Kaichen Zhou, Jia-Xing Zhong, Sangyun Shin, Kai Lu, Yiyuan Yang, Andrew Markham, Niki Trigoni

https://openreview.net/forum?id=j7U4pFkCYB

Keywords: View Synthesis, Monocular Video

Compressor summary: DynPoint is an algorithm that accelerates view synthesis for unconstrained monocular videos by predicting 3D correspondence between frames and using hierarchical neural point clouds.


Scaling Data-Constrained Language Models

Niklas Muennighoff, Alexander M Rush, Boaz Barak, Teven Le Scao, Nouamane Tazi, Aleksandra Piktus, Sampo Pyysalo, Thomas Wolf, Colin Raffel

https://openreview.net/forum?id=j5BuTrEj35

Keywords: large language models, scaling laws, data engineering

Compressor summary: The paper explores how to scale language models when there is limited text data and proposes a scaling law for compute optimality that takes into account repeated tokens and excess parameters.


What You See is What You Read? Improving Text-Image Alignment Evaluation

Michal Yarom, Yonatan Bitton, Soravit Changpinyo, Roee Aharoni, Jonathan Herzig, Oran Lang, Eran Ofek, Idan Szpektor

https://openreview.net/forum?id=j5AoleAIru

Keywords: Vision-and-language, Image-text alignment, Text-to-image generation, Image-to-text generation, Multi-modal models, Synthetic images, Meta-evaluation, Visual-question-answering

Compressor summary: The authors propose two methods for automatic evaluation of text-image alignment and introduce SeeTRUE, a dataset with human judgements for semantically aligned text-image pairs.


Resolving the Tug-of-War: A Separation of Communication and Learning in Federated Learning

Junyi Li, Heng Huang

https://openreview.net/forum?id=j4QVhftpYM

Keywords: Federated Learning

Compressor summary: FedSep is a novel two-layer federated learning framework that separates communication and learning layers for better privacy and efficiency, and outperforms existing methods in communication-efficient and heterogeneous-model FL tasks.


Active Vision Reinforcement Learning under Limited Visual Observability

Jinghuan Shang, Michael S Ryoo

https://openreview.net/forum?id=j2oYaFpbrB

Keywords: Reinforcement Learning, Active Reinforcement Learning, Visual Reinforcement Learning, Active Vision, Active Perception, Partial Observability, Sensorimotor

Compressor summary: The paper introduces SUGARL, a framework that learns motor and sensory policies together using an intrinsic sensorimotor reward, which enables agents to control their visual observations in partially observable environments.


Versatile Energy-Based Probabilistic Models for High Energy Physics

Taoli Cheng, Aaron Courville

https://openreview.net/forum?id=j0U6XJubbP

Keywords: Generative modeling, Energy-based models, Out-of-distribution detection, Sciences, Application, Physics

Compressor summary: The authors propose a flexible energy-based probabilistic model for High Energy Physics events that can generate simulations, detect anomalies, and identify particles.


Privacy Amplification via Compression: Achieving the Optimal Privacy-Accuracy-Communication Trade-off in Distributed Mean Estimation

Wei-Ning Chen, Dan Song, Ayfer Ozgur, Peter Kairouz

https://openreview.net/forum?id=izNfcaHJk0

Keywords: Differential Privacy, Federated Learning, Communication

Compressor summary: The paper studies how to achieve the best balance of privacy, communication, and accuracy in federated learning and analytics, and proposes two methods that use compression and random data selection to improve this trade-off.


Near-Optimal Algorithms for Gaussians with Huber Contamination: Mean Estimation and Linear Regression

Ilias Diakonikolas, Daniel Kane, Ankit Pensia, Thanasis Pittas

https://openreview.net/forum?id=iyweRIXAeH

Keywords: robust statistics, high-dimensional inference, regression, nearly linear time algorithms

Compressor summary: The paper presents novel algorithms for Gaussian mean estimation and linear regression with Huber contamination, achieving near-optimal sample and time complexity with optimal error guarantees.


GPEX, A Framework For Interpreting Artificial Neural Networks

Amir Akbarnejad, Gilbert Bigras, Nilanjan Ray

https://openreview.net/forum?id=iy4Of0w8ML

Keywords: Gaussian processes, Explainable AI

Compressor summary: The authors propose a method that uses Gaussian processes to match and explain the outputs of deep neural networks without strict assumptions, and demonstrate its effectiveness on various datasets using computational techniques for scalability and GPU acceleration.


Non-adversarial training of Neural SDEs with signature kernel scores

Zacharia Issa, Blanka Horvath, Maud Lemercier, Cristopher Salvi

https://openreview.net/forum?id=ixcsBZw5pl

Keywords: Neural SDEs, score-based generative models, signature kernels, time series

Compressor summary: This paper proposes a new way to train Neural SDEs non-adversarially using kernel scores, which improves stability, avoids mode collapse, and allows for generating spatiotemporal data.


Open Visual Knowledge Extraction via Relation-Oriented Multimodality Model Prompting

Hejie Cui, Xinyu Fang, Zihan Zhang, Ran Xu, Xuan Kan, Xin Liu, Yue Yu, Manling Li, Yangqiu Song, Carl Yang

https://openreview.net/forum?id=ixVAXsdtJO

Keywords: Visual Knowledge Extraction, Multimodality, Large Model Prompting

Compressor summary: OpenVik is a new method for extracting open visual knowledge from images using an open relational region detector and a multimodal model, with enhanced data techniques to diversify the results.


Distilling Out-of-Distribution Robustness from Vision-Language Foundation Models

Andy Zhou, Jindong Wang, Yu-Xiong Wang, Haohan Wang

https://openreview.net/forum?id=iwp3H8uSeK

Keywords: robustness, knowledge distillation, adversarial training, data augmentation, generalization

Compressor summary: The paper introduces a framework that improves vision models' robustness by combining knowledge distillation and data augmentation using a robust teacher and Discrete Adversarial Distillation (DAD).


Patch Diffusion: Faster and More Data-Efficient Training of Diffusion Models

Zhendong Wang, Yifan Jiang, Huangjie Zheng, Peihao Wang, Pengcheng He, Zhangyang Wang, Weizhu Chen, Mingyuan Zhou

https://openreview.net/forum?id=iv2sTQtbst

Keywords: diffusion models, training efficiency, data efficiency

Compressor summary: Patch Diffusion is a faster and more efficient way to train diffusion models that uses patches with random sizes and coordinates, improving the quality of generated images even on small datasets.


Saddle-to-Saddle Dynamics in Diagonal Linear Networks

Scott Pesme, Nicolas Flammarion

https://openreview.net/forum?id=iuqCXg1Gng

Keywords: gradient flow, saddle-to-saddle, diagonal linear network, incremental learning

Compressor summary: The paper describes how gradient flow in two-layer linear networks for regression finds the minimum l1-norm solution by jumping between saddles using a recursive algorithm and arc-length time-reparametrization, with no assumptions on data or network size.


Improving Compositional Generalization using Iterated Learning and Simplicial Embeddings

Yi Ren, Samuel Lavoie, Mikhail Galkin, Danica J. Sutherland, Aaron Courville

https://openreview.net/forum?id=irRHgjePdR

Keywords: compositional generalization, systematic generalization, iterated learning, representation learning, graph neural networks

Compressor summary: The authors propose iterated learning on models with simplicial embeddings to improve deep networks' compositional generalization, inspired by human language development and Kolmogorov complexity analysis.


Punctuation-level Attack: Single-shot and Single Punctuation Can Fool Text Models

Wenqiang Wang, Chongyang Du, Tao Wang, Kaihao Zhang, Wenhan Luo, Lin Ma, Wei Liu, Xiaochun Cao

https://openreview.net/forum?id=ir6WWkFR80

Keywords: Punctuation-level Attack, Textual Adversarial attack, Natural Language Processing

Compressor summary: The paper introduces a new way of textual attack using single punctuation changes and proposes a search method to find the best position for the attack without exhaustive search.


Adaptive recurrent vision performs zero-shot computation scaling to unseen difficulty levels

Vijay Veerabadran, Srinivas Ravishankar, Yuan Tang, Ritik Raina, Virginia R. de Sa

https://openreview.net/forum?id=iqezE0EyXq

Keywords: cognitive science, recurrent neural networks, adaptive computation time, visual reasoning

Compressor summary: The study investigates how adaptive recurrent neural networks can help vision models solve difficult visual reasoning problems by adjusting their computational resources based on the input's needs.


Generalized Belief Transport

Junqi Wang, PEI WANG, Patrick Shafto

https://openreview.net/forum?id=iohoef1bfM

Keywords: Bayesian theory, Belief transport, Unbalanced optimal transport, parametrization, asymptotic behavior, environment drift detection

Compressor summary: The paragraph discusses a new mathematical framework called Generalized Belief Transport that unifies different learning models by considering them as points in a broader space of possibilities, and explores its properties and behavior.


History Filtering in Imperfect Information Games: Algorithms and Complexity

Christopher Solinas, Doug Rebstock, Nathan R. Sturtevant, Michael Buro

https://openreview.net/forum?id=inIONNg8Sq

Keywords: search, game theory, multi-agent, learning, markov chain monte carlo, complexity

Compressor summary: The authors analyze the computational aspects of filtering histories for subgame decomposition, a key technique for AI in imperfect information games, and introduce a novel Markov Chain Monte Carlo-based generation algorithm for trick-taking card games.


AIMS: All-Inclusive Multi-Level Segmentation for Anything

Lu Qi, Jason Kuen, Weidong Guo, Jiuxiang Gu, Zhe Lin, Bo Du, Yu Xu, Ming-Hsuan Yang

https://openreview.net/forum?id=ikkdTD3hQJ

Keywords: Image Segmentation

Compressor summary: The paper introduces AIMS, a new task for image segmentation that divides visual regions into three levels, and proposes a unified model to tackle challenges like annotation inconsistency and task correlation.


Semi-Supervised Contrastive Learning for Deep Regression with Ordinal Rankings from Spectral Seriation

Weihang Dai, Yao DU, Hanru Bai, Kwang-Ting Cheng, Xiaomeng Li

https://openreview.net/forum?id=ij3svnPLzG

Keywords: Semi-supervised learning, deep regression, contrastive learning

Compressor summary: The authors propose a new semi-supervised contrastive learning method for deep regression that uses unlabeled data and achieves better results than existing methods.


Frequency-domain MLPs are More Effective Learners in Time Series Forecasting

Kun Yi, Qi Zhang, Wei Fan, Shoujin Wang, Pengyang Wang, Hui He, Ning An, Defu Lian, Longbing Cao, Zhendong Niu

https://openreview.net/forum?id=iif9mGCTfy

Keywords: time series forecasting, multi-layer perceptrons, frequency domain

Compressor summary: The paper proposes FreTS, a time series forecasting method that uses frequency-domain MLPs to learn global dependencies and concentrate on key parts of frequency components, outperforming existing methods.


GNNEvaluator: Evaluating GNN Performance On Unseen Graphs Without Labels

Xin Zheng, Miao Zhang, Chunyang Chen, Soheila Molaei, Chuan Zhou, Shirui Pan

https://openreview.net/forum?id=ihlT8yvQ2I

Keywords: graph neural networks, GNN model evaluation, node classification accuracy

Compressor summary: The paper proposes a two-stage framework to evaluate graph neural network (GNN) models by estimating their performance on unseen graphs without labels using a DiscGraph set and a GNNEvaluator.


Maximum Independent Set: Self-Training through Dynamic Programming

Lorenzo Brusca, Lars C.P.M. Quaedvlieg, Stratis Skoulakis, Grigorios Chrysos, Volkan Cevher

https://openreview.net/forum?id=igE3Zbxvws

Keywords: Maximum Independent Set, Combinatorial Optimization, Graph Neural Networks, Dynamic Programming

Compressor summary: The paper introduces a GNN-based algorithm for solving the MIS problem using a DP-like recursive approach and shows its effectiveness on various datasets.


EvoPrompting: Language Models for Code-Level Neural Architecture Search

Angelica Chen, David Dohan, David So

https://openreview.net/forum?id=ifbF4WdT8f

Keywords: language models, evolution, prompting, neural architecture search, code generation

Compressor summary: The paper explores using language models as mutation and crossover operators for evolutionary neural architecture search (NAS), proposing EvoPrompting which combines prompt engineering with soft prompt-tuning, and shows its effectiveness in finding diverse and high performing models on various tasks.


Interpretable Prototype-based Graph Information Bottleneck

Sangwoo Seo, Sungwon Kim, Chanyoung Park

https://openreview.net/forum?id=icWwBKyVMs

Keywords: Graph neural network, Explainable AI, Interpretability

Compressor summary: The paper introduces interpretable Prototype-based Graph Information Bottleneck (PGIB), a novel framework for explainable GNNs that learns prototypes from key subgraphs important for predictions, improving both performance and interpretability.


DELIFFAS: Deformable Light Fields for Fast Avatar Synthesis

YoungJoong Kwon, Lingjie Liu, Henry Fuchs, Marc Habermann, Christian Theobalt

https://openreview.net/forum?id=iajxrSgOSX

Keywords: DELIFFAS: Avatar Modeling, Avatar Synthesis, Animatable Human, Light Fields, Human Performance Capture

Compressor summary: The DELIFFAS method creates realistic human avatars with fast inference speed by using a surface light field attached to a deformable mesh model, achieving state-of-the-art results in appearance synthesis and speed.


Similarity-based cooperative equilibrium

Caspar Oesterheld, Johannes Treutlein, Roger Baker Grosse, Vincent Conitzer, Jakob Nicolaus Foerster

https://openreview.net/forum?id=ia4AL3QnOv

Keywords: Program equilibrium, multi-agent learning, game theory, opponent shaping, superrationality, decision theory, cooperative AI, Newcomb's problem

Compressor summary: The paper proposes a realistic setting in which machine learning agents can achieve cooperative outcomes in social dilemmas by only observing a single number indicating their similarity, and shows that this can be learned using simple ML methods.


PCF-GAN: generating sequential data via the characteristic function of measures on the path space

Hang Lou, Siran Li, Hao Ni

https://openreview.net/forum?id=iWWLgcUTZU

Keywords: Generative adversarial networks, time series generation, rough path theory, Lie group

Compressor summary: The paragraph introduces PCF-GAN, a novel GAN that uses path characteristic functions to represent time series distributions and improve their generation and reconstruction quality.


Provable benefits of annealing for estimating normalizing constants: Importance Sampling, Noise-Contrastive Estimation, and beyond

Omar Chehab, Aapo Hyvarinen, Andrej Risteski

https://openreview.net/forum?id=iWGC0Nsq9i

Keywords: noise-contrastive estimation, annealed importance sampling

Compressor summary: The paragraph discusses different design choices for Monte Carlo methods that estimate the normalization constant using annealing and their impact on estimation error, showing that some choices are more efficient than others depending on the context.


On the Identifiability and Interpretability of Gaussian Process Models

Jiawen Chen, Wancen Mu, Yun Li, Didong Li

https://openreview.net/forum?id=iVYInarGXg

Keywords: Gaussian process, Identifiability, Interpretability, Mixture kernel, Separable kernel

Compressor summary: The paper analyzes additive and multiplicative mixtures of Mat\'ern kernels in single- and multi-output Gaussian process models, showing that the latter are more suitable for multi-output tasks due to better identifiability properties.


Adversarial Training from Mean Field Perspective

Soichiro Kumano, Hiroshi Kera, Toshihiko Yamasaki

https://openreview.net/forum?id=iT9MOAZqsb

Keywords: adversarial training; mean field theory

Compressor summary: The study analyzes adversarial training in random deep neural networks using a new mean field theory framework, showing its effects on network capacity, width, and dimensions.


A Trichotomy for Transductive Online Learning

Steve Hanneke, Shay Moran, Jonathan Shafer

https://openreview.net/forum?id=iSd8g75QvP

Keywords: Online Learning, Transductive Online Learning, Offline Learning, Mistake Bound

Compressor summary: The paper presents new bounds on learner mistakes in transductive online learning, showing a trichotomy of possible mistake rates depending on VC and Littlestone dimensions.


Uncertainty-Aware Alignment Network for Cross-Domain Video-Text Retrieval

Xiaoshuai Hao, Wanqian Zhang

https://openreview.net/forum?id=iQlK3VJxV7

Keywords: video-text retrieval; cross-domain;Unsupervised Domain Adaptation Video-text Retrieval;

Compressor summary: The paper presents UAN, a novel method to improve unsupervised domain adaptation video-text retrieval by addressing multimodal mutual information and uncertainty-aware alignment.


Learning To Dive In Branch And Bound

Max B. Paulus, Andreas Krause

https://openreview.net/forum?id=iPTF2hON1C

Keywords: Generative Modeling, Combinatorial Optimization, Mixed Integer Programming, Graph Neural Networks, Diving Heuristics

Compressor summary: L2Dive is a diving heuristic that uses graph neural networks to predict variable assignments and improve the performance of mixed integer linear programs on various combinatorial optimization problems.


Parallel Submodular Function Minimization

Deeparnab Chakrabarty, Andrei Graur, Haotian Jiang, Aaron Sidford

https://openreview.net/forum?id=iMfPFPMsZo

Keywords: parallel computation, convex optimization, submodular function minimization

Compressor summary: The paragraph discusses two new methods for parallel submodular function minimization with different trade-offs between depth and query complexity, and also introduces a highly-parallel algorithm for minimizing an $\ell_\infty$-Lipschitz function over the hypercube.


A Unified Model and Dimension for Interactive Estimation

Nataly Brukhim, Miroslav Dudík, Aldo Pacchiano, Robert E. Schapire

https://openreview.net/forum?id=iM0MWWBr4W

Keywords: Interactive learning, bandits, statistical queries

Compressor summary: The paper proposes a new interactive learning framework with a combinatorial measure called Dissimilarity dimension that captures learnability and unifies two classic models, with improved analyses.


Bicriteria Approximation Algorithms for the Submodular Cover Problem

Wenjing Chen, Victoria G. Crawford

https://openreview.net/forum?id=iKarSI2a73

Keywords: submodular, combinatorial optimization, approximation algorithms

Compressor summary: The paper presents scalable algorithms for solving submodular cover problems with different variants and applications in data summarization and graph cut.


Joint Prompt Optimization of Stacked LLMs using Variational Inference

Alessandro Sordoni, Xingdi Yuan, Marc-Alexandre Côté, Matheus Pereira, Adam Trischler, Ziang Xiao, Arian Hosseini, Friederike Niedtner, Nicolas Le Roux

https://openreview.net/forum?id=iImnbUVhok

Keywords: deep prompt optimization, llm, variational inference, graphical model, chaining

Compressor summary: The text describes how to optimize prompts for two-layer deep language networks (DLNs) using latent variable inference and shows their potential to match or surpass GPT-4's performance with smaller models.


Simplifying Neural Network Training Under Class Imbalance

Ravid Shwartz-Ziv, Micah Goldblum, Yucen Lily Li, C. Bayan Bruss, Andrew Gordon Wilson

https://openreview.net/forum?id=iGmDQn4CRj

Keywords: Class Imbalance, Hyperparameters, Long-Tailed Distributions

Compressor summary: The authors show that by adjusting some standard deep learning components, one can achieve excellent results on class-imbalanced datasets without using specialized techniques.


Stochastic Collapse: How Gradient Noise Attracts SGD Dynamics Towards Simpler Subnetworks

Feng Chen, Daniel Kunin, Atsushi Yamamura, Surya Ganguli

https://openreview.net/forum?id=iFxWrxDekd

Keywords: Implicit Bias, SGD Dynamics, Implicit regularization, Learning rate schedule, Stochastic Gradient Descent, Invariant set, Attractive saddle points, Stochastic collapse, Permutation invariance, Simplicity bias, Teacher-student

Compressor summary: The paper shows that stochastic gradient descent (SGD) often simplifies neural networks during training by converging to simpler subnetworks, improving generalization.


MultiMoDN—Multimodal, Multi-Task, Interpretable Modular Networks

Vinitra Swamy, Malika Satayeva, Jibril Frej, Thierry Bossy, Thijs Vogels, Martin Jaggi, Tanja Käser, Mary-Anne Hartley

https://openreview.net/forum?id=iB3Ew6z4WL

Keywords: Deep Learning, Multimodal Learning, Multi-task learning, Missingness, Interpretability

Compressor summary: MultiModN is a multimodal network that fuses latent representations in a sequence of any modality type, providing interpretable feedback on real-world tasks while being resistant to missing not-at-random data.


Sharp Bounds for Generalized Causal Sensitivity Analysis

Dennis Frauen, Valentyn Melnychuk, Stefan Feuerriegel

https://openreview.net/forum?id=iAcEmyhwk2

Keywords: Causal machine learning, treatment effect estimation, sensitivity analysis, unobserved confounding, uncertainty estimation

Compressor summary: The paper presents a general framework for causal sensitivity analysis under unobserved confounding, extending the marginal sensitivity model to various types of treatments and outcomes, and providing sharp bounds for different causal effects using observational data.


CAST: Cross-Attention in Space and Time for Video Action Recognition

Dongho Lee, Jongseo Lee, Jinwoo Choi

https://openreview.net/forum?id=iATY9W5Xw7

Keywords: action recognition, video understanding, cross-attention, balanced spatio-temporal understanding

Compressor summary: Recognizing human actions in videos requires spatial and temporal understanding. Most existing action recognition models lack a balanced spatio-temporal understanding of videos. In this work, we propose a novel two-stream architecture, called Cross-Attention in Space and Time (CAST), that achieves a balanced spatio-temporal understanding of videos using only RGB input. Our proposed bottleneck cross-attention mechanism enables the spatial and temporal expert models to exchange information and make synergistic predictions, leading to improved performance. We validate the proposed method with extensive experiments on public benchmarks with different characteristics: EPIC-Kitchens-100, Something-Something-V2, and Kinetics-400. Our method consistently shows favorable performance across these datasets, while the performance of existing methods fluctuates depending on the dataset characteristics. The code is available at https://github.com/KHU-VLL/CAST.


Geodesic Multi-Modal Mixup for Robust Fine-Tuning

Changdae Oh, Junhyuk So, Hoyoon Byun, YongTaek Lim, Minchul Shin, Jong-June Jeon, Kyungwoo Song

https://openreview.net/forum?id=iAAXq60Bw1

Keywords: multi-modal learning, robustness, fine-tuning, contrastive learning, CLIP, Mixup

Compressor summary: The authors study CLIP's multi-modal embeddings and propose a fine-tuning method to improve alignment and uniformity for better transferability and robustness.


Cinematic Mindscapes: High-quality Video Reconstruction from Brain Activity

Zijiao Chen, Jiaxin Qing, Juan Helen Zhou

https://openreview.net/forum?id=i913TUOvTK

Keywords: Video Reconstruction from Brain Activities, Diffusion Model, Contrastive Learning

Compressor summary: The paragraph describes a new method called Mind-Video that can reconstruct high-quality videos from brain activities using various techniques and outperforms previous methods in accuracy and interpretability.


ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision Transformer

Haoran You, Huihong Shi, Yipin Guo, Yingyan Celine Lin

https://openreview.net/forum?id=i6mMWNcTfu

Keywords: Efficient Vision Transformer; Multiplication-reduced networks; Hardware acceleration

Compressor summary: The authors propose a new multiplication-reduced model called ShiftAddViT, which uses bitwise shifts and additions instead of dense multiplications to speed up ViTs on GPUs without losing much accuracy.


Decentralized Matrix Sensing: Statistical Guarantees and Fast Convergence

Marie Maros, Gesualdo Scutari

https://openreview.net/forum?id=i5sSWKbF3b

Keywords: Distributed non-convex optimization, Low-rank matrix recovery

Compressor summary: The paper studies how agents in a network can estimate a low-rank matrix using a decentralized gradient algorithm and shows that it converges to the correct solution with good performance guarantees.


Segment Any Point Cloud Sequences by Distilling Vision Foundation Models

Youquan Liu, Lingdong Kong, Jun CEN, Runnan Chen, Wenwei Zhang, Liang Pan, Kai Chen, Ziwei Liu

https://openreview.net/forum?id=i39yXaUKuF

Keywords: autonomous driving, point cloud segmentation, self-supervised learning, 3D scene understanding

Compressor summary: Seal is a novel framework that leverages vision foundation models for efficient and versatile segmentation of diverse automotive point clouds, with properties such as scalability, consistency, and generalizability.


A Unified Fast Gradient Clipping Framework for DP-SGD

Weiwei Kong, Andres Munoz medina

https://openreview.net/forum?id=i2H2sEiq2T

Keywords: differential privacy, dp-sgd, gradient clipping, computational complexity

Compressor summary: The paper proposes a framework that enables fast norm computations for gradients in DP-SGD with arbitrary intermediate operations and shows improvements for specific operations using components of the framework.


GloptiNets: Scalable Non-Convex Optimization with Certificates

Gaspard Beugnot, Julien Mairal, Alessandro Rudi

https://openreview.net/forum?id=i28zCSsQIc

Keywords: non-convex optimization, polynomial optimization, kernel sum-of-squares

Compressor summary: The authors propose a novel optimization method that leverages the regularity of smooth functions in the Fourier spectrum and neural network techniques, leading to better performance and scalability.


State-space models with layer-wise nonlinearity are universal approximators with exponential decaying memory

Shida Wang, Beichen Xue

https://openreview.net/forum?id=i0OmcF14Kf

Keywords: state space, approximation theory, sequence modelling

Compressor summary: The paper shows how adding nonlinear activation to state-space models improves their ability to learn complex sequence patterns, but does not solve the problem of exponential decaying memory.


Learning to Influence Human Behavior with Offline Reinforcement Learning

Joey Hong, Sergey Levine, Anca Dragan

https://openreview.net/forum?id=hzND3ZEFg2

Keywords: offline reinforcement learning, human-aware reinforcement learning, multi-agent influence

Compressor summary: The text discusses how AI agents can use offline reinforcement learning to influence suboptimal humans towards better performance by learning from a dataset of human-human interactions.


CLeAR: Continual Learning on Algorithmic Reasoning for Human-like Intelligence

Bong Gyun Kang, HyunGi Kim, Dahuin Jung, Sungroh Yoon

https://openreview.net/forum?id=hz33V7Tb2O

Keywords: Continual learning, Knowledge transfer, Algorithmic reasoning

Compressor summary: CLeAR is a novel algorithmic reasoning methodology that enables continual learning of abstract logical concepts with minimal forgetting and backward transfer.


CWCL: Cross-Modal Transfer with Continuously Weighted Contrastive Loss

Rakshith Sharma Srinivasa, Jaejin Cho, Chouchang Yang, Yashas Malur Saidutta, Ching-Hua Lee, Yilin Shen, Hongxia Jin

https://openreview.net/forum?id=hz10oiVMNE

Keywords: contrastive loss, multimodal representation learning, zero-shot learning, intent classification, pre-trained models, modality alignment, cross-modal transfer

Compressor summary: The paper introduces CWCL, a new contrastive loss function that improves cross-modal 0-shot transfer by using a continuous measure of similarity, achieving better results than existing methods on various tasks and datasets.


A polar prediction model for learning to represent visual transformations

Pierre-Étienne H Fiquet, Eero P Simoncelli

https://openreview.net/forum?id=hyPUZX03Ks

Keywords: video prediction, neural coding, symmetry discovery, self-supervised representation-learning

Compressor summary: The paper proposes a self-supervised learning method that uses natural video patterns to predict future frames and discover simple transformations, achieving comparable performance to conventional deep networks while being interpretable and fast.


Focus on Query: Adversarial Mining Transformer for Few-Shot Segmentation

Yuan Wang, Naisong Luo, Tianzhu Zhang

https://openreview.net/forum?id=hxJu0386if

Keywords: Computer Vision, Few-shot Segmentation

Compressor summary: The paper introduces AMFormer, a new few-shot segmentation model that focuses on query-centric learning and achieves state-of-the-art results with weak support labels.


Analyzing Vision Transformers for Image Classification in Class Embedding Space

Martina G. Vilas, Timothy Schaumlöffel, Gemma Roig

https://openreview.net/forum?id=hwjmEZ8561

Keywords: transformers, computer vision, image classification, mechanistic interpretability, explainability

Compressor summary: This paper presents a method to analyze how Vision Transformers learn image representations for classification tasks and use attention and context to build categorical features.


NPCL: Neural Processes for Uncertainty-Aware Continual Learning

Saurav Jha, Dong Gong, He Zhao, Lina Yao

https://openreview.net/forum?id=huh0XmSdBK

Keywords: Continual Learning, Neural Process, Uncertainty, Incremental Learning

Compressor summary: The paper proposes a new approach to continuous learning using neural processes, which can handle tasks with less interference, reduce forgetting, and provide reliable uncertainty estimates for deep neural networks.


$p$-value Adjustment for Monotonous, Unbiased, and Fast Clustering Comparison

Kai Klede, Thomas Altstidl, Dario Zanca, Bjoern Eskofier

https://openreview.net/forum?id=htkdwc6jDB

Keywords: Clustering, (Other) Machine Learning Topics

Compressor summary: The paper introduces a new clustering comparison method called p-value adjusted Rand Index (PMI_2) that is unbiased, monotonous, and efficient for selecting better clustering and community detection algorithms.


AMDP: An Adaptive Detection Procedure for False Discovery Rate Control in High-Dimensional Mediation Analysis

Jiarong Ding, Xuehu Zhu

https://openreview.net/forum?id=htM8yp2EwX

Keywords: Mediation analysis, Composite null hypothesis, Local false discovery rate, Optimal ranking rule, High-dimensional

Compressor summary: The paper introduces a new method for detecting significant mediators in high-dimensional mediation analysis that controls the false discovery rate by ranking and selecting hypotheses based on p-values and composite null probabilities.


UniPC: A Unified Predictor-Corrector Framework for Fast Sampling of Diffusion Models

Wenliang Zhao, Lujia Bai, Yongming Rao, Jie Zhou, Jiwen Lu

https://openreview.net/forum?id=hrkmlPhp1u

Keywords: diffusion models, fast sampling, predictor-corrector, training-free

Compressor summary: The paper introduces UniPC, a framework for fast and high-quality sampling of diffusion probabilistic models in image synthesis, improving on previous methods especially with few steps.


Boosting Verification of Deep Reinforcement Learning via Piece-Wise Linear Decision Neural Networks

Jiaxu Tian, Dapeng Zhi, Si Liu, Peixin Wang, Cheng Chen, Min Zhang

https://openreview.net/forum?id=hpYb4eUinX

Keywords: Deep reinforcement learning, Reachability analysis, Hybrid system, State abstraction

Compressor summary: The paper proposes a new method to verify deep reinforcement learning systems using piece-wise linear decision neural networks, which can reduce overestimation and speed up verification by up to 438 times.


Macro Placement by Wire-Mask-Guided Black-Box Optimization

Yunqi Shi, Ke Xue, Lei Song, Chao Qian

https://openreview.net/forum?id=hoyL1Ypjoo

Keywords: Black Box Optimization, Macro Placement, Electronic Design Automation, Reinforcement Learning, Application

Compressor summary: The paper proposes a new optimization framework, WireMask-BBO, for improving chip floorplanning by minimizing wirelength and avoiding overlapping, achieving significant improvements in quality and efficiency over previous methods.


Computing Approximate $\ell_p$ Sensitivities

Swati Padmanabhan, David Woodruff, Qiuyi Zhang

https://openreview.net/forum?id=hn1oJO7lg6

Keywords: $\ell_p$ sensitivities, Lewis weights, leverage scores, approximation algorithms, total sensitivity

Compressor summary: The paper introduces efficient algorithms for computing $\ell_p$ sensitivities and other summary statistics of a matrix, which can help reduce dimensionality in regression tasks and improve data quality after removing low-sensitivity points.


Global-correlated 3D-decoupling Transformer for Clothed Avatar Reconstruction

Zechuan Zhang, Li Sun, Zongxin Yang, Ling Chen, Yi Yang

https://openreview.net/forum?id=hlkhPdhuAO

Keywords: 3D human avatar reconstruction, vision transformer, parametric body model, tri-plane representation

Compressor summary: The GTA model uses a transformer-based architecture to reconstruct 3D clothed human avatars from monocular images, overcoming limitations of previous methods by capturing global-correlated features and using a hybrid prior fusion strategy.


Fantastic Weights and How to Find Them: Where to Prune in Dynamic Sparse Training

Aleksandra Nowak, Bram Grooten, Decebal Constantin Mocanu, Jacek Tabor

https://openreview.net/forum?id=hkPn7M9k1W

Keywords: Dynamic Sparse Training, Pruning, Deep Learning

Compressor summary: The paragraph discusses the importance of studying pruning criteria for Dynamic Sparse Training (DST) and how simplicity works best in the low-density regime.


Iteratively Learn Diverse Strategies with State Distance Information

Wei Fu, Weihua Du, Jingwei Li, Sunli Chen, Jingzhao Zhang, Yi Wu

https://openreview.net/forum?id=hiwF7aG1dt

Keywords: diverse behavior, multi-agent reinforcement learning, deep reinforcement learning

Compressor summary: The paper proposes a new diversity-driven reinforcement learning algorithm (SIPO) that incorporates state-space distance information and shows improved performance in discovering diverse and interpretable strategies across various domains.


Test-Time Amendment with a Coarse Classifier for Fine-Grained Classification

Kanishk Jain, Shyamgopal Karthik, Vineet Gandhi

https://openreview.net/forum?id=hiQG8qGxso

Keywords: Hierarchical Classification, Fine-grained Classification, Ensembles, Mistake Severity

Compressor summary: The paper proposes HiE, a method that uses label hierarchy to correct coarse-grained predictions and improve fine-grained classification performance, achieving state-of-the-art results and practical applicability.


Learning Topology-Agnostic EEG Representations with Geometry-Aware Modeling

Ke Yi, Yansen Wang, Kan Ren, Dongsheng Li

https://openreview.net/forum?id=hiOUySN0ub

Keywords: Electroencephalogram, EEG Pre-training, EEG-based emotion recognition

Compressor summary: The authors propose a new EEG pre-training framework, MMM, which uses a unified channel topology and multi-dimensional position encoding to improve performance on emotional recognition tasks.


Path following algorithms for $\ell_2$-regularized $M$-estimation with approximation guarantee

Yunzhang Zhu, Renxiong Liu

https://openreview.net/forum?id=hgLMht2Z3L

Keywords: regularization, optimization, tuning parameter selection

Compressor summary: The paper proposes a method for choosing grid points and an adaptive stopping criterion to efficiently compute approximation solution paths for regularized M-estimation problems in machine learning.


Taylor TD-learning

Michele Garibbo, Maxime Robeyns, Laurence Aitchison

https://openreview.net/forum?id=hcXDbbzgoh

Keywords: Reinforcement learning, TD-learning, model-based, variance reduction

Compressor summary: Taylor TD is a model-based RL framework that reduces variance in continuous state-action settings by using a first-order Taylor series expansion of TD updates, integrating over stochasticity, and providing stable learning guarantees.


Exact Generalization Guarantees for (Regularized) Wasserstein Distributionally Robust Models

Waïss Azizian, Franck Iutzeler, Jérôme Malick

https://openreview.net/forum?id=haniyY7zm1

Keywords: robust optimization, distributionally robust optimization, optimization under uncertainty, generalization, Wasserstein distance, optimal transport

Compressor summary: Wasserstein distributionally robust estimators are powerful models for prediction and decision-making under uncertainty with improved generalization guarantees on various classes of models without the curse of dimensionality or distribution shifts.


$SE(3)$ Equivariant Convolution and Transformer in Ray Space

Yinshuang Xu, Jiahui Lei, Kostas Daniilidis

https://openreview.net/forum?id=haHIji0yFt

Keywords: equivariance, light field, equivariant convolution over homogeneous space

Compressor summary: The paper proposes an SE(3)-equivariant convolution and transformer for learning geometric priors from multiple views, which can improve 3D reconstruction and novel view rendering tasks.


InsActor: Instruction-driven Physics-based Characters

Jiawei Ren, Mingyuan Zhang, Cunjun Yu, Xiao Ma, Liang Pan, Ziwei Liu

https://openreview.net/forum?id=hXevuspQnX

Keywords: Physics-based Animation; Human Motion Generation

Compressor summary: InsActor is a framework that uses diffusion policies and skill discovery to generate physics-based character animations from high-level human instructions.


A new perspective on building efficient and expressive 3D equivariant graph neural networks

weitao Du, Yuanqi Du, Limei Wang, Dieqiao Feng, Guifeng Wang, Shuiwang Ji, Carla P Gomes, Zhi-Ming Ma

https://openreview.net/forum?id=hWPNYWkYPN

Keywords: Graph Neural Networks, Geometric Deep Learning, Equivariance, Symmetry

Compressor summary: The paper introduces local hierarchy of 3D isomorphism to evaluate and improve the expressiveness of equivariant GNNs for 3D object modeling, and presents LEFTNet that achieves state-of-the-art results on molecular property prediction tasks.


Hardware Resilience Properties of Text-Guided Image Classifiers

Syed Talal Wasim, Kabila Haile Soboka, Abdulrahman Mahmoud, Salman Khan, David Brooks, Gu-Yeon Wei

https://openreview.net/forum?id=hVVp8TXIPs

Keywords: Hardware Resilience, Reliability, Image Classification, CLIP, Vision-Language, Multimodal

Compressor summary: The paper proposes a method to improve image classification model reliability using GPT-3 text embeddings as an initialization for the classification layer, resulting in significant hardware reliability improvements with minimal accuracy loss.


A Pseudo-Semantic Loss for Autoregressive Models with Logical Constraints

Kareem Ahmed, Kai-Wei Chang, Guy Van den Broeck

https://openreview.net/forum?id=hVAla2O73O

Keywords: Neuro-Symbolic, Tractable Models, Learning with Constraints

Compressor summary: The authors propose a neuro-symbolic method for enforcing constraints on auto-regressive distributions that approximates the likelihood with pseudolikelihood, improving performance on Sudoku, shortest-path prediction, and detoxifying language models.


A Hierarchical Training Paradigm for Antibody Structure-sequence Co-design

Fang Wu, Stan Z. Li

https://openreview.net/forum?id=hV52oj0Sik

Keywords: Antibody Design

Compressor summary: The paper proposes HTP, a hierarchical training paradigm that uses GNNs and protein language models to design therapeutic antibodies with desired binding specificity by incorporating evolutionary information from various databases.


Adversarial Examples Are Not Real Features

Ang Li, Yifei Wang, Yiwen Guo, Yisen Wang

https://openreview.net/forum?id=hSkEcIFi3o

Keywords: Adversarial Examples, Adversarial Training, Generalization, Generative Models

Compressor summary: The paper re-examines the theory of adversarial examples by showing that non-robust features are not as useful or transferable across different learning paradigms, suggesting they are more like shortcuts and robustness alone is insufficient for model reliability.


Imagine That! Abstract-to-Intricate Text-to-Image Synthesis with Scene Graph Hallucination Diffusion

Shengqiong Wu, Hao Fei, Hanwang Zhang, Tat-Seng Chua

https://openreview.net/forum?id=hSTaTBIUCj

Keywords: Image Synthesis, Scene Graph, Diffusion Model

Compressor summary: The text describes a new method for generating detailed images from simple text prompts using scene graph hallucination and diffusion models.


Rubik's Cube: High-Order Channel Interactions with a Hierarchical Receptive Field

Naishan Zheng, Man Zhou, Chong Zhou, Chen Change Loy

https://openreview.net/forum?id=hOOOvOMok5

Keywords: Image restoration, low-light image enhancement, image de-noising

Compressor summary: The paper introduces a new high-order channel-wise convolution operator called Rubik's cube convolution that improves image restoration performance efficiently and simply by expanding first-order interactions to arbitrary high orders.


Near Optimal Reconstruction of Spherical Harmonic Expansions

Amir Zandieh, Insu Han, Haim Avron

https://openreview.net/forum?id=hNpedVWwoe

Keywords: spherical harmonic expansion, Gegenbauer polynomials, interpolation, leverage score sampling

Compressor summary: The paragraph describes an efficient algorithm to recover the spherical harmonic expansion of functions defined on the unit sphere using a near-optimal number of function evaluations and kernel regression.


Game Solving with Online Fine-Tuning

Ti-Rong Wu, Hung Guei, Ting Han Wei, Chung-Chin Shih, Jui-Te Chin, I-Chen Wu

https://openreview.net/forum?id=hN4qpvGzWn

Keywords: Game Solving, Computer Games, AlphaZero, Online Fine-Tuning, Monte Carlo Tree Search, Deep Reinforcement Learning

Compressor summary: The paper proposes using online fine-tuning and tailor-designed heuristics to improve the AlphaZero algorithm's performance in solving games, especially those with poor lines of play from the losing side.


Parts of Speech–Grounded Subspaces in Vision-Language Models

James Oldfield, Christos Tzelepis, Yannis Panagakis, Mihalis Nicolaou, Ioannis Patras

https://openreview.net/forum?id=hLoanbRrjM

Keywords: generative models, text-to-image, vision-language models, interpretability

Compressor summary: The paper proposes a method to separate different visual properties in CLIP's image representations using part-of-speech associations and apply it to text-to-image synthesis and style transfer tasks.


Self-Supervised Motion Magnification by Backpropagating Through Optical Flow

Zhaoying Pan, Daniel Geng, Andrew Owens

https://openreview.net/forum?id=hLPJ7xLbNF

Keywords: Video Processing, Motion Processing, Motion Magnification, Optical Flow

Compressor summary: The paper introduces a self-supervised technique to magnify subtle motions in video by adjusting the optical flow based on a given factor and training a model with a loss function that minimizes the deviation from the desired magnification.


Unified Embedding: Battle-Tested Feature Representations for Web-Scale ML Systems

Benjamin Coleman, Wang-Cheng Kang, Matthew Fahrbach, Ruoxi Wang, Lichan Hong, Ed H. Chi, Derek Zhiyuan Cheng

https://openreview.net/forum?id=hJzEoQHfCe

Keywords: embedding learning; recommendation systems; representation learning

Compressor summary: The text introduces Feature Multiplexing, a framework that uses one representation space for multiple features, which improves embedding efficiency and effectiveness for web-scale machine learning systems.


Unleashing the Power of Graph Data Augmentation on Covariate Distribution Shift

Yongduo Sui, Qitian Wu, Jiancan Wu, Qing Cui, Longfei Li, JUN ZHOU, Xiang Wang, Xiangnan He

https://openreview.net/forum?id=hIGZujtOQv

Keywords: Graph Neural Network, Graph Data Augmentation, Distribution Shift

Compressor summary: The text discusses the problem of distribution shifts in graph representation learning, proposing a new data augmentation strategy called Adversarial Invariant Augmentation (AIA) to handle covariate shift on graphs.


Face Reconstruction from Facial Templates by Learning Latent Space of a Generator Network

Hatef Otroshi Shahreza, Sébastien Marcel

https://openreview.net/forum?id=hI6EPhq70A

Keywords: Face Recognition (FR), Face reconstruction, Generative Adversarial Network (GAN), Privacy, Security, Template Inversion (TI) attack, Transferability

Compressor summary: The paper proposes a method to attack face recognition systems by reconstructing realistic face images from facial templates using a GAN-based framework, which can be applied in different scenarios and is transferable between systems.


Block Broyden's Methods for Solving Nonlinear Equations

Chengchang Liu, Cheng Chen, Luo Luo, John C.S. Lui

https://openreview.net/forum?id=hHv3UuffXV

Keywords: Broyden's method, nonlinear equations

Compressor summary: The paper introduces block variants of good and bad Broyden's methods for nonlinear equation solving, with improved convergence rates and reduced computational cost compared to existing methods.


Equivariant Flow Matching with Hybrid Probability Transport for 3D Molecule Generation

Yuxuan Song, Jingjing Gong, Minkai Xu, Ziyao Cao, Yanyan Lan, Stefano Ermon, Hao Zhou, Wei-Ying Ma

https://openreview.net/forum?id=hHUZ5V9XFu

Keywords: Drug Design, Molecule Generation, Deep Learning, Computational Biology

Compressor summary: The paper introduces geometric flow matching, a hybrid probability path method that improves the stability and efficiency of diffusion models for generating 3D molecules.


Creating a Public Repository for Joining Private Data

James Cook, Milind Shyani, Nina Mishra

https://openreview.net/forum?id=hExFOGZTSt

Keywords: Differential privacy, machine learning, linear queries

Compressor summary: The paper proposes a non-interactive privacy-preserving method to publish datasets with sensitive attributes and enable joins with other datasets on those attributes, using private sketches and satisfying pure differential privacy.


A Finite-Sample Analysis of Payoff-Based Independent Learning in Zero-Sum Stochastic Games

Zaiwei Chen, Kaiqing Zhang, Eric Mazumdar, Asuman E. Ozdaglar, Adam Wierman

https://openreview.net/forum?id=hElNdYMs8Z

Keywords: Zero-sum stochastic games, payoff-based independent learning, best-response-type dynamics, finite-sample analysis

Compressor summary: The authors propose a novel learning dynamics for two-player stochastic games that combines matrix game and stochastic game methods, and provide theoretical guarantees on its convergence, rationality, and symmetry.


UniTSFace: Unified Threshold Integrated Sample-to-Sample Loss for Face Recognition

Qiufu Li, Xi Jia, Jiancan Zhou, Linlin Shen, Jinming Duan

https://openreview.net/forum?id=hE7PG1lUZx

Keywords: Face Recognition, Unified Threshold, USS Loss

Compressor summary: The paper proposes a new loss function (USS loss) for face recognition that uses a unified threshold to distinguish positive from negative facial pairs and improves performance compared to existing methods.


Distributionally Robust Linear Quadratic Control

Bahar Taskesen, Dan Andrei Iancu, Çağıl Koçyiğit, Daniel Kuhn

https://openreview.net/forum?id=hE5RWzQyvf

Keywords: linear quadratic control, distributionally robust optimization, optimal transport, Wasserstein distance

Compressor summary: The paper proposes an optimal control policy for a generalization of the discrete-time, finite-horizon LQG problem with unknown noise distributions, using the Frank-Wolfe algorithm and Kalman filter estimation.


Online Adaptive Policy Selection in Time-Varying Systems: No-Regret via Contractive Perturbations

Yiheng Lin, James A Preiss, Emile Timothy Anand, Yingying Li, Yisong Yue, Adam Wierman

https://openreview.net/forum?id=hDajsofjRM

Keywords: Online policy selection, online control, online learning

Compressor summary: The GAPS algorithm optimizes online policy selection in systems with varying costs and dynamics by approximating an ideal gradient descent algorithm, achieving optimal or local regret depending on convexity, and outperforming current methods in adaptability.


Knowledge Distillation for High Dimensional Search Index

Zepu Lu, Jin Chen, Defu Lian, ZAIXI ZHANG, Yong Ge, Enhong Chen

https://openreview.net/forum?id=hCg4w8L8Dt

Keywords: Approximate Nearest Neighbor Search, Knowledge Distillation, Product Quantization, Inverted Index

Compressor summary: The paper proposes a new learning algorithm, KDindex, that improves retrieval performance of compressed search indices by distilling knowledge from high-precision models and using additional supervision signals between queries and documents.


Optimal Guarantees for Algorithmic Reproducibility and Gradient Complexity in Convex Optimization

Liang Zhang, Junchi YANG, Amin Karbasi, Niao He

https://openreview.net/forum?id=hCdqDkA25J

Keywords: Reproducibility, Convex Optimization, Minimax Optimization, Saddle-Point Problem

Compressor summary: The paper shows how to achieve both optimal reproducibility and near-optimal convergence for machine learning algorithms under various error-prone oracles, challenging the previous perception of a trade-off between these factors.


On the Generalization Properties of Diffusion Models

Puheng Li, Zhong Li, Huishuai Zhang, Jiang Bian

https://openreview.net/forum?id=hCUG1MCFk5

Keywords: diffusion models, training dynamics, generalization gap, modes shift

Compressor summary: This paper explores the generalization capabilities of diffusion models and provides theoretical and empirical evidence for their performance on different scenarios, such as early stopping and mode shifts.


Learning Modulated Transformation in GANs

Ceyuan Yang, Qihang Zhang, Yinghao Xu, Jiapeng Zhu, Yujun Shen, Bo Dai

https://openreview.net/forum?id=h8vJVABiBP

Keywords: generative adversarial network, image synthesis, video synthesis

Compressor summary: The paper introduces a new module (MTM) for GANs that can handle geometry deformation by predicting spatial offsets based on latent codes, improving performance on various generative tasks.


Towards robust and generalizable representations of extracellular data using contrastive learning

Ankit Vishnubhotla, Charlotte Loh, Akash Srivastava, Liam Paninski, Cole Lincoln Hurwitz

https://openreview.net/forum?id=h6WUKM7PCI

Keywords: contrastive learning, representations, extracellular, high-density, spike sorting, cell-type classification, transformers, invariance

Compressor summary: The authors propose a new contrastive learning framework, CEED, for analyzing high-density extracellular recordings in neuroscience and show that it outperforms current methods.


VideoComposer: Compositional Video Synthesis with Motion Controllability

Xiang Wang, Hangjie Yuan, Shiwei Zhang, Dayou Chen, Jiuniu Wang, Yingya Zhang, Yujun Shen, Deli Zhao, Jingren Zhou

https://openreview.net/forum?id=h4r00NGkjR

Keywords: Video Synthesis, Video Diffusion Model, Compositional Synthesis

Compressor summary: VideoComposer is a tool that enables users to create videos with various temporal conditions by using motion vectors from compressed videos and a Spatio-Temporal Condition encoder.


Large language models implicitly learn to straighten neural sentence trajectories to construct a predictive representation of natural language.

Eghbal A. Hosseini, Evelina Fedorenko

https://openreview.net/forum?id=h3lTrt4Ftb

Keywords: (Cognitive/Neuroscience) Language, Structured Prediction, (Application) Natural Language and Text Processing

Compressor summary: The authors investigate how predicting the next word in a sentence affects the way transformer models represent language, and find that these models tend to favor straighter trajectories for making predictions.


Front-door Adjustment Beyond Markov Equivalence with Limited Graph Knowledge

Abhin Shah, Karthikeyan Shanmugam, Murat Kocaoglu

https://openreview.net/forum?id=h3kuB4z2G9

Keywords: Causal Effect Estimation, Front-door Adjustment, Limited Graph Knowledge

Compressor summary: The text discusses a new method to estimate causal effects without knowing the graph structure, using testable conditional independence statements and front-door-like adjustment with limited side information.


Customizable Image Synthesis with Multiple Subjects

Zhiheng Liu, Yifei Zhang, Yujun Shen, Kecheng Zheng, Kai Zhu, Ruili Feng, Yu Liu, Deli Zhao, Jingren Zhou, Yang Cao

https://openreview.net/forum?id=h3QNH3qeC3

Keywords: text2img, diffusion model, customization

Compressor summary: This paper proposes an efficient method to synthesize images with multiple user-specified subjects using text embeddings and layout guidance, achieving better results than existing approaches.


Scattering Vision Transformer: Spectral Mixing Matters

Badri Narayana Patro, Vijay Srinivas Agneeswaran

https://openreview.net/forum?id=h3MShWMxNt

Keywords: Vision Transformer, Scatter Network, Spectral Transformer, Token Mixing, Channel Mixing, and Einstein Blending Method

Compressor summary: The Scattering Vision Transformer (SVT) is a novel approach that captures intricate image details and reduces attention complexity in vision transformers using spectrally scattering networks and spectral gating.


Multi-modal Queried Object Detection in the Wild

Yifan Xu, Mengdan Zhang, Chaoyou Fu, Peixian Chen, Xiaoshan Yang, Ke Li, Changsheng Xu

https://openreview.net/forum?id=h3CGHf7457

Keywords: open-world object detection, multi-modal query, vision-language pre-training

Compressor summary: MQ-Det is a method to improve object detection using both text and visual information, which can be easily applied to existing detectors and boosts performance on various tasks.


Faster Differentially Private Convex Optimization via Second-Order Methods

Arun Ganesh, MAHDI HAGHIFAM, Thomas Steinke, Abhradeep Guha Thakurta

https://openreview.net/forum?id=h2lkx9SQCD

Keywords: differential privacy, logistic regression, optimization, Newton's method, second-order methods

Compressor summary: The paper proposes a private version of Newton's method that can accelerate differentially private optimization, and tests it on logistic regression with promising results.


Improved Convergence in High Probability of Clipped Gradient Methods with Heavy Tailed Noise

Ta Duy Nguyen, Thien Hang Nguyen, Alina Ene, Huy Nguyen

https://openreview.net/forum?id=h1FhXVM0cB

Keywords: convex optimization, non-convex optimization, high probability convergence, heavy-tailed noise, clipped stochastic gradient descent, clipped stochastic mirror descent

Compressor summary: The authors propose a new analysis approach for clipped gradient methods with heavy-tailed noise that improves convergence guarantees, allows time-varying step sizes and clipping parameters, and does not need problem constants.


Resilient Constrained Learning

Ignacio Hounie, Alejandro Ribeiro, Luiz F. O. Chamon

https://openreview.net/forum?id=h0RVoZuUl6

Keywords: Constrained Learning, Relaxation, Lagrangian duality, Primal-Dual, Machine Learning, Federated Learning, Invariance

Compressor summary: The paper proposes a constrained learning approach that adapts requirements while solving the task, balancing performance gains against the cost of relaxing constraints, called resilient constrained learning.


Voicebox: Text-Guided Multilingual Universal Speech Generation at Scale

Matthew Le, Apoorv Vyas, Bowen Shi, Brian Karrer, Leda Sari, Rashel Moritz, Mary Williamson, Vimal Manohar, Yossi Adi, Jay Mahadeokar, Wei-Ning Hsu

https://openreview.net/forum?id=gzCS252hCO

Keywords: speech generation, flow-matching, diffusion, in-context learning, text-to-speech

Compressor summary: Voicebox is a versatile text-guided generative model for speech that can perform various tasks, such as zero-shot TTS, noise removal, and style conversion, with better performance and faster speed than VALL-E.


Query-based Temporal Fusion with Explicit Motion for 3D Object Detection

Jinghua Hou, Zhe Liu, dingkang liang, Zhikang Zou, Xiaoqing Ye, Xiang Bai

https://openreview.net/forum?id=gySmwdmVDF

Keywords: 3D Object Detection, Temporal, LiDAR-only, Multi-modality, Autonomous Driving

Compressor summary: The paper introduces QTNet, a simple and effective method for 3D detection in autonomous driving that exploits object queries in previous frames to enhance current ones using Motion-guided Temporal Modeling (MTM).


Emergent Communication for Rules Reasoning

Yuxuan Guo, Yifan Hao, Rui Zhang, Enshuai Zhou, Zidong Du, Xishan Zhang, Xinkai Song, Yuanbo Wen, Yongwei Zhao, Xuehai Zhou, Jiaming Guo, Qi Yi, Shaohui Peng, Di Huang, Ruizhi Chen, Qi Guo, Yunji Chen

https://openreview.net/forum?id=gx20B4ItIw

Keywords: Emergent communication, Multi-agent communication, Raven's Progressive Matrices, Representation learning

Compressor summary: The paper introduces a new game for training deep-learning-based agents to reason and communicate high-level rules, and presents a dataset and training method for this purpose.


Composable Coresets for Determinant Maximization: Greedy is Almost Optimal

Siddharth Gollapudi, Sepideh Mahabadi, Varun Sivashankar

https://openreview.net/forum?id=gwvwbsnTps

Keywords: Determinant Maximization, Composable Coresets, Greedy Algorithm, DPP

Compressor summary: The paper proposes and analyzes a Greedy algorithm for determinant maximization with an almost optimal approximation factor and shows its practicality and local optimality on real data sets.


Use perturbations when learning from explanations

Juyeon Heo, Vihari Piratla, Matthew Robert Wicker, Adrian Weller

https://openreview.net/forum?id=guyhQMSp2F

Keywords: Learning from explanation, Robustness, Interpretability, Shortcuts, Explanations

Compressor summary: The paper proposes a new approach to machine learning from explanations that improves performance by using robustness techniques instead of model smoothing.


NVFi: Neural Velocity Fields for 3D Physics Learning from Dynamic Videos

Jinxi Li, Ziyang Song, Bo Yang

https://openreview.net/forum?id=gsi9lJ3994

Keywords: Physics Learning, Velocity Field, Dynamic Radiance Field, Future Frame Extrapolation

Compressor summary: The paper proposes a method to learn geometry, appearance, and velocity of 3D scenes from videos for various applications such as future frame prediction and 3D scene analysis.


Flow-Based Feature Fusion for Vehicle-Infrastructure Cooperative 3D Object Detection

Haibao Yu, Yingjuan Tang, Enze Xie, Jilei Mao, Ping Luo, Zaiqing Nie

https://openreview.net/forum?id=gsglrhvQxX

Keywords: vehicle-infrastructure cooperative autonomous driving, 3D object detection, feature flow, self-supervised learning

Compressor summary: The Feature Flow Net (FFNet) is a novel cooperative detection framework for autonomous driving that predicts future features and uses feature flow transmission to address temporal asynchrony and limited communication conditions in traffic environments, achieving better performance with low transmission cost and latency.


Diffusion with Forward Models: Solving Stochastic Inverse Problems Without Direct Supervision

Ayush Tewari, Tianwei Yin, George Cazenavette, Semon Rezchikov, Joshua B. Tenenbaum, Fredo Durand, William T. Freeman, Vincent Sitzmann

https://openreview.net/forum?id=gq4xkwQZ1l

Keywords: 3D generative models, neural rendering, neural scene representations, NeRF, diffusion models, differentiable rendering, inverse graphics, inverse problems

Compressor summary: The text introduces a new type of generative models that can learn from indirect observations through a known forward model, enabling conditional generation of hidden signals and applications in computer vision tasks like inverse graphics.


Neural Relation Graph: A Unified Framework for Identifying Label Noise and Outlier Data

Jang-Hyun Kim, Sangdoo Yun, Hyun Oh Song

https://openreview.net/forum?id=gpyeRyc858

Keywords: Dataset cleaning, Label error detection, Outlier detection, Neural Networks, Robustness

Compressor summary: The paper proposes a method to identify and clean data problems using the relational structure of data in the feature-embedded space, and demonstrates its effectiveness on various image, speech, and language tasks.


Spectral Evolution and Invariance in Linear-width Neural Networks

Zhichao Wang, Andrew William Engel, Anand Sarwate, Ioana Dumitriu, Tony Chiang

https://openreview.net/forum?id=gpqBGyKeKH

Keywords: Random matrix theory, Heavy tails, Feature learning, Linear-width neural networks, Spike phase transition

Compressor summary: The spectral properties of high-dimensional linear neural networks change under different training strategies, affecting their generalization and feature learning abilities.


Contrastive Retrospection: honing in on critical steps for rapid learning and generalization in RL

Chen Sun, Wannan Yang, Thomas Jiralerspong, Dane Malenfant, Benjamin Alsbury-Nealy, Yoshua Bengio, Blake Aaron Richards

https://openreview.net/forum?id=gpJw8f4tIU

Keywords: Reinforcement learning, long term credit assignment, rapid credit assignment, contrastive learning, few-shot learning in RL

Compressor summary: The paper introduces Contrastive Retrospection (ConSpec), an RL algorithm that uses offline contrastive learning to identify critical steps in a task and provides intrinsic rewards for matching them, improving learning in various RL tasks.


Coop: Memory is not a Commodity

Jianhao Zhang, Shihan Ma, Peihong Liu, Jinhui Yuan

https://openreview.net/forum?id=gmmXyAq8TI

Keywords: Tensor Rematerialization; Gradient Checkpointing; Activation Recomputing; Deep Learning; Deep Learning Frameworks; Memory Allocator

Compressor summary: Tensor rematerialization technique optimizes memory allocation and saves resources for deep neural networks by using a sliding window and cheap tensor partitioning.


Structure of universal formulas

Dmitry Yarotsky

https://openreview.net/forum?id=gmVoaAxB1R

Keywords: VC-dimension, neural networks, activation functions, approximation, polynomials, algebraic functions

Compressor summary: The paper studies expressive models like neural networks and their limitations in approximating continuous functions on different sets using a hierarchy of classes and a new family of functions with polynomial constraints.


Creating Multi-Level Skill Hierarchies in Reinforcement Learning

Joshua Benjamin Evans, Özgür Şimşek

https://openreview.net/forum?id=gjBk6IQofa

Keywords: Reinforcement Learning, Hierarchical Reinforcement Learning, RL, HRL, Skill Discovery, Skill Hierarchies, Graph-Based, Graphs, Graph Clustering, Graph Partitioning

Compressor summary: The text proposes an automatic method to generate a skill hierarchy for autonomous agents based on modularity maximisation and interaction graphs, which improves their learning performance in various environments.


SimMTM: A Simple Pre-Training Framework for Masked Time-Series Modeling

Jiaxiang Dong, Haixu Wu, Haoran Zhang, Li Zhang, Jianmin Wang, Mingsheng Long

https://openreview.net/forum?id=ginTcBUnL8

Keywords: Time-series analysis, pre-training, masked time-series modeling

Compressor summary: SimMTM is a simple framework that improves masked modeling for time series by aggregating complementary information from neighbors and uncovering local structure.


Scaling Laws for Hyperparameter Optimization

Arlind Kadra, Maciej Janowski, Martin Wistuba, Josif Grabocka

https://openreview.net/forum?id=ghzEUGfRMD

Keywords: hyperparameter optimization, multi-fidelity hyperparameter optimization, multi-fidelity hpo, power laws, deep neural networks, deep power laws, deep ensemble, deep learning, large language models, scaling laws, llm

Compressor summary: The authors propose Deep Power Laws (DPL), a neural network ensemble that optimizes hyperparameters using power law scaling and gray-box evaluations, outperforming seven other methods on various datasets.


Hierarchical Semi-Implicit Variational Inference with Application to Diffusion Model Acceleration

Longlin Yu, Tianyu Xie, Yu Zhu, Tong Yang, Xiangyu Zhang, Cheng Zhang

https://openreview.net/forum?id=ghIBaprxsV

Keywords: Hierarchical semi-implicit variational inference, Score based training, Diffusion model

Compressor summary: HSIVI is a hierarchical extension of SIVI that improves expressiveness and allows acceleration of diffusion models using pre-trained score networks.


Reducing Shape-Radiance Ambiguity in Radiance Fields with a Closed-Form Color Estimation Method

Qihang Fang, Yafei Song, Keqiang Li, Liefeng Bo

https://openreview.net/forum?id=gh9JNeqjzo

Keywords: Neural radiance field, Novel-view synthesis, Regularization

Compressor summary: The paper proposes a more adaptive method to reduce shape-radiance ambiguity in neural radiance fields by estimating the color field based on the density field and using it to regularize NeRF's rendering process.


Learning to Configure Separators in Branch-and-Cut

Sirui Li, Wenbin Ouyang, Max B. Paulus, Cathy Wu

https://openreview.net/forum?id=gf5xJVQS5p

Keywords: Combinatorial Optimization, Branch-and-Cut, Learning Guided Optimization, Deep Learning

Compressor summary: This paper proposes a novel data-driven strategy for selecting combinatorial separators in mixed integer linear programs, which can improve the solve time by up to 72% and 37% on synthetic and real-world benchmarks.


Learning From Biased Soft Labels

Hua Yuan, Yu Shi, Ning Xu, Xu Yang, Xin Geng, Yong Rui

https://openreview.net/forum?id=gevmGxsTSI

Keywords: Soft label; knowledge distillation; weakly-supervised learning; Machine learning.

Compressor summary: The paper explores how biased soft labels can still be effective for knowledge distillation, proposes indicators to measure their effectiveness, and applies the framework to three weakly-supervised learning scenarios.


Combating Representation Learning Disparity with Geometric Harmonization

Zhihan Zhou, Jiangchao Yao, Feng Hong, Ya Zhang, Bo Han, Yanfeng Wang

https://openreview.net/forum?id=geLARFEK8O

Keywords: Long-tailed learning, self-supervised learning

Compressor summary: The text introduces Geometric Harmonization (GH), a method that improves self-supervised learning (SSL) under long-tailed distribution by ensuring category-level uniformity in representation learning, avoiding collapse of tail classes and dominance of head classes.


How do Minimum-Norm Shallow Denoisers Look in Function Space?

Chen Zeno, Greg Ongie, Yaniv Blumenfeld, Nir Weinberger, Daniel Soudry

https://openreview.net/forum?id=gdzxWGGxWE

Keywords: Denoiser, Denoising, Neural network, Function space

Compressor summary: The paper studies the theoretical properties of shallow ReLU neural network denoisers for image reconstruction tasks, deriving their functions and showing they generalize better than an alternative method in low noise settings.


xTrimoGene: An Efficient and Scalable Representation Learner for Single-Cell RNA-Seq Data

Jing Gong, Minsheng Hao, Xingyi Cheng, Xin Zeng, Chiming Liu, Jianzhu Ma, Xuegong Zhang, Taifeng Wang, Le Song

https://openreview.net/forum?id=gdwcoBCMVi

Keywords: pre-train, encoder-decoder, scRNA-seq, scalable

Compressor summary: The xTrimoGene model is a novel transformer architecture that can efficiently process large single-cell RNA-seq data and achieve state-of-the-art performance on various downstream tasks, while reducing computational and memory costs significantly.


Finding Safe Zones of Markov Decision Processes Policies

Lee Cohen, Yishay Mansour, Michal Moshkovitz

https://openreview.net/forum?id=gdVcFOvxT3

Keywords: Theoretical guarantees, algorithms, learning theory, MDP, computational complexity, Interpretability

Compressor summary: The paper studies safe zones in Markov decision processes, which are subsets of states that most trajectories stay within, and presents an algorithm to find approximate safe zones with low complexity.


Towards Optimal Caching and Model Selection for Large Model Inference

Banghua Zhu, Ying Sheng, Lianmin Zheng, Clark Barrett, Michael Jordan, Jiantao Jiao

https://openreview.net/forum?id=gd20oaZqqF

Keywords: caching, model selection, large language models, foundation models, inference, bandit, regret

Compressor summary: The paper proposes two methods to reduce resource consumption and latency of large foundation models during inference: caching previous queries and selecting the best model for each query, and shows their theoretical and empirical benefits.


Synthetic Combinations: A Causal Inference Framework for Combinatorial Interventions

Abhineet Agarwal, Anish Agarwal, Suhas Vijaykumar

https://openreview.net/forum?id=gbhixjg2dX

Keywords: Causal Inference, Matrix Completion, Combinatorial Learning, Ranking

Compressor summary: The paper proposes Synthetic Combinations, a method to learn causal parameters from observational data by exploiting latent structure across units and interventions.


Object-Centric Slot Diffusion

Jindong Jiang, Fei Deng, Gautam Singh, Sungjin Ahn

https://openreview.net/forum?id=gbOukzirpK

Keywords: Object-Centric Representation Learning, Diffusion Models, Unsupervised Representation Learning

Compressor summary: The paper introduces Latent Slot Diffusion, a novel model that combines diffusion models with object-centric learning for better image generation, especially in complex scenes, without supervised annotations.


Slot-guided Volumetric Object Radiance Fields

DI QI, Tong Yang, Xiangyu Zhang

https://openreview.net/forum?id=ganlU27uvj

Keywords: 3D object-centric representation learning, NeRF, 3D-aware slot

Compressor summary: The paper introduces a new method called sVORF that learns 3D object-centric representations from images by decomposing scenes into objects using volumetric radiance fields and object slots.


Semi-Implicit Denoising Diffusion Models (SIDDMs)

yanwu xu, Mingming Gong, Shaoan Xie, Wei Wei, Matthias Grundmann, kayhan Batmanghelich, Tingbo Hou

https://openreview.net/forum?id=gaktiSjatl

Keywords: Diffusion Model, GAN, Semi-implicit Modeling

Compressor summary: The paragraph introduces a novel approach to improve sample quality and diversity in generative models by combining an implicit model with the forward diffusion process, enabling larger steps during inference without compromising performance.


On Private and Robust Bandits

Yulian Wu, Xingyu Zhou, Youming Tao, Di Wang

https://openreview.net/forum?id=gaXAjtHic2

Keywords: Bandits, privacy, robustness

Compressor summary: The paper investigates how to balance privacy, robustness, and accuracy in multi-armed bandits with heavy-tailed rewards under a limited budget.


Best Arm Identification with Fixed Budget: A Large Deviation Perspective

Po-An Wang, Ruo-Chun Tzeng, Alexandre Proutiere

https://openreview.net/forum?id=gYetLsNO8x

Keywords: Best arm identification, Large deviation

Compressor summary: The paper studies how to find the best arm in stochastic Multi-Armed Bandits with a fixed sampling budget, and presents a new adaptive algorithm called Continuous Rejects that outperforms existing methods using Large Deviation techniques.


Differentiable sorting for censored time-to-event data.

Andre Vauvelle, Benjamin Wild, Roland Eils, Spiros Denaxas

https://openreview.net/forum?id=gYWjI7wLhc

Keywords: Survival Analysis, Censored Data, Semi-supervised Learning, Time-to-event-data, Algorithmic Supervision, Sorting, Risk Prediction, Weakly-supervised Learning, Machine Learning, Cox's Partial Likelihood, Differentiable Sorting Networks, Transitive Inductive Bias, Ranking Losses, Listwise Ranking, Healthcare Applications, Deep Learning, Neural Networks, Top-k Risk Prediction

Compressor summary: Diffsurv is a new survival analysis method that uses differentiable sorting networks to handle censored data and improve risk prediction in healthcare applications.


Global Convergence Analysis of Local SGD for Two-layer Neural Network without Overparameterization

Yajie Bao, Amarda Shehu, Mingrui Liu

https://openreview.net/forum?id=gVLKXT9JwG

Keywords: convolutional neural network, gaussian input, local SGD, global convergence, non-convex optimization

Compressor summary: This paper presents the first global convergence analysis of local SGD for two-layer neural networks without overparameterization or noise, using a self-correction mechanism and an exact recursive characterization of the direction of global parameters.


Learning Provably Robust Estimators for Inverse Problems via Jittering

Anselm Krainovic, Mahdi Soltanolkotabi, Reinhard Heckel

https://openreview.net/forum?id=gUlcyeHzw1

Keywords: Adversarial Training, Jittering, Denoising, Deconvolution, Compressive Sensing, Inverse Problems, Robustness

Compressor summary: The paper explores if adding noise during training makes neural networks more resistant to worst-case perturbations in inverse problems like denoising and finds that jittering improves robustness but may not work for all inverse problems.


Test-time Adaptation of Discriminative Models via Diffusion Generative Feedback

Mihir Prabhudesai, Tsung-Wei Ke, Alexander Cong Li, Deepak Pathak, Katerina Fragkiadaki

https://openreview.net/forum?id=gUTVpByfVX

Keywords: test-time adaptation, diffusion models, generative models, classification, segmentation, depth prediction

Compressor summary: Diffusion-TTA uses generative models to adapt pre-trained discriminative models for different test examples, improving their accuracy and performance in online settings.


BiSLS/SPS: Auto-tune Step Sizes for Stable Bi-level Optimization

Chen Fan, Gaspard Choné-Ducasse, Mark Schmidt, Christos Thrampoulidis

https://openreview.net/forum?id=gUEekxYr6D

Keywords: Adaptive step sizes, bi-level optimization, convergence rates, line searches

Compressor summary: The authors propose adaptive step-size methods for bi-level optimization in deep learning to improve convergence speed without requiring fine-tuning.


Rethinking Conditional Diffusion Sampling with Progressive Guidance

Anh-Dung Dinh, Daochang Liu, Chang Xu

https://openreview.net/forum?id=gThGBHhqcU

Keywords: Diffusion model, conditional generative model, guidance diffusion, generative models, classifier guidance

Compressor summary: The paper presents Progressive Guidance, a generalized method for diffusion generative models that addresses diversity and adversarial effects issues in classifier guidance by refining image details progressively.


Optimistic Rates for Multi-Task Representation Learning

Austin Watkins, Enayat Ullah, Thanh Nguyen-Tang, Raman Arora

https://openreview.net/forum?id=gQ4h6WvME0

Keywords: Learning Theory, Multi-task and Transfer Learning, Classification

Compressor summary: The paper investigates how multi-task representation learning can improve transfer learning by providing new statistical rates on the excess risk of the target task depending on the difficulty of the learning task.


Partial Matrix Completion

Elad Hazan, Adam Tauman Kalai, Varun Kanade, Clara Mohri, Y. Jennifer Sun

https://openreview.net/forum?id=gPylY8sCbw

Keywords: matrix completion, online learning

Compressor summary: The paragraph introduces the Partial Matrix Completion problem, where the goal is to accurately complete a subset of matrix entries, and presents an algorithm that handles complex sampling distributions and performs well in online settings.


Content-based Unrestricted Adversarial Attack

Zhaoyu Chen, Bo Li, Shuang Wu, Kaixun Jiang, Shouhong Ding, Wenqiang Zhang

https://openreview.net/forum?id=gO60SSGOMy

Keywords: unrestricted attack, adversarial example, diffusion model, black-box attack, adversarial transferability

Compressor summary: The proposed Content-based Unrestricted Adversarial Attack framework optimizes images along an adversarial direction to create photorealistic adversarial examples that deceive both human perception and deep neural networks with high attack performance.


AdaVAE: Bayesian Structural Adaptation for Variational Autoencoders

Paribesh Regmi, Rui Li

https://openreview.net/forum?id=gMjIUZBKH8

Keywords: nonparametric Bayes, variational autoencoders

Compressor summary: The paragraph introduces a Bayesian inference framework for variational autoencoders (VAEs) that adapts network structures to data and prevents overfitting, leading to improved generative performance.


One Fits All: Power General Time Series Analysis by Pretrained LM

Tian Zhou, Peisong Niu, Xue Wang, Liang Sun, Rong Jin

https://openreview.net/forum?id=gMS6FVZvmF

Keywords: general time series analysis, time series forecasting, cross modality knowledge transfer; pretrained language model;

Compressor summary: The paper presents Frozen Pretrained Transformer (FPT), a model that uses pre-trained language or image models for time series analysis, and shows its comparable or state-of-the-art performance on various tasks.


Triangulation Residual Loss for Data-efficient 3D Pose Estimation

Jiachen Zhao, Tao Yu, Liang An, Yipeng Huang, Fang Deng, Qionghai Dai

https://openreview.net/forum?id=gLwjBDsE3G

Keywords: 3D pose estimation, triangulation, animal pose estimation

Compressor summary: The paper proposes a self-supervised method called Triangulation Residual loss for 3D pose estimation that uses global multiview geometric consistency and achieves state-of-the-art performance with limited labeled data.


Labeling Neural Representations with Inverse Recognition

Kirill Bykov, Laura Kopf, Shinichi Nakajima, Marius Kloft, Marina MC Höhne

https://openreview.net/forum?id=gLfgyIWiWW

Keywords: Explainable AI, Mechanistic Interpretability, Machine Learning, Deep Neural Networks

Compressor summary: INVERT is a scalable method to link DNNs' complex data representations to human-interpretable concepts without relying on segmentation masks or high computational demands, enabling interpretable explanations and assessing their statistical significance.


Fed-GraB: Federated Long-tailed Learning with Self-Adjusting Gradient Balancer

Zikai Xiao, Zihan Chen, Songshang Liu, Hualiang Wang, YANG FENG, Jin Hao, Joey Tianyi Zhou, Jian Wu, Howard Hao Yang, Zuozhu Liu

https://openreview.net/forum?id=gJewjFjfN2

Keywords: Federated learning, Long-tailed learning, Data heterogeneity

Compressor summary: The paper introduces $\texttt{Fed-GraB}$, a method for federated long-tailed learning that re-weights client gradients based on global distribution feedback to improve performance on minority classes while preserving majority class performance.


Pengi: An Audio Language Model for Audio Tasks

Soham Deshmukh, Benjamin Elizalde, Rita Singh, Huaming Wang

https://openreview.net/forum?id=gJLAfO4KUq

Keywords: audio language model, audio representation learning, audio and speech processing, multi-task and transfer learning

Compressor summary: Pengi is an audio language model that uses transfer learning to enable open-ended and close-ended tasks without extra fine-tuning, achieving state-of-the-art performance on 21 downstream tasks.


NeRF Revisited: Fixing Quadrature Instability in Volume Rendering

Mikaela Angelina Uy, Kiyohiro Nakayama, Guandao Yang, Rahul Krishna Thomas, Leonidas Guibas, Ke Li

https://openreview.net/forum?id=gJHAT79cZU

Keywords: neural radiance fields, volumetric rendering, nerfs, numerical quadrature, importance sampling

Compressor summary: The paragraph discusses quadrature instability in Neural Radiance Fields (NeRF) and proposes a solution that improves texture, geometry, and depth supervision by reformulating the sample-based rendering equation to match piecewise linear volume density.


How Does Adaptive Optimization Impact Local Neural Network Geometry?

Kaiqi Jiang, Dhruv Malik, Yuanzhi Li

https://openreview.net/forum?id=gIG8LvTLuc

Keywords: optimization, adaptive algorithms, neural networks

Compressor summary: Adaptive optimization methods improve neural network training by biasing iterate trajectories towards regions with lower $R^{\text{OPT}}\_{\text{med}}$, a statistic similar to the condition number of the loss Hessian, compared to vanilla gradient methods like SGD.


Generalizing Nonlinear ICA Beyond Structural Sparsity

Yujia Zheng, Kun Zhang

https://openreview.net/forum?id=gI1SOgW3kw

Keywords: Latent variable models, nonlinear independent component analysis

Compressor summary: The paragraph discusses new identifiability results for nonlinear ICA under different conditions, such as undercompleteness, partial sparsity, source dependence, and flexible grouping structures, addressing limitations of previous methods.


Theoretical and Practical Perspectives on what Influence Functions Do

Andrea Schioppa, Katja Filippova, Ivan Titov, Polina Zablotskaia

https://openreview.net/forum?id=gGl0n7Onug

Keywords: Explainable AI, Influence Functions, Training Data Attribution

Compressor summary: Influence functions are useful for debugging and correcting deep neural networks but their predictive power is limited by parameter divergence, which causes influence to fade over training time.


SOL: Sampling-based Optimal Linear bounding of arbitrary scalar functions

Yuriy Biktairov, Jyotirmoy Deshmukh

https://openreview.net/forum?id=gAQCx61chN

Keywords: Neural network verification, Robustness, Linear bounding

Compressor summary: The paper proposes optimal methods for finding tight linear bounds for activation functions in neural networks, which are important for robustness certification tools, using a new sampling-based approach that is efficient and practical.


Inverse Preference Learning: Preference-based RL without a Reward Function

Joey Hejna, Dorsa Sadigh

https://openreview.net/forum?id=gAP52Z2dar

Keywords: preference learning, preference-based reinforcement learning, human-in-the-loop reinforcement learning

Compressor summary: The paper proposes Inverse Preference Learning (IPL), a new simple and efficient algorithm for learning from offline preference data in preference-based RL, which eliminates the need for a learned reward function by using the $Q$-function instead.


Rethinking Semi-Supervised Medical Image Segmentation: A Variance-Reduction Perspective

Chenyu You, Weicheng Dai, Yifei Min, Fenglin Liu, David A. Clifton, S Kevin Zhou, Lawrence Hamilton Staib, James s Duncan

https://openreview.net/forum?id=g9gjpFOiO4

Keywords: Long-tailed Medical Image Segmentation, Contrastive Learning, Variance Reduction, Imbalanced Learning, Semi-Supervised Learning

Compressor summary: The paper introduces $\texttt{ARCO}$, a semi-supervised contrastive learning framework that improves medical image segmentation by using stratified group theory, variance-reduced estimation, and universal sampling techniques.


Where Did I Come From? Origin Attribution of AI-Generated Images

Zhenting Wang, Chen Chen, Yi Zeng, Lingjuan Lyu, Shiqing Ma

https://openreview.net/forum?id=g8bjq0qxOl

Keywords: Origin attribution of generated images

Compressor summary: The paper proposes an alteration-free, model-agnostic method for determining if an image was generated by a specific image generation model using reverse-engineering and reconstruction loss.


Prioritizing Samples in Reinforcement Learning with Reducible Loss

Shiva Kanth Sujit, Somjit Nath, Pedro Braga, Samira Ebrahimi Kahou

https://openreview.net/forum?id=g78QqvhnDU

Keywords: reinforcement learning, sample efficiency, experience replay

Compressor summary: The paper proposes a method to prioritize reinforcement learning samples based on their learn-ability, which is defined as the decrease of training loss over time, to improve sample efficiency and robustness.


BLIP-Diffusion: Pre-trained Subject Representation for Controllable Text-to-Image Generation and Editing

Dongxu Li, Junnan Li, Steven Hoi

https://openreview.net/forum?id=g6We1SwaY9

Keywords: Diffusion-based Models, Text-to-Image Generation, Image Editing, Vision-and-Language, Multimodal

Compressor summary: BLIP-Diffusion is a new model that generates images of a given subject based on both image and text inputs, using a pre-trained multimodal encoder and a diffusion model. It enables efficient fine-tuning and can be combined with other techniques for more advanced applications.


Transformers over Directed Acyclic Graphs

Yuankai Luo, Veronika Thost, Lei Shi

https://openreview.net/forum?id=g49s1N5nmO

Keywords: Graph Neural Networks, Transformers, Graph Classification, Node Classification, Scalability

Compressor summary: This paper proposes efficient and effective architecture adaptations for transformer models on directed acyclic graphs (DAGs), enabling them to outperform graph neural networks tailored for DAGs.


Incentivizing Honesty among Competitors in Collaborative Learning and Optimization

Florian E. Dorner, Nikola Konstantinov, Georgi Stoyanov Pashaliev, Martin Vechev

https://openreview.net/forum?id=g2ROKOASiv

Keywords: Game Theory, Federated Learning, Optimization, Strategic Behavior, Economics, Mechanisms

Compressor summary: This paper studies how to design mechanisms that encourage honest communication between competitors in collaborative machine learning, despite their incentive to manipulate their updates for personal gain, and demonstrates their effectiveness on a non-convex federated learning task.


LART: Neural Correspondence Learning with Latent Regularization Transformer for 3D Motion Transfer

Haoyu Chen, Hao Tang, Radu Timofte, Luc Van Gool, Guoying Zhao

https://openreview.net/forum?id=g27BggUT3L

Keywords: 3D motion transfer, 3D Transformer, geometric preservation, 3D generation, correspondence learning

Compressor summary: The paper introduces LART, a 3D Transformer framework for transferring realistic and high-fidelity motions between meshes without key point annotations or predefined correspondence, using latent metric regularization to ensure motion quality.


MIMEx: Intrinsic Rewards from Masked Input Modeling

Toru Lin, Allan Jabri

https://openreview.net/forum?id=g1dMYenhe4

Keywords: reinforcement learning, exploration, intrinsic reward, intrinsic motivation, masked autoencoder

Compressor summary: The paper proposes MIMEx, a framework for deriving intrinsic rewards in high-dimensional environments using conditional prediction and masked input modeling, which can improve exploration in sparse-reward visuomotor tasks.


Supported Value Regularization for Offline Reinforcement Learning

Yixiu Mao, Hongchang Zhang, Chen Chen, Yi Xu, Xiangyang Ji

https://openreview.net/forum?id=fze7P9oy6l

Keywords: offline reinforcement learning

Compressor summary: SVR is a value regularization method for offline RL that penalizes Q-values for out-of-distribution actions and maintains optimal convergence results while achieving state-of-the-art performance on continuous control tasks.


Template-free Articulated Neural Point Clouds for Reposable View Synthesis

Lukas Uzolas, Elmar Eisemann, Petr Kellnhofer

https://openreview.net/forum?id=fyfmHi8ay3

Keywords: Radiance Fields, View Synthesis, Kinematics, Reposing, NeRF

Compressor summary: The paper introduces a new method for synthesizing realistic dynamic scenes using point-based representation, Linear Blend Skinning, and multi-view video, improving visual quality, learning time, and generalization.


Neural (Tangent Kernel) Collapse

Mariia Seleznova, Dana Weitzner, Raja Giryes, Gitta Kutyniok, Hung-Hsu Chou

https://openreview.net/forum?id=fyLvHzEssH

Keywords: Neural Collapse, Neural Tangent Kernel, NTK alignment, Local Elasticity, Gradient Flow

Compressor summary: This work connects two ideas in deep learning, shows how class labels affect neural network training dynamics, and proves that well-trained networks develop structure based on these assumptions.


Text-to-Image Diffusion Models are Zero Shot Classifiers

Kevin Clark, Priyank Jaini

https://openreview.net/forum?id=fxNQJVMwK2

Keywords: diffusion models, zero-shot, text-to-image, generative models, foundation models, stable diffusion

Compressor summary: The authors propose a method to evaluate text-to-image diffusion models as zero-shot classifiers, showing they perform competitively with CLIP on image classification tasks and have advantages in attribute binding. They suggest generative pre-training could be useful for vision and vision-language problems.


Learning Score-based Grasping Primitive for Human-assisting Dexterous Grasping

Tianhao Wu, Mingdong Wu, Jiyao Zhang, Yunchong Gan, Hao Dong

https://openreview.net/forum?id=fwvfxDbUFw

Keywords: Human-asissting Dexterous Grasping, Score-matching, Reinforcement Learning

Compressor summary: The paper proposes a new task for training robotic hands to assist users in grasping objects, using a hand-object conditioned grasping primitive and a history-conditional policy.


Sensitivity in Translation Averaging

Lalit Manam, Venu Madhav Govindu

https://openreview.net/forum?id=fvm9jVcpBn

Keywords: sensitivity, translation averaging, structure from motion, 3D computer vision

Compressor summary: The paper studies sensitivity in 3D computer vision's translation averaging under uncertainty, provides a criterion for well-conditioned problems, and offers an algorithm to remove ill-conditioned directions, improving 3D reconstructions.


Demo2Code: From Summarizing Demonstrations to Synthesizing Code via Extended Chain-of-Thought

Huaxiaoyue Wang, Gonzalo Gonzalez-Pumariega, Yash Sharma, Sanjiban Choudhury

https://openreview.net/forum?id=ftPoVcm821

Keywords: Large Language Model, Code Generation, Robotics

Compressor summary: The paper introduces Demo2Code, a framework that generates code for robotic tasks from demonstrations using a recursive summarization and synthesis process.


UE4-NeRF:Neural Radiance Field for Real-Time Rendering of Large-Scale Scene

Jiaming Gu, Minchao Jiang, Hongsheng Li, Xiaoyuan Lu, Guangming Zhu, Syed Afaq Ali Shah, Liang Zhang, Mohammed Bennamoun

https://openreview.net/forum?id=fsCcGr8YFR

Keywords: NeRF, UE4, Large-scale scenes, Real-time rendering, Rasterization

Compressor summary: The paper introduces UE4-NeRF, a system for real-time rendering of large-scale scenes using NeRF and Unreal Engine 4, achieving high quality results at 4K resolution with fast frame rates.


Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task

Maya Okawa, Ekdeep Singh Lubana, Robert P. Dick, Hidenori Tanaka

https://openreview.net/forum?id=frVo9MzRuU

Keywords: Diffusion model; Science of deep learning; Mechanistic interpretability

Compressor summary: The paragraph discusses a study of how conditional diffusion models generate and reason over novel samples, focusing on the factors that affect their compositional generalization abilities.


Meek Separators and Their Applications in Targeted Causal Discovery

Kirankumar Shiragur, Jiaqi Zhang, Caroline Uhler

https://openreview.net/forum?id=frSfSaRGXY

Keywords: Causality, Graphical Models

Compressor summary: The text discusses a new technique called Meek separator, which helps in learning part of the causal graph with fewer interventions and presents two randomized algorithms that achieve logarithmic approximation for subset search and causal matching.


Agnostically Learning Single-Index Models using Omnipredictors

Aravind Gollakota, Parikshit Gopalan, Adam Klivans, Konstantinos Stavropoulos

https://openreview.net/forum?id=frHPeRedHo

Keywords: generalized linear models, single-index models, agnostic learning, pac learning, logistic regression, omnipredictors, multiaccuracy, calibration

Compressor summary: The paper presents a new algorithm to learn Single-Index Models with arbitrary activations and weaker distributional assumptions, based on Omniprediction and Bregman divergences.


Efficient Adversarial Contrastive Learning via Robustness-Aware Coreset Selection

Xilie Xu, Jingfeng Zhang, Feng Liu, Masashi Sugiyama, Mohan Kankanhalli

https://openreview.net/forum?id=fpzA8uRA95

Keywords: robust pre-training, adversarial contrastive learning, coreset selection

Compressor summary: The paper proposes a fast method called robustness-aware coreset selection (RCS) for adversarial contrastive learning (ACL), which improves the efficiency of generating adversarial variants and achieves a good balance between speed and robustness.


Tackling Heavy-Tailed Rewards in Reinforcement Learning with Function Approximation: Minimax Optimal and Instance-Dependent Regret Bounds

Jiayi Huang, Han Zhong, Liwei Wang, Lin Yang

https://openreview.net/forum?id=fpHfRD3f4N

Keywords: machine learning, reinforcement learning, linear bandits, heavy-tailed rewards, instance-dependent regret

Compressor summary: The paper proposes efficient algorithms for reinforcement learning with heavy-tailed rewards, achieving near-optimal regret bounds in both linear bandits and linear function approximation settings.


Data-Informed Geometric Space Selection

Shuai Zhang, Wenqi Jiang

https://openreview.net/forum?id=fpElyckKkd

Keywords: Geometric representation learning

Compressor summary: The paper proposes a method to automatically choose and combine different geometric spaces for each input data point to improve machine learning tasks without human intervention.


Permutation Equivariant Neural Functionals

Allan Zhou, Kaien Yang, Kaylee Burns, Adriano Cardace, Yiding Jiang, Samuel Sokota, J Zico Kolter, Chelsea Finn

https://openreview.net/forum?id=fmYmXNPmhv

Keywords: equivariance, permutation, implicit neural representation, generalization

Compressor summary: The text introduces a framework for building permutation equivariant neural functionals that process the weights of other neural networks using NF-Layers with an inductive bias based on symmetry.


Are Diffusion Models Vision-And-Language Reasoners?

Benno Krojer, Elinor Poole-Dayan, Vikram Voleti, Christopher Pal, Siva Reddy

https://openreview.net/forum?id=fmJv8Hj0yo

Keywords: diffusion model, automatic evaluation, vision-and-language, compositionality

Compressor summary: The authors present a method to adapt diffusion-based image generation models for vision-and-language tasks and introduce a new benchmark to evaluate their performance, finding that the proposed approach outperforms CLIP on compositional tasks and reduces bias in Stable Diffusion.


Prototypical Variational Autoencoder for 3D Few-shot Object Detection

Weiliang Tang, Biqi YANG, Xianzhi Li, Yun-Hui Liu, Pheng-Ann Heng, Chi-Wing Fu

https://openreview.net/forum?id=fljrZsJ2I8

Keywords: 3D Point Cloud Object Detection, Few Shot Learning, Computer Vision, Geometric Prototype

Compressor summary: The paper proposes a VAE-based method (GP-VAE) for few-shot 3D object detection, which enhances feature diversity and distinctiveness using probabilistic latent space and improves performance over existing methods.


Policy Finetuning in Reinforcement Learning via Design of Experiments using Offline Data

Ruiqi Zhang, Andrea Zanette

https://openreview.net/forum?id=fjXTcUUgaC

Keywords: offline RL, online RL, exploration, non-reactive, fine-tuning

Compressor summary: The paper proposes an algorithm that uses an offline dataset to create a single non-reactive exploration policy with guaranteed performance.


Bootstrapped Training of Score-Conditioned Generator for Offline Design of Biological Sequences

Minsu Kim, Federico Berto, Sungsoo Ahn, Jinkyoo Park

https://openreview.net/forum?id=fj0ZeRtUTU

Keywords: Biological sequence design, offline model based optimization, conditional generation, bootstrapping, ensemble

Compressor summary: The BootGen algorithm optimizes biological sequences by training a generator with rank-based weights and then bootstrapping it with self-generated data labeled by a proxy score function, resulting in diverse and high-scoring designs.


Rotating Features for Object Discovery

Sindy Löwe, Phillip Lippe, Francesco Locatello, Max Welling

https://openreview.net/forum?id=fg7iyNK81W

Keywords: Object Discovery, Object-Centric Representations, Structured Representation Learning

Compressor summary: The paper proposes Rotating Features, a method to represent objects in distributed representations, and demonstrates its applicability to real-world data.


Neural Foundations of Mental Simulation: Future Prediction of Latent Representations on Dynamic Scenes

Aran Nayebi, Rishi Rajalingham, Mehrdad Jazayeri, Guangyu Robert Yang

https://openreview.net/forum?id=ffOhY40Nrh

Keywords: neural coding, mental simulation, foundation models, primate frontal cortex

Compressor summary: The authors use various machine learning models to investigate the neural mechanisms behind human and animal understanding of the physical world and future prediction, finding that models trained on dynamic video scenes in latent space best match the data.


RS-Del: Edit Distance Robustness Certificates for Sequence Classifiers via Randomized Deletion

Zhuoqun Huang, Neil G Marchant, Keane Lucas, Lujo Bauer, Olga Ohrimenko, Benjamin I. P. Rubinstein

https://openreview.net/forum?id=ffFcRPpnWx

Keywords: certified robustness, randomized smoothing, malware detection, sequence classification, edit distance

Compressor summary: The paper proposes randomized deletion (RS-Del), a randomized smoothing technique for discrete sequence classifiers, which provides certified robustness against edit distance-bounded adversaries and is applied to malware detection.


Analysis of Variance of Multiple Causal Networks

Zhongli Jiang, Dabao Zhang

https://openreview.net/forum?id=fezV91IJIo

Keywords: causal inference, large graphs, multi-task learning, structural model, directed cyclic graph

Compressor summary: The authors propose a new method for constructing and comparing multiple directed cyclic graphs (DCGs) using a structural model, limited-information, and parallel computation, which can handle algorithmic difficulty, computational burden, and variational causalities.


Representation Learning via Consistent Assignment of Views over Random Partitions

Thalles Santos Silva, Adín Ramírez Rivera

https://openreview.net/forum?id=fem6BIJkdv

Keywords: representation learning, unsupervised learning, self-supervised learning, computer vision

Compressor summary: CARP is a self-supervised clustering method that learns visual feature representations using random partitions and improves downstream task performance.


LEPARD: Learning Explicit Part Discovery for 3D Articulated Shape Reconstruction

Di Liu, Anastasis Stathopoulos, Qilong Zhangli, Yunhe Gao, Dimitris N. Metaxas

https://openreview.net/forum?id=fcYObrixSS

Keywords: 3D computer vision, deep learning

Compressor summary: LEPARD is a framework that uses learning to reconstruct 3D animal shapes from single images by breaking down the object into simpler, robust 3D parts.


A fast heuristic to optimize time-space tradeoff for large models

Akifumi Imanishi, Zijian Xu, Masayuki Takagi, Sixue Wang, Emilio Castillo

https://openreview.net/forum?id=fbpTObq6TW

Keywords: Recomputation, Gradient checkpointing, Memory reduction, Computational graph optimization

Compressor summary: The paper introduces FastSA, a new simulated annealing-based recomputation algorithm that significantly reduces GPU memory usage and is faster than current methods for optimizing large neural networks.


Multi Time Scale World Models

Vaisakh Shaj, Saleh GHOLAM ZADEH, Ozan Demir, Luiz Ricardo Douat, Gerhard Neumann

https://openreview.net/forum?id=fY7dShbtmo

Keywords: Hierarchical Models; Multi Time Scale Learning; World Models

Compressor summary: The Multi Time Scale State Space (MTS3) model is a probabilistic framework for learning world models that can make accurate long-horizon predictions and uncertainty estimates over multiple time scales.


Sample based Explanations via Generalized Representers

Che-Ping Tsai, Chih-Kuan Yeh, Pradeep Kumar Ravikumar

https://openreview.net/forum?id=fX64q0SNfL

Keywords: explainable machine learning, sample based explanation, representer point

Compressor summary: The paper introduces a new class of sample-based explanations for machine learning models that use both global and local importance measures and show how existing methods can be considered as special cases.


Rethinking Tokenizer and Decoder in Masked Graph Modeling for Molecules

Zhiyuan Liu, Yaorui Shi, An Zhang, Enzhi Zhang, Kenji Kawaguchi, Xiang Wang, Tat-Seng Chua

https://openreview.net/forum?id=fWLf8DV0fI

Keywords: Molecular Representation Learning, Masked Graph Modeling, Graph Tokenizer

Compressor summary: The paragraph discusses a new approach to learning representations of molecular graphs, focusing on improving the tokenizer and decoder in the masked graph modeling process, and introduces a novel method called SimSGT that outperforms existing methods.


Learning Domain-Aware Detection Head with Prompt Tuning

Haochen Li, Rui Zhang, Hantao Yao, Xinkai Song, Yifan Hao, Yongwei Zhao, Ling Li, Yunji Chen

https://openreview.net/forum?id=fW5ZUSVTkv

Keywords: domain adaptation, object detection, prompt tuning

Compressor summary: The paper proposes a method to improve domain adaptive object detection by using a domain-aware detection head with prompt tuning, which generates a dynamic detection head for each domain using learnable tokens and textual descriptions.


Two Sides of One Coin: the Limits of Untuned SGD and the Power of Adaptive Methods

Junchi YANG, Xiang Li, Ilyas Fatkhullin, Niao He

https://openreview.net/forum?id=fUZUoSLXw3

Keywords: Nonconvex optimization, Stochastic Gradient Descent, Adaptive methods

Compressor summary: Untuned Stochastic Gradient Descent (SGD) has a fast convergence rate but an exponential dependency on smoothness, while adaptive methods prevent this issue and are better for practical use.


Asynchronous Proportional Response Dynamics: Convergence in Markets with Adversarial Scheduling

Yoav Kolumbus, Menahem Levy, Noam Nisan

https://openreview.net/forum?id=fU9U7OYxfE

Keywords: Asynchronous Dynamics, Fisher Markets, Proportional Response, Best Response, Game Dynamics, Competitive Equilibrium, Convergence

Compressor summary: In this study, researchers explore how Proportional Response Dynamics (PRD) lead to competitive equilibria in linear Fisher markets with asynchronous bidding and reveal new properties of these markets.


Curriculum Learning for Graph Neural Networks: Which Edges Should We Learn First

Zheng Zhang, Junxiang Wang, Liang Zhao

https://openreview.net/forum?id=fTyGT5fulj

Keywords: Graph neural networks, Curriculum learning, Graph structure learning

Compressor summary: The paper proposes a novel curriculum learning strategy for graph neural networks that gradually incorporates edges based on their difficulty, improving representation quality and generalization.


Provable Guarantees for Nonlinear Feature Learning in Three-Layer Neural Networks

Eshaan Nichani, Alex Damian, Jason D. Lee

https://openreview.net/forum?id=fShubymWrc

Keywords: Deep Learning Theory, Feature Learning, Three-Layer Neural Network, Depth Separation, Gradient Descent, Representation Learning

Compressor summary: This paper explores how three-layer neural networks have better feature learning capabilities than two-layer networks, and provides theoretical and empirical evidence for this claim.


In Defense of Softmax Parametrization for Calibrated and Consistent Learning to Defer

Yuzhou Cao, Hussein Mozannar, Lei Feng, Hongxin Wei, Bo An

https://openreview.net/forum?id=fPAAgjISu0

Keywords: Classification, Learning to Defer, Probability Estimation

Compressor summary: The paper proposes a new loss function for learning to defer in machine learning classifiers that avoids uncalibrated estimates and improves classification accuracy.


Dynamics of Finite Width Kernel and Prediction Fluctuations in Mean Field Neural Networks

Blake Bordelon, Cengiz Pehlevan

https://openreview.net/forum?id=fKwG6grp8o

Keywords: Deep Learning Theory, Feature Learning, Dynamics, Ensembles

Compressor summary: The paper studies how finite width affects feature learning neural networks, showing how it influences kernel fluctuations, prediction variance, and online learning in different network depths and training regimes.


Reduced Policy Optimization for Continuous Control with Hard Constraints

Shutong Ding, Jingya Wang, Yali Du, Ye Shi

https://openreview.net/forum?id=fKVEMNmWqU

Keywords: Reinforcement Learning, Hard Constraint, Generalized Reduced Gradient

Compressor summary: The paper proposes a reduced policy optimization (RPO) algorithm that combines reinforcement learning with generalized reduced gradient method to handle general hard constraints in continuous control tasks, and introduces three new benchmarks to test its performance.


Learning Energy-Based Prior Model with Diffusion-Amortized MCMC

Peiyu Yu, Yaxuan Zhu, Sirui Xie, Xiaojian Ma, Ruiqi Gao, Song-Chun Zhu, Ying Nian Wu

https://openreview.net/forum?id=fKQEmHoLb6

Keywords: Energy-Based Model, Denoising Diffusion Probabilistic Model, MCMC

Compressor summary: The paper introduces a diffusion-based amortization method for latent space EBMs to improve sampling quality and generation performance, and provides theoretical and experimental evidence for its effectiveness.


Entropic Neural Optimal Transport via Diffusion Processes

Nikita Gushchin, Alexander Kolesov, Alexander Korotin, Dmitry P. Vetrov, Evgeny Burnaev

https://openreview.net/forum?id=fHyLsfMDIs

Keywords: Optimal transport, Schrödinger Bridge, Entropy regularized OT, Neural Networks, Unpaired Learning

Compressor summary: The authors propose a new neural algorithm for computing optimal transport between distributions based on a dynamic Schrödinger Bridge problem, which is faster and more flexible than previous methods.


Calibrated Stackelberg Games: Learning Optimal Commitments Against Calibrated Agents

Nika Haghtalab, Chara Podimata, Kunhe Yang

https://openreview.net/forum?id=fHsBNNDroC

Keywords: calibration, Stackelberg games, learning in repeated games, strategic agents, best response, strategic classification, Stackelberg Security Games

Compressor summary: The paper proposes Calibrated Stackelberg Games, a generalization of standard Stackelberg Games where an agent does not directly observe the principal's actions but responds to calibrated forecasts, and shows that this model can capture real-life applications and achieve optimal utility.


Switching Autoregressive Low-rank Tensor Models

Hyun Dong Lee, Andrew Warrington, Joshua I Glaser, Scott Linderman

https://openreview.net/forum?id=fFJThJ94rY

Keywords: switching, autoregressive, low-rank tensor, time-series, probabilistic, neural, neuroscience, behavioral, arhmm, slds

Compressor summary: SALT models are a new probabilistic approach to time-series analysis that balance the benefits of ARHMMs and SLDSs by using low-rank factorization to control parameters and capture long-range dependencies.


Diffusion Model is an Effective Planner and Data Synthesizer for Multi-Task Reinforcement Learning

Haoran He, Chenjia Bai, Kang Xu, Zhuoran Yang, Weinan Zhang, Dong Wang, Bin Zhao, Xuelong Li

https://openreview.net/forum?id=fAdMly4ki5

Keywords: multi-task reinforcement learning, diffusion models, planning, data synthesis

Compressor summary: The paper proposes a diffusion model that can handle multiple tasks with diverse and multimodal data in offline settings, using Transformer backbones and prompt learning for planning and synthesis.


Sequential Memory with Temporal Predictive Coding

Mufeng Tang, Helen Barron, Rafal Bogacz

https://openreview.net/forum?id=f8zIs2IB6Q

Keywords: Predictive coding, sequential memory, hippocampus

Compressor summary: The authors propose a novel predictive coding-based model for sequential memory called temporal predictive coding, which can accurately memorize and retrieve sequential inputs with a biologically plausible neural implementation and exhibits properties consistent with neuroscience theories.


Learning Re-sampling Methods with Parameter Attribution for Image Super-resolution

Xiaotong Luo, Yuan Xie, Yanyun Qu

https://openreview.net/forum?id=f7wFwPJwBe

Keywords: image super-resolution, long-tail distribution, re-sampling, integrated gradient

Compressor summary: The paper proposes a bi-sampling method for single image super-resolution, which balances the training data and improves the model's ability to reconstruct hard regions.


Provable convergence guarantees for black-box variational inference

Justin Domke, Robert M. Gower, Guillaume Garrigos

https://openreview.net/forum?id=f71xXsoG1v

Keywords: optimization, variational inference

Compressor summary: The text discusses how to provide theoretical guarantees for black-box variational inference by improving gradient estimators with unusual noise bounds and a composite non-smooth objective.


Reward Finetuning for Faster and More Accurate Unsupervised Object Discovery

Katie Z Luo, Zhenzhen Liu, Xiangyu Chen, Yurong You, Sagie Benaim, Cheng Perng Phoo, Mark Campbell, Wen Sun, Bharath Hariharan, Kilian Q Weinberger

https://openreview.net/forum?id=f6rQJ83ycb

Keywords: Self Driving, Self-Supervised Object Discovery, Reward Ranked Finetuning

Compressor summary: The paper proposes a machine learning method for detecting objects from LiDAR points using heuristics as human feedback, leading to faster and more accurate results than previous methods.


SA-Solver: Stochastic Adams Solver for Fast Sampling of Diffusion Models

Shuchen Xue, Mingyang Yi, Weijian Luo, Shifeng Zhang, Jiacheng Sun, Zhenguo Li, Zhi-Ming Ma

https://openreview.net/forum?id=f6a9XVFYIo

Keywords: Diffusion Model Sampler, Multi-step SDE Solver

Compressor summary: The paper proposes SA-Solver, an efficient stochastic method for solving diffusion SDEs and generating high-quality data using a variance-controlled approach.


Norm-guided latent space exploration for text-to-image generation

Dvir Samuel, Rami Ben-Ari, Nir Darshan, Haggai Maron, Gal Chechik

https://openreview.net/forum?id=f56xMRb7Vt

Keywords: diffusion models, few-shot learning, long-tail learning

Compressor summary: The paper proposes a new method for interpolating between text-to-image diffusion seeds that considers their norm values, which improves the generation of rare concepts and performance on few-shot and long-tail learning tasks.


The Learnability of In-Context Learning

Noam Wies, Yoav Levine, Amnon Shashua

https://openreview.net/forum?id=f3JNQd7CHM

Keywords: in-context, PAC, language models, foundation models, LLMs

Compressor summary: This paper proposes a theoretical framework for in-context learning, showing how it can efficiently identify and learn tasks from input examples without modifying large language models' weights.


Privacy Auditing with One (1) Training Run

Thomas Steinke, Milad Nasr, Matthew Jagielski

https://openreview.net/forum?id=f38EY21lBw

Keywords: Differential privacy, membership inference attacks, privacy auditing

Compressor summary: The proposed auditing scheme for differentially private machine learning systems uses parallelism and statistical generalization to assess privacy with minimal assumptions and fewer training runs than existing methods.


Iterative Reachability Estimation for Safe Reinforcement Learning

Milan Ganai, Zheng Gong, Chenning Yu, Sylvia Lee Herbert, Sicun Gao

https://openreview.net/forum?id=f2U4HCY8bg

Keywords: Constraints, Safety, Hamilton Jacobi Reachability, Deep Reinforcement Learning, Robotics

Compressor summary: The paragraph introduces a new framework called RESPO that optimizes for rewards and safety in reinforcement learning and shows its effectiveness on various environments.


HubRouter: Learning Global Routing via Hub Generation and Pin-hub Connection

Xingbo Du, Chonghua Wang, Ruizhe Zhong, Junchi Yan

https://openreview.net/forum?id=f0Jj3C3Pnp

Keywords: global routing, generative models

Compressor summary: The paper proposes a novel hub-based approach for global routing in VLSI systems, using deep generative models for hub generation and an actor-critic model for pin-hub connection, which improves efficiency and quality over existing methods.


CROMA: Remote Sensing Representations with Contrastive Radar-Optical Masked Autoencoders

Anthony Fuller, Koreen Millard, James R Green

https://openreview.net/forum?id=ezqI5WgGvY

Keywords: Remote Sensing, Earth Observation, Self-supervised learning, Multimodal

Compressor summary: CROMA is a framework that uses self-supervised learning to learn rich unimodal and multimodal representations from spatially aligned multimodal data in remote sensing, outperforming the current SoTA multispectral model on various benchmarks.


$\texttt{TACO}$: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning

Ruijie Zheng, Xiyao Wang, Yanchao Sun, Shuang Ma, Jieyu Zhao, Huazhe Xu, Hal Daumé III, Furong Huang

https://openreview.net/forum?id=ezCsMOy1w9

Keywords: Deep Reinforcement Learning, Visual Reinforcement Learning, Online Visual RL, Offline Visual RL, Action Representation

Compressor summary: $\texttt{TACO}$ is a temporal contrastive learning method that improves sample efficiency in reinforcement learning by learning concurrent state and action representations.


Continual Learning for Instruction Following from Realtime Feedback

Alane Suhr, Yoav Artzi

https://openreview.net/forum?id=ez6Cb0ZGzG

Keywords: continual learning, interaction, instruction following, user feedback, natural language processing, language grounding, situated interaction, collaboration

Compressor summary: The paper presents an approach that trains an agent from user feedback in natural language and binary form, using contextual bandit learning to improve instruction execution accuracy and show robustness and equivalence to supervised data.


An Adaptive Algorithm for Learning with Unknown Distribution Drift

Alessio Mazzetto, Eli Upfal

https://openreview.net/forum?id=exiXmAfuDK

Keywords: statistical learning, learning theory, machine learning, supervised learning, non-stationary, transfer learning, distribution drift

Compressor summary: The paper presents a method for learning from data with unknown distribution changes, without needing information about the change magnitude, and shows its effectiveness in binary classification and linear regression tasks.


Model Spider: Learning to Rank Pre-Trained Models Efficiently

Yi-Kai Zhang, Ting-Ji Huang, Yao-Xiang Ding, De-Chuan Zhan, Han-Jia Ye

https://openreview.net/forum?id=exg62lfHrB

Keywords: Pre-trained Model Ranking, Transfer Learning

Compressor summary: Model Spider is a method that uses vector representations to efficiently select and enrich Pre-Trained Models for various tasks, including visual models and Large Language Models.


Learning Dictionary for Visual Attention

Yingjie Liu, Xuan Liu, Hui Yu, XUAN TANG, Xian Wei

https://openreview.net/forum?id=exPzwOhBgx

Keywords: dictionary learning, attention, transformer, computer vision, point cloud

Compressor summary: The authors propose a novel attention module called Dic-Attn, which uses dictionary learning to decompose and reconstruct input data, enabling efficient visual attention construction for computer vision tasks.


Geometry-Aware Adaptation for Pretrained Models

Nicholas Roberts, Xintong Li, Dyah Adila, Sonia Cromp, Tzu-Heng Huang, Jitian Zhao, Frederic Sala

https://openreview.net/forum?id=exGOXqxR0L

Keywords: structured prediction, learning on graphs, partially observed label spaces, high cardinality label spaces

Compressor summary: The proposed technique Loki improves the performance of machine learning models by using external or self-derived metrics to adapt to new classes without additional training, achieving significant improvements over existing methods.


VPP: Efficient Conditional 3D Generation via Voxel-Point Progressive Representation

Zekun Qi, Muzhou Yu, Runpei Dong, Kaisheng Ma

https://openreview.net/forum?id=etd0ebzGOG

Keywords: Text to shape generation, 3D shape generation, Efficient inference, Representation Learning

Compressor summary: The proposed Voxel-Point Progressive Representation (VPP) method improves 3D generation quality and efficiency by combining structured voxel representation and sparse point representation, enabling diverse and high-fidelity 3D shape creation for various downstream tasks.


Causal Discovery from Subsampled Time Series with Proxy Variables

Mingzhou Liu, Xinwei Sun, Lingjing Hu, Yizhou Wang

https://openreview.net/forum?id=etYk6TeO2q

Keywords: causal discovery, time series, subsampling, proxy variables

Compressor summary: The paper proposes a new method to infer causality from time series data with missing measurements, using proxies for hidden variables.


Doubly-Robust Self-Training

Banghua Zhu, Mingyu Ding, Philip Jacobson, Ming Wu, Wei Zhan, Michael Jordan, Jiantao Jiao

https://openreview.net/forum?id=esy7pkZmKn

Keywords: semi-supervised learning, self-training, auto-labeling, self-labeling, doubly robust

Compressor summary: The paper proposes a semi-supervised learning algorithm called doubly-robust self-training that adapts to the quality of pseudo-labels and outperforms the standard self-training method in image classification and 3D object detection tasks.


Supply-Side Equilibria in Recommender Systems

Meena Jagadeesan, Nikhil Garg, Jacob Steinhardt

https://openreview.net/forum?id=eqyhjLG5Nr

Keywords: content creator incentives, Nash equilibria, specialization, economic aspects of recommender systems

Compressor summary: The paragraph discusses how algorithmic recommender systems affect producer behavior and content diversity, and presents a model that analyzes the conditions for specialization and its impact on profitability in such systems.


Understanding and Improving Feature Learning for Out-of-Distribution Generalization

Yongqiang Chen, Wei Huang, Kaiwen Zhou, Yatao Bian, Bo Han, James Cheng

https://openreview.net/forum?id=eozEoAtjG8

Keywords: Out-of-Distribution Generalization, Feature Learning, Invariant Risk Minimization

Compressor summary: The paragraph discusses how ERM can learn both spurious and invariant features, but tends to favor spurious ones if they are stronger, making it harder for models to generalize to out-of-distribution data. It proposes Feature Augmented Training (FeAT) to improve feature learning for OOD tasks.


Black-Box Differential Privacy for Interactive ML

Haim Kaplan, Yishay Mansour, Shay Moran, Kobbi Nissim, Uri Stemmer

https://openreview.net/forum?id=eoDNaH3pfB

Keywords: Differential privacy, online learning

Compressor summary: The authors propose an interactive variant of joint differential privacy that can handle online processes more effectively than traditional forms, allowing for private online classification with less error.


Train Once and Explain Everywhere: Pre-training Interpretable Graph Neural Networks

Jun Yin, Chaozhuo Li, Hao Yan, Jianxun Lian, Senzhang Wang

https://openreview.net/forum?id=enfx8HM4Rp

Keywords: Intrinsic Interpretability, Graph Neural Networks, Pre-training and Fine-tuning

Compressor summary: The Pre-training Interpretable Graph Neural Network (π-GNN) is a new model that learns universal patterns of graph structure and local interactions to provide transparent predictions by identifying influential parts of the input graph.


Getting ViT in Shape: Scaling Laws for Compute-Optimal Model Design

Ibrahim Alabdulmohsin, Xiaohua Zhai, Alexander Kolesnikov, Lucas Beyer

https://openreview.net/forum?id=en4LGxpd9E

Keywords: Vision transformer, scaling laws, compute-optimal model design, vision

Compressor summary: The paper presents SoViT, a shape-optimized vision transformer that achieves competitive results with much larger models while using less compute resources, and evaluates its performance across various tasks.


LaFTer: Label-Free Tuning of Zero-shot Classifier using Language and Unlabeled Image Collections

Muhammad Jehanzeb Mirza, Leonid Karlinsky, Wei Lin, Horst Possegger, Mateusz Kozinski, Rogerio Feris, Horst Bischof

https://openreview.net/forum?id=elPtHcfjpH

Keywords: VL Models

Compressor summary: The paper proposes a label-free method to improve zero-shot visual classification using an unlabeled image collection and auto-generated texts describing the categories, achieving significant performance improvements and outperforming some few-shot prompting methods.


Simultaneous embedding of multiple attractor manifolds in a recurrent neural network using constrained gradient optimization

Haggai Agmon, Yoram Burak

https://openreview.net/forum?id=ekMLUoC2sq

Keywords: Theoretical Neuroscience, Computational Neuroscience, Recurrent Neural Networks, Attractor models

Compressor summary: The paragraph discusses how synaptic weight adjustment can reduce interference in recurrent neural networks that store multiple continuous variables in working memory, leading to better memory storage.


Resilient Multiple Choice Learning: A learned scoring scheme with application to audio scene analysis

Victor Letzelter, Mathieu Fontaine, Mickael Chen, Patrick Perez, Slim Essid, Gaël Richard

https://openreview.net/forum?id=eibTaY6qGI

Keywords: Multiple Choice Learning, Audio processing.

Compressor summary: Resilient Multiple Choice Learning (rMCL) is a new method for estimating conditional distributions in regression settings with multiple targets per input, using a novel scoring scheme based on Voronoi tessellations to preserve diversity in predictions.


Learning and processing the ordinal information of temporal sequences in recurrent neural circuits

Xiaolong Zou, Zhikun Chu, Qinghai Guo, Jie Cheng, Bo Ho, Si Wu, Yuanyuan Mi

https://openreview.net/forum?id=eeeqORvJbf

Keywords: temporal sequence processing, temporal order structure, tree-structured attractor

Compressor summary: The paragraph discusses how recurrent neural circuits can learn to represent the abstract order structure of temporal sequences, allowing for flexible and robust processing of temporal sequences, and suggests that understanding this mechanism could help develop brain-inspired algorithms.


DiffVL: Scaling Up Soft Body Manipulation using Vision-Language Driven Differentiable Physics

Zhiao Huang, Feng Chen, Yewen Pu, Chunru Lin, Hao Su, Chuang Gan

https://openreview.net/forum?id=ecRaDicXxw

Keywords: Differentiable physics; Soft body manipulation

Compressor summary: DiffVL is a method that lets non-experts create soft-body manipulation tasks using vision and natural language, making it easier for differential physics solvers to find solutions.


Semantic HELM: A Human-Readable Memory for Reinforcement Learning

Fabian Paischer, Thomas Adler, Markus Hofmarcher, Sepp Hochreiter

https://openreview.net/forum?id=ebMPmx5mr7

Keywords: Reinforcement Learning, Language Models, History Compression, Partial Observability, Foundation Models, Interpretability, Explainable AI

Compressor summary: The authors propose a novel memory mechanism for reinforcement learning agents in partially observable environments that uses CLIP to associate visual inputs with language tokens, which are then fed to a pretrained language model to provide a human-readable representation of the past, improving interpretability and performance on tasks requiring memory.


Graph-Structured Gaussian Processes for Transferable Graph Learning

Jun Wu, Lisa Ainsworth, Andrew Leakey, Haixun Wang, Jingrui He

https://openreview.net/forum?id=eZbqD9BoXe

Keywords: graph learning, transfer learning, Gaussian process

Compressor summary: GraphGP is a framework for adaptively transferring knowledge across graphs with different assumptions, addressing challenges in transferable graph learning by using a novel graph structure-aware neural network and showing its effectiveness through experiments.


Unleash the Potential of Image Branch for Cross-modal 3D Object Detection

Yifan Zhang, Qijian Zhang, Junhui Hou, Yixuan Yuan, Guoliang Xing

https://openreview.net/forum?id=eYCGrGdKf3

Keywords: 3D object detection, 3D point cloud

Compressor summary: The paper introduces UPIDet, a new cross-modal 3D object detector for autonomous vehicles that leverages image domain information to improve LiDAR-based detectors' performance and achieves top rank in the KITTI benchmark.


Penguin: Parallel-Packed Homomorphic Encryption for Fast Graph Convolutional Network Inference

Ran Ran, Nuo Xu, Tao Liu, Wei Wang, Gang Quan, Wujie Wen

https://openreview.net/forum?id=eXubleMT0q

Keywords: Cryptographic inference, Graph Convolutional Network, Parallel Packing

Compressor summary: The authors propose Penguin, a ciphertext packing technique for enhancing the efficiency of homomorphic encryption-based graph convolutional network inference on encrypted data.


Flat Seeking Bayesian Neural Networks

Van-Anh Nguyen, Long Tung Vuong, Hoang Phan, Thanh-Toan Do, Dinh Phung, Trung Le

https://openreview.net/forum?id=eX6xDto3Ed

Keywords: Bayesian, sharpness-aware, posterior

Compressor summary: The paper proposes a method to improve the generalization ability of Bayesian Neural Networks by making the posterior distribution aware of the model's sharpness/flatness.


Practical and Asymptotically Exact Conditional Sampling in Diffusion Models

Luhuan Wu, Brian L. Trippe, Christian A Naesseth, David Blei, John Patrick Cunningham

https://openreview.net/forum?id=eWKqr1zcRv

Keywords: diffusion models; conditional sampling; sequential monte carlo methods; generative models; protein design

Compressor summary: The Twisted Diffusion Sampler (TDS) is a sequential Monte Carlo algorithm that improves conditional generation by incorporating heuristic approximations without compromising exactness, achieving better results in simulation and protein design tasks.


Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline

Zangwei Zheng, Xiaozhe Ren, Fuzhao Xue, Yang Luo, Xin Jiang, Yang You

https://openreview.net/forum?id=eW233GDOpm

Keywords: large language models, inference optimization, batch processing

Compressor summary: The paper proposes an efficient inference pipeline for large language models that groups queries with similar response lengths into micro-batches, improving throughput without sacrificing effectiveness.


Inferring the Future by Imagining the Past

Kartik Chandra, Tony Chen, Tzu-Mao Li, Jonathan Ragan-Kelley, Joshua B. Tenenbaum

https://openreview.net/forum?id=eVrmcOvJV4

Keywords: cognitive science, cogsci, inverse planning, Bayesian inference, theory of mind, Monte Carlo, inverse reinforcement learning

Compressor summary: The paper proposes a Monte Carlo algorithm that models how humans infer complex sequences of events from static comic book panels or other dynamic scenes, by connecting its inference problem to Monte Carlo path tracing in computer graphics.


XAGen: 3D Expressive Human Avatars Generation

Zhongcong Xu, Jianfeng Zhang, Jun Hao Liew, Jiashi Feng, Mike Zheng Shou

https://openreview.net/forum?id=eUf0CaS5AP

Keywords: Human Avatar, 3D-aware GAN

Compressor summary: XAGen is a 3D generative model for human avatars that can control facial expressions, jaw poses, hand poses, and other expressive attributes, achieving high realism and diversity.


Algorithmic Regularization in Tensor Optimization: Towards a Lifted Approach in Matrix Sensing

Ziye Ma, Javad Lavaei, Somayeh Sojoudi

https://openreview.net/forum?id=eU6P4aUdCA

Keywords: non-convex optimization, low-rank matrix optimization, matrix sensing, implicit bias, tensor, over-parametrization

Compressor summary: The paper investigates how gradient descent helps generalization in machine learning models by inducing implicit regularization for tensor optimization in the lifted matrix sensing framework.


ASPEN: Breaking Operator Barriers for Efficient Parallelization of Deep Neural Networks

Jongseok Park, Kyungmin Bin, Gibum Park, Sangtae Ha, Kyunghan Lee

https://openreview.net/forum?id=eTp4RetK74

Keywords: Deep Neural Network, Deep Learning, Parallel Execution Algorithm, Parallelization, Deep Learning Parallelism, Dynamic, Asynchronous, Scheduling, Dynamic Scheduling, Dynamic Execution, tile, tiling, dataflow, dataflow graph, tile-based dataflow graph, opportunistic parallelism

Compressor summary: ASPEN is a novel parallel computation solution for DNNs that removes operator barriers, exposes parallelism across operators, and achieves high resource utilization and memory reuse, outperforming TorchScript and TVM by up to 4.3$\times$.


Double Randomized Underdamped Langevin with Dimension-Independent Convergence Guarantee

Yuanshi Liu, Cong Fang, Tong Zhang

https://openreview.net/forum?id=eTMHsUp3Ii

Keywords: Langevin, Dimension dependence, Acceleration

Compressor summary: The paper presents a double randomization method for fast high-dimensional sampling of log-concave distributions with composite structures, achieving dimension-independent convergence guarantees and outperforming previous results by a $d^{1/3}$ factor.


A Rigorous Link between Deep Ensembles and (Variational) Bayesian Methods

Veit David Wild, Sahra Ghalebikesabi, Dino Sejdinovic, Jeremias Knoblauch

https://openreview.net/forum?id=eTHawKFT4h

Keywords: Wasserstein gradient flow, generalised variational inference, deep ensembles, Bayesian deep learning, variational Bayes

Compressor summary: The paper presents a unified theory linking Bayesian, variational Bayesian, and ensemble methods for uncertainty quantification in deep learning by reformulating the non-convex optimization problem as a convex one using Wasserstein gradient flows.


Direct Training of SNN using Local Zeroth Order Method

Bhaskar Mukhoty, Velibor Bojkovic, William de Vazelhes, Xiaohan Zhao, Giulia De Masi, Huan Xiong, Bin Gu

https://openreview.net/forum?id=eTF3VDH2b6

Keywords: Spiking Neural Network, Zeroth Order, Surrogate Gradient

Compressor summary: The paragraph describes a new technique for training spiking neural networks that uses zeroth-order methods, surrogate approximations of the Heaviside function, and efficient backpropagation to achieve better accuracy and faster training.


Spatially Resolved Gene Expression Prediction from Histology Images via Bi-modal Contrastive Learning

Ronald Xie, Kuan Pang, Sai W Chung, Catia Perciani, Sonya MacParland, BO WANG, Gary Bader

https://openreview.net/forum?id=eT1tMdAUoc

Keywords: BLEEP, Histology, H&E, Gene Expression Prediction, Spatial Transcriptomics, Contrastive Learning

Compressor summary: BLEEP is a framework that predicts gene expression from histology images, providing insights into tissue architecture and potentially reducing the time and cost of diagnosis and research.


Strategyproof Voting under Correlated Beliefs

Daniel Halpern, Rachel Li, Ariel D. Procaccia

https://openreview.net/forum?id=eT1QOsssRB

Keywords: social choice, strategyproof, voting

Compressor summary: The paper explores how relaxing strategyproofness criteria for selecting winners in ranked voting scenarios can still lead to impossibility theorems, and argues that plurality rule is a promising choice due to its strategyproofness under certain belief models.


Classification of Heavy-tailed Features in High Dimensions: a Superstatistical Approach

Urte Adomaityte, Gabriele Sicuro, Pierpaolo Vivo

https://openreview.net/forum?id=eR7PrfJe9o

Keywords: Classification, Gaussian Mixture Model, Superstatistics, Empirical Risk Minimization, Replica theory, Power-law distribution

Compressor summary: The paper investigates how a mixture of two clouds of data points with random variances learns in high dimensions under different assumptions and distributions, testing Gaussian universality claims and analysing generalisation and separability.


Combating Bilateral Edge Noise for Robust Link Prediction

Zhanke Zhou, Jiangchao Yao, Jiaxu Liu, Xiawei Guo, quanming yao, LI He, Liang Wang, Bo Zheng, Bo Han

https://openreview.net/forum?id=ePkLqJh5kw

Keywords: Robust link prediction, Edge noise

Compressor summary: The text proposes a method called Robust Graph Information Bottleneck (RGIB) to improve link prediction on graphs by extracting reliable supervision signals and avoiding representation collapse under edge noise.


TRIAGE: Characterizing and auditing training data for improved regression

Nabeel Seedat, Jonathan Crabbé, Zhaozhi Qian, Mihaela van der Schaar

https://openreview.net/forum?id=eP6cDDwBNC

Keywords: data-centric AI, data characterization, data quality

Compressor summary: TRIAGE is a novel framework for characterizing training data in regression tasks, using conformal predictive distributions to score samples as under-, over-, or well-estimated by the model, and enabling data sculpting/filtering and dataset selection improvements.


Contextual Gaussian Process Bandits with Neural Networks

Haoting Zhang, Jinghai He, Rhonda Righter, Zuo-Jun Shen, Zeyu Zheng

https://openreview.net/forum?id=eNhW9UnlGG

Keywords: contextual bandit, Gaussian process, neural network

Compressor summary: The paper proposes a neural network-accompanied Gaussian process (NN-AGP) model for solving contextual decision-making problems with high accuracy and uncertainty quantification, and proves its theoretical guarantees and shows its empirical performance.


Diversify \& Conquer: Outcome-directed Curriculum RL via Out-of-Distribution Disagreement

Daesol Cho, Seungjae Lee, H. Jin Kim

https://openreview.net/forum?id=eMR57voMz1

Keywords: Curriculum learning, Out-of-distribution disagreement, Underspecification, Outcome-directed RL

Compressor summary: The paper proposes a new curriculum RL method called D2C that uses diversification of classifiers and bipartite matching to explore and conquer uninformed search problems without access to domain knowledge or environment characteristics.


Equivariant flow matching

Leon Klein, Andreas Krämer, Frank Noe

https://openreview.net/forum?id=eLH2NFOO1B

Keywords: Normalizing Flows, Flow Matching, Equivariance, Boltzmann Generators, Molecular Dynamics, Optimal Transport

Compressor summary: Equivariant flow matching is a new training method for normalizing flows in physics that uses optimal transport and exploits physical symmetries to improve sampling efficiency and scalability.


Frequency-Enhanced Data Augmentation for Vision-and-Language Navigation

Keji He, Chenyang Si, Zhihe Lu, Yan Huang, Liang Wang, Xinchao Wang

https://openreview.net/forum?id=eKFrXWb0sT

Keywords: Vision-and-Language Navigation; High-Frequency; Data Augmentation

Compressor summary: The paper proposes a new technique to improve visual-textual matching in vision-and-language navigation tasks by using high-frequency information and data augmentation, leading to better performance with simple and efficient models.


What Planning Problems Can A Relational Neural Network Solve?

Jiayuan Mao, Tomás Lozano-Pérez, Joshua B. Tenenbaum, Leslie Pack Kaelbling

https://openreview.net/forum?id=eJZ5vJEaaa

Keywords: Planning, Relational Neural Network, Circuit Complexity

Compressor summary: The paper analyzes how relational neural networks can be used to learn goal-conditioned policies for planning problems, depending on the number of objects and planning horizon.


AD-PT: Autonomous Driving Pre-Training with Large-scale Point Cloud Dataset

Jiakang Yuan, Bo Zhang, Xiangchao Yan, Botian Shi, Tao Chen, Yikang LI, Yu Qiao

https://openreview.net/forum?id=eIFZtkshgH

Keywords: 3D Object Detection, 3D Pre-training, Autonomous Driving

Compressor summary: The paper introduces a semi-supervised point cloud pre-training approach for autonomous driving that improves generalizability across various tasks and benchmarks.


RGMIL: Guide Your Multiple-Instance Learning Model with Regressor

Zhaolong Du, Shasha Mao, Yimeng Zhang, Shuiping Gou, Licheng Jiao, Lin Xiong

https://openreview.net/forum?id=eGoE9CVRPc

Keywords: Video Analysis, Multiple-Instance Learning, Representation learning

Compressor summary: The authors propose RGMIL, a method for video analysis with limited annotations, which uses Regressor-Guided Pooling (RGP) to produce discriminative instance-level representations and achieve near ideal performance compared to supervised models.


Revisiting Visual Model Robustness: A Frequency Long-Tailed Distribution View

Zhiyu Lin, Yifei Gao, Yunfan Yang, Jitao Sang

https://openreview.net/forum?id=eE5L1RkxW0

Keywords: visual models, robustness, frequency domain, long-tailed distribution

Compressor summary: The paper proposes a new strategy to improve visual models' robustness and accuracy by addressing the under-fitting behavior on high-frequency components (HFC) in images, which is caused by limited information content in HFC.


How to Select Which Active Learning Strategy is Best Suited for Your Specific Problem and Budget

Guy Hacohen, Daphna Weinshall

https://openreview.net/forum?id=eDDZh8C4W4

Keywords: Deep Active learning, Low budget, High budget, Deep learning

Compressor summary: The paper proposes a practical method for selecting the best active learning query strategy based on the problem characteristics and available budget, and demonstrates its effectiveness in computer vision tasks.


Evaluating Post-hoc Explanations for Graph Neural Networks via Robustness Analysis

Junfeng Fang, Wei Liu, Yuan Gao, Zemin Liu, An Zhang, Xiang Wang, Xiangnan He

https://openreview.net/forum?id=eD534mPhAg

Keywords: Post-hoc Explainability, Explanation Evaluation, Graph Neural Network, Robustness Analysis

Compressor summary: The paper proposes OAR and SimOAR, novel evaluation metrics for explaining graph neural networks that address the out-of-distribution issue and improve credibility by measuring adversarial robustness.


On Sparse Modern Hopfield Model

Jerry Yao-Chieh Hu, Donglin Yang, Dennis Wu, Chenwei Xu, Bo-Yu Chen, Han Liu

https://openreview.net/forum?id=eCgWNU2Imw

Keywords: Hopfield Models; Modern Hopfield Networks; Sparse Attention; Memory Networks

Compressor summary: The paper introduces a sparse version of the modern Hopfield model with improved theoretical properties and empirical performance.


POP-3D: Open-Vocabulary 3D Occupancy Prediction from Images

Antonín Vobecký, Oriane Siméoni, David Hurych, Spyros Gidaris, Andrei Bursuc, Patrick Perez, Josef Sivic

https://openreview.net/forum?id=eBXM62SqKY

Keywords: open-vocabulary segmentation, voxel occupancy prediction, semantic segmentation, autonomous driving, language-image alignment

Compressor summary: The paper presents a new model for predicting 3D semantic occupancy maps from 2D images using self-supervised learning without 3D annotations.


Automatic Clipping: Differentially Private Deep Learning Made Easier and Stronger

Zhiqi Bu, Yu-Xiang Wang, Sheng Zha, George Karypis

https://openreview.net/forum?id=e8i7OaPj0q

Keywords: deep learning, differential privacy, optimization, hyper-parameter tuning

Compressor summary: Automatic clipping is a technique that simplifies differential privacy training for deep learning models by eliminating the need to tune a clipping threshold and achieving similar or better performance than existing methods.


Using Imperfect Surrogates for Downstream Inference: Design-based Supervised Learning for Social Science Applications of Large Language Models

Naoki Egami, Musashi Hinck, Brandon M. Stewart, Hanying Wei

https://openreview.net/forum?id=e8RZwixcE4

Keywords: Computational Social Science, Large Language Models, Statistical Inference, Causal Inference

Compressor summary: The authors propose a new algorithm (DSL) that uses imperfect annotations from large language models for computational social science research, ensuring unbiased and accurate statistical analysis by combining them with a smaller number of high-quality labels.


DynGFN: Towards Bayesian Inference of Gene Regulatory Networks with GFlowNets

Lazar Atanackovic, Alexander Tong, BO WANG, Leo J Lee, Yoshua Bengio, Jason Hartford

https://openreview.net/forum?id=e7MK5Vq44Q

Keywords: Bayesian Structure Learning, Generative Flow Networks, Single-cell, Dynamical Systems

Compressor summary: The paper proposes a novel approach for inferring gene regulatory networks using RNA velocity techniques and Generative Flow Networks, addressing both the challenges of cyclicity and uncertainty in the data.


Accessing Higher Dimensions for Unsupervised Word Translation

Sida Wang

https://openreview.net/forum?id=e5srDjF9l7

Keywords: co-occurrences, unsupervised word translation, bilingual lexicon induction, robust statistics, unsupervised machine translation

Compressor summary: The authors develop a method for unsupervised word translation that uses high-dimensional signals instead of low-dimensional word vectors, achieving better results with less data and computation.


Gacs-Korner Common Information Variational Autoencoder

Michael Kleinman, Alessandro Achille, Stefano Soatto, Jonathan Kao

https://openreview.net/forum?id=e4XidX6AHd

Keywords: Common Information, Gacs-Korner, Variational Autoencoder

Compressor summary: The paragraph introduces a concept called common information that separates shared and unique aspects of two random variables, which can be estimated from data using variational autoencoders and measured against ground-truth factors.


CLadder: A Benchmark to Assess Causal Reasoning Capabilities of Language Models

Zhijing Jin, Yuen Chen, Felix Leeb, Luigi Gresele, Ojasv Kamal, Zhiheng LYU, Kevin Blin, Fernando Gonzalez Adauto, Max Kleiman-Weiner, Mrinmaya Sachan, Bernhard Schölkopf

https://openreview.net/forum?id=e2wtjx0Yqu

Keywords: Large Language Models, Causal Reasoning, Causal Inference, Benchmark Dataset, Natural Language Processing

Compressor summary: The paper introduces a new natural language processing task and dataset to assess whether large language models can perform causal inference based on formal rules.


Estimating and Controlling for Equalized Odds via Sensitive Attribute Predictors

Beepul Bharti, Paul Yi, Jeremias Sulam

https://openreview.net/forum?id=e2aCgjtjMR

Keywords: fairness, sensitive attributes, equalized odds, missing data, proxies

Compressor summary: This paper proposes a new method for controlling fairness in machine learning models without access to sensitive attributes by using post-processing corrections and provides theoretical guarantees and experimental validation.


Likelihood-Based Diffusion Language Models

Ishaan Gulrajani, Tatsunori Hashimoto

https://openreview.net/forum?id=e2MCL6hObn

Keywords: diffusion, language, model

Compressor summary: The authors develop and release a diffusion-based language model (Plaid 1B) that surpasses GPT-2 124M in likelihood on benchmarks and produces coherent samples.


Multinomial Logistic Regression: Asymptotic Normality on Null Covariates in High-Dimensions

Kai Tan, Pierre C Bellec

https://openreview.net/forum?id=e1oe8F2tjV

Keywords: High-dimensional statistics, statistical inference, multi-class classification, asymptotic normality, multinomial logistic regression

Compressor summary: The paper explores how to accurately estimate and test the importance of features in high-dimensional multinomial logistic models.


MathNAS: If Blocks Have a Role in Mathematical Architecture Design

Wang Qinsi, Jinghan Ke, Zhi Liang, Sihai Zhang

https://openreview.net/forum?id=e1l4ZYprQH

Keywords: Neural Architecture Search

Compressor summary: MathNAS is a novel NAS framework that predicts network performance training-free by calculating the performances of all possible building blocks first, reducing the complexity of evaluation and achieving state-of-the-art results on large-scale datasets.


Large Language Models of Code Fail at Completing Code with Potential Bugs

Tuan Dinh, Jinman Zhao, Samson Tan, Renato Negrinho, Leonard Lausen, Sheng Zha, George Karypis

https://openreview.net/forum?id=e1WgjvFGWp

Keywords: language model of code; code completion; language model; software engineering; machine learning for code

Compressor summary: The paragraph discusses code completion with buggy code contexts and introduces two datasets to study the problem, showing that current Code-LLMs struggle with this scenario and there is room for improvement in mitigating bugs.


Aligning Gradient and Hessian for Neural Signed Distance Function

Ruian Wang, Zixiong Wang, Yunxiao Zhang, Shuangmin Chen, Shiqing Xin, Changhe Tu, Wenping Wang

https://openreview.net/forum?id=e0tt2G8hqf

Keywords: implicit neural representation, signed distance function, shape operator

Compressor summary: The paper introduces a method to learn the Signed Distance Function (SDF) from point clouds without normals, which improves gradient accuracy and reduces ghost geometry in surface reconstruction.


Private (Stochastic) Non-Convex Optimization Revisited: Second-Order Stationary Points and Excess Risks

Daogao Liu, Arun Ganesh, Sewoong Oh, Abhradeep Guha Thakurta

https://openreview.net/forum?id=e0pRF9tOtm

Keywords: Differential Privacy, Non-convex optimization, Stationary points, Exponential Mechanism

Compressor summary: The paper proposes a novel framework for non-convex optimization under differential privacy constraint that uses two types of gradient oracles to improve accuracy and identifies second-order stationary points.


Hypernetwork-based Meta-Learning for Low-Rank Physics-Informed Neural Networks

Woojin Cho, Kookjin Lee, Donsub Rim, Noseong Park

https://openreview.net/forum?id=dzqKAM2sKa

Keywords: Scientific machine learning, Physics-informed neural networks, Meta learning, Hypernetworks

Compressor summary: The study suggests using lightweight low-rank physics-informed neural networks with meta-learning to solve partial differential equations for various input parameters quickly and effectively.


Characterizing Out-of-Distribution Error via Optimal Transport

Yuzhe Lu, Yilong Qin, Runtian Zhai, Andrew Shen, Ketong Chen, Zhenlin Wang, Soheil Kolouri, Simon Stepputtis, Joseph Campbell, Katia P. Sycara

https://openreview.net/forum?id=dz5X8hnfJc

Keywords: Distribution Shift, OOD Error Prediction, Optimal Transport, Deep Learning

Compressor summary: The paper proposes a new method for estimating model performance on out-of-distribution data using optimal transport theory and an empirically-motivated variant that improves accuracy and robustness.


GMSF: Global Matching Scene Flow

Yushan Zhang, Johan Edstedt, Bastian Wandt, Per-Erik Forssen, Maria Magnusson, Michael Felsberg

https://openreview.net/forum?id=dybrsuNAB9

Keywords: Scene flow, point clouds, transformers

Compressor summary: The paper proposes a simple one-shot global matching method for scene flow estimation from point clouds, which achieves state-of-the-art results on various benchmarks and outperforms previous methods.


Compositional Foundation Models for Hierarchical Planning

Anurag Ajay, Seungwook Han, Yilun Du, Shuang Li, Abhi Gupta, Tommi S. Jaakkola, Joshua B. Tenenbaum, Leslie Pack Kaelbling, Akash Srivastava, Pulkit Agrawal

https://openreview.net/forum?id=dyXNh5HLq3

Keywords: Foundation Models, Composition, Hierarchical Planning

Compressor summary: The HiP model combines language, vision, and action models to create symbolic plans for long-horizon tasks involving table-top manipulation.


Equivariant Single View Pose Prediction Via Induced and Restriction Representations

Owen Lewis Howell, David Klee, Ondrej Biza, Linfeng Zhao, Robin Walters

https://openreview.net/forum?id=dxVN2fZjx6

Keywords: Equivarient Machine Learning, Pose Prediction, Computer Vision

Compressor summary: The text discusses how to design neural networks for 3D object recognition from 2D images, considering constraints and consistency properties, and achieving state-of-the-art results on some tasks.


Few-shot Generation via Recalling Brain-Inspired Episodic-Semantic Memory

Zhibin Duan, Lv Zhiyi, Chaojie Wang, Bo Chen, Bo An, Mingyuan Zhou

https://openreview.net/forum?id=dxPcdEeQk9

Keywords: Generative Model, Memory-augmented Generative Model

Compressor summary: The authors propose a variational structured memory module (VSM) to enhance existing generative models for few-shot generation by mimicking human memory mechanisms.


Deep Patch Visual Odometry

Zachary Teed, Lahav Lipson, Jia Deng

https://openreview.net/forum?id=dwfHbm8g66

Keywords: SLAM, Simultaneous Localization and Mapping, Visual Odometry, Structure from motion, SfM

Compressor summary: Deep Patch Visual Odometry (DPVO) is a new system that uses sparse patch-based matching instead of dense flow for monocular visual odometry, achieving better accuracy and efficiency than previous methods.


SmoothHess: ReLU Network Feature Interactions via Stein's Lemma

Max Torop, Aria Masoomi, Davin Hill, Kivanc Kose, Stratis Ioannidis, Jennifer Dy

https://openreview.net/forum?id=dwIeEhbaD0

Keywords: Interpretability, Feature Interactions, Stein's Lemma

Compressor summary: SmoothHess is a method for estimating feature interactions in ReLU neural networks by using Stein's Lemma and efficient sampling, without needing network modifications or changing the architecture.


Actively Testing Your Model While It Learns: Realizing Label-Efficient Learning in Practice

Dayou Yu, Weishi Shi, Qi Yu

https://openreview.net/forum?id=du0hvEpgj8

Keywords: active learning, active testing

Compressor summary: The paper proposes a novel active testing while learning (ATL) framework that integrates active learning with active testing, reducing data annotation costs and improving model performance using an "active feedback" mechanism.


Counterfactual-Augmented Importance Sampling for Semi-Offline Policy Evaluation

Shengpu Tang, Jenna Wiens

https://openreview.net/forum?id=dsH244r9fA

Keywords: healthcare, reinforcement learning, offline RL, off-policy evaluation, counterfactuals

Compressor summary: The paper proposes a semi-offline evaluation framework for reinforcement learning that uses counterfactual trajectories annotated by humans to reduce bias and variance in off-policy evaluation.


The Memory-Perturbation Equation: Understanding Model's Sensitivity to Data

Peter Nickl, Lu Xu, Dharmesh Tailor, Thomas Möllenhoff, Mohammad Emtiyaz Khan

https://openreview.net/forum?id=dqS1GuoG2V

Keywords: model interpretability, model understanding, bayesian learning, robustness, adaptive learning

Compressor summary: The Memory-Perturbation Equation (MPE) simplifies the study of model sensitivity to training data perturbations, using Bayesian principles, and shows its usefulness in predicting generalization to unseen test data.


Large-Scale Distributed Learning via Private On-Device LSH

Tahseen Rabbani, Marco Bornstein, Furong Huang

https://openreview.net/forum?id=dpdbbN7AKr

Keywords: distributed learning, locality-sensitive hashing, recommender systems, compression

Compressor summary: The paper proposes a new family of hash functions for locality-sensitive hashing that enables privacy, personalization, and memory efficiency for pruning dense hidden layers in large-scale recommender networks.


Revisit Weakly-Supervised Audio-Visual Video Parsing from the Language Perspective

Yingying Fan, Yu Wu, Bo Du, Yutian Lin

https://openreview.net/forum?id=doWqIXcRlq

Keywords: Weakly-Supervised Audio-Visual Video Parsing, Language Guided Segment-Level Label Denoising, Dynamic Re-weighting

Compressor summary: The paper proposes a language-based method for weakly-supervised audio-visual video parsing that tackles segment-level label noise and outperforms existing approaches.


Fully Dynamic $k$-Clustering in $\tilde O(k)$ Update Time

Sayan Bhattacharya, Martin Costa, Silvio Lattanzi, Nikos Parotsidis

https://openreview.net/forum?id=dnGEPkmnzO

Keywords: clustering, k-median, k-means, dynamic algorithms, amortized analysis

Compressor summary: The paper presents a fast and approximate dynamic algorithm for the k-median and k-means problems on metric spaces, compares it to previous methods, and provides a lower bound analysis.


The Rank-Reduced Kalman Filter: Approximate Dynamical-Low-Rank Filtering In High Dimensions

Jonathan Schmidt, Philipp Hennig, Jörg Nick, Filip Tronarp

https://openreview.net/forum?id=dnB71DMyDD

Keywords: Gaussian, filtering, smoothing, bayesian, state-space models, dynamic-low-rank, high-dimensional, spatio-temporal, Gaussian processes, regression, low rank, state estimation

Compressor summary: The paper proposes a new low-rank Gaussian filtering and smoothing method that is faster and more accurate than existing ensemble-based methods for high-dimensional dynamical systems.


Trust Your $\nabla$: Gradient-based Intervention Targeting for Causal Discovery

Mateusz Olko, Michał Zając, Aleksandra Nowak, Nino Scherrer, Yashas Annadani, Stefan Bauer, Łukasz Kuciński, Piotr Miłoś

https://openreview.net/forum?id=dmD63sv0TZ

Keywords: causal discovery, experimental design, active learning, neural networks

Compressor summary: The paper proposes a method called GIT that uses gradient estimators to minimize the number of required experiments for acquiring interventional data and inferring causal structures from data.


Sample Complexity of Forecast Aggregation

Tao Lin, Yiling Chen

https://openreview.net/forum?id=dlDFakG6kJ

Keywords: information aggregation, sample complexity, distribution learning, Bayesian forecast aggregation

Compressor summary: A Bayesian model for forecast aggregation requires exponentially more samples when experts' signals depend on each other, but only linearly more samples if their signals are independent given the event outcome.


Scalable Transformer for PDE Surrogate Modeling

Zijie Li, Dule Shu, Amir Barati Farimani

https://openreview.net/forum?id=djyn8Q0anK

Keywords: Efficient attention, Neural PDE solver

Compressor summary: FactFormer is a new model that uses an axial factorized kernel integral to improve the efficiency and stability of applying Transformer models to problems with many grid points, such as simulating 2D and 3D fluid dynamics.


Improving Adversarial Transferability via Intermediate-level Perturbation Decay

Qizhang Li, Yiwen Guo, Wangmeng Zuo, Hao Chen

https://openreview.net/forum?id=dikH9tdPi2

Keywords: adversarial examples, black-box attack, adversarial transferability

Compressor summary: The ILPD method improves intermediate-level adversarial attacks by crafting effective and strong perturbations in one stage of optimization, leading to better performance against various victim models on ImageNet and CIFAR-10.


Fast Exact Leverage Score Sampling from Khatri-Rao Products with Applications to Tensor Decomposition

Vivek Bharadwaj, Osman Asif Malik, Riley Murray, Laura Grigori, Aydin Buluc, James Demmel

https://openreview.net/forum?id=deaHiTb6Cu

Keywords: Tensor Decomposition, Leverage Scores, Randomized Linear Algebra, Sketching, Khatri-Rao Product, Sparse Tensors

Compressor summary: The paper introduces a fast sampler for the Khatri-Rao product that can handle large matrices and outperforms existing methods in linear least-squares problems.


Functional Equivalence and Path Connectivity of Reducible Hyperbolic Tangent Networks

Matthew Farrugia-Roberts

https://openreview.net/forum?id=ddKCg3OhGw

Keywords: theory, neural network theory, structural redundancy, functional equivalence, functional equivalence class, partial identifiability, parameter canonicalisation, parameter space, piecewise-linear, connectivity

Compressor summary: The paper characterizes unit redundancies and reducible functional equivalence classes in a neural network architecture using an algorithm and shows that these classes are path-connected sets with a small maximum diameter.


Differentially Private Decoupled Graph Convolutions for Multigranular Topology Protection

Eli Chien, Wei-Ning Chen, Chao Pan, Pan Li, Ayfer Ozgur, Olgica Milenkovic

https://openreview.net/forum?id=dd3KNayGFz

Keywords: Graph Neural Networks, Differential Privacy, Multigranular Topology Protection

Compressor summary: The authors propose a new framework called Graph Differential Privacy (GDP) for graph learning that ensures provably private model parameters and predictions, and compare it to existing Differentially Private Decoupled Graph Convolutions (DPDGCs).


Deep Gaussian Markov Random Fields for Graph-Structured Dynamical Systems

Fiona Lippert, Bart Kranstauber, E. Emiel van Loon, Patrick Forré

https://openreview.net/forum?id=dcw7qRUuD8

Keywords: probabilistic inference, graphical models, spatiotemporal dynamical systems, state-space models

Compressor summary: The paper proposes a computationally efficient method for estimating and learning state variables in graph-structured models with unknown dynamics and limited data by combining deep learning and Gaussian Markov random fields.


Finding Order in Chaos: A Novel Data Augmentation Method for Time Series in Contrastive Learning

Berken Utku Demirel, Christian Holz

https://openreview.net/forum?id=dbVRDk2wt7

Keywords: Contrastive learning, Time-series, Augmentation

Compressor summary: The paper proposes a new data augmentation method for time-series tasks that connects samples to find order in the latent space and improves performance on three tasks, including heart rate estimation, human activity recognition, and cardiovascular disease detection.


Epistemic Neural Networks

Ian Osband, Zheng Wen, Seyed Mohammad Asghari, Vikranth Dwaracherla, Morteza Ibrahimi, Xiuyuan Lu, Benjamin Van Roy

https://openreview.net/forum?id=dZqcC1qCmB

Keywords: Uncertainty, Deep Learning, Neural Networks

Compressor summary: The epinet is an architecture that improves the uncertainty estimation of conventional neural networks by supplementing them with modest computation, enabling them to outperform large ensembles.


Causal Discovery in Semi-Stationary Time Series

Shanyun Gao, Raghavendra Addanki, Tong Yu, Ryan A. Rossi, Murat Kocaoglu

https://openreview.net/forum?id=dYeUvLUxBQ

Keywords: time-series causal discovery, constraint-based causal discovery

Compressor summary: The paper proposes a method for discovering causal relations from non-stationary time series using a structural causal model that accounts for periodic changes, and tests it on simulated and real-world data.


Self-Correcting Bayesian Optimization through Bayesian Active Learning

Carl Hvarfner, Erik Orm Hellsten, Frank Hutter, Luigi Nardi

https://openreview.net/forum?id=dX9MjUtP1A

Keywords: Bayesian Optimization, Bayesian Active Learning, Gaussian Processes

Compressor summary: The authors propose two methods to improve Gaussian processes for Bayesian optimization and active learning by better learning hyperparameters and incorporating self-correction.


Train Faster, Perform Better: Modular Adaptive Training in Over-Parameterized Models

Yubin Shi, Yixuan Chen, Mingzhi Dong, Xiaochen Yang, Dongsheng Li, Yujiang Wang, Robert P. Dick, Qin Lv, Yingying Zhao, Fan Yang, Tun Lu, Ning Gu, Li Shang

https://openreview.net/forum?id=dWDEBW2raJ

Keywords: Modular Adaptive Training, Efficient Training, Over-parameterized Model, Neural Tangent Kernel.

Compressor summary: This paper introduces modular neural tangent kernel (mNTK) to study learning dynamics of over-parameterized models, proposes Modular Adaptive Training (MAT) to improve efficiency and performance by selectively updating modules based on their mNTK eigenvalues, and shows that MAT reduces computational costs and outperforms baselines.


Human-like Few-Shot Learning via Bayesian Reasoning over Natural Language

Kevin Ellis

https://openreview.net/forum?id=dVnhdm9MIg

Keywords: Cognitive science, Bayesian, Language model, Induction, Psychology, Reasoning

Compressor summary: The model uses natural language hypotheses and Bayesian reasoning to learn a broad range of human-like concepts efficiently.


RECKONING: Reasoning through Dynamic Knowledge Encoding

Zeming Chen, Gail Weiss, Eric Mitchell, Asli Celikyilmaz, Antoine Bosselut

https://openreview.net/forum?id=dUAcAtCuKk

Keywords: natural language processing, multi-hop reasoning, knowledge memorisation

Compressor summary: RECKONING teaches language models to reason more robustly by encoding contextual knowledge into their parameters before answering questions, improving performance and generalization.


Does a sparse ReLU network training problem always admit an optimum ?

TUNG QUOC LE, Rémi Gribonval, Elisa Riccietti

https://openreview.net/forum?id=dTj5tH94xv

Keywords: Topology, best approximation property, closedness, function space, sparse neural networks

Compressor summary: The paper investigates the existence of an optimal solution for sparse ReLU neural networks and shows that it is not always guaranteed, while providing conditions for its existence.


Slow and Weak Attractor Computation Embedded in Fast and Strong E-I Balanced Neural Dynamics

Xiaohan Lin, Liyuan Li, Boxin Shi, Tiejun Huang, Yuanyuan Mi, Si Wu

https://openreview.net/forum?id=dSRyKIYRnP

Keywords: Continuous attractor neural network; Excitation inhibition balance; Brain-inspired algorithms; Object tracking;

Compressor summary: The study explores how a neural circuit can have both attractor states and irregular firings by using two sets of synapses with different strengths and speeds, leading to improved performance in a tracking problem.


NeuralGF: Unsupervised Point Normal Estimation by Learning Neural Gradient Function

Qing Li, Huifang Feng, Kanle Shi, Yue Gao, Yi Fang, Yu-Shen Liu, Zhizhong Han

https://openreview.net/forum?id=dR6p49RYLq

Keywords: Point Clouds, Normal Estimation, Neural Gradient

Compressor summary: The authors propose a new neural network approach to estimate oriented normals from point clouds without using ground truth normals, achieving robust and accurate results on benchmarks.


Safe Exploration in Reinforcement Learning: A Generalized Formulation and Algorithms

Akifumi Wachi, Wataru Hashimoto, Xun Shen, Kazumune Hashimoto

https://openreview.net/forum?id=dQLsvKNwZC

Keywords: Reinforcement Learning, Safety Exploration

Compressor summary: MASE is a meta-algorithm for safe exploration in reinforcement learning that combines an unconstrained RL algorithm with an uncertainty quantifier to guarantee safety while penalizing unsafe actions.


Locally Invariant Explanations: Towards Stable and Unidirectional Explanations through Local Invariant Learning

Amit Dhurandhar, Karthikeyan Natesan Ramamurthy, Kartik Ahuja, Vijay Arya

https://openreview.net/forum?id=dOxm4FnMFu

Keywords: Explainable AI, Game theory, Invariance

Compressor summary: The paper proposes a novel method for explaining black-box models using game theory and invariant risk minimization, which can produce high fidelity, stable, and unidirectional explanations with simple and efficient training.


From Discrete Tokens to High-Fidelity Audio Using Multi-Band Diffusion

Robin San Roman, Yossi Adi, Antoine Deleforge, Romain Serizel, Gabriel Synnaeve, Alexandre Défossez

https://openreview.net/forum?id=dOanKg3jKS

Keywords: diffusion, audio, compression

Compressor summary: The proposed high-fidelity multi-band diffusion model generates any audio type from low-bitrate discrete representations, outperforming other methods in perceptual quality at equal bit rate.


Constrained Policy Optimization with Explicit Behavior Density For Offline Reinforcement Learning

Jing Zhang, Chi Zhang, Wenjia Wang, Bingyi Jing

https://openreview.net/forum?id=dLmDPVv19z

Keywords: Offline Reinforcement Learning, GAN, Flow Model, Policy Control

Compressor summary: CPED is a new method for offline RL that uses a flow-GAN model to estimate behavior policy density, allowing it to safely explore and achieve better performance than existing methods.


Double and Single Descent in Causal Inference with an Application to High-Dimensional Synthetic Control

Jann Spiess, Guido Imbens, Amar Venugopal

https://openreview.net/forum?id=dL0GM9Wwtq

Keywords: Double descent, interpolating regression, synthetic control, causal inference

Compressor summary: The authors study how over-parameterized models in causal inference, such as high-dimensional linear regression and synthetic control with many control units, can perform better than simpler ones, and provide a unified theoretical perspective on their improvement.


SwiFT: Swin 4D fMRI Transformer

Peter Yongho Kim, Junbeom Kwon, Sunghwan Joo, Sangyoon Bae, Donggyu Lee, Yoonho Jung, Shinjae Yoo, Jiook Cha, Taesup Moon

https://openreview.net/forum?id=dKeWh6EzBB

Keywords: fMRI, Swin Transformer, 4D, neuroscience

Compressor summary: SwiFT is a Swin Transformer architecture that learns brain dynamics directly from fMRI volumes using a 4D window multi-head self-attention mechanism and absolute positional embeddings, achieving better performance than existing models on predicting sex, age, and cognitive intelligence.


Keypoint-Augmented Self-Supervised Learning for Medical Image Segmentation with Limited Annotation

Zhangsihao Yang, Mengwei Ren, Kaize Ding, Guido Gerig, Yalin Wang

https://openreview.net/forum?id=dK0Ew3kkVf

Keywords: Keypoints, Medical image, self-supervised learning, transformer, segmentation

Compressor summary: Keypoint-augmented fusion layer improves medical image segmentation by enhancing CNN features with long-range spatial self-attention, and incorporates global and local self-supervised pretraining for better representations.


Causal-structure Driven Augmentations for Text OOD Generalization

Amir Feder, Yoav Wald, Claudia Shi, Suchi Saria, David Blei

https://openreview.net/forum?id=dJZ3MvDw86

Keywords: Counterfactually Augmented Data, Invariant Learning, Out-of-distribution Generalization, Clinical NLP

Compressor summary: The authors propose using counterfactual data augmentation to improve the robustness of text classifiers in cases where the label is spuriously correlated with an attribute, and show improved OOD accuracy on medical diagnosis tasks.


On the Convergence of Black-Box Variational Inference

Kyurae Kim, Jisu Oh, Kaiwen Wu, Yian Ma, Jacob R. Gardner

https://openreview.net/forum?id=dHQ2av9NzO

Keywords: black-box variational inference, stochastic gradient descent, Bayesian inference, variational inference, probabilistic machine learning, Bayesian machine learning, variational Bayes

Compressor summary: The paper shows that using a specific algorithm called proximal stochastic gradient descent can improve the performance of black-box variational inference, a method for Bayesian inference, compared to other common approaches.


LMC: Large Model Collaboration with Cross-assessment for Training-Free Open-Set Object Recognition

Haoxuan Qu, Xiaofei Hui, Yujun Cai, Jun Liu

https://openreview.net/forum?id=dHF3Im8Aic

Keywords: Deep learning, Open-set object recognition, Large models, Training-free

Compressor summary: The paper proposes a method called Large Model Collaboration (LMC) that uses multiple pre-trained models to improve open-set object recognition without additional training and by extracting implicit knowledge from the models.


Cookie Consent Has Disparate Impact on Estimation Accuracy

Erik Miehling, Rahul Nair, Elizabeth M. Daly, Karthikeyan Natesan Ramamurthy, Robert Nelson Redmond

https://openreview.net/forum?id=dFtpRphNb3

Keywords: fairness, cookies, recommender systems

Compressor summary: The text discusses how user consent on cookie sharing affects the accuracy of personalized ads and recommender systems, and suggests new notions of fairness to address privacy concerns.


CP-SLAM: Collaborative Neural Point-based SLAM System

Jiarui Hu, Mao Mao, Hujun Bao, Guofeng Zhang, Zhaopeng Cui

https://openreview.net/forum?id=dFSeZm6dTC

Keywords: Collaborative SLAM; Neural Point Field; Keyframe-based SLAM; Pose Graph Optimization

Compressor summary: The paper introduces a collaborative implicit neural SLAM system using RGB-D image sequences, with a novel 3D scene representation and learning strategy that improves accuracy and cooperation.


Separable Physics-Informed Neural Networks

Junwoo Cho, Seungtae Nam, Hyunmo Yang, Seok-Bae Yun, Youngjoon Hong, Eunbyung Park

https://openreview.net/forum?id=dEySGIcDnI

Keywords: partial differential equations, scientific machine learning, physics-informed neural networks, fluid dynamics

Compressor summary: The proposed method, SPINN, improves the performance of PINNs by reducing computational costs and increasing accuracy in solving multi-dimensional PDEs using a per-axis basis network architecture and forward-mode automatic differentiation.


Fed-CO$_{2}$: Cooperation of Online and Offline Models for Severe Data Heterogeneity in Federated Learning

Zhongyi Cai, Ye Shi, Wei Huang, Jingya Wang

https://openreview.net/forum?id=dEDdRWunxU

Keywords: Federated Learning, Data Heterogeneity, Model Cooperation, Mutual Learning, Knowledge Transfer

Compressor summary: Fed-CO$_2$ is a framework for distributed learning that handles data heterogeneity issues like label and feature skew using cooperation between online and offline models, improving performance over existing algorithms.


Online Inventory Problems: Beyond the i.i.d. Setting with Online Convex Optimization

Massil HIHAT, Stéphane Gaïffas, Guillaume Garrigos, Simon Bussy

https://openreview.net/forum?id=dDk6URGRXP

Keywords: online convex optimization, inventory control, newsvendor, online learning, regret analysis

Compressor summary: The paper presents an online algorithm, MaxCOSD, for inventory control problems with general demands, losses, and dynamics, and requires non-degeneracy assumptions on the demand process for learning.


Supervised Pretraining Can Learn In-Context Reinforcement Learning

Jonathan Lee, Annie Xie, Aldo Pacchiano, Yash Chandak, Chelsea Finn, Ofir Nachum, Emma Brunskill

https://openreview.net/forum?id=dCYBAGQXLo

Keywords: decision making, reinforcement learning, in-context learning, bandits, transformers, offline reinforcement learning, exploration, reinforcement learning theory

Compressor summary: The paper introduces Decision-Pretrained Transformer (DPT), a method for supervised pretraining of transformers to solve decision-making problems using in-context learning, and demonstrates its surprising capabilities in solving RL problems, generalizing to new tasks, and adapting strategies.


This Looks Like Those: Illuminating Prototypical Concepts Using Multiple Visualizations

Chiyu Ma, Brandon Zhao, Chaofan Chen, Cynthia Rudin

https://openreview.net/forum?id=dCAk9VlegR

Keywords: deep learning, interpretability, prototype-based neural network, case-based reasoning

Compressor summary: ProtoConcepts is a method for interpretable image classification using multiple image patches to represent prototypical concepts, improving visual explanations and understanding.


SmooSeg: Smoothness Prior for Unsupervised Semantic Segmentation

Mengcheng Lan, Xinjiang Wang, Yiping Ke, Jiaxing Xu, Litong Feng, Wayne Zhang

https://openreview.net/forum?id=dB4lvScPIj

Keywords: unsupervised semantic segmentation; self-supervised learning; smoothness prior

Compressor summary: The paper proposes a new method called SmooSeg for unsupervised semantic segmentation that uses smoothness prior to simplify the task, improving performance on three datasets.


Contrastive Sampling Chains in Diffusion Models

Junyu Zhang, Daochang Liu, Shichao Zhang, Chang Xu

https://openreview.net/forum?id=dAbGv5Jz5U

Keywords: diffusion models, contrastive loss, discretization error, contrastive sampling chain

Compressor summary: The text explains a method that combines contrastive loss and score matching to reduce discretization error in diffusion models, improving image quality and sample efficiency.


A*Net: A Scalable Path-based Reasoning Approach for Knowledge Graphs

Zhaocheng Zhu, Xinyu Yuan, Mikhail Galkin, Sophie Xhonneux, Ming Zhang, Maxime Gazeau, Jian Tang

https://openreview.net/forum?id=dAJrxQz1lk

Keywords: Knowledge Graph Reasoning, Path-based Methods, Scalability, A* Algorithm

Compressor summary: A\*Net is a scalable path-based method for knowledge graph reasoning that learns a priority function to select important nodes and edges, achieving competitive performance with less computation.


ZipLM: Inference-Aware Structured Pruning of Language Models

Eldar Kurtic, Elias Frantar, Dan Alistarh

https://openreview.net/forum?id=d8j3lsBWpV

Keywords: LLMs, pruning, compression, inference

Compressor summary: ZipLM is a novel compression method for large language models that achieves high accuracy and speedup, outperforming prior techniques in various settings.


3D Copy-Paste: Physically Plausible Object Insertion for Monocular 3D Detection

Yunhao Ge, Hong-Xing Yu, Cheng Zhao, Yuliang Guo, Xinyu Huang, Liu Ren, Laurent Itti, Jiajun Wu

https://openreview.net/forum?id=d86B6Mdweq

Keywords: 3D Copy-Paste, Object Insertion, Monocular 3D Object Detection

Compressor summary: The paper presents a method to insert virtual objects into real scenes for monocular 3D object detection, ensuring physical plausibility in locations, appearances, and shadows.


Meta-AdaM: An Meta-Learned Adaptive Optimizer with Momentum for Few-Shot Learning

Siyuan Sun, Hongyang Gao

https://openreview.net/forum?id=d85pPNBHLt

Keywords: Few shot learning, Meta Learning

Compressor summary: Meta-AdaM is a meta-learned adaptive optimizer that uses weight-update history and momentum for faster few-shot learning, outperforming existing methods on benchmark datasets.


How to Fine-tune the Model: Unified Model Shift and Model Bias Policy Optimization

Hai Zhang, Hang Yu, Junqiao Zhao, Di Zhang, Chang Huang, Hongtu Zhou, Xiao Zhang, Chen Ye

https://openreview.net/forum?id=d7a5TpePV7

Keywords: model-based reinforcement learning, model shift, model bias, fine-tuning, performance difference bound

Compressor summary: The paper proposes USB-PO, an algorithm that adapts model updates to improve MBRL performance while accounting for model shift and bias, achieving state-of-the-art results on multiple benchmark tasks.


Unsupervised Image Denoising with Score Function

Yutong Xie, Mingze Yuan, Bin Dong, Quanzheng Li

https://openreview.net/forum?id=d6LShzSTOP

Keywords: unsupervised learning, image denoising, score function

Compressor summary: The paper introduces a new unsupervised learning approach for single image denoising that can handle various noise models and outperforms current methods in complex cases.


Structural Pruning for Diffusion Models

Gongfan Fang, Xinyin Ma, Xinchao Wang

https://openreview.net/forum?id=d4f40zJJIS

Keywords: Diffusion Model, Network Pruning, Model Compression, Efficient Deep Learning

Compressor summary: Diff-Pruning is an efficient compression method that reduces FLOPs and training time for diffusion models, preserving their generative behavior.


Towards Test-Time Refusals via Concept Negation

Peiran Dong, Song Guo, Junxiao Wang, Bingjie WANG, Jiewei Zhang, Ziming Liu

https://openreview.net/forum?id=d4X0QWS2Ln

Keywords: Diffusion models, test-time refusal, concept negation, safety in generative models

Compressor summary: The paper introduces ProtoRe, a framework to improve concept negation for generating refusals in image synthesis by using CLIP's knowledge to identify negative concepts and purify outputs.


On the Gini-impurity Preservation For Privacy Random Forests

XinRan Xie, Man-Jie Yuan, Xuetong Bai, Wei Gao, Zhi-Hua Zhou

https://openreview.net/forum?id=d47iuwOt3j

Keywords: classification, random forests, privacy-preserving machine learng, data encrytion

Compressor summary: This paper proposes a new encryption method that preserves the Gini impurity of data for random forests, using modified binary search trees and homomorphic encryption.


Unbalanced Low-rank Optimal Transport Solvers

Meyer Scetbon, Michal Klein, Giovanni Palla, marco cuturi

https://openreview.net/forum?id=d2WsCmoITF

Keywords: Optimal Transport, Unbalanced

Compressor summary: The paper proposes new methods to improve optimal transport in machine learning by addressing its computational and modelling limitations, and applies them to solve real-world spatial transcriptomics problems.


Zero-Shot Anomaly Detection via Batch Normalization

Aodong Li, Chen Qiu, Marius Kloft, Padhraic Smyth, Maja Rudolph, Stephan Mandt

https://openreview.net/forum?id=d1wjMBYbP1

Keywords: deep anomaly detection, zero-shot learning, batch normalization

Compressor summary: The paper proposes ACR, a method that helps deep anomaly detectors adapt to new data distributions without training data, achieving good zero-shot performance on tabular and image data.


Optimality of Message-Passing Architectures for Sparse Graphs

Aseem Baranwal, Kimon Fountoulakis, Aukosh Jagannath

https://openreview.net/forum?id=d1knqWjmNt

Keywords: graph neural networks, message passing, bayesian inference, node classification, contextual stochastic block model

Compressor summary: The paper proposes an optimal node classification method for sparse feature-decorated graphs using a message-passing graph neural network and compares its performance with existing methods on a statistical model.


VRA: Variational Rectified Activation for Out-of-distribution Detection

Mingyu Xu, Zheng Lian, Bin Liu, Jianhua Tao

https://openreview.net/forum?id=d0VItRE2ZH

Keywords: Out-of-distribution Detection

Compressor summary: The paper introduces a new technique called Variational Rectified Activation (VRA) that improves out-of-distribution detection in machine learning models by simulating suppression and amplification operations using piecewise functions, and shows its effectiveness on multiple benchmark datasets.


On the Role of Randomization in Adversarially Robust Classification

Lucas Gnecco Heredia, Muni Sreenivas Pydi, Laurent Meunier, benjamin negrevergne, Yann Chevaleyre

https://openreview.net/forum?id=d0IEd3VgBh

Keywords: adversarial attacks, robustness, adversarial, attacks, deep learning, randomization, randomized ensembles

Compressor summary: The paper investigates how randomization affects adversarial robustness in classifiers and shows conditions for randomized ensembles to outperform deterministic ones, as well as examples of deterministic classifiers that outperform probabilistic ones.


Exploring Diverse In-Context Configurations for Image Captioning

Xu Yang, Yongliang Wu, Mingzhuo Yang, Haokun Chen, Xin Geng

https://openreview.net/forum?id=czwZnNf60r

Keywords: Image Caption; Few-shot Prompt; Vision Language Model;

Compressor summary: The authors explore different strategies for configuring image-text pairs in vision-language tasks, finding that they can significantly improve performance in image captioning compared to random sampling.


Efficient Neural Music Generation

Max W. Y. Lam, Qiao Tian, Tang Li, Zongyu Yin, Siyuan Feng, Ming Tu, Yuliang Ji, Rui Xia, Mingbo Ma, Xuchen Song, Jitong Chen, Yuping Wang, Yuxuan Wang

https://openreview.net/forum?id=cxazQGSsQa

Keywords: Music Generation, Language Model, Diffusion Model, MusicLM

Compressor summary: MeLoDy is a new music generation model that uses less computing power than MusicLM while maintaining high quality and adaptability.


Privacy Assessment on Reconstructed Images: Are Existing Evaluation Metrics Faithful to Human Perception?

Xiaoxiao Sun, Nidham Gazagnadou, Vivek Sharma, Lingjuan Lyu, Hongdong Li, Liang Zheng

https://openreview.net/forum?id=cx9a4Xvb3l

Keywords: Privacy Assessment, Reconstructed Images, Evaluation Metrics, Human Perception

Compressor summary: The paper studies how well hand-crafted metrics reflect human judgement of model privacy leakage in reconstructed images and proposes a learning-based measure called SemSim to better evaluate semantic similarity between original and reconstructed images.


GPT4Tools: Teaching Large Language Model to Use Tools via Self-instruction

Rui Yang, Lin Song, Yanwei Li, Sijie Zhao, Yixiao Ge, Xiu Li, Ying Shan

https://openreview.net/forum?id=cwjh8lqmOL

Keywords: multimodality, foundation models, tool usage

Compressor summary: The paper proposes GPT4Tools, a method to enable open-source LLMs to use multi-modal tools efficiently and effectively using instruction-following datasets and Low-Rank Adaptation optimization.


On the Learnability of Multilabel Ranking

Vinod Raman, UNIQUE SUBEDI, Ambuj Tewari

https://openreview.net/forum?id=cwBeRBe9hq

Keywords: Multilabel Ranking, PAC Learning, Online Learning

Compressor summary: This paper studies how easy it is to learn multilabel ranking problems with relevance-score feedback in different settings and introduces two classes of ranking losses based on learnability.


Demystifying Softmax Gating Function in Gaussian Mixture of Experts

Huy Nguyen, TrungTin Nguyen, Nhat Ho

https://openreview.net/forum?id=cto6jIIbMZ

Keywords: Mixture of Experts, Maximum Likelihood Estimation, Voronoi Loss Function, Algebraic Geometry.

Compressor summary: The text discusses challenges in estimating parameters for softmax gating Gaussian mixture of experts models and proposes new Voronoi loss functions to address them, connecting convergence rates to solvability problems of polynomial equations.


Counterfactual Generation with Identifiability Guarantees

Hanqi Yan, Lingjing Kong, Lin Gui, Yuejie Chi, Eric Xing, Yulan He, Kun Zhang

https://openreview.net/forum?id=cslnCXE9XA

Keywords: Causal Representation Learning, Identifiability, Counterfactual Generation, Latent variable models, Disentanglement.

Compressor summary: The paper proposes a method to generate counterfactual images or texts by identifying content and style latent variables that vary across domains, using theoretical insights from relative sparsity of influences, and achieving state-of-the-art performance in unsupervised style transfer tasks.


Efficient RL with Impaired Observability: Learning to Act with Delayed and Missing State Observations

Minshuo Chen, Yu Bai, H. Vincent Poor, Mengdi Wang

https://openreview.net/forum?id=csdEeUn0ve

Keywords: Delayed and missing observations, MDPs, efficient regret bounds

Compressor summary: The paper studies how reinforcement learning can be efficient in control systems when agents have delayed or missing state observations and provides regret bounds and comparisons with full observability.


RayDF: Neural Ray-surface Distance Fields with Multi-view Consistency

Zhuoman Liu, Bo Yang, Yan Luximon, Ajay Kumar, Jinxi Li

https://openreview.net/forum?id=crZlhMnfeO

Keywords: implicit shape representations, multi-view consistency, novel view synthesis

Compressor summary: The paper introduces RayDF, a fast and accurate method for representing 3D shapes using ray-based neural functions that maintain multi-view geometry consistency, outperforming existing coordinate- and ray-based approaches.


No-regret Algorithms for Fair Resource Allocation

Abhishek Sinha, Ativ Joshi, Rajarshi Bhattacharjee, Cameron N Musco, Mohammad Hajiesmaili

https://openreview.net/forum?id=crNAh1EZKo

Keywords: Online Learning, Bandit Algorithms, Learning Theory

Compressor summary: The paper proposes an efficient online resource allocation policy that achieves sublinear approximate regret in a fairness problem with a surprising phase transition phenomenon at the critical exponent $\alpha=\frac{1}{2}.$


Back-Modality: Leveraging Modal Transformation for Data Augmentation

Zhi Li, Yifan Liu, Yin Zhang

https://openreview.net/forum?id=cr99foBDPV

Keywords: data augmentation, cross-modal

Compressor summary: Back-Modality is a new way to improve data augmentation by transforming data between different modalities and then back again, making it easier to use various techniques for different types of data.


Group Fairness in Peer Review

Haris Aziz, Evi Micha, Nisarg Shah

https://openreview.net/forum?id=cpUuSV8kRw

Keywords: peer review; group fairness; core; stable

Compressor summary: The paper proposes a group fairness concept called the core for large AI conferences, which ensures that every community is treated fairly in the peer review process, and develops an efficient algorithm to find optimal assignments using real data from CVPR and ICLR.


Implicit Manifold Gaussian Process Regression

Bernardo Fichera, Viacheslav Borovitskiy, Andreas Krause, Aude Billard

https://openreview.net/forum?id=co4p15OMoc

Keywords: Gaussian process, manifolds, manifold learning, uncertainty, regression, graph Laplacian

Compressor summary: The paper proposes a new Gaussian process regression method that can infer implicit structure from data without explicit input of a manifold, potentially improving its performance in high-dimensional settings.


Context-PIPs: Persistent Independent Particles Demands Context Features

Weikang BIAN, Zhaoyang Huang, Xiaoyu Shi, Yitong Dong, Yijin Li, Hongsheng Li

https://openreview.net/forum?id=cnpkzQZaLU

Keywords: Optical Flow; Video Correspondence; Computer Vision;

Compressor summary: The paper introduces Context-PIPs, a framework that improves point trajectory accuracy in videos by using spatial context features from the source and target points.


Bayesian Learning of Optimal Policies in Markov Decision Processes with Countably Infinite State-Space

Saghar Adler, Vijay Subramanian

https://openreview.net/forum?id=cm53OBkctM

Keywords: Thompson Sampling, Reinforcement Learning, Queueing theory

Compressor summary: The paper proposes a Bayesian algorithm using Thompson sampling for optimal control of countably-infinite state-space Markov Decision Processes, and shows its applicability in two queueing models.


Self-Supervised Visual Acoustic Matching

Arjun Somayazulu, Changan Chen, Kristen Grauman

https://openreview.net/forum?id=clKbFMt29V

Keywords: Audio-Visual learning, Visual Acoustic Matching

Compressor summary: The paper proposes a self-supervised method for acoustic matching that uses only target scene images and audio to re-synthesize audio in a different environment, outperforming existing methods.


Hierarchical Vector Quantized Transformer for Multi-class Unsupervised Anomaly Detection

Ruiying Lu, YuJie Wu, Long Tian, Dongsheng Wang, Bo Chen, Xiyang Liu, Ruimin Hu

https://openreview.net/forum?id=clJTNssgn6

Keywords: Anomaly Detection, Transformer, Vector Quantization, Unsupervised Anomaly Detection

Compressor summary: The paper proposes a hierarchical vector quantized prototype-oriented Transformer for unsupervised image anomaly detection that preserves normal patterns as discrete prototypes and uses optimal transport to evaluate abnormal scores.


Convergent Bregman Plug-and-Play Image Restoration for Poisson Inverse Problems

Samuel Hurault, Ulugbek Kamilov, Arthur Leclaire, Nicolas Papadakis

https://openreview.net/forum?id=clCELP8zFb

Keywords: Plug-and-Play, Poisson Inverse Problems, Bregman distance, Proximal Gradient Descent, nonconvex and nonsmooth optimization, Poisson inverse problems

Compressor summary: The paper proposes a new method called Bregman Score Denoiser for solving Poisson inverse problems using Plug-and-Play algorithms that better capture the problem's smoothness properties.


A-NeSI: A Scalable Approximate Method for Probabilistic Neurosymbolic Inference

Emile van Krieken, Thiviyan Thanapalasingam, Jakub M. Tomczak, Frank Van Harmelen, Annette Ten Teije

https://openreview.net/forum?id=chlTA9Cegc

Keywords: Neurosymbolic Learning, Generative Modeling, Approximate Inference

Compressor summary: A-NeSI is a new framework for probabilistic neurosymbolic learning that allows scalable approximate inference, symbolic explanations of predictions, and guarantees logical constraint satisfaction while maintaining performance.


Expressive probabilistic sampling in recurrent neural networks

Shirui Chen, Linxing Preston Jiang, Rajesh P. N. Rao, Eric Todd SheaBrown

https://openreview.net/forum?id=ch1buUOGa3

Keywords: neural coding, probabilistic sampling, neural dynamics, recurrent neural network

Compressor summary: The paper explores how recurrent neural circuits can sample from arbitrary probability distributions, which could improve Bayesian brain models.


Fine-Grained Cross-View Geo-Localization Using a Correlation-Aware Homography Estimator

Xiaolong Wang, Runsen Xu, Zhuofan Cui, Zeyu Wan, Yu Zhang

https://openreview.net/forum?id=cgiP4cMBP9

Keywords: Fine-Grained Cross-View Geo-Localization, Homography Estimation

Compressor summary: The paper presents a new method for accurately aligning ground images with satellite images using a differentiable spherical transform and a correlation-aware homography estimator, achieving sub-pixel resolution and meter-level GPS accuracy, and outperforming existing techniques in localization error reduction.


On Separate Normalization in Self-supervised Transformers

Xiaohui Chen, Yinkai Wang, Yuanqi Du, Soha Hassoun, Liping Liu

https://openreview.net/forum?id=cezKbXsT3V

Keywords: Transformer, Self-supervised Learning, Normalization

Compressor summary: The paper proposes a simple modification to self-supervised training methods for transformers by using separate normalization layers for tokens and the [CLS] symbol, which improves downstream task performance in various domains.


IBA: Towards Irreversible Backdoor Attacks in Federated Learning

Dung Thuy Nguyen, Tuan Minh Nguyen, Anh Tuan Tran, Khoa D Doan, KOK SENG WONG

https://openreview.net/forum?id=cemEOP8YoC

Keywords: Backdoor Attacks, Federated Learning, Durability, Imperceptibility, Stealthiness

Compressor summary: The paper proposes a novel backdoor attack method for federated learning that is stealthy, efficient, and durable, and bypasses existing defenses.


Max-Sliced Mutual Information

Dor Tsur, Ziv Goldfeld, Kristjan Greenewald

https://openreview.net/forum?id=ce9B2x3zQa

Keywords: CCA, dimensionality reduction, information theory, mutual information, neural estimation, slicing

Compressor summary: The paper introduces a new method called max-sliced mutual information (mSMI) that combines the strengths of CCA and Shannon's mutual information for quantifying dependencies in high-dimensional data.


Semi-Supervised Domain Generalization with Known and Unknown Classes

Lei Zhang, Ji-Fu Li, Wei Wang

https://openreview.net/forum?id=ce59j806df

Keywords: Domain Generalization, Semi-Supervised Learning, Out-of-Distribution Detection, Deep Learning

Compressor summary: The CWAEE method is a semi-supervised domain generalization approach that uses class-wise adaptive thresholds and Fourier Transformation to handle unknown classes in unlabeled data, achieving better performance on real-world datasets.


Learning non-Markovian Decision-Making from State-only Sequences

Aoyang Qin, Feng Gao, Qing Li, Song-Chun Zhu, Sirui Xie

https://openreview.net/forum?id=cdlmsnQkZ9

Keywords: Sequential Decision Making, Generative Model, Imitation Learning

Compressor summary: The paper proposes a deep generative model for imitation learning in non-Markovian settings, using energy-based policy prior and maximum likelihood estimation for inference and planning.


DeepACO: Neural-enhanced Ant Systems for Combinatorial Optimization

Haoran Ye, Jiarui Wang, Zhiguang Cao, Helan Liang, Yong Li

https://openreview.net/forum?id=cd5D1DD923

Keywords: Neural Combinatorial Optimization, Ant Colony Optimization, Evolutionary algorithm, Meta-heuristic, Deep reinforcement learning, Learned heuristic measure, Neural local search, Generalization

Compressor summary: DeepACO is a framework that uses deep reinforcement learning to improve Ant Colony Optimization algorithms without manual design of heuristics, achieving better results on various Combinatorial Optimization Problems.


Towards Distribution-Agnostic Generalized Category Discovery

Jianhong Bai, Zuozhu Liu, Hualiang Wang, Ruizhe Chen, Lianrui Mu, Xiaomeng Li, Joey Tianyi Zhou, YANG FENG, Jian Wu, Haoji Hu

https://openreview.net/forum?id=cczH4Xl7Zo

Keywords: Generalized Category Discovery, Open-world Recognition, Long-tail Learning, Contrastive Learning

Compressor summary: The paper introduces a novel framework, BaCon, that tackles the realistic task of distribution-agnostic generalized category discovery (DA-GCD) in computer vision, addressing data imbalance and open-endedness with contrastive learning and pseudo-labeling.


Empowering Convolutional Neural Nets with MetaSin Activation

Farnood Salehi, Tunc Ozan Aydin, André Gaillard, Guglielmo Camporese, Yuxuan Wang

https://openreview.net/forum?id=cay8LnKSro

Keywords: sin activation, image prediction, image resampling, monte-carlo denoising, knowledge distillation

Compressor summary: The authors propose MetaSin activation, a novel ensemble function that improves performance and reliability in training sin networks for visual data tasks like denoising and resampling.


Augmentation-free Dense Contrastive Distillation for Efficient Semantic Segmentation

Jiawei Fan, Chao Li, Xiaolong Liu, Meina Song, Anbang Yao

https://openreview.net/forum?id=caUhYUVsLl

Keywords: Knowledge distillation, semantic segmentation, contrastive learning

Compressor summary: Af-DCD is a new contrastive distillation method for semantic segmentation that uses masked feature mimicking and clever feature partitions to train efficient student models without data augmentation or memory buffer.


Bayesian Extensive-Rank Matrix Factorization with Rotational Invariant Priors

Farzad Pourkamali, Nicolas Macris

https://openreview.net/forum?id=ca2QmdOlIh

Keywords: Matrix factorization, Bayesian inference, rotation invariant estimators, random matrix theory, spherical integrals, replica method

Compressor summary: The paper proposes Rotation Invariant Estimators for matrix factorization with additive noise, which are conjectured to be optimal in high dimensions based on a Bayesian approach using random matrix theory, spherical integrals, and the replica method.


Guarantees for Self-Play in Multiplayer Games via Polymatrix Decomposability

Revan MacQueen, James R. Wright

https://openreview.net/forum?id=cZVBRg59eb

Keywords: Algorithmic Game Theory, Self-Play, Regret-Minimization, Multi-agent RL, Multiplayer Games, General-Sum Games

Compressor summary: The paragraph discusses how self-play can generate strategies with performance guarantees in some multiplayer games if they approximate constant-sum subgames with bounded subgame stability.


Data Minimization at Inference Time

Cuong Tran, Ferdinando Fioretto

https://openreview.net/forum?id=cZS5X3PLOR

Keywords: Privacy; data minimization

Compressor summary: The paper explores whether using only a small fraction of features for inference in high-stakes domains is sufficient to achieve accurate predictions, and proposes an efficient algorithm to determine which attributes are needed for each individual.


Context-guided Embedding Adaptation for Effective Topic Modeling in Low-Resource Regimes

Yishi Xu, Jianqiao Sun, Yudi Su, Xinyang Liu, Zhibin Duan, Bo Chen, Mingyuan Zhou

https://openreview.net/forum?id=cYkSt7jqlx

Keywords: Few-shot generative model; topic modeling;

Compressor summary: The paragraph describes a method for adapting word embeddings based on contextual information to improve topic modeling in low-resourced tasks and achieve better performance than existing methods.


Discovering Hierarchical Achievements in Reinforcement Learning via Contrastive Learning

Seungyong Moon, Junyoung Yeom, Bumsoo Park, Hyun Oh Song

https://openreview.net/forum?id=cUuXVaMmmv

Keywords: reinforcement learning, hierarchical reinforcement learning, contrastive learning, procedurally generated environments

Compressor summary: The paper proposes an improved model-free approach using PPO and contrastive learning to discover hierarchical achievements in procedurally generated environments effectively.


Riemannian Residual Neural Networks

Isay Katsman, Eric Ming Chen, Sidhanth Holalkere, Anna Asch, Aaron Lou, Ser-Nam Lim, Christopher De Sa

https://openreview.net/forum?id=cRzt1umRNx

Keywords: neural network, riemannian, manifold, resnet

Compressor summary: The authors extend ResNets to work on general Riemannian manifolds, achieving better performance than existing manifold neural networks for graphs with hierarchical structures or manifold-valued data.


Precise asymptotic generalization for multiclass classification with overparameterized linear models

David Xing Wu, Anant Sahai

https://openreview.net/forum?id=cRGINXQWem

Keywords: overparameterized, multiclass, classification, theory, generalization, interpolation, bi-level, Gaussian model

Compressor summary: This paper analyzes the generalization performance of a linear model for multiclass and multi-label classification with growing data dimensions, resolves a conjecture from a previous paper, and introduces a new variant of the Hanson-Wright inequality for these problems.


GraphMP: Graph Neural Network-based Motion Planning with Efficient Graph Search

Xiao Zang, Miao Yin, Jinqi Xiao, Saman Zonouz, Bo Yuan

https://openreview.net/forum?id=cQdc9Dyk4i

Keywords: graph neural network, deep learning

Compressor summary: GraphMP is a neural motion planner that improves graph information extraction and search processing, achieving better path quality and planning speed than existing methods in various environments.


The Curious Price of Distributional Robustness in Reinforcement Learning with a Generative Model

Laixi Shi, Gen Li, Yuting Wei, Yuxin Chen, Matthieu Geist, Yuejie Chi

https://openreview.net/forum?id=cOQH8YO255

Keywords: distributionally robust reinforcement learning, robust Markov decision processes, sample complexity

Compressor summary: The paper studies how robust reinforcement learning is to different uncertainty sets and shows that learning with total variation is more efficient than standard MDPs.


A Scalable Neural Network for DSIC Affine Maximizer Auction Design

Zhijian Duan, Haoran Sun, Yurong Chen, Xiaotie Deng

https://openreview.net/forum?id=cNb5hkTfGC

Keywords: Automated Mechanism Design, Auction Design, Affine Maximizer Auctions, Deep Learning, Game Theory

Compressor summary: AMenuNet is a scalable neural network that designs dominant strategy incentive compatible auctions by constructing allocation menus from bidder and item representations.


PanoGen: Text-Conditioned Panoramic Environment Generation for Vision-and-Language Navigation

Jialu Li, Mohit Bansal

https://openreview.net/forum?id=cNObl6QQEH

Keywords: Vision-and-Language Navigation, diffusion models, image inpainting for panorama generation

Compressor summary: PanoGen is a method to create diverse and realistic 3D environments based on text descriptions, which can improve Vision-and-Language Navigation performance and help agents generalize better to new situations.


Variational Imbalanced Regression: Fair Uncertainty Quantification via Probabilistic Smoothing

Ziyan Wang, Hao Wang

https://openreview.net/forum?id=cMUBkkTrMo

Keywords: probabilistic methods, imbalanced regression, variational inference

Compressor summary: The paper introduces a probabilistic deep learning model, VIR, that performs well in imbalanced regression and naturally estimates uncertainty using data borrowing and reweighting techniques.


HIQL: Offline Goal-Conditioned RL with Latent States as Actions

Seohong Park, Dibya Ghosh, Benjamin Eysenbach, Sergey Levine

https://openreview.net/forum?id=cLQCCtVDuW

Keywords: reinforcement learning

Compressor summary: The paper proposes a hierarchical algorithm for goal-conditioned RL from offline data that leverages the structure of reaching distant goals through subgoals, and shows its effectiveness on various benchmarks.


On the Role of Entanglement and Statistics in Learning

Srinivasan A, Vojtěch Havlíček, Louis Schatzki

https://openreview.net/forum?id=cGeLeh995N

Keywords: Quantum Computing, Statistical Learning, Quantum learning theory, Entanglement

Compressor summary: The paper investigates how entangled, separable, and statistical measurements affect learning in the quantum statistical query model. It reveals different trade-offs depending on the task and introduces a new quantum dimensionality concept to lower bound the complexity of learning various quantum states and functions. It also gives unconditional separation results for error mitigation methods.


NeuroGF: A Neural Representation for Fast Geodesic Distance and Path Queries

Qijian Zhang, Junhui Hou, Yohanes Yudhi Adikusuma, Wenping Wang, Ying He

https://openreview.net/forum?id=cGdGh3Mp2W

Keywords: geodesic distance, implicit representation, 3D geometry

Compressor summary: The paper introduces NeuroGF, a neural implicit representation of geodesics on 3D mesh models that can efficiently answer queries and encode both geometry and geodesics in a unified format.


The geometry of hidden representations of large transformer models

Lucrezia Valeriani, Diego Doimo, Francesca Cuturello, Alessandro Laio, Alessio ansuini, Alberto Cazzaniga

https://openreview.net/forum?id=cCYvakU5Ek

Keywords: Representations, transformers, geometry, interpretability

Compressor summary: Large transformers have similar representation evolution across various data types and can be used to identify layers with high semantic content.


Conditional Score Guidance for Text-Driven Image-to-Image Translation

Hyunsoo Lee, Minsoo Kang, Bohyung Han

https://openreview.net/forum?id=cBS5CU96Jq

Keywords: Diffusion, Image-to-Image Translation

Compressor summary: The paper proposes a text-driven image-to-image translation method that edits regions of interest in a source image based on a modifying text and preserves other parts, using a novel score function and mixup technique for better guidance and fusion.


Strategic Apple Tasting

Keegan Harris, Chara Podimata, Steven Wu

https://openreview.net/forum?id=cBIPcZKFdw

Keywords: strategic classification, strategic learning, apple tasting, bandit feedback, learning with incentives

Compressor summary: The text presents algorithms for decision-making under incentives and one-sided feedback, aiming to minimize strategic regret compared to the optimal policy. One algorithm works for stochastic agents and the other for adversarial ones, with different performance guarantees. The algorithms also apply to bandit feedback settings.


Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture

Daniel Y Fu, Simran Arora, Jessica Grogan, Isys Johnson, Sabri Eyuboglu, Armin W Thomas, Benjamin Frederick Spector, Michael Poli, Atri Rudra, Christopher Re

https://openreview.net/forum?id=cB0BImqSS9

Keywords: structured matrices, transformers, efficiency

Compressor summary: Monarch Mixer (M2) is a new architecture that uses Monarch matrices, a sub-quadratic primitive, to scale in sequence length and model dimension, achieving high performance and efficiency in various domains.


Enhancing Motion Deblurring in High-Speed Scenes with Spike Streams

Shiyan Chen, Jiyuan Zhang, Yajing Zheng, Tiejun Huang, Zhaofei Yu

https://openreview.net/forum?id=cAyLnMxiTl

Keywords: spike camera, neuromorphic vision sensors, motion deblurring, high speed imaging

Compressor summary: The paper proposes a new method to remove motion blur from high-speed scenes using spike camera data and traditional RGB images, improving visual quality over existing methods.


Explainable Brain Age Prediction using coVariance Neural Networks

Saurabh Sihag, Gonzalo Mateos, Corey McMillan, Alejandro Ribeiro

https://openreview.net/forum?id=cAhJF87GN0

Keywords: graph neural networks, brain age, Alzheimer's disease, interpretability, explainability, computational neuroscience

Compressor summary: The paper proposes a framework using coVariance neural networks for predicting brain age based on cortical thickness features, which allows for anatomical interpretability in identifying regions related to elevated brain age gap in Alzheimer's disease.


Estimating the Rate-Distortion Function by Wasserstein Gradient Descent

Yibo Yang, Stephan Eckstein, Marcel Nutz, Stephan Mandt

https://openreview.net/forum?id=cAaTbLa3ad

Keywords: information theory, rate-distortion function, optimal transport

Compressor summary: The paper proposes a new method to estimate the rate-distortion function using optimal transport, which learns the support of the optimal reproduction distribution and has advantages over neural network methods in terms of computational effort and tuning.


Fair, Polylog-Approximate Low-Cost Hierarchical Clustering

Marina Knittel, Max Springer, John P Dickerson, MohammadTaghi Hajiaghayi

https://openreview.net/forum?id=cAPMmCl2f3

Keywords: Fair machine learning, hierarchical clustering, clustering

Compressor summary: The paper introduces a new algorithm for fair hierarchical clustering that breaks the polynomial-approximate barrier and achieves low cost.


Characterization and Learning of Causal Graphs with Small Conditioning Sets

Murat Kocaoglu

https://openreview.net/forum?id=cANkPsVtsw

Keywords: causal discovery

Compressor summary: The paper proposes a new constraint-based causal discovery algorithm, the k-PC algorithm, which uses conditional independence tests with an upper bound on the conditioning set size and graphically characterizes the resulting k-Markov equivalence class of causal graphs for more robust causal discovery in small data settings.


Sequential Subset Matching for Dataset Distillation

Jiawei Du, Qin Shi, Joey Tianyi Zhou

https://openreview.net/forum?id=c9fXCzR5fK

Keywords: Dataset distillation, gradients matching

Compressor summary: The paper introduces Sequential Subset Matching (SeqMatch), a new method for dataset distillation that adaptively optimizes synthetic data to improve performance by addressing the coupling issue caused by static optimization.


Maximum Average Randomly Sampled: A Scale Free and Non-parametric Algorithm for Stochastic Bandits

Masoud Moravej Khorasani, Erik Weyer

https://openreview.net/forum?id=c8nIdZ5HJJ

Keywords: Stochastic Multi-armed bandit, Online Learning, Upper Confidence Bound

Compressor summary: MARS is a data-dependent UCB algorithm for multi-armed bandits that doesn't need scaling, uses maximum average of randomly sampled rewards, and performs similarly to $\psi$-UCB without correction factors.


EMMA-X: An EM-like Multilingual Pre-training Algorithm for Cross-lingual Representation Learning

Ping Guo, Xiangpeng Wei, Yue Hu, Baosong Yang, Dayiheng Liu, Fei Huang, jun xie

https://openreview.net/forum?id=c5dRV9tA3K

Keywords: cross-lingual pretraining;language-agnostic representation

Compressor summary: The paper introduces Emma-X, a pre-training algorithm to learn cross-lingual universals using non-parallel data, and shows its effectiveness on xrete benchmark tasks.


PLASTIC: Improving Input and Label Plasticity for Sample Efficient Reinforcement Learning

Hojoon Lee, Hanseul Cho, Hyunseung Kim, Daehoon Gwak, Joonkee Kim, Jaegul Choo, Se-Young Yun, Chulhee Yun

https://openreview.net/forum?id=c5WOU7p4ES

Keywords: Reinforcement Learning, Sharpness Minimization, Generalization, Plasticity, Deep Learning

Compressor summary: The PLASTIC algorithm, which balances input and label plasticity in off-policy reinforcement learning, enhances sample efficiency by preserving the model's ability to adapt to changing data and goals.


Tight Bounds for Volumetric Spanners and Applications

Aditya Bhaskara, Sepideh Mahabadi, Ali Vakilian

https://openreview.net/forum?id=c4Xc0uTLXW

Keywords: volumetric spanner, well-conditioned basis, determinant maximization, minimum volume enclosing ellipsoid

Compressor summary: A volumetric spanner is a subset of points that can express all other points with small coefficients, and this paper provides almost optimal bounds on its size and a simple construction method for all $\ell_p$ norms, with applications to finding coresets for the MVEE problem.


Tight Risk Bounds for Gradient Descent on Separable Data

Matan Schliserman, Tomer Koren

https://openreview.net/forum?id=c2eedxSlPJ

Keywords: Convex optimization, Gradient Descent, separable data, generalization bounds, Stochastic Gradient Descent.

Compressor summary: This paper investigates the generalization abilities of unregularized gradient methods in separable linear classification, provides tighter risk bounds depending on the tail decay rate of the loss function, and simplifies the proof technique compared to previous work.


BIOT: Biosignal Transformer for Cross-data Learning in the Wild

Chaoqi Yang, M Brandon Westover, Jimeng Sun

https://openreview.net/forum?id=c2LZyTyddi

Keywords: biological signal, transformer, cross-data learning, in-the-wild learning

Compressor summary: The paper introduces Biosignal Transformer, a flexible encoder for biosignals that can handle different formats and enable cross-data learning, outperforming baselines and improving performance on seizure detection task.


Language Models Don't Always Say What They Think: Unfaithful Explanations in Chain-of-Thought Prompting

Miles Turpin, Julian Michael, Ethan Perez, Samuel R. Bowman

https://openreview.net/forum?id=bzs4uPLXvi

Keywords: Natural language processing, large language models, XAI, explainability

Compressor summary: CoT explanations from LLMs can be misleading due to added biasing features, impacting accuracy and trust in the models.


Federated Linear Bandits with Finite Adversarial Actions

Li Fan, Ruida Zhou, Chao Tian, Cong Shen

https://openreview.net/forum?id=bzXpQUnule

Keywords: Federated bandits, contextual bandits, regret analysis

Compressor summary: FedSupLinUCB is an order-optimal algorithm for federated linear bandits with adversarial finite action sets, achieving low regret and controlled communication cost in asynchronous and synchronous cases.


Natural Language Instruction-following with Task-related Language Development and Translation

Jing-Cheng Pang, Xinyu Yang, Si-Hang Yang, Xiong-Hui Chen, Yang Yu

https://openreview.net/forum?id=bx0SDRVDzF

Keywords: Reinforcement learning, instruction-following, autonomous agent

Compressor summary: TALAR simplifies natural language instructions into task-related language for efficient policy training in reinforcement learning.


Global Structure-Aware Diffusion Process for Low-light Image Enhancement

Jinhui HOU, Zhiyu Zhu, Junhui Hou, Hui LIU, Huanqiang Zeng, Hui Yuan

https://openreview.net/forum?id=bv9mmH0LGF

Keywords: Image enhancement, diffusion models

Compressor summary: The paper proposes a diffusion-based method for low-light image enhancement, using curvature and uncertainty regularization to improve detail preservation, contrast, and noise suppression, outperforming existing methods.


Chatting Makes Perfect: Chat-based Image Retrieval

Matan Levy, Rami Ben-Ari, Nir Darshan, Dani Lischinski

https://openreview.net/forum?id=bt7pQ7o7zG

Keywords: Image Retrieval, Multi-modal learning

Compressor summary: The paragraph introduces ChatIR, a chat-based image retrieval system that uses dialog with the user to clarify their search intent and leverages Large Language Models to generate follow-up questions, achieving significant gains in image retrieval compared to existing approaches.


Learning Dynamic Attribute-factored World Models for Efficient Multi-object Reinforcement Learning

Fan Feng, Sara Magliacane

https://openreview.net/forum?id=bsNslV3Ahe

Keywords: multi-object RL, compositional generalization, factored representations

Compressor summary: DAFT-RL is a framework that improves object-centric representation learning and generalization in reinforcement learning by factorizing dynamics and rewards according to object attributes.


Self-Chained Image-Language Model for Video Localization and Question Answering

Shoubin Yu, Jaemin Cho, Prateek Yadav, Mohit Bansal

https://openreview.net/forum?id=brOMKBEGXP

Keywords: Video Question Answering, Video Localization, Image-Language Model

Compressor summary: The text describes a new video question answering method called SeViLA that uses a single image-language model to perform both keyframe localization and answer prediction, improving performance on various benchmarks.


Parallel Sampling of Diffusion Models

Andy Shih, Suneel Belkhale, Stefano Ermon, Dorsa Sadigh, Nima Anari

https://openreview.net/forum?id=bpzwUfX1UP

Keywords: diffusion models, parallel sampling

Compressor summary: ParaDiGMS is a new method that speeds up diffusion model sampling by running multiple denoising steps in parallel, achieving 2-4x faster sampling without sacrificing quality.


SyncTREE: Fast Timing Analysis for Integrated Circuit Design through a Physics-informed Tree-based Graph Neural Network

Yuting Hu, Jiajie Li, Florian Klemme, Gi-Joon Nam, Tengfei Ma, Hussam Amrouch, Jinjun Xiong

https://openreview.net/forum?id=bprclnHNvm

Keywords: Graph Neural Networks, Integrated Circuits, Circuit Timing Analysis, Physics-guided Deep Learning

Compressor summary: The paper proposes SyncTREE, a tree-based graph neural network that uses AI to improve the speed and accuracy of timing analysis for interconnects in integrated circuits.


EDGI: Equivariant Diffusion for Planning with Embodied Agents

Johann Brehmer, Joey Bose, Pim De Haan, Taco Cohen

https://openreview.net/forum?id=bpmM6SkDUy

Keywords: Planning, Diffusion models, Equivariance, Equivariant generative models

Compressor summary: EDGI is an algorithm that uses a new equivariant diffusion model to efficiently plan and learn in structured environments with various symmetries.


Learning Universal Policies via Text-Guided Video Generation

Yilun Du, Sherry Yang, Bo Dai, Hanjun Dai, Ofir Nachum, Joshua B. Tenenbaum, Dale Schuurmans, Pieter Abbeel

https://openreview.net/forum?id=bo8q5MRcwy

Keywords: sequential decision making, general-purpose agent, video diffusion

Compressor summary: The paper explores using text-guided image synthesis to create more general-purpose agents that can plan and execute various tasks by generating realistic videos of their actions.


Active Reasoning in an Open-World Environment

Manjie Xu, Guangyuan Jiang, Wei Liang, Chi Zhang, Yixin Zhu

https://openreview.net/forum?id=bo5oIoL95U

Keywords: Visual Reasoning, Abductive Reasoning, Active Reasoning

Compressor summary: The authors introduce Conan, an interactive environment for assessing active reasoning abilities in vision-language models, which differ from humans' ability to explore and reason with incomplete information.


Hierarchical Open-vocabulary Universal Image Segmentation

Xudong Wang, Shufan Li, Konstantinos Kallidromitis, Yusuke Kato, Kazuki Kozuka, Trevor Darrell

https://openreview.net/forum?id=bn4qZxltsH

Keywords: Universal Image Segmentation, Hierarchical, Open-vocabulary

Compressor summary: HIPIE is a model that uses text descriptions and hierarchical representation to perform open-vocabulary image segmentation, achieving state-of-the-art results in various image comprehension tasks.


Bringing regularized optimal transport to lightspeed: a splitting method adapted for GPUs

Jacob Lindbäck, Zesen Wang, Mikael Johansson

https://openreview.net/forum?id=bmdnWIuypV

Keywords: optimal transport, domain adaptation, splitting methods, gpu computations

Compressor summary: The algorithm efficiently solves regularized optimal transport problems using a Douglas-Rachford splitting technique with global convergence guarantees, low cost, and GPU parallelization, making it fast for various applications like domain adaptation and generative modeling.


Paxion: Patching Action Knowledge in Video-Language Foundation Models

Zhenhailong Wang, Ansel Blume, Sha Li, Genglin Liu, Jaemin Cho, Zineng Tang, Mohit Bansal, Heng Ji

https://openreview.net/forum?id=blm1pqiOXe

Keywords: video-language model, action knowledge benchmarking, action understanding, temporal understanding

Compressor summary: The paper introduces ActionBench, a benchmark to assess video-language models' action knowledge, and Paxion, a framework that improves their performance by training them with DVDM objective.


Adaptive SGD with Polyak stepsize and Line-search: Robust Convergence and Variance Reduction

Xiaowen Jiang, Sebastian U Stich

https://openreview.net/forum?id=blC2kbzvNC

Keywords: Convex Optimization, SGD, Adaptive Methods, Variance Reduction, Polyak Stepsize, Line-Search

Compressor summary: The authors propose new variants of SPS and SLS that achieve optimal rates in various settings and a novel VR method that works well with them.


Kiki or Bouba? Sound Symbolism in Vision-and-Language Models

Morris Alper, Hadar Averbuch-Elor

https://openreview.net/forum?id=bfmSc1ETT9

Keywords: multimodal learning, computer vision, NLP, cognitive science

Compressor summary: This paper investigates whether sound symbolism, the idea that certain sounds are associated with specific meanings across languages, is reflected in vision-and-language models like CLIP and Stable Diffusion, and finds strong evidence that they do show this pattern.


Contrastive Lift: 3D Object Instance Segmentation by Slow-Fast Contrastive Fusion

Yash Sanjay Bhalgat, Iro Laina, Joao F. Henriques, Andrea Vedaldi, Andrew Zisserman

https://openreview.net/forum?id=bbbbbov4Xu

Keywords: Neural Radiance Fields, Instance Segmentation, Metric Learning, Clustering, 3D Computer Vision

Compressor summary: The paper presents a novel 3D instance segmentation method that uses 2D pre-trained models and a slow-fast clustering objective function to handle large numbers of objects in complex scenes, achieving state-of-the-art results on several datasets.


Anonymous and Copy-Robust Delegations for Liquid Democracy

Markus Utke, Ulrike Schmidt-Kraepelin

https://openreview.net/forum?id=bbL20Oupi4

Keywords: liquid democracy, directed trees, parameterized markov chain, matrix tree theorem, axiomatic method

Compressor summary: The paragraph discusses a new voting system that combines representative and direct democracy, and two rules that balance anonymity and copy-robustness, which are shown to be equivalent using a polynomial-time algorithm with applications in other fields.


Lie Point Symmetry and Physics-Informed Networks

Tara Akhound-Sadegh, Laurence Perreault-Levasseur, Johannes Brandstetter, Max Welling, Siamak Ravanbakhsh

https://openreview.net/forum?id=ba4boN3W1n

Keywords: PDE, Lie point symmetry, Symmetry, Neural PDE solver, PINNs

Compressor summary: The text discusses how incorporating PDE symmetries into physics-informed neural networks can improve their ability to learn solutions and neighboring solutions efficiently.


Hybrid Search for Efficient Planning with Completeness Guarantees

Kalle Kujanpää, Joni Pajarinen, Alexander Ilin

https://openreview.net/forum?id=bY0c46ZtXa

Keywords: Planning, Subgoal search, Reinforcement learning, Hierarchical Imitation Learning, Hierarchical planning, Hierarchical reinforcement learning

Compressor summary: The paper proposes complete subgoal search, an efficient hybrid method that combines high-level and low-level actions to achieve completeness in solving complex planning problems.


Kissing to Find a Match: Efficient Low-Rank Permutation Representation

Hannah Dröge, Zorah Lähner, Yuval Bahat, Onofre Martorell Nadal, Felix Heide, Michael Moeller

https://openreview.net/forum?id=bXvmnpCMmq

Keywords: low rank, permutation, kissing number, matrix factorization, assigment problem

Compressor summary: The paper proposes an efficient method for approximating large permutation matrices using low-rank matrix factorization and reducing memory costs, enabling the solution of previously infeasible problems.


Model-Based Reparameterization Policy Gradient Methods: Theory and Practical Algorithms

Shenao Zhang, Boyi Liu, Zhaoran Wang, Tuo Zhao

https://openreview.net/forum?id=bUgqyyNo8j

Keywords: Reinforcement Learning, Model-Based Reinforcement Learning, Policy Gradient

Compressor summary: The paper analyzes the challenges of applying ReParameterization Policy Gradient Methods to long-term reinforcement learning problems and proposes a spectral normalization method to improve gradient estimation and reduce variance.


The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter

AJAY KUMAR JAISWAL, Shiwei Liu, Tianlong Chen, Zhangyang Wang

https://openreview.net/forum?id=bU9hwbsVcy

Keywords: Pre-trained Models, Sparsity, Emergence, Transformers, Pruning

Compressor summary: The paper studies sparse patterns in large pre-trained transformers, proposes a concept of essential sparsity with a sharp dropping point, and finds that self-supervised learning triggers stronger emergent sparsification than supervised learning.


Sample-Efficient and Safe Deep Reinforcement Learning via Reset Deep Ensemble Agents

Woojun Kim, Yongjae Shin, Jongeui Park, Youngchul Sung

https://openreview.net/forum?id=bTidcHIK2t

Keywords: deep reinforcement learning, primacy bais, reset, deep ensemble learning

Compressor summary: The paper proposes a new reset method for deep reinforcement learning that uses deep ensembles to address primacy bias, overfitting, and performance collapse issues while improving sample efficiency.


Seeing is not Believing: Robust Reinforcement Learning against Spurious Correlation

Wenhao Ding, Laixi Shi, Yuejie Chi, Ding Zhao

https://openreview.net/forum?id=bTL5SNOpfa

Keywords: reinforcement learning, robustness, causality, spurious correlation

Compressor summary: The paper proposes a new robust reinforcement learning framework that avoids learning spurious correlations caused by unobserved confounders and shows improved performance in real-world tasks.


Primal-Attention: Self-attention through Asymmetric Kernel SVD in Primal Representation

Yingyi Chen, Qinghua Tao, Francesco Tonin, Johan Suykens

https://openreview.net/forum?id=bRyduWAAVT

Keywords: Self-attention, primal-dual representations, SVD, kernel method, asymmetry, transformer

Compressor summary: The paper introduces Primal-Attention, a new attention mechanism based on asymmetric Kernel Singular Value Decomposition (KSVD), which improves efficiency and promotes low-rank property in self-attention.


Distributionally Robust Bayesian Optimization with $\varphi$-divergences

Hisham Husain, Vu Nguyen, Anton van den Hengel

https://openreview.net/forum?id=bRlEwWd7Vy

Keywords: Bayesian Optimization, Distributionally Robust Optimization, φ-divergences

Compressor summary: The authors propose a new algorithm for Bayesian Optimization that is robust against data-shift in various divergences and achieve sublinear regret bounds with experimental validation.


Efficiently incorporating quintuple interactions into geometric deep learning force fields

Zun Wang, Guoqing Liu, Yichi Zhou, Tong Wang, Bin Shao

https://openreview.net/forum?id=bPJmu1PbZD

Keywords: Machine learning force field, graph neural network, many-body interactions

Compressor summary: The quintuple network (QuinNet) is a graph neural network that efficiently represents many-body interactions up to five-body interactions with high accuracy for molecular dynamics simulations.


Online Control for Meta-optimization

Xinyi Chen, Elad Hazan

https://openreview.net/forum?id=bOQNd7tWAp

Keywords: online learning, control, hyperparameter optimization

Compressor summary: The authors propose a novel meta-optimization approach based on control theory to learn the best optimization algorithm online, overcoming nonconvexity challenges and achieving regret guarantees.


Structured Neural Networks for Density Estimation and Causal Inference

Asic Q Chen, Ruian Shi, Xiang Gao, Ricardo Baptista, Rahul G Krishnan

https://openreview.net/forum?id=bNXVRJjmOl

Keywords: generative models, density estimation, normalizing flows, binary matrix factorization, causal inference

Compressor summary: The Structured Neural Network (StrNN) injects structure into neural networks by masking pathways based on binary matrix factorization, enabling learning functions with desired independencies and improving data-efficient generative modeling and causal effect estimation.


Empowering Collaborative Filtering with Principled Adversarial Contrastive Loss

An Zhang, Leheng Sheng, Zhibo Cai, Xiang Wang, Tat-Seng Chua

https://openreview.net/forum?id=bNNIf8F9OU

Keywords: Collaborative filtering, Contrastive loss, Recommendation, Generalization ability

Compressor summary: The paragraph discusses the challenges of using contrastive learning in collaborative filtering and proposes a new loss function called Adversarial InfoNCE that improves generalization ability for out-of-distribution tasks.


Learning Layer-wise Equivariances Automatically using Gradients

Tycho F.A. van der Ouderaa, Alexander Immer, Mark van der Wilk

https://openreview.net/forum?id=bNIHdyunFC

Keywords: learning layer-wise relaxed equivariances bayesian symmetry discovery marginal likelihood

Compressor summary: The authors propose a method to learn flexible symmetry constraints in neural networks from data using gradients and differentiable Laplace approximations, which improves generalization performance on image classification tasks.


Sequential Predictive Two-Sample and Independence Testing

Aleksandr Podkopaev, Aaditya Ramdas

https://openreview.net/forum?id=bN1ZBSOV2f

Keywords: two-sample testing, independence testing, testing by betting, sequential testing

Compressor summary: The text discusses sequential nonparametric two-sample and independence testing using prediction-based betting strategies that work well on structured high-dimensional data and adapt to changing data distributions.


Function Space Bayesian Pseudocoreset for Bayesian Neural Networks

Balhae Kim, Hyungi Lee, Juho Lee

https://openreview.net/forum?id=bM6mynsusR

Keywords: Bayesian pseudocoresets, Function space variational inference

Compressor summary: The paper proposes a new method to create a compact synthetic dataset for scalable Bayesian inference that works on function space instead of weight space, improving uncertainty quantification and robustness.


Greatness in Simplicity: Unified Self-Cycle Consistency for Parser-Free Virtual Try-On

Chenghu Du, junyin Wang, Shuqing Liu, Shengwu Xiong

https://openreview.net/forum?id=bLB4vTwSbC

Keywords: parser-free virtual try-on, self-cycle consistency, human analysis and understanding, fashion synthesis, Markov Random Field

Compressor summary: The paragraph discusses a new parser-free virtual try-on network (USC-PFN) that improves garment deformation modeling and alignment with the human body using self-cycle consistency and Markov Random Field, achieving state-of-the-art performance.


Uncovering the Hidden Dynamics of Video Self-supervised Learning under Distribution Shifts

Pritam Sarkar, Ahmad Beirami, Ali Etemad

https://openreview.net/forum?id=bKqrWLCMrX

Keywords: computer vision, self-supervised learning, video self-supervised learning, natural distribution shift, video learning, out-of-distribution generalization

Compressor summary: The paper investigates how six self-supervised video methods handle different types of distribution shifts and reveals their strengths and weaknesses in various scenarios.


Deep Optimal Transport: A Practical Algorithm for Photo-realistic Image Restoration

Theo Joseph Adrai, Guy Ohayon, Michael Elad, Tomer Michaeli

https://openreview.net/forum?id=bJJY9TFfe0

Keywords: Computer Vision, Image Restoration, Deep Learning, Perceptual Quality

Compressor summary: The algorithm improves image restoration by using optimal transport in the latent space of a variational auto-encoder to control the trade-off between perceptual quality and MSE, and it works well with few training images.


Neural Lad: A Neural Latent Dynamics Framework for Times Series Modeling

Ting Li, Jianguo Li, Zhanxing Zhu

https://openreview.net/forum?id=bISkJSa5Td

Keywords: Neural CDE, Time-series forecasting, Latent Dynamic

Compressor summary: Neural Lad is a new neural ODE framework that addresses issues in existing models by incorporating local change and seasonality in latent dynamics and achieving better performance in various time series forecasting tasks.


Riemannian Laplace approximations for Bayesian neural networks

Federico Bergamin, Pablo Moreno-Muñoz, Søren Hauberg, Georgios Arvanitidis

https://openreview.net/forum?id=bHS7qjLOAy

Keywords: Riemannian geometry, Laplace approximation, Approximate inference, Bayesian neural networks

Compressor summary: The paper proposes a Riemannian Laplace approximation for Bayesian neural networks that adapts to the true posterior's shape and improves performance across tasks.


Asymmetric Certified Robustness via Feature-Convex Neural Networks

Samuel Pfrommer, Brendon G. Anderson, Julien Piet, Somayeh Sojoudi

https://openreview.net/forum?id=bH4LVNVXUo

Keywords: asymmetric certified robustness, input-convex neural networks

Compressor summary: The paper introduces a new neural network architecture for machine learning models to resist asymmetric adversarial attacks, which induce false negatives, and proves its effectiveness theoretically and empirically.


FourierGNN: Rethinking Multivariate Time Series Forecasting from a Pure Graph Perspective

Kun Yi, Qi Zhang, Wei Fan, Hui He, Liang Hu, Pengyang Wang, Ning An, Longbing Cao, Zhendong Niu

https://openreview.net/forum?id=bGs1qWQ1Fx

Keywords: multivariate time series forecasting, fourier space

Compressor summary: The authors propose a novel graph neural network architecture, Fourier Graph Neural Network (FouderGNN), that uses a hypervariate graph data structure to capture spatiotemporal dynamics in multivariate time series forecasting and outperforms existing methods with lower complexity.


ATTA: Anomaly-aware Test-Time Adaptation for Out-of-Distribution Detection in Segmentation

Zhitong Gao, Shipeng Yan, Xuming He

https://openreview.net/forum?id=bGcdjXrU2w

Keywords: ood detection, semantic segmentation, anomaly segmentation, test-time adaptation

Compressor summary: The paper proposes a dual-level framework to detect out-of-distribution images that can handle both domain shift and semantic shift, improving OOD detection performance on various benchmarks.


Differentiable Neuro-Symbolic Reasoning on Large-Scale Knowledge Graphs

CHEN SHENGYUAN, YUNFENG CAI, Huang Fang, Xiao Huang, Mingming Sun

https://openreview.net/forum?id=bETvUctiTR

Keywords: Neuro-Symbolic Reasoning, Knowledge graph embedding, Probabilistic soft logic

Compressor summary: DiffLogic is a neuro-symbolic reasoning framework that uses a tailored filter to select essential triples and probabilistic soft logic to assess agreement among rules, weights, and observed triples for effective and efficient knowledge graph reasoning.


A Bayesian Take on Gaussian Process Networks

Enrico Giudice, Jack Kuipers, Giusi Moffa

https://openreview.net/forum?id=bBIHqoZ3OR

Keywords: Bayesian networks, structure learning, graphical models, gaussian processes, Bayesian inference, MCMC sampling, importance sampling

Compressor summary: The paragraph explains how Gaussian Process Networks (GPNs) model continuous distributions with flexible Graphical Models, but their Bayesian structure learning is hard to compute, so the authors propose a Monte Carlo method that samples from the posterior distribution and performs well in simulations.


Marich: A Query-efficient Distributionally Equivalent Model Extraction Attack

Pratik Karmakar, Debabrota Basu

https://openreview.net/forum?id=bAI21VEMvM

Keywords: Differential Privacy, Model Extraction Attacks, Active Sampling, Max-Information Attack

Compressor summary: The authors propose a novel black-box model extraction attack called Marich that uses active sampling to select queries from public data and create an accurate replica of the target ML model.


A case for reframing automated medical image classification as segmentation

Sarah Hooper, Mayee F Chen, Khaled Kamal Saab, Kush Bhatia, Curtis Langlotz, Christopher Re

https://openreview.net/forum?id=b8xowIlZ7v

Keywords: Medical imaging, segmentation, classification

Compressor summary: The paper compares classification and segmentation models for medical image analysis, proposes methods to use segmentation models for classification, and discusses the benefits of switching from segmentation to classification in terms of performance, robustness, and interpretability.


One-Step Diffusion Distillation via Deep Equilibrium Models

Zhengyang Geng, Ashwini Pokle, J Zico Kolter

https://openreview.net/forum?id=b6XvK2de99

Keywords: Deep Equilibrium Models, Diffusion Models, Distillation, Generative Models

Compressor summary: The paper introduces the GET model, which distills diffusion models directly from noise to image using an offline training approach with better performance than existing one-step methods.


Convergence of Alternating Gradient Descent for Matrix Factorization

Rachel Ward, Tamara G. Kolda

https://openreview.net/forum?id=b6FeLpKKjl

Keywords: matrix factorization; gradient descent; global convergence; concentration; optimization

Compressor summary: The paper studies alternating gradient descent with fixed step size for asymmetric matrix factorization, showing that a specific number of iterations is enough to reach an approximate solution with high probability, starting from a certain random initialization, and experiments confirm the theoretical results.


On the Robustness of Removal-Based Feature Attributions

Chris Lin, Ian Connick Covert, Su-In Lee

https://openreview.net/forum?id=b60wLlkBta

Keywords: explainable artificial intelligence, interpretable machine learning, feature attributions, removal-based feature attributions, robustness

Compressor summary: The paper studies how removal-based feature attribution methods can be more robust to input and model perturbations and provides a unified analysis with theoretical and empirical results.


A Heavy-Tailed Algebra for Probabilistic Programming

Feynman T. Liang, Liam Hodgkinson, Michael W. Mahoney

https://openreview.net/forum?id=b5R8mbqo9Q

Keywords: probabilistic programming, static analysis, heavy tails, monte carlo, mcmc, variational inference

Compressor summary: The paper proposes a method to improve tail behavior accuracy in probabilistic models using noise in neural networks, based on an algebra using the generalized Gamma distribution.


LightSpeed: Light and Fast Neural Light Fields on Mobile Devices

Aarush Gupta, Junli Cao, Chaoyang Wang, Ju Hu, Sergey Tulyakov, Jian Ren, Laszlo Attila Jeni

https://openreview.net/forum?id=b2wSODM7iG

Keywords: light field, neural radiance field, novel view synthesis

Compressor summary: The authors propose a new method for real-time image synthesis on mobile devices using the light slab representation, which improves both quality and speed compared to existing neural light field methods.


Stable Nonconvex-Nonconcave Training via Linear Interpolation

Thomas Pethick, Wanyun Xie, Volkan Cevher

https://openreview.net/forum?id=b1JPBGJhUi

Keywords: Minimax optimization, Lookahead, Generative adversarial networks, Stability, Nonconvex-nonconcave, Cohypomonotone

Compressor summary: The paper proposes a new optimization method called RAPP that uses linear interpolation to stabilize neural network training, and extends it to constrained and regularized settings, as well as to Lookahead algorithms, with experiments on generative adversarial networks showing its benefits.


Sharp Recovery Thresholds of Tensor PCA Spectral Algorithms

Michael Jacob Feldman, David Donoho

https://openreview.net/forum?id=b1BhHjBxsx

Keywords: tensor PCA, spectral algorithms, random matrix theory

Compressor summary: The paper studies different ways to convert noisy tensors into matrices and uses spectral methods to recover the low-rank signal, proving that some methods work better than others in certain conditions.


On the Ability of Graph Neural Networks to Model Interactions Between Vertices

Noam Razin, Tom Verbin, Nadav Cohen

https://openreview.net/forum?id=ayZpFoAu5c

Keywords: Graph Neural Networks, Expressivity, Interactions, Edge Sparsification

Compressor summary: The paper proposes a measure to quantify interaction strength in graph neural networks (GNNs) and introduces an edge sparsification method that preserves interaction ability while reducing computation.


Experiment Planning with Function Approximation

Aldo Pacchiano, Jonathan Lee, Emma Brunskill

https://openreview.net/forum?id=axmY49ahVI

Keywords: regret, model selection, planning, static, lower bound

Compressor summary: The paper proposes two experiment planning strategies for function approximation in contextual bandits using large datasets of contexts but no rewards, and analyzes their optimality guarantees and differences from adaptive learning.


Greedy Poisson Rejection Sampling

Gergely Flamich

https://openreview.net/forum?id=axRMkinASf

Keywords: channel simulation, relative entropy coding, reverse channel coding, rejection sampling, Poisson process

Compressor summary: The paper presents a new algorithm (GPRS) for efficient one-shot channel simulation in one-dimensional problems with unimodal density ratio and shows its superior performance over existing methods.


Characterization of Overfitting in Robust Multiclass Classification

Jingyuan Xu, Weiwei Liu

https://openreview.net/forum?id=awbWWO0nb6

Keywords: Learning Theory

Compressor summary: The paper investigates how adaptive algorithms may overfit the test dataset in multiclass classification, with bounds on the overfitting bias.


LEACE: Perfect linear concept erasure in closed form

Nora Belrose, David Schneider-Joseph, Shauli Ravfogel, Ryan Cotterell, Edward Raff, Stella Biderman

https://openreview.net/forum?id=awIpKpwTwF

Keywords: interpretability, fairness, concept erasure, representation, adversarial, robustness

Compressor summary: LEACE is a method to remove specific features from representations while minimizing changes, and it can be applied to language models for fairness and interpretability tasks.


Risk-Averse Active Sensing for Timely Outcome Prediction under Cost Pressure

Yuchao Qin, Mihaela van der Schaar, Changhee Lee

https://openreview.net/forum?id=aw1vLo7TE7

Keywords: active sensing, value of information, risk-averse learning

Compressor summary: The paper proposes RAS, a novel risk-averse active sensing approach for cost-efficient acquisition of patient covariates in healthcare to enable timely and accurate outcome predictions.


Discovering Intrinsic Spatial-Temporal Logic Rules to Explain Human Actions

Chengzhi Cao, Chao Yang, Ruimao Zhang, Shuang Li

https://openreview.net/forum?id=avuRopYsCg

Keywords: Logic rule, human actions, sports analyze

Compressor summary: The paper proposes a model that uses spatial-temporal logic rules to capture human movement patterns based on intentions and environmental factors, with potential applications in sports analytics, robotics, and autonomous cars.


Concept Distillation: Leveraging Human-Centered Explanations for Model Improvement

Avani Gupta, Saurabh Saini, P J Narayanan

https://openreview.net/forum?id=arkmhtYLL6

Keywords: Human Centered Concepts, ML interpretability, XAI based Model Improvement, Debiasing

Compressor summary: The authors propose CAVs for ante-hoc training with Concept Loss and Concept Distillation to make models more sensitive and robust to concepts, which enhances interpretability and knowledge transfer.


LICO: Explainable Models with Language-Image COnsistency

Yiming Lei, Zilong Li, Yangyang Li, Junping Zhang, Hongming Shan

https://openreview.net/forum?id=apjOYp3mOa

Keywords: Language-image consistency, prompt learning, image classification, CNN interpretation

Compressor summary: The paper proposes LICO, a model that uses linguistic prompts and optimal transport theory to generate more accurate and interpretable attention maps for image classification tasks.


FlowCam: Training Generalizable 3D Radiance Fields without Camera Poses via Pixel-Aligned Scene Flow

Cameron Omid Smith, Yilun Du, Ayush Tewari, Vincent Sitzmann

https://openreview.net/forum?id=apFDDJOYf5

Keywords: Pose Estimation, Scene Flow Estimation, Scene Representation Learning, Computer Vision, Neural Implicit Representations, Neural Radiance Fields, View Synthesis, Self-Supervised Representation Learning

Compressor summary: The authors propose a self-supervised method for learning 3D neural scenes from videos by jointly estimating camera poses and reconstructing the scene in a single forward pass.


Factorized Contrastive Learning: Going Beyond Multi-view Redundancy

Paul Pu Liang, Zihao Deng, Martin Q. Ma, James Zou, Louis-Philippe Morency, Russ Salakhutdinov

https://openreview.net/forum?id=alLs7EtRJP

Keywords: multimodal learning, contrastive learning, self-supervised learning, information theory

Compressor summary: FactorCL is a new method for learning multimodal representations that captures both shared and unique information relevant to downstream tasks using factorization, maximizing mutual information lower bounds, minimizing mutual information upper bounds, and data augmentations.


Decompose a Task into Generalizable Subtasks in Multi-Agent Reinforcement Learning

Zikang Tian, Ruizhi Chen, Xing Hu, Ling Li, Rui Zhang, Fan Wu, Shaohui Peng, Jiaming Guo, Zidong Du, Qi Guo, Yunji Chen

https://openreview.net/forum?id=aky0dKv9ip

Keywords: Multi-Agent Reinforcement Learning, Transfer Learning, Zero-Shot Generalization

Compressor summary: The paper proposes a novel framework called DT2GS that decomposes MARL tasks into generalizable subtasks using a scalable encoder and an adaptive semantic module to achieve zero-shot generalization and task transferability.


Improving Robustness with Adaptive Weight Decay

Amin Ghiasi, Ali Shafahi, Reza Ardekani

https://openreview.net/forum?id=ajnThDhuq6

Keywords: Adaptive weight decay, adversarial robustness, weight decay, robust overfitting, overfitting, adversarial attacks, noisy label

Compressor summary: The paper proposes adaptive weight decay that adjusts the regularization strength during training to improve adversarial robustness without extra data or tuning.


Learning Mixtures of Gaussians Using the DDPM Objective

Kulin Shah, Sitan Chen, Adam Klivans

https://openreview.net/forum?id=aig7sgdRfI

Keywords: Mixtures of Gaussians, score-based generative models, provable learning of score, Expectation-Maximization, DDPM generative model

Compressor summary: The paper proves that gradient descent on a specific diffusion model can efficiently learn Gaussian mixture models with different levels of initialization and separation.


Cross-Episodic Curriculum for Transformer Agents

Lucy Xiaoyang Shi, Yunfan Jiang, Jake Grigsby, Linxi Fan, Yuke Zhu

https://openreview.net/forum?id=afKnrwJBAl

Keywords: Transformers, In-context Learning, Reinforcement Learning, Robotics

Compressor summary: CEC is a new algorithm that improves Transformer agents' learning efficiency and generalization by sequentially structuring online learning trials and mixed-quality demonstrations, resulting in better performance and strong generalization across episodes.


RanPAC: Random Projections and Pre-trained Models for Continual Learning

Mark McDonnell, Dong Gong, Amin Parvaneh, Ehsan Abbasnejad, Anton van den Hengel

https://openreview.net/forum?id=aec58UfBzA

Keywords: continual learning, class incremental learning, domain incremental learning, pre-trained models, parameter-efficient transfer learning

Compressor summary: The paper proposes a new approach for continual learning with pre-trained models that uses random projectors and class-prototype accumulation to avoid forgetting without rehearsal memory.


Tree Variational Autoencoders

Laura Manduchi, Moritz Vandenhirtz, Alain Ryser, Julia E Vogt

https://openreview.net/forum?id=adq0oXb9KM

Keywords: hierarchical clustering, hierarchical VAE, representation learning, VAE, deep clustering

Compressor summary: TreeVAE is a new model that learns a tree structure to encode latent variables and discover hidden clusters in data, while improving generative performance and allowing conditional sampling.


Adapting to Continuous Covariate Shift via Online Density Ratio Estimation

Yu-Jie Zhang, Zhen-Yu Zhang, Peng Zhao, Masashi Sugiyama

https://openreview.net/forum?id=ad3JNoR2np

Keywords: Covariate Shift, Density Ratio Estimation, Online Convex Optimization, Dynamic Regret, Logistic Regression

Compressor summary: The paper studies continuous covariate shift in machine learning, proposes an online density ratio estimation method to adaptively train predictors, and provides theoretical and empirical guarantees.


Mixed Samples as Probes for Unsupervised Model Selection in Domain Adaptation

Dapeng Hu, Jian Liang, Jun Hao Liew, Chuhui Xue, Song Bai, Xinchao Wang

https://openreview.net/forum?id=ackajXqei2

Keywords: Unsupervised Domain Adaptation; Model Selection; Hyperparameter Selection; Unsupervised Validation;

Compressor summary: MixVal is a novel model selection method for unsupervised domain adaptation that uses mixed target samples with pseudo labels to evaluate the performance of different UDA models on unlabeled target data.


MarioGPT: Open-Ended Text2Level Generation through Large Language Models

Shyam Sudhakaran, Miguel González-Duque, Matthias Freiberger, Claire Glanois, Elias Najarro, Sebastian Risi

https://openreview.net/forum?id=aa8KsqfTPa

Keywords: Large Language Models, Procedural Content Generation, Open-endedness, Novelty Search

Compressor summary: MarioGPT is a text-prompted, fine-tuned GPT2 model that generates diverse Super Mario Bros levels and can be controlled by user input to address challenges in procedural content generation.


Anchor Data Augmentation

Nora Schneider, Shirin Goshtasbpour, Fernando Perez-Cruz

https://openreview.net/forum?id=aZ9hvpnp0k

Keywords: Data Augmentation, Regression, Deep Learning

Compressor summary: The paper introduces a new data augmentation method for nonlinear regression called Anchor Data Augmentation, which uses causality and distributionally robust Anchor regression to generate more training examples and improve prediction accuracy.


Reproducibility in Multiple Instance Learning: A Case For Algorithmic Unit Tests

Edward Raff, James Holt

https://openreview.net/forum?id=aZ44Na3l9p

Keywords: reproducibility; multiple instance learning

Compressor summary: The paragraph discusses how five deep-MIL models do not follow the standard assumption of Multiple Instance Learning (MIL) and may lead to incorrect predictions in various applications like healthcare and cyber security.


Language Models Can Improve Event Prediction by Few-Shot Abductive Reasoning

Xiaoming Shi, Siqiao Xue, Kangrui Wang, Fan Zhou, James Y. Zhang, JUN ZHOU, Chenhao Tan, Hongyuan Mei

https://openreview.net/forum?id=aW9BqtRQkh

Keywords: event sequences, irregular time series, event prediction, large language model, reasoning, few-shot prompting

Compressor summary: The paper proposes LAMP, a framework that uses a large language model for abductive reasoning to improve event prediction performance on real-world datasets.


Drift doesn't Matter: Dynamic Decomposition with Diffusion Reconstruction for Unstable Multivariate Time Series Anomaly Detection

Chengsen Wang, Zirui Zhuang, Qi Qi, Jingyu Wang, Xingyu Wang, Haifeng Sun, Jianxin Liao

https://openreview.net/forum?id=aW5bSuduF1

Keywords: Anomaly Detection, Time Series, Diffusion, Transformer

Compressor summary: D$^3$R is an anomaly detection network for unstable data that uses dynamic decomposition and reconstruction to handle drift and achieve better performance than existing methods.


A Dynamical System View of Langevin-Based Non-Convex Sampling

Mohammad Reza Karimi Jaghargh, Ya-Ping Hsieh, Andreas Krause

https://openreview.net/forum?id=aRBa0lSxEB

Keywords: Non-Convex Sampling, Langevin Dynamics, Dynamical Systems

Compressor summary: The paragraph discusses a novel framework for addressing non-convex sampling challenges in machine learning using tools from dynamical systems theory to improve convergence guarantees.


Scale-Space Hypernetworks for Efficient Biomedical Image Analysis

Jose Javier Gonzalez Ortiz, John Guttag, Adrian V Dalca

https://openreview.net/forum?id=aN0llPIbdg

Keywords: hypernetworks, amortized learning, computer vision, rescaling, convolutional neural networks, pareto efficiency

Compressor summary: The authors propose Scale-Space HyperNetworks (SSHN), a method that learns a spectrum of CNNs with varying internal rescaling factors, allowing for better accuracy-efficiency trade-offs in medical image analysis tasks.


The emergence of clusters in self-attention dynamics

Borjan Geshkovski, Cyril Letrouit, Yury Polyanskiy, Philippe Rigollet

https://openreview.net/forum?id=aMjaEkkXJx

Keywords: Transformers, Self-Attention, Clustering, Interacting Particle Systems, Continuous Time

Compressor summary: The paragraph discusses how learned representations in Transformers form clusters and limiting objects depending on the value matrix spectrum, and proves convergence to low-rank Boolean matrices in one dimension.


FourierHandFlow: Neural 4D Hand Representation Using Fourier Query Flow

Jihyun Lee, Junbong Jang, Donghwan Kim, Minhyuk Sung, Tae-Kyun Kim

https://openreview.net/forum?id=aMTiwdK3y8

Keywords: 4D representation, hand reconstruction, implicit representation

Compressor summary: FourierHandFlow is a novel method for learning spatio-temporally continuous representations of human hands from RGB videos using Fourier series and implicit shape priors.


Inference-Time Intervention: Eliciting Truthful Answers from a Language Model

Kenneth Li, Oam Patel, Fernanda Viégas, Hanspeter Pfister, Martin Wattenberg

https://openreview.net/forum?id=aLLuYpn83y

Keywords: Large Language Model, AI Safety

Compressor summary: ITI is a technique that improves the truthfulness of large language models by shifting model activations during inference based on a learned set of directions.


Learning to Modulate pre-trained Models in RL

Thomas Schmied, Markus Hofmarcher, Fabian Paischer, Razvan Pascanu, Sepp Hochreiter

https://openreview.net/forum?id=aIpGtPwXny

Keywords: Reinforcement Learning, Transformer, Decision Transformer, Multi-task learning, Continual learning, NLP, Fine-tuning, Prompt Tuning, Parameter efficient Fine-tuning

Compressor summary: The paper investigates catastrophic forgetting in reinforcement learning, compares various fine-tuning methods, and proposes a novel method called Learning-to-Modulate (L2M) that preserves pre-training skills while adapting to new tasks.


Feature Adaptation for Sparse Linear Regression

Jonathan Kelner, Frederic Koehler, Raghu Meka, Dhruv Rohatgi

https://openreview.net/forum?id=aIUnoHuENG

Keywords: theory, sparse linear regression, feature adaptation, lasso

Compressor summary: The paper presents an algorithm that adapts the Lasso to handle sparse linear regression with correlated design and ill-conditioned covariates, achieving near-optimal sample complexity.


CS4ML: A general framework for active learning with arbitrary data based on Christoffel functions

Juan M Cardenas, Ben Adcock, Nick Dexter

https://openreview.net/forum?id=aINqoP32cb

Keywords: active learning, regression, arbitrary data, leverage scores, Christoffel functions, generative models, Magnetic Resonance Imaging (MRI), Physics-Informed Neural Networks (PINNs)

Compressor summary: The paper presents a general active learning framework for regression problems that can handle different types of data and shows its effectiveness in scientific computing applications.


Enhancing Adaptive History Reserving by Spiking Convolutional Block Attention Module in Recurrent Neural Networks

Qi Xu, Yuyuan Gao, Jiangrong Shen, Yaxin Li, Xuming Ran, Huajin Tang, Gang Pan

https://openreview.net/forum?id=aGZp61S9Lj

Keywords: Spiking neural networks (SNNs), Recurrent spiking neural network (RSNN), Dynamic Vision Sensor (DVS), Spiking convolutional block attention module (SCBAM)

Compressor summary: The paper proposes a recurrent spiking neural network with an attention module to process spatio-temporal patterns in time series data by combining both spatial and temporal features efficiently.


Optimal Unbiased Randomizers for Regression with Label Differential Privacy

Ashwinkumar Badanidiyuru, Badih Ghazi, Pritish Kamath, Ravi Kumar, Ethan Jacob Leeman, Pasin Manurangsi, Avinash V Varadarajan, Chiyuan Zhang

https://openreview.net/forum?id=aG6xOP9QY7

Keywords: label differential privacy

Compressor summary: We propose a new family of label randomizers for training _regression_ models under the constraint of label differential privacy (DP). In particular, we leverage the trade-offs between bias and variance to construct better label randomizers depending on a privately estimated prior distribution over the labels. We demonstrate that these randomizers achieve state-of-the-art privacy-utility trade-offs on several datasets, highlighting the importance of reducing bias when training neural networks with label DP. We also provide theoretical results shedding light on the structural properties of the optimal unbiased randomizers.


Spectral Entry-wise Matrix Estimation for Low-Rank Reinforcement Learning

Stefan Stojanovic, Yassir Jedra, Alexandre Proutiere

https://openreview.net/forum?id=aDLmRMb0K9

Keywords: Low-rank matrix estimation; low rank bandits; low rank MDP; spectral methods

Compressor summary: The paper explores matrix estimation methods with low entry-wise prediction error in low-rank reinforcement learning problems, and presents two new algorithms with near-optimal performance.


Polynomial-Time Linear-Swap Regret Minimization in Imperfect-Information Sequential Games

Gabriele Farina, Charilaos Pipis

https://openreview.net/forum?id=aCOKUvqHtD

Keywords: online learning, algorithmic game theory, extensive form games, correlated equilibrium, swap regret, linear swap regret

Compressor summary: The paragraph discusses how no-linear-swap regret learners can achieve sublinear regret in sequential games by considering all linear transformations of the mixed strategy space, and shows this leads to a stronger notion of hindsight rationality than previous ones.


On the Convergence of CART under Sufficient Impurity Decrease Condition

Rahul Mazumder, Haoyue Wang

https://openreview.net/forum?id=a2svOXTVgO

Keywords: decision tree, CART

Compressor summary: The paper studies the convergence rate and error bound of CART, a machine-learning model, under a regression setting and introduces sufficient conditions for the SID condition.


Students Parrot Their Teachers: Membership Inference on Model Distillation

Matthew Jagielski, Milad Nasr, Katherine Lee, Christopher A. Choquette-Choo, Nicholas Carlini, Florian Tramèr

https://openreview.net/forum?id=a2Yg9Za6Rb

Keywords: model distillation, membership inference, privacy, dark knowledge

Compressor summary: The paper studies how well knowledge distillation protects privacy, and shows that it is not very effective against membership inference attacks, especially if the student and teacher models are similar or the attacker can tamper with the teacher data.


Training Chain-of-Thought via Latent-Variable Inference

Matthew Douglas Hoffman, Du Phan, david dohan, Sholto Douglas, Tuan Anh Le, Aaron T Parisi, Pavel Sountsov, Charles Sutton, Sharad Vikram, Rif A. Saurous

https://openreview.net/forum?id=a147pIS2Co

Keywords: Large language models, latent-variable models, control variates, chain-of-thought, MCMC

Compressor summary: The paragraph discusses a new method to improve large language models' performance by combining chain-of-thought prompting and supervised fine-tuning, using a Markov-chain Monte Carlo expectation-maximization algorithm.


3D molecule generation by denoising voxel grids

Pedro O. Pinheiro, Joshua Rackers, joseph Kleinhenz, Michael Maser, Omar Mahmood, Andrew Martin Watkins, Stephen Ra, Vishnu Sresht, Saeed Saremi

https://openreview.net/forum?id=Zyzluw0hC4

Keywords: generative model, molecule generation, drug discovery

Compressor summary: The paragraph describes a new method called VoxMol that generates 3D molecules as atomic densities on grids by training a neural network to denoise noisy molecule distributions.


Rehearsal Learning for Avoiding Undesired Future

Tian Qin, Tian-Zuo Wang, Zhi-Hua Zhou

https://openreview.net/forum?id=ZwQJRXLjVm

Keywords: decision-making, structural rehearsal model, Bayesian inference, probabilistic graphical model

Compressor summary: The paper introduces a rehearsal learning framework that helps find decisions to avoid undesired outcomes using structural rehearsal models, probabilistic graphical models, and a Bayesian approach with a risk quantification bound.


Thought Cloning: Learning to Think while Acting by Imitating Human Thinking

Shengran Hu, Jeff Clune

https://openreview.net/forum?id=ZvDmna23r3

Keywords: Reinforcement learning, Imitation Learning, AI Safety, Interpretability

Compressor summary: Thought Cloning is a novel Imitation Learning framework that trains AI agents to not only mimic human behaviors but also the thoughts behind them, improving performance, safety, and interpretability.


Injecting Multimodal Information into Rigid Protein Docking via Bi-level Optimization

Ruijia Wang, YiWu Sun, Yujie Luo, Shaochuan Li, Cheng Yang, Xingyi Cheng, Hui Li, Chuan Shi, Le Song

https://openreview.net/forum?id=ZuaVKlWdD2

Keywords: complex structure prediction, rigid docking, protein docking, antibody-antigen docking

Compressor summary: BiDock is a novel rigid protein docking model that uses both sequence and structure information to predict inter-protein distances for better predictions.


Differentially Private Approximate Near Neighbor Counting in High Dimensions

Alexandr Andoni, Piotr Indyk, Sepideh Mahabadi, Shyam Narayanan

https://openreview.net/forum?id=Zt9RzHjSEy

Keywords: Differential Privacy, Near Neighbor Search, Locality Sensitive Hashing, Data Structures, Range Query

Compressor summary: The paper proposes an efficient algorithm for range counting under differential privacy that balances additive and multiplicative errors without depending on the dimension.


WalkLM: A Uniform Language Model Fine-tuning Framework for Attributed Graph Embedding

Yanchao Tan, Zihao Zhou, Hang Lv, Weiming Liu, Carl Yang

https://openreview.net/forum?id=ZrG8kTbt70

Keywords: Arributed graph, unsupervised graph learning, language models, representation learning

Compressor summary: The text describes an approach to create unsupervised graph embeddings using language models and random walks, which can handle complex attributes and structures in real-world graphs and improve performance on various downstream tasks.


Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing

Josh Alman, Jiehao Liang, Zhao Song, Ruizhe Zhang, Danyang Zhuo

https://openreview.net/forum?id=ZqSx5vXOgC

Keywords: Training neural network, Dynamic activated neuron detection, Sparsity, Fine-grained complexity, Data structure

Compressor summary: This paper proposes a preprocessing method to reduce the time needed to process each data point in deep neural networks by storing weight-data correlation in a tree structure.


SE(3) Diffusion Model-based Point Cloud Registration for Robust 6D Object Pose Estimation

Haobo Jiang, Mathieu Salzmann, Zheng Dang, Jin Xie, Jian Yang

https://openreview.net/forum?id=Znpz1sv4IP

Keywords: 6D object pose estimation, Point cloud registration, Diffusion probabilistic model

Compressor summary: The paper presents a new point cloud registration method using an SE(3) diffusion model for accurate 6D object pose estimation in real-world scenarios.


Time Series as Images: Vision Transformer for Irregularly Sampled Time Series

Zekun Li, Shiyang Li, Xifeng Yan

https://openreview.net/forum?id=ZmeAoWQqe0

Keywords: irregularly sampled time series, vision transformer, healthcare, time series classification

Compressor summary: The paper proposes a new method to classify irregularly sampled time series by converting them into images and using pre-trained vision transformers, which shows strong performance and robustness against missing data.


Optimistic Meta-Gradients

Sebastian Flennerhag, Tom Zahavy, Brendan O'Donoghue, Hado van Hasselt, András György, Satinder Singh

https://openreview.net/forum?id=ZmSg4f16uo

Keywords: meta-learning, online optimisation, convex optimisation

Compressor summary: The paper explores how gradient-based meta-learning relates to convex optimization, proves convergence rates for meta learning with momentum, and shows that the Bootstrapped Meta-Gradient method captures optimism in meta-learning.


Collapsed Inference for Bayesian Deep Learning

Zhe Zeng, Guy Van den Broeck

https://openreview.net/forum?id=Zi1KKzh5Aj

Keywords: Bayesian Model Averaging, Weighted Model Integration, Bayesian Deep Learning, Collapsed Inference

Compressor summary: The paper introduces a novel inference method for Bayesian neural networks that improves both uncertainty estimation and predictive performance by using collapsed samples and volume computation solvers.


A Unified Conditional Framework for Diffusion-based Image Restoration

Yi Zhang, Xiaoyu Shi, Dasong Li, Xiaogang Wang, Jian Wang, Hongsheng Li

https://openreview.net/forum?id=ZgVJvaAS2h

Keywords: image restoration, diffusion model, denoising, deblurring, JPEG restoration

Compressor summary: The paper introduces a conditional framework for image restoration using diffusion probabilistic models that integrates guidance from a lightweight UNet and produces high-quality results on challenging tasks.


Leveraging the two-timescale regime to demonstrate convergence of neural networks

Pierre Marion, Raphaël Berthier

https://openreview.net/forum?id=ZfFR4d5gUM

Keywords: neural networks, non-convex optimization, gradient flow, convergence proof, two-timescale algorithm

Compressor summary: The study analyzes how shallow neural networks trained with different step sizes converge to a global optimum in a simple setting, showing that stochastic gradient descent works well when the stepsizes follow a two-timescale regime.


TrojLLM: A Black-box Trojan Prompt Attack on Large Language Models

Jiaqi Xue, Mengxin Zheng, Ting Hua, Yilin Shen, Yepeng Liu, Ladislau Bölöni, Qian Lou

https://openreview.net/forum?id=ZejTutd7VY

Keywords: Large Language Model, Trojan Attack, Adversary Attack, Prompt Injection, GPT-4, Black-box

Compressor summary: The paper introduces TrojLLM, a framework that can generate stealthy triggers to manipulate the outputs of large language models like GPT-3 and GPT-4 for malicious purposes.


Local Convergence of Gradient Methods for Min-Max Games: Partial Curvature Generically Suffices

Guillaume Wang, Lénaïc Chizat

https://openreview.net/forum?id=ZeRiLBvIps

Keywords: Gradient methods, min-max optimization, spectral analysis, last-iterate convergence

Compressor summary: The paper studies how gradient methods for two-player zero-sum differentiable games converge to local Nash equilibria and shows that they are faster when the game has partial curvature, which is related to the average eigenvalues of the symmetric part of the Jacobian.


SimFBO: Towards Simple, Flexible and Communication-efficient Federated Bilevel Learning

Yifan Yang, Peiyao Xiao, Kaiyi Ji

https://openreview.net/forum?id=ZdxGmJGKOo

Keywords: Federated bilevel optimization, federated hypergradient, communication efficiency, system-level heterogeneity, linear speedup

Compressor summary: The paper introduces SimFBO, a simple and efficient federated bilevel optimization framework, and its variant ShroFBO, which improve convergence speed, communication efficiency, and robustness to heterogeneity in machine learning and edge computing.


Optimal testing using combined test statistics across independent studies

Lasse Vuursteen, Botond Szabo, Aad van der Vaart, Harry van Zanten

https://openreview.net/forum?id=ZcuFDaMTYw

Keywords: testing, meta-analysis, p-values, e-values, optimal, combining trials

Compressor summary: The paper investigates the power and limitations of combining test statistics from independent trials in high-dimensional models, and introduces methods to improve meta-analysis.


Large Language Models Are Semi-Parametric Reinforcement Learning Agents

Danyang Zhang, Lu Chen, Situo Zhang, Hongshen Xu, Zihan Zhao, Kai Yu

https://openreview.net/forum?id=ZcJa1R6j3v

Keywords: Learning from Experiences, LLM, Reinforcement Learning, Decision Making, Experience Memory

Compressor summary: The paragraph describes a novel evolvable LLM-based agent framework called Rememberer, which uses a long-term memory to learn from past experiences and improve its performance in different tasks without fine-tuning the parameters of the LLM, achieving superior results on two RL task sets.


Continuous Parametric Optical Flow

Jianqin Luo, Zhexiong Wan, yuxin mao, Bo Li, Yuchao Dai

https://openreview.net/forum?id=ZZgfS1DbmO

Keywords: optical flow, point trajectories, continuous motion, neural ordinary differential equation

Compressor summary: The paper introduces a new method to represent continuous and dense motion in videos using B-splines and neural ODEs, and proposes a synthetic dataset and evaluation metrics to measure its performance.


Generalizable Lightweight Proxy for Robust NAS against Diverse Perturbations

Hyeonjeong Ha, Minseon Kim, Sung Ju Hwang

https://openreview.net/forum?id=ZZWg9jJQ1j

Keywords: neural architecture search, generalization, efficiency, zero-cost proxy

Compressor summary: The paper proposes a novel method to find optimal neural architectures that can learn generalizable features and robustness against various perturbations using a lightweight and efficient proxy.


A Definition of Continual Reinforcement Learning

David Abel, Andre Barreto, Benjamin Van Roy, Doina Precup, Hado van Hasselt, Satinder Singh

https://openreview.net/forum?id=ZZS9WEWYbD

Keywords: Continual Reinforcement Learning, Reinforcement Learning, Lifelong Reinforcement Learning, Continual Learning

Compressor summary: The paper defines the continual reinforcement learning problem as a setting where agents never stop learning and always search for better policies through an implicit process, and shows that traditional views of multi-task reinforcement learning and continual supervised learning are special cases of this definition.


TART: A plug-and-play Transformer module for task-agnostic reasoning

Kush Bhatia, Avanika Narayan, Christopher De Sa, Christopher Re

https://openreview.net/forum?id=ZXbgVm3PSt

Keywords: In-context learning, task-agnostic methods, large language models

Compressor summary: The paper proposes TART, a method to improve large language models' reasoning abilities using synthetic tasks, which enhances their in-context learning performance across different models, sizes, tasks, and modalities.


Structured Semidefinite Programming for Recovering Structured Preconditioners

Arun Jambulapati, Jerry Li, Christopher Musco, Kirankumar Shiragur, Aaron Sidford, Kevin Tian

https://openreview.net/forum?id=ZViPzk1sUI

Keywords: preconditioning, semidefinite programming, numerical linear algebra, linear regression, semi-random models

Compressor summary: The authors present a framework and algorithms for finding approximate optimal preconditioners and solving linear systems faster than previous methods.


Beyond Confidence: Reliable Models Should Also Consider Atypicality

Mert Yuksekgonul, Linjun Zhang, James Zou, Carlos Guestrin

https://openreview.net/forum?id=ZVRG3toCTT

Keywords: trustworthy machine learning, reliable machine learning, uncertainty

Compressor summary: The authors study how atypical inputs or classes affect model reliability, accuracy, and uncertainty, and show incorporating atypicality improves performance in various tasks, including skin lesion classification.


Self-Supervised Learning with Lie Symmetries for Partial Differential Equations

Grégoire Mialon, Quentin Garrido, Hannah Lawrence, Danyal Rehman, Yann LeCun, Bobak Kiani

https://openreview.net/forum?id=ZULq9QV8rH

Keywords: Self-supervised learning, partial differential equations, Lie symmetries, data augmentation

Compressor summary: The text discusses using self-supervised learning to learn representations of partial differential equations (PDEs) from heterogeneous data, improving both coefficient regression and neural solver performance.


Revisiting Area Convexity: Faster Box-Simplex Games and Spectrahedral Generalizations

Arun Jambulapati, Kevin Tian

https://openreview.net/forum?id=ZRBGwpeewz

Keywords: Optimization, optimal transport, linear programming, semidefinite programming

Compressor summary: The paper explores area convexity to create a fast first-order algorithm for solving box-simplex games and improve the complexity of other combinatorial optimization problems.


Rethinking the Role of Token Retrieval in Multi-Vector Retrieval

Jinhyuk Lee, Zhuyun Dai, Sai Meher Karthik Duddu, Tao Lei, Iftekhar Naim, Ming-Wei Chang, Vincent Y Zhao

https://openreview.net/forum?id=ZQzm0Z47jz

Keywords: information retrieval, document retrieval, natural language processing

Compressor summary: XTR simplifies multi-vector retrieval by rethinking token retrieval and improves ranking efficiency without sacrificing performance on information retrieval benchmarks.


Normalizing flow neural networks by JKO scheme

Chen Xu, Xiuyuan Cheng, Yao Xie

https://openreview.net/forum?id=ZQMlfNijY5

Keywords: Normalizing flow, invertible neural networks, JKO scheme

Compressor summary: JKO-iFlow is a neural ODE flow network for efficient sampling and likelihood estimation in high dimensions, which avoids SDE trajectories and score matching, reducing memory load and improving accuracy.


Learning Adversarial Low-rank Markov Decision Processes with Unknown Transition and Full-information Feedback

Canzhe Zhao, Ruofeng Yang, Baoxiang Wang, Xuezhou Zhang, Shuai Li

https://openreview.net/forum?id=ZPtzwr2SwJ

Keywords: adversarial low-rank mdps

Compressor summary: This paper studies low-rank MDPs with changing losses and proposes POLO, an oracle-efficient algorithm with a sublinear regret guarantee, which is the first of its kind for this problem setting.


PyNeRF: Pyramidal Neural Radiance Fields

Haithem Turki, Michael Zollhöfer, Christian Richardt, Deva Ramanan

https://openreview.net/forum?id=ZPj7ey5fXa

Keywords: view synthesis, 3d reconstruction, scene representation, 3d deep learning

Compressor summary: The paper introduces a simple modification to grid-based Neural Radiance Fields that improves rendering quality and reduces error rates by using different spatial grid resolutions at render time.


Encoding Human Behavior in Information Design through Deep Learning

Guanghui Yu, Wei Tang, Saumik Narayanan, Chien-Ju Ho

https://openreview.net/forum?id=ZOKhtz2Z9X

Keywords: Information design; Human behavior; Behavioral experiments

Compressor summary: The authors study how to design optimal information policies using a neural network (HAIDNet) that adapts to different human behavior patterns, and show its effectiveness in simulations and real experiments.


Depth-discriminative Metric Learning for Monocular 3D Object Detection

Wonhyeok Choi, Mingyu Shin, Sunghoon Im

https://openreview.net/forum?id=ZNBblMEP16

Keywords: Monocular 3D object detection, Autonomous driving, Recognition, Regression, Metric learning

Compressor summary: The text describes a new method for monocular 3D object detection that uses metric learning and an auxiliary head to improve depth estimation without increasing model size or inference time.


PanoGRF: Generalizable Spherical Radiance Fields for Wide-baseline Panoramas

Zheng Chen, Yan-Pei Cao, Yuan-Chen Guo, Chen Wang, Ying Shan, Song-Hai Zhang

https://openreview.net/forum?id=ZKVxABGJ6r

Keywords: neural rendering, neural radiance field, novel view synthesis, panorama, 360-degree image

Compressor summary: PanoGRF is a method to generate realistic views of virtual environments using wide-baseline panoramas by incorporating 360-degree scene priors and depth information.


ResShift: Efficient Diffusion Model for Image Super-resolution by Residual Shifting

Zongsheng Yue, Jianyi Wang, Chen Change Loy

https://openreview.net/forum?id=ZIyAHaLlsn

Keywords: Super-resolution; Diffusion model; Efficient

Compressor summary: A novel diffusion model for image super-resolution reduces the number of steps needed, improves efficiency, and maintains or exceeds performance compared to existing methods.


Sparse Parameterization for Epitomic Dataset Distillation

Xing Wei, Anjia Cao, Funing Yang, Zhiheng Ma

https://openreview.net/forum?id=ZIfhYAE2xg

Keywords: Dataset Distillation, Dataset Condensation, Sparse Coding, Dictionary Learning

Compressor summary: The SPEED framework uses dictionary learning, sparse coding, and feature recurrence to create synthetic datasets that capture the essential information of large and diverse original datasets.


A Graph-Theoretic Framework for Understanding Open-World Semi-Supervised Learning

Yiyou Sun, Zhenmei Shi, Yixuan Li

https://openreview.net/forum?id=ZITOHWeAy7

Keywords: open-world learning, clustering, spectral analysis

Compressor summary: The paper presents a graph-theoretic framework for open-world semi-supervised learning and proposes a new algorithm, SORL, that provides theoretical guarantees and empirical performance on clustering known and novel classes.


On kernel-based statistical learning theory in the mean field limit

Christian Fiedler, Michael Herty, Sebastian Trimpe

https://openreview.net/forum?id=ZGElmTRk3w

Keywords: Reproducing Kernel Hilbert Spaces, Kernel Methods, Mean Field Limit, Interacting Particle Systems, Support Vector Machines, Statistical Learning Theory

Compressor summary: The paper studies how machine learning kernels and Support Vector Machines behave when the number of input variables goes to infinity, providing new theoretical tools and insights for large-scale problems.


Learning to Reason and Memorize with Self-Notes

Jack Lanchantin, Shubham Toshniwal, Jason E Weston, Arthur Szlam, Sainbayar Sukhbaatar

https://openreview.net/forum?id=ZFwNdsDCRL

Keywords: Memory, Reasoning, Language Models

Compressor summary: The authors propose Self-Notes, a method for improving large language models' multi-step reasoning and memory retention by allowing them to deviate from the input context and write down their thoughts.


Human-in-the-Loop Optimization for Deep Stimulus Encoding in Visual Prostheses

Jacob Granley, Tristan Fauvel, Matthew Chalk, Michael Beyeler

https://openreview.net/forum?id=ZED5wdGous

Keywords: Brain Computer Interfaces, BCI, Stimulus Encoding, Visual Prostheses, Bayesian Optimization, Preferential Bayesian Optimization, Human-in-the-loop Optimization, Sensory Neuroprostheses, Neuroprostheses, Patient-Specific Optimization, Latent Space Bayesian Optimization

Compressor summary: The paragraph discusses a new approach to optimize stimuli for neuroprosthetic devices that improves the quality of restored vision by combining deep learning and Bayesian optimization.


On Learning Latent Models with Multi-Instance Weak Supervision

Kaifu Wang, Efthymia Tsamoura, Dan Roth

https://openreview.net/forum?id=ZD65F3x1jU

Keywords: weak supervision, partial label learning, neuro-symbolic learning, latent structural learning

Compressor summary: The paper introduces a new problem called multi-instance Partial Label Learning and provides the first theoretical analysis of its learnability, error bounds, and experiments.


AdANNS: A Framework for Adaptive Semantic Search

Aniket Rege, Aditya Kusupati, Sharan Ranjit S, Alan Fan, Qingqing Cao, Sham M. Kakade, Prateek Jain, Ali Farhadi

https://openreview.net/forum?id=ZBzYWP2Gpl

Keywords: Semantic Search, Approximate Nearest Neighbor Search, Large-scale search, Representation Learning

Compressor summary: The paper proposes AdANNS, a framework that uses adaptive representations with varying capacities to improve accuracy-compute trade-offs in web-scale search systems based on Matryoshka Representations.


Stability Guarantees for Feature Attributions with Multiplicative Smoothing

Anton Xue, Rajeev Alur, Eric Wong

https://openreview.net/forum?id=ZBxycYCuEL

Keywords: Feature Attribution, Smoothing, Explainable, Interpretable, Provable Guarantees

Compressor summary: The paper proposes a smoothing method called Multiplicative Smoothing (MuS) that improves the stability of feature attribution methods for machine learning models by making the model sufficiently Lipschitz.


Aiming towards the minimizers: fast convergence of SGD for overparametrized problems

Chaoyue Liu, Dmitriy Drusvyatskiy, Misha Belkin, Damek Davis, Yian Ma

https://openreview.net/forum?id=ZBB8EFO7ma

Keywords: Polyak-Lojasiewicz condition, SGD, interpolation, fast convergence

Compressor summary: This paper introduces a regularity condition for interpolation regime in deep learning, allowing stochastic gradient method to have the same worst-case complexity as deterministic gradient method using only one sampled gradient per iteration.


Generalized Bayesian Inference for Scientific Simulators via Amortized Cost Estimation

Richard Gao, Michael Deistler, Jakob H. Macke

https://openreview.net/forum?id=ZARAiV25CW

Keywords: simulation-based inference, generalized bayesian inference, neural network, machine learning for science

Compressor summary: ACE is a new method that uses neural networks to approximate the cost function for generalized Bayesian inference, enabling efficient and robust parameter estimation for complex simulators, especially when the model is misspecified.


Characterizing the Impacts of Semi-supervised Learning for Weak Supervision

Jeffrey Li, Jieyu Zhang, Ludwig Schmidt, Alexander Ratner

https://openreview.net/forum?id=Z8TjsPFBSx

Keywords: Weak Supervision, Semi-supervised Learning, Learning From Limited Labels

Compressor summary: The paper proposes a simple design space to study how semi-supervised learning techniques can improve programmatic weak supervision, finding that simpler methods often perform as well as complex ones, except in certain cases where smaller end models or fewer labeled examples are beneficial.


NEO-KD: Knowledge-Distillation-Based Adversarial Training for Robust Multi-Exit Neural Networks

Seokil Ham, Jungwuk Park, Dong-Jun Han, Jaekyun Moon

https://openreview.net/forum?id=Z7Cz9un2Fy

Keywords: Multi-exit Neural Network, Adversarial Training, Knowledge Distillation, Adversarial Transferability

Compressor summary: NEO-KD is a new adversarial training strategy for multi-exit neural networks that improves robustness against attacks by using neighbor and exit-wise orthogonal knowledge distillation.


Puzzlefusion: Unleashing the Power of Diffusion Models for Spatial Puzzle Solving

Sepidehsadat Hosseini, Mohammad Amin Shabani, Saghar Irandoust, Yasutaka Furukawa

https://openreview.net/forum?id=Z764QxwETf

Keywords: Diffusion, Jigsaw, puzzle solving

Compressor summary: The paper proposes a neural architecture using Diffusion Models for solving jigsaw puzzles of room layouts, and introduces new datasets for training an end-to-end system that outperforms other methods.


Topology-Aware Uncertainty for Image Segmentation

Saumya Gupta, Yikai Zhang, Xiaoling Hu, Prateek Prasanna, Chao Chen

https://openreview.net/forum?id=Z6eexoCy7W

Keywords: Topological Representation, Discrete Morse Theory, Structural Uncertainty, Image Segmentation

Compressor summary: This paper proposes a method to estimate uncertainty in the units of topological structures for segmentation tasks, using tools from topological data analysis and a joint prediction model.


Tree-Rings Watermarks: Invisible Fingerprints for Diffusion Images

Yuxin Wen, John Kirchenbauer, Jonas Geiping, Tom Goldstein

https://openreview.net/forum?id=Z57JrmubNl

Keywords: Diffusion Model, Watermark, Privacy and Security

Compressor summary: Tree-Ring Watermarking is a novel technique for fingerprinting diffusion models' outputs by embedding a pattern in the initial noise vector, making it invisible to humans and resistant to common image manipulations.


Meta-Learning with Neural Bandit Scheduler

Yunzhe Qi, Yikun Ban, Tianxin Wei, Jiaru Zou, Huaxiu Yao, Jingrui He

https://openreview.net/forum?id=Z2L7F0nekb

Keywords: Meta Learning, Contextual Bandits

Compressor summary: The paper introduces BASS, a novel task scheduling framework for meta-learning that optimizes the task schedule based on the meta-model's status and balances exploitation and exploration.


Optimal Extragradient-Based Algorithms for Stochastic Variational Inequalities with Separable Structure

Angela Yuan, Chris Junchi Li, Gauthier Gidel, Michael Jordan, Quanquan Gu, Simon Shaolei Du

https://openreview.net/forum?id=Z28nPtAVxx

Keywords: Stochastic variational inequalities, convex-concave separable saddle-point optimization, extragradient-based algorithm, Nesterov's acceleration, scheduled restarting, scaling reduction

Compressor summary: The paper proposes a new algorithm, AG-EG, for solving stochastic monotone variational inequalities with a separable structure and proves its optimal convergence rate.


Revisiting Logistic-softmax Likelihood in Bayesian Meta-Learning for Few-Shot Classification

Tianjun Ke, Haoqun Cao, Zenan Ling, Feng Zhou

https://openreview.net/forum?id=Z1W0u3Cr74

Keywords: Few-shot learning, Gaussian processes, Conditional conjugate

Compressor summary: The paper proposes a redesigned logistic-softmax likelihood for Bayesian meta-learning that controls the confidence level and improves uncertainty estimates, leading to better results on few-shot classification tasks.


Path Regularization: A Convexity and Sparsity Inducing Regularization for Parallel ReLU Networks

Tolga Ergen, Mert Pilanci

https://openreview.net/forum?id=Z1Aj59LoZD

Keywords: Convex optimization, deep learning theory, path norm, group sparsity, polynomial-time training, ReLU networks, parallel architectures, global optimality, computational complexity

Compressor summary: The authors study deep neural networks and introduce an analytic method to reveal hidden convexity in their optimization landscape, leading to a parsimonious convex model that can be trained efficiently.


A Unified Framework for Rank-based Loss Minimization

Rufeng Xiao, Yuze Ge, Rujun Jiang, Yifan Yan

https://openreview.net/forum?id=Z16jo3d6OD

Keywords: rank-based loss, ADMM, nonconvex nonsmooth optimization, conditional Value-at-Risk, human-aligned risk, ranked range loss

Compressor summary: The paper presents a new optimization algorithm for rank-based loss, which is used in various machine learning models, and shows its effectiveness and efficiency through experiments.


Compositional Policy Learning in Stochastic Control Systems with Formal Guarantees

Đorđe Žikelić, Mathias Lechner, Abhinav Verma, Krishnendu Chatterjee, Thomas A Henzinger

https://openreview.net/forum?id=Yx8Sw2H5Q7

Keywords: Verification, Compositional learning

Compressor summary: The authors propose a method to learn neural network policies for stochastic environments with formal guarantees, using SpectRL's logical specifications and reach-avoid supermartingales.


Text Promptable Surgical Instrument Segmentation with Vision-Language Models

Zijian Zhou, Oluwatosin Alabi, Meng Wei, Tom Vercauteren, Miaojing Shi

https://openreview.net/forum?id=YwgA3avHrP

Keywords: Surgical Instrument Segmentation, Vision Language Models, Text Promptable Segmentation

Compressor summary: The paper presents a text promptable surgical instrument segmentation approach that uses vision-language models and multiple text prompts to improve accuracy and adaptability in minimally invasive surgeries.


Integration-free Training for Spatio-temporal Multimodal Covariate Deep Kernel Point Processes

YIXUAN ZHANG, Quyu Kong, Feng Zhou

https://openreview.net/forum?id=Yvpenkym8A

Keywords: Spatio-temporal Point Processes, Deep Kernel, Covariate, Integration-free

Compressor summary: The study introduces a new model called Deep Kernel Mixture Point Processes (DKMPP) that uses a deep kernel to better capture relationships between events and multimodal covariate data, and improves efficiency by using score matching methods.


Online Map Vectorization for Autonomous Driving: A Rasterization Perspective

Gongjie Zhang, Jiahao Lin, Shuang Wu, Yilin Song, Zhipeng Luo, Yang Xue, Shijian Lu, Zuoguan Wang

https://openreview.net/forum?id=YvO5yTVv5Y

Keywords: Online HD Map Construction, Map Vectorization, Autonomous Driving, Evaluation Metric, Rasterization, Differentiable Rasterization, Bird's-Eye-View Perception

Compressor summary: The paper introduces MapVR, a novel framework that applies rasterization to vectorized maps for more accurate perception and safer autonomous driving.


Diversifying Spatial-Temporal Perception for Video Domain Generalization

Kun-Yu Lin, Jia-Run Du, Yipeng Gao, Jiaming Zhou, Wei-Shi Zheng

https://openreview.net/forum?id=YsZTDcIQwQ

Keywords: video understanding and analysis, video domain generalization

Compressor summary: The paper introduces Spatial-Temporal Diversification Network (STDN), a model that improves video domain generalization by discovering diverse spatial-temporal cues in videos, to defend against heavy reliance on domain-specific cues.


Deep Fractional Fourier Transform

Hu Yu, Jie Huang, Lingzhi Li, Man Zhou, Feng Zhao

https://openreview.net/forum?id=YsYKv95jy9

Keywords: Fractional Fourier Transform, image restoration

Compressor summary: The paper introduces a new spatial-frequency analysis tool, FRFT, which provides comprehensive unified perspectives for image processing and proposes a simple yet effective operator, MFRFC, that significantly improves performance on various computer vision tasks.


Federated Learning with Client Subsampling, Data Heterogeneity, and Unbounded Smoothness: A New Algorithm and Lower Bounds

Michael Crawshaw, Yajie Bao, Mingrui Liu

https://openreview.net/forum?id=Yq6GKgN3RC

Keywords: federated learning, client subsampling, nonconvex optimization, relaxed smoothness, data heterogeneity, lower bound

Compressor summary: EPISODE++ is a new algorithm for Federated Learning with client subsampling and data heterogeneity that achieves linear speedup, reduces communication rounds, and works well on RNNs for text classification.


Pareto Frontiers in Deep Feature Learning: Data, Compute, Width, and Luck

Benjamin L. Edelman, Surbhi Goel, Sham M. Kakade, eran malach, Cyril Zhang

https://openreview.net/forum?id=Ypbke6biDm

Keywords: deep learning, feature learning, parity, grokking, lottery tickets, scaling

Compressor summary: This work studies how algorithm choices affect resource tradeoffs in deep learning for sparse parity learning, and shows that width and sparsity improve sample efficiency.


Diffusion-SS3D: Diffusion Model for Semi-supervised 3D Object Detection

Cheng-Ju Ho, Chen-Hsuan Tai, Yen-Yu Lin, Ming-Hsuan Yang, Yi-Hsuan Tsai

https://openreview.net/forum?id=YoghyvSG0H

Keywords: Semi-supervised learning, 3D object detection, diffusion model

Compressor summary: The paper proposes Diffusion-SS3D, a method that improves semi-supervised 3D object detection by using a diffusion model to denoise corrupted labels and enhance pseudo-label generation in a teacher-student framework.


GraphAdapter: Tuning Vision-Language Models With Dual Knowledge Graph

Xin Li, Dongze Lian, Zhihe Lu, Jiawang Bai, Zhibo Chen, Xinchao Wang

https://openreview.net/forum?id=YmEDnMynuO

Keywords: Efficient transfer learning, vision-language model, adapter-style tuning

Compressor summary: GraphAdapter is an efficient tuning strategy for vision-language models that leverages dual-modality knowledge graphs to capture task-specific structure and inter-class relationships, improving performance on downstream tasks.


STEVE-1: A Generative Model for Text-to-Behavior in Minecraft

Shalev Lifshitz, Keiran Paster, Harris Chan, Jimmy Ba, Sheila A. McIlraith

https://openreview.net/forum?id=YkBDJWerKg

Keywords: minecraft, instruction following, foundation models, sequence models, reinforcement learning, sequential decision making, goal conditioned reinforcement learning, text conditioned reinforcement learning, transformers, deep learning

Compressor summary: The text introduces STEVE-1, a cost-effective AI model that follows instructions in Minecraft by adapting pretrained models and using self-supervised learning techniques.


Boosting Learning for LDPC Codes to Improve the Error-Floor Performance

Hee-Youl Kwak, Dae-Young Yun, Yongjune Kim, Sang-Hyo Kim, Jong-Seon No

https://openreview.net/forum?id=Yj3lFEyfnl

Keywords: Error-floor, Low-density parity-check codes, Boosting learning, Training shcedule, weight sharing, Neural decoders, Min-sum

Compressor summary: The authors propose new training methods for neural min-sum decoders to eliminate the error-floor effect in low-density parity-check codes, achieving better error-floor performance than other methods without extra hardware costs.


Evaluating Neuron Interpretation Methods of NLP Models

Yimin Fan, Fahim Dalvi, Nadir Durrani, Hassan Sajjad

https://openreview.net/forum?id=YiwMpyMdPX

Keywords: Neuron interpretation, NLP, Interpretability, Machine Learning

Compressor summary: The paper proposes a voting-based framework to compare different neuron interpretation methods and finds that they agree on important neurons and focus on last layer representations.


Near-Optimal Bounds for Learning Gaussian Halfspaces with Random Classification Noise

Ilias Diakonikolas, Jelena Diakonikolas, Daniel Kane, Puqian Wang, Nikos Zarifis

https://openreview.net/forum?id=YifKp5b15e

Keywords: PAC Learning, Random Classification Noise

Compressor summary: The paper studies learning general halfspaces with random classification noise and shows an information-computation gap between algorithmic and statistical query complexity results.


Anytime Model Selection in Linear Bandits

Parnian Kassraie, Nicolas Emmenegger, Andreas Krause, Aldo Pacchiano

https://openreview.net/forum?id=YiRX7nQ77Q

Keywords: bandits, model selection, online learning

Compressor summary: ALEXP is a model selection algorithm for bandit optimization that uses online learning with emulated full-information feedback to achieve exponentially improved regret compared to existing methods.


Autodecoding Latent 3D Diffusion Models

Evangelos Ntavelis, Aliaksandr Siarohin, Kyle Olszewski, Chaoyang Wang, Luc Van Gool, Sergey Tulyakov

https://openreview.net/forum?id=YhAZqWhOnS

Keywords: 3D Generation, Diffusion Models

Compressor summary: The proposed 3D autodecoder approach generates high-quality 3D assets from 2D images or monocular videos using a latent space that embeds properties learned from the target data, achieving superior results compared to existing methods on various benchmarks.


Bandit Social Learning under Myopic Behavior

Kiarash Banihashem, MohammadTaghi Hajiaghayi, Suho Shin, Aleksandrs Slivkins

https://openreview.net/forum?id=YeP8osxOht

Keywords: multi-armed bandits, greedy algorithm, social learning, myopic behavior, learning failures, algorithmic game theory

Compressor summary: The paper examines how social learning agents collectively follow a multi-armed bandit protocol but fail to explore due to myopic behaviors.


Your representations are in the network: composable and parallel adaptation for large scale models

Yonatan Dukler, Alessandro Achille, Hao Yang, Varsha Vivek, Luca Zancato, Benjamin Bowman, Avinash Ravichandran, Charless Fowlkes, Ashwin Swaminathan, Stefano Soatto

https://openreview.net/forum?id=Ydxnan4P2G

Keywords: Efficient learning, Compute-efficient deep learning, Deep Learning Theory, class-incremental-learning, downstream adaptation

Compressor summary: InCA is a transfer learning framework that uses lightweight cross-attention modules to adapt large models efficiently and identify useful representations for downstream tasks.


Learning Trajectories are Generalization Indicators

Jingwen Fu, Zhizheng Zhang, Dacheng Yin, Yan Lu, Nanning Zheng

https://openreview.net/forum?id=YdfcKb4Wif

Keywords: Generalization, Learning Trajectory

Compressor summary: The paper examines how the learning path of DNNs affects their ability to generalize, proposing a new bound that considers trajectory complexity and training set diversity.


Real-Time Motion Prediction via Heterogeneous Polyline Transformer with Relative Pose Encoding

Zhejun Zhang, Alexander Liniger, Christos Sakaridis, Fisher Yu, Luc Van Gool

https://openreview.net/forum?id=YcmGuwdLoU

Keywords: Motion Prediction, Autonomous Driving, Transformer

Compressor summary: The paper introduces a novel attention mechanism (KNARPE) and a hierarchical framework (HPTR) for motion prediction in autonomous driving systems, improving efficiency and scalability while achieving superior performance on public benchmarks.


Fast and Regret Optimal Best Arm Identification: Fundamental Limits and Low-Complexity Algorithms

Qining Zhang, Lei Ying

https://openreview.net/forum?id=Yc9bqbnrbs

Keywords: stochastic multi-armed bandits, regret optimal best arm identification, commitment

Compressor summary: The paper proposes ROBAI, a method for solving a stochastic MAB problem with dual objectives of quick optimal arm identification and reward maximization in a limited number of rounds.


BiMatting: Efficient Video Matting via Binarization

Haotong Qin, Lei Ke, Xudong Ma, Martin Danelljan, Yu-Wing Tai, Chi-Keung Tang, Xianglong Liu, Fisher Yu

https://openreview.net/forum?id=YbYQ0JEQ80

Keywords: Video Matting, Model Binarization, Deep Learning

Compressor summary: BiMatting is a binarized video matting model that overcomes representation degradation and redundant computations, achieving high accuracy, efficiency, and resource savings for real-time video applications.


Toolformer: Language Models Can Teach Themselves to Use Tools

Timo Schick, Jane Dwivedi-Yu, Roberto Dessi, Roberta Raileanu, Maria Lomeli, Eric Hambro, Luke Zettlemoyer, Nicola Cancedda, Thomas Scialom

https://openreview.net/forum?id=Yacmpz84TH

Keywords: Language Models, Zero-Shot Learning, Tool Use, APIs

Compressor summary: The paper introduces Toolformer, a language model that learns to use external APIs to improve its performance on various tasks.


Contrast, Attend and Diffuse to Decode High-Resolution Images from Brain Activities

Jingyuan Sun, Mingxiao Li, Yunhao Zhang, Marie-Francine Moens, Zijiao Chen, Shaonan Wang

https://openreview.net/forum?id=YZSLDEE0mw

Keywords: Neural decoding, brain machine interface, mind reader, visual reconstruction, vision decoding

Compressor summary: The paper introduces a two-phase fMRI representation learning framework that uses denoising and image reconstruction to improve visual decoding from neural responses, achieving superior results compared to previous methods.


GAN You See Me? Enhanced Data Reconstruction Attacks against Split Inference

Ziang Li, Mengda Yang, Yaxin Liu, Juan Wang, Hongxin Hu, Wenzhe Yi, Xiaoyang Xu

https://openreview.net/forum?id=YZGWhs1H7F

Keywords: deep learning, split inference, data reconstruction attack

Compressor summary: GLASS is a novel GAN-based attack that effectively reconstructs private data from Split Inference models, while GLASS++ enhances its stability and both are resistant to common defense mechanisms.


Structured Prediction with Stronger Consistency Guarantees

Anqi Mao, Mehryar Mohri, Yutao Zhong

https://openreview.net/forum?id=YZ7ip645Ra

Keywords: structured prediction, consistency, learning theory, natural language processing

Compressor summary: The authors study surrogate losses for structured prediction and propose new ones with better learning guarantees and efficient algorithms.


Fixing the NTK: From Neural Network Linearizations to Exact Convex Programs

Rajat Vadiraj Dwaraknath, Tolga Ergen, Mert Pilanci

https://openreview.net/forum?id=YWsPN0EMZr

Keywords: neural tangent kernel, NTK, ReLU activations, neural networks, gated ReLU, convex optimization, kernel, multiple kernel learning, MKL, group lasso, iterative reweighting, group norm

Compressor summary: This paper connects neural network training with gradient flow to multiple kernel learning using a gated ReLU network and shows how to improve the weights for better predictions.


Predicting a Protein's Stability under a Million Mutations

Jeffrey Ouyang-Zhang, Daniel Jesus Diaz, Adam Klivans, Philipp Kraehenbuehl

https://openreview.net/forum?id=YWSOpYjyG4

Keywords: stability, proteins, biology, physical

Compressor summary: The authors propose a deep learning algorithm called Mutate Everything that predicts the effect of mutations on protein stability quickly and accurately, using existing models as its base.


Offline Reinforcement Learning with Differential Privacy

Dan Qiao, Yu-Xiang Wang

https://openreview.net/forum?id=YVMc3KiWBQ

Keywords: Differential privacy, offline reinforcement learning, reinforcement learning theory

Compressor summary: The paper proposes offline reinforcement learning algorithms with differential privacy guarantees that protect sensitive information while maintaining good learning performance and utility.


MoVie: Visual Model-Based Policy Adaptation for View Generalization

Sizhe Yang, Yanjie Ze, Huazhe Xu

https://openreview.net/forum?id=YV1MYtj2AR

Keywords: visual reinforcement learning, visual generalization

Compressor summary: The paper proposes MoVie, a method to improve visual reinforcement learning agents' ability to generalize to unseen views without explicit rewards, achieving significant improvements in 18 real-world robotics tasks.


CELLE-2: Translating Proteins to Pictures and Back with a Bidirectional Text-to-Image Transformer

Emaad Khwaja, Yun S. Song, Aaron Agarunov, Bo Huang

https://openreview.net/forum?id=YSMLVffl5u

Keywords: text-to-image, protein localization, protein engineering, transformers

Compressor summary: CELL-E 2 is a transformer that can generate and design protein subcellular localization images and sequences from amino acid sequences, capturing spatial complexity and producing NLS.


Implicit Regularization in Over-Parameterized Support Vector Machine

Yang Sui, Xin HE, Yang Bai

https://openreview.net/forum?id=YSFQRVkkl0

Keywords: Over-parameterization, SVM, Sparsity, Lasso

Compressor summary: The paper proposes a regularization-free algorithm for high-dimensional SVMs using over-parameterization and Nesterov's method, which improves computational efficiency and achieves near-oracle convergence rate.


3D-LLM: Injecting the 3D World into Large Language Models

Yining Hong, Haoyu Zhen, Peihao Chen, Shuhong Zheng, Yilun Du, Zhenfang Chen, Chuang Gan

https://openreview.net/forum?id=YQA28p7qNz

Keywords: 3D Visual Reasoning, 3D Large Language Model

Compressor summary: The paper introduces 3D-LLMs, a new type of language models that can perform various 3D tasks using point clouds and features. The paper collects large amounts of 3D-language data and uses a novel training method to achieve superior performance on several benchmarks.


Harnessing Hard Mixed Samples with Decoupled Regularizer

Zicheng Liu, Siyuan Li, Ge Wang, Lirong Wu, Cheng Tan, Stan Z. Li

https://openreview.net/forum?id=YPQg2RTFD8

Keywords: mixup, data augmentation, classification, data efficiency

Compressor summary: Decoupled mixup is an efficient data augmentation method that smooths decision boundaries and mines discriminative features without extra computation, outperforming dynamic mixup methods.


Spatio-Angular Convolutions for Super-resolution in Diffusion MRI

Matthew Lyon, Paul Armitage, Mauricio A Álvarez

https://openreview.net/forum?id=YPHIrNKI0d

Keywords: Diffusion MRI, super-resolution, image synthesis, conditional image synthesis, continuous convolution, parametric continuous convolution

Compressor summary: The paragraph introduces a novel method for improving diffusion MRI resolution using a parametric continuous convolution network (PCCNN) that performs well and requires fewer parameters than existing models.


On Measuring Fairness in Generative Models

Christopher T.H Teo, Milad Abdollahzadeh, Ngai-man Cheung

https://openreview.net/forum?id=YOZaej0ZC7

Keywords: Fairness, Generative models, GAN, Calibration

Compressor summary: This paper presents CLEAM, a new framework to measure fairness in generative models that reduces errors and has minimal overhead.


Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness

Suraj Srinivas, Sebastian Bordt, Himabindu Lakkaraju

https://openreview.net/forum?id=YMMlHBSQdC

Keywords: robustness, generative models, perceptually aligned gradients, bayes optimality, manifold hypothesis

Compressor summary: This paper explains how perceptually-aligned gradients (PAGs) in robust computer vision models arise from off-manifold robustness and identifies three regimes of robustness affecting both alignment and accuracy.


Connecting Pre-trained Language Model and Downstream Task via Properties of Representation

Chenwei Wu, Holden Lee, Rong Ge

https://openreview.net/forum?id=YLOJ4aKAka

Keywords: language model representation, downstream performance, deep learning theory

Compressor summary: The paper investigates how pre-trained language models' representations depend on downstream tasks' properties and structures, and proposes an "anchor vector" concept to improve performance transfer understanding.


ContiFormer: Continuous-Time Transformer for Irregular Time Series Modeling

Yuqi Chen, Kan Ren, Yansen Wang, Yuchen Fang, Weiwei Sun, Dongsheng Li

https://openreview.net/forum?id=YJDz4F2AZu

Keywords: Irregular Time Series Modeling, Transformer, Neural Ordinary Differential Equation

Compressor summary: ContiFormer combines Neural ODEs and Transformers to model continuous-time dynamics on irregular time series, capturing both input relationships and dynamic changes.


Conformal Prediction Sets for Ordinal Classification

PRASENJIT DEY, Srujana Merugu, Sivaramakrishnan R Kaveri

https://openreview.net/forum?id=YI4bn6aAmz

Keywords: Ordinal Classification, Conformal Predictions, Unimodal modelling

Compressor summary: The paper proposes a framework to adapt conformal prediction methods for ordinal classification, using a non-parametric approach to generate contiguous sets with guaranteed coverage and minimal cardinality.


NICE: NoIse-modulated Consistency rEgularization for Data-Efficient GANs

Yao Ni, Piotr Koniusz

https://openreview.net/forum?id=YFW6MVGVTn

Keywords: Image Generation, limited dataset, Generative Adversarial Networks

Compressor summary: NICE is a novel approach that uses adaptive noise to modulate the discriminator's latent features, preventing overfitting and improving GAN training stability on limited data.


Inverse Reinforcement Learning with the Average Reward Criterion

Feiyang Wu, Jingyang Ke, Anqi Wu

https://openreview.net/forum?id=YFSrf8aciU

Keywords: Machine Learning, Reinforcement Learning, Inverse Reinforcement Learning, Markov Decision Process, stochastic optimization, complexity analysis

Compressor summary: This paper presents a novel inverse reinforcement learning method that works under an average-reward setting and has lower complexity than previous methods, along with experiments on various tasks.


Model-Free Active Exploration in Reinforcement Learning

Alessio Russo, Alexandre Proutiere

https://openreview.net/forum?id=YEtstXIpP3

Keywords: reinforcement learning; best policy identification; model free; exploration; sample complexity

Compressor summary: The paper proposes a new model-free exploration method for Reinforcement Learning that can find near-optimal policies more quickly than existing methods.


$\textbf{A}^2\textbf{CiD}^2$: Accelerating Asynchronous Communication in Decentralized Deep Learning

Adel Nabli, Eugene Belilovsky, Edouard Oyallon

https://openreview.net/forum?id=YE04aRkeZb

Keywords: Decentralized Optimization for Deep Learning, Asynchronous Optimization, Distributed Training, Data-Parallel

Compressor summary: The paper introduces a new gossip-based optimization algorithm called $\textbf{A}^2\textbf{CiD}^2$ that speeds up distributed training of deep learning models by using continuous local momentum and reducing idle time.


Finding Counterfactually Optimal Action Sequences in Continuous State Spaces

Stratis Tsirtsis, Manuel Gomez Rodriguez

https://openreview.net/forum?id=YDCpf85eXc

Keywords: Counterfactual reasoning, Markov decision process, Structural causal model, A* search

Compressor summary: The paper presents a new method to analyze continuous state environments for retrospective analysis of treatment decisions using causal inference and reinforcement learning techniques.


A Unified Algorithm Framework for Unsupervised Discovery of Skills based on Determinantal Point Process

Jiayu Chen, Vaneet Aggarwal, Tian Lan

https://openreview.net/forum?id=Y8p3ThNDmK

Keywords: Reinforcement Learning, Unsupervised Skill Discovery, Determinantal Point Process, Options

Compressor summary: The paper proposes a new algorithm that combines diversity and coverage in unsupervised option discovery using Determinantal Point Process, and shows its effectiveness on Mujoco and Atari tasks.


Model Shapley: Equitable Model Valuation with Black-box Access

Xinyi Xu, Thanh Lam, Chuan-Sheng Foo, Bryan Kian Hsiang Low

https://openreview.net/forum?id=Y6IGTNMdLT

Keywords: Model Valuation, Dirichlet Abstraction, Shapley Value

Compressor summary: The paper proposes a new method for valuing machine learning models in AI marketplaces, called model Shapley, which is fair and can be used to predict the values of many models efficiently.


Quantum Bayesian Optimization

Zhongxiang Dai, Gregory Kang Ruey Lau, Arun Verma, Yao Shu, Bryan Kian Hsiang Low, Patrick Jaillet

https://openreview.net/forum?id=Y44NurSDjq

Keywords: quantum bandits, kernelized bandits

Compressor summary: The Q-GP-UCB algorithm optimizes non-linear reward functions with quantum computing, achieving better regret upper bounds than classical methods and previous quantum approaches.


PTQD: Accurate Post-Training Quantization for Diffusion Models

Yefei He, Luping Liu, Jing Liu, Weijia Wu, Hong Zhou, Bohan Zhuang

https://openreview.net/forum?id=Y3g1PV5R9l

Keywords: Diffusion models, Post-training quantization, Mixed precision

Compressor summary: The paragraph discusses a new post-training quantization method for diffusion models that reduces computational costs and improves sample quality by correcting quantization noise and adjusting bitwidths for each denoising step.


Detection Based Part-level Articulated Object Reconstruction from Single RGBD Image

Yuki Kawana, Tatsuya Harada

https://openreview.net/forum?id=Y3NjoeO4Q1

Keywords: articulated objects, shape reconstruction, 3D reconstruction

Compressor summary: The paper presents a novel method for reconstructing multiple man-made articulated objects from a single RGBD image using part-level representation and proposes three techniques to overcome challenges, achieving better shape reconstruction and kinematics estimation on synthetic and real data.


Beyond NTK with Vanilla Gradient Descent: A Mean-Field Analysis of Neural Networks with Polynomial Width, Samples, and Time

Arvind Venkat Mahankali, Jeff Z. HaoChen, Kefan Dong, Margalit Glasgow, Tengyu Ma

https://openreview.net/forum?id=Y2hnMZvVDm

Keywords: Nonconvex Optimization, Mean-Field Analysis, Beyond NTK, Deep Learning Theory

Compressor summary: The paper analyzes projected gradient flow on two-layer neural networks without modifications and shows they achieve better sample complexity than kernel methods.


Expert load matters: operating networks at high accuracy and low manual effort

Sara Sangalli, Ertunc Erdil, Ender Konukoglu

https://openreview.net/forum?id=Y2VQWfi7Vc

Keywords: human-ai collaboration system, optimization

Compressor summary: The paper proposes a new loss function for training deep neural networks that considers both model accuracy and expert load, aiming to improve classification and reduce human intervention in critical applications.


A Unified Approach for Maximizing Continuous DR-submodular Functions

Mohammad Pedramfar, Christopher John Quinn, Vaneet Aggarwal

https://openreview.net/forum?id=Y1sJJW3pID

Keywords: Stochastic optimization, submodular maximization, Frank-Wolfe algorithm

Compressor summary: The paper proposes a unified method to maximize continuous DR-submodular functions across various settings and oracle access types, improving or matching existing results in most cases.


Posterior Contraction Rates for Matérn Gaussian Processes on Riemannian Manifolds

Paul Rosa, Viacheslav Borovitskiy, Alexander Terenin, Judith Rousseau

https://openreview.net/forum?id=Y18r0xWkSh

Keywords: Gaussian processes, posterior contraction, manifolds, kernels

Compressor summary: The paper investigates whether intrinsic models of Gaussian processes on Riemannian manifolds lead to better performance than Euclidean ones, by proving optimal contraction rates for both types and showing empirical results that support the theoretical findings.


Towards Higher Ranks via Adversarial Weight Pruning

Yuchuan Tian, Hanting Chen, Tianyu Guo, Chao Xu, Yunhe Wang

https://openreview.net/forum?id=Y17N9B0vXn

Keywords: Weight Pruning, Matrix Rank

Compressor summary: The paper introduces RPG, a pruning method for CNNs that maintains the ranks of sparse weights in an adversarial manner to improve performance on edge devices with high sparsity.


Minimax Risks and Optimal Procedures for Estimation under Functional Local Differential Privacy

Bonwoo Lee, Jeongyoun Ahn, Cheolwoo Park

https://openreview.net/forum?id=XzTM9gVRT4

Keywords: Data privacy, Functional local differential privacy, Gaussian mechanism, Minimax risks, Statistical utility

Compressor summary: This paragraph discusses how functional differential privacy (DP), a type of local DP, balances data privacy and utility by analyzing its performance in estimating mean and density, and suggests it as a reliable standard for local DP.


Look Ma, No Hands! Agent-Environment Factorization of Egocentric Videos

Matthew Chang, Aditya Prakash, Saurabh Gupta

https://openreview.net/forum?id=Xyj46OxEhK

Keywords: Inpainting, Diffusion, Robot Learning, Egocentric Vision

Compressor summary: This paper presents a method to separate human hand and environment in egocentric videos using a diffusion model and attention, improving performance on various robotics tasks.


An Empirical Study Towards Prompt-Tuning for Graph Contrastive Pre-Training in Recommendations

Haoran Yang, Xiangyu Zhao, Yicong Li, Hongxu Chen, Guandong Xu

https://openreview.net/forum?id=XyAP8ScqLV

Keywords: graph contrastive learning, prompt tuning, recommendation system

Compressor summary: CPTPP is a novel framework for graph learning that uses personalized prompts to bridge the gap between pre-training and downstream tasks, achieving better results and modeling user preferences more effectively.


CAP: Correlation-Aware Pruning for Highly-Accurate Sparse Vision Models

Denis Kuznedelev, Eldar Kurtic, Elias Frantar, Dan Alistarh

https://openreview.net/forum?id=Xy7DoWSNZX

Keywords: neural network pruning, vision transformer, sparsity, model compression

Compressor summary: The Correlation Aware Pruner (CAP) is a new framework that significantly improves the compressibility of state-of-the-art computer vision models by handling complex weight correlations and providing efficient post-compression recovery.


ReSync: Riemannian Subgradient-based Robust Rotation Synchronization

Huikang Liu, Xiao Li, Anthony Man-Cho So

https://openreview.net/forum?id=Xxllzjt6T5

Keywords: Manifold optimization, Riemannian subgradient method, rotation synchronization

Compressor summary: ReSync is a Riemannian subgradient algorithm for synchronizing robust rotations in various engineering applications, with strong theoretical guarantees and experimentally proven effectiveness.


Non-Rigid Shape Registration via Deep Functional Maps Prior

Puhua Jiang, Mingze Sun, Ruqi Huang

https://openreview.net/forum?id=XvfEYqEbIb

Keywords: shape registration; functional maps; unsupervised learning

Compressor summary: The paper introduces a learning-based method for registering non-rigid shapes without correspondence supervision using deep functional maps and a trained orientation regressor, achieving state-of-the-art results on several benchmarks.


Self-supervised Graph Neural Networks via Low-Rank Decomposition

Liang Yang, Runjie Shi, Qiuliang Zhang, Bingxin Niu, Zhen Wang, Xiaochun Cao, Chuan Wang

https://openreview.net/forum?id=XvGQ6F3sG8

Keywords: Graph neural network, Self-supervised learning, Low-Rank recovery

Compressor summary: This paper proposes Low-Rank Decomposition-based methods for graph neural networks to address issues with propagation-based approaches and improve local and long-distance node representation.


LayoutGPT: Compositional Visual Planning and Generation with Large Language Models

Weixi Feng, Wanrong Zhu, Tsu-Jui Fu, Varun Jampani, Arjun Reddy Akula, Xuehai He, S Basu, Xin Eric Wang, William Yang Wang

https://openreview.net/forum?id=Xu8aG5Q8M3

Keywords: Large Language Models, Compositional Image Generation, 3D scene synthesis

Compressor summary: The paragraph discusses LayoutGPT, a method that uses LLMs to generate layouts from text inputs and enhances the visual planning skills of LLMs, achieving superior performance in text-to-image generation and 3D indoor scene synthesis.


Is This Loss Informative? Faster Text-to-Image Customization by Tracking Objective Dynamics

Anton Voronov, Mikhail Khoroshikh, Artem Babenko, Max Ryabinin

https://openreview.net/forum?id=Xs6Xwc0Glj

Keywords: text-to-image generation, diffusion models, early stopping

Compressor summary: The researchers propose a simple early stopping criterion that speeds up text-to-image adaptation by eight times without sacrificing quality.


A Randomized Approach to Tight Privacy Accounting

Jiachen T. Wang, Saeed Mahloujifar, Tong Wu, Ruoxi Jia, Prateek Mittal

https://openreview.net/forum?id=XrqqPDAsRE

Keywords: Differential Privacy; Privacy Accounting

Compressor summary: The paper introduces a new differential privacy approach (EVR) that verifies and releases query outputs based on an estimated privacy parameter, using Monte Carlo techniques to improve accuracy and efficiency in privacy accounting.


Locality-Aware Generalizable Implicit Neural Representation

Doyup Lee, Chiheon Kim, Minsu Cho, Wook-Shin Han

https://openreview.net/forum?id=XqcXf7ix5q

Keywords: implicit neural representations, representation learning, neural fields

Compressor summary: The proposed framework combines a transformer encoder with a locality-aware INR decoder to learn generalizable implicit neural representations that capture fine-grained details in spatial and spectral aspects, improving performance on downstream tasks like image generation.


Multi-Prompt Alignment for Multi-Source Unsupervised Domain Adaptation

Haoran Chen, Xintong Han, Zuxuan Wu, Yu-Gang Jiang

https://openreview.net/forum?id=Xq2s5yxzd2

Keywords: multi source unsupervised domain adaptation; transfer learning; computer vision

Compressor summary: The text introduces Multi-Prompt Alignment (MPA), a method for unsupervised domain adaptation that uses prompt learning and auto-encoding to align source and target domains, achieving state-of-the-art results.


Regularized Behavior Cloning for Blocking the Leakage of Past Action Information

Seokin Seo, HyeongJoo Hwang, Hongseok Yang, Kee-Eung Kim

https://openreview.net/forum?id=XpmJNP8BVA

Keywords: Imitation learning, Information leakage, Causal Confusion

Compressor summary: The paper proposes a new regularization method called Past Action Leakage Regularization (PALR) for behavior cloning in offline imitation learning, which uses conditional independence to reduce the leakage of past actions into observation histories and improves performance.


SQ Lower Bounds for Non-Gaussian Component Analysis with Weaker Assumptions

Ilias Diakonikolas, Daniel Kane, Lisheng Ren, Yuxin Sun

https://openreview.net/forum?id=Xp68yXQiRk

Keywords: Non-Gaussian Component Analysis

Compressor summary: The paper investigates whether the chi-squared condition is necessary for proving SQ lower bounds for Non-Gaussian Component Analysis and shows that it is not.


Res-Tuning: A Flexible and Efficient Tuning Paradigm via Unbinding Tuner from Backbone

Zeyinzi Jiang, Chaojie Mao, Ziyuan Huang, Ao Ma, Yiliang Lv, Yujun Shen, Deli Zhao, Jingren Zhou

https://openreview.net/forum?id=XmpthbaJql

Keywords: Parameter-efficient Transfer Learning, Memory-efficient Transfer Learning, Residual Network, Vision Transformer, Foundation Model

Compressor summary: Res-Tuning is a new tuning paradigm that separates tuners from the backbone, allowing for flexible combination of various tuning strategies and improved efficiency in foundation models.


Distributed Inference and Fine-tuning of Large Language Models Over The Internet

Alexander Borzunov, Max Ryabinin, Artem Chumachenko, Dmitry Baranchuk, Tim Dettmers, Younes Belkada, Pavel Samygin, Colin Raffel

https://openreview.net/forum?id=XmN7ZNbUAe

Keywords: volunteer computing, distributed deep learning, distributed inference, efficient inference, large language models

Compressor summary: The authors propose methods to efficiently run large language models on geodistributed devices, addressing reliability and load-balancing issues, and demonstrating their approach with Petals, a decentralized system that speeds up inference and fine-tuning of LLMs.


Exact Representation of Sparse Networks with Symmetric Nonnegative Embeddings

Sudhanshu Chanpuriya, Ryan A. Rossi, Anup Rao, Tung Mai, Nedim Lipka, Zhao Song, Cameron N Musco

https://openreview.net/forum?id=XlvsieCnAX

Keywords: graph, network, embeddings, arboricity, factorization, model, community, nonnegative

Compressor summary: The paragraph describes a new graph factorization model that captures both similar and dissimilar links between nodes and can represent low arboricity graphs with good performance on community detection and link prediction tasks.


Structure-free Graph Condensation: From Large-scale Graphs to Condensed Graph-free Data

Xin Zheng, Miao Zhang, Chunyang Chen, Quoc Viet Hung Nguyen, Xingquan Zhu, Shirui Pan

https://openreview.net/forum?id=XkcufOcgUc

Keywords: graph neural networks (GNNs), graph condensation, training trajectory meta-matching, graph neural feature score

Compressor summary: The paper introduces SFGC, a new method for reducing the size of large-scale graphs by synthesizing small-scale graph-free data that implicitly encodes topology structure information without explicit graph structures, and evaluates its effectiveness and generalization ability through two components: meta-matching and graph neural feature score.


ASIF: Coupled Data Turns Unimodal Models to Multimodal without Training

Antonio Norelli, Marco Fumero, Valentino Maiorca, Luca Moschella, Emanuele Rodolà, Francesco Locatello

https://openreview.net/forum?id=XjOj3ZmWEl

Keywords: Representation learning, Multimodal models, Analogy, Sparsity, Nonparametric, Relative representations, Language, Semiotics

Compressor summary: The paper proposes a new method to create a common space between visual and language domains without training, using single-domain encoders and fewer image-text pairs, and shows its advantages in interpretability, fast deployment, and zero-shot transfer ability.


Batch Bayesian Optimization For Replicable Experimental Design

Zhongxiang Dai, Quoc Phong Nguyen, Sebastian Shenghong Tay, Daisuke Urano, Richalynn Leong, Bryan Kian Hsiang Low, Patrick Jaillet

https://openreview.net/forum?id=Xj4LJiXvlX

Keywords: Bayesian optimization, Gaussian processes, AI4Science

Compressor summary: The BTS-RED framework proposes three algorithms to tackle trade-offs in experimental design problems with large, heteroscedastic noise by adaptively choosing replications and balancing average performance and variability.


Deep Insights into Noisy Pseudo Labeling on Graph Data

Botao WANG, Jia Li, Yang Liu, Jiashun Cheng, Yu Rong, Wenjia Wang, Fugee Tsung

https://openreview.net/forum?id=XhNlBvb4XV

Keywords: Pseudo labeling, Graph data, Error analysis, Cautious

Compressor summary: Pseudo labeling is a technique to expand labeled data for graph learning models, but it can introduce errors; this paper analyzes these errors and proposes a cautious methodology to improve performance.


On the Minimax Regret for Online Learning with Feedback Graphs

Khaled Eldowa, Emmanuel Esposito, Tommaso Cesari, Nicolò Cesa-Bianchi

https://openreview.net/forum?id=XfYpIaKDb6

Keywords: Online learning, Feedback graphs, Multiarmed bandits

Compressor summary: The paper presents an improved upper and lower bound for the regret of online learning with different types of feedback graphs, using FTRL with $q$-Tsallis entropy and new techniques for time-varying graphs.


Towards Robust and Expressive Whole-body Human Pose and Shape Estimation

Hui En Pang, Zhongang Cai, Lei Yang, Qingyi Tao, Zhonghua Wu, Tianwei Zhang, Ziwei Liu

https://openreview.net/forum?id=XfKnoW4Zef

Keywords: Whole-body, SMPLX Model, Human Pose and Shape Estimation, Human Mesh Recovery

Compressor summary: The paper proposes a new framework to improve whole-body pose and shape estimation by addressing challenges related to bounding box quality, robustness to augmentations, and pixel alignment.


Deep Recurrent Optimal Stopping

NIRANJAN DAMERA VENKATA, Chiranjib Bhattacharyya

https://openreview.net/forum?id=XetXfkYZ6i

Keywords: optimal stopping, recurrent neural networks, probabilistic graphical models, policy gradient methods

Compressor summary: The paper introduces an OSPG algorithm that uses RNNs to optimize value functions in non-Markovian optimal stopping problems without recursion, reducing the curse of non-Markovianity and the need for costly Monte Carlo simulations.


Hierarchical Integration Diffusion Model for Realistic Image Deblurring

Zheng Chen, Yulun Zhang, Ding Liu, Bin Xia, Jinjin Gu, Linghe Kong, Xin Yuan

https://openreview.net/forum?id=XeMryhpniy

Keywords: image deblurring, diffusion model

Compressor summary: The Hierarchical Integration Diffusion Model (HI-Diff) is a new image deblurring method that uses a compact latent space and multiple scales to improve efficiency, accuracy, and generalization compared to previous diffusion models.


Bayesian Learning via Q-Exponential Process

Shuyi Li, Michael O'Connor, Shiwei Lan

https://openreview.net/forum?id=XddoUFpjkP

Keywords: Functional Regularization, Besov Process, $Q$-Exponential Distribution, Elliptic Contour Distribution

Compressor summary: The paper introduces a new stochastic process, Q-exponential (Q-EP), for high-dimensional optimization and modeling that generalizes the $q$-exponential distribution and improves upon Gaussian process in terms of penalty strength and flexibility.


On Calibrating Diffusion Probabilistic Models

Tianyu Pang, Cheng Lu, Chao Du, Min Lin, Shuicheng YAN, Zhijie Deng

https://openreview.net/forum?id=XcQzXeF7fX

Keywords: Diffusion Probabilistic Models, Model Calibration

Compressor summary: The paper proposes a calibration method for diffusion probabilistic models that reduces the score matching loss and increases the model likelihood, with experiments on various datasets to validate the approach.


Holistic Transfer: Towards Non-Disruptive Fine-Tuning with Partial Target Data

Cheng-Hao Tu, Hong-You Chen, Zheda Mai, Jike Zhong, Vardaan Pahuja, Tanya Berger-Wolf, Song Gao, Charles Stewart, Yu Su, Wei-Lun Chao

https://openreview.net/forum?id=XbVnNXaIQY

Keywords: Fine-tuning, Transfer learning, Domain adaptation, Continual learning, Robustness, Personalization

Compressor summary: The paragraph discusses a learning problem involving adapting a model to a new domain using limited data and presents challenges and solutions to improve classification accuracy for all classes.


DiViNeT: 3D Reconstruction from Disparate Views using Neural Template Regularization

Aditya Vora, Akshay Gadi Patil, Hao Zhang

https://openreview.net/forum?id=XbInLmYLDr

Keywords: Multi-view Neural 3D Reconstruction, Sparse and Disparate Views, Neural Rendering, Volume Rendering

Compressor summary: DiViNet is a method that uses neural templates to reconstruct 3D surface details from few RGB images by regularizing the ill-posed problem.


Making Scalable Meta Learning Practical

Sang Keun Choe, Sanket Vaibhav Mehta, Hwijeen Ahn, Willie Neiswanger, Pengtao Xie, Emma Strubell, Eric Xing

https://openreview.net/forum?id=Xazhn0JoNx

Keywords: meta learning, bilevel optimization, large-scale learning, implicit differentiation

Compressor summary: SAMA is a meta learning framework that improves scalability, reduces computational burden, and supports adaptive optimizers while achieving state-of-the-art results in text and image classification tasks.


Provable Training for Graph Contrastive Learning

Yue Yu, Xiao Wang, Mengmei Zhang, Nian Liu, Chuan Shi

https://openreview.net/forum?id=Xasl21tSOf

Keywords: Graph Contrastive Learning, Graph Neural Networks, Bound Propagation

Compressor summary: The paper proposes a metric called "node compactness" and a regularization method named PrOvable Training (POT) to improve Graph Contrastive Learning by addressing its imbalanced training across nodes.


Binary Radiance Fields

Seungjoo Shin, Jaesik Park

https://openreview.net/forum?id=XY6BnwIh4q

Keywords: neural radiance fields, inverse rendering, binarization

Compressor summary: The paper introduces BiRF, a compact way to store radiance fields using binary encoding and 2D-3D grids, which outperforms existing efficient models in static scene reconstruction with minimal storage space.


Learning-to-Rank Meets Language: Boosting Language-Driven Ordering Alignment for Ordinal Classification

Rui Wang, Pei Pei Li, Huaibo Huang, Chunshui Cao, Ran He, Zhaofeng He

https://openreview.net/forum?id=XXagS1RQH0

Keywords: Ordinal Classification, Representation Learning, Vision-Language, Prompt Learning

Compressor summary: The paper introduces L2RCLIP, a method that uses language-driven ordering alignment and vision-language models to improve ordinal classification tasks such as age estimation, HCI classification, and aesthetic assessment.


Have it your way: Individualized Privacy Assignment for DP-SGD

Franziska Boenisch, Christopher Mühl, Adam Dziedzic, Roy Rinberg, Nicolas Papernot

https://openreview.net/forum?id=XXPzBhOs4f

Keywords: privacy, machine learning, differential privacy, DP-SGD, individualized privacy

Compressor summary: The paper proposes Individualized DP-SGD (IDP-SGD), a variant of Differentially Private Stochastic Gradient Descent that allows for tailored privacy budgets based on individual user preferences, improving privacy-utility trade-offs.


On the Size and Approximation Error of Distilled Datasets

Alaa Maalouf, Murad Tukan, Noel Loo, Ramin Hasani, Mathias Lechner, Daniela Rus

https://openreview.net/forum?id=XWYv4BNShP

Keywords: Dataset Distillation, Size and Approximation Error

Compressor summary: The paper analyzes the theoretical limitations and guarantees of dataset distillation using kernel ridge regression methods and provides the first proof of existence of small distilled datasets with corresponding excess risk for shift-invariant kernels.


Implicit Contrastive Representation Learning with Guided Stop-gradient

Byeongchan Lee, Sehyun Lee

https://openreview.net/forum?id=XUu2GloTXb

Keywords: representation learning, self-supervised learning, contrastive learning

Compressor summary: The paper introduces a novel method for self-supervised representation learning using an asymmetric network architecture, which stabilizes training, boosts performance, and works well with small batch sizes and no predictor.


A Bayesian Approach To Analysing Training Data Attribution In Deep Learning

Elisa Nguyen, Minjoon Seo, Seong Joon Oh

https://openreview.net/forum?id=XSCYxDp3yE

Keywords: training data attribution, interpretability, explainability, data-driven xai

Compressor summary: This paper introduces a Bayesian perspective on training data attribution techniques for deep models, highlighting the challenges of noise and suggesting to only use TDA for predictions consistently influenced by specific training data.


Aleatoric and Epistemic Discrimination: Fundamental Limits of Fairness Interventions

Hao Wang, Luxi He, Rui Gao, Flavio Calmon

https://openreview.net/forum?id=XRy4YQYLe0

Keywords: information theory, fair machine learning

Compressor summary: The text discusses two types of discrimination in ML models and provides methods to measure and reduce them, especially when dealing with missing data.


Block-State Transformers

Jonathan Pilault, Mahan Fathi, Orhan Firat, Christopher Pal, Pierre-Luc Bacon, Ross Goroshin

https://openreview.net/forum?id=XRTxIBs2eu

Keywords: State Space Models, Efficient Transformers, Long Range Language Modeling, Language Modeling

Compressor summary: The Block-State Transformer combines an SSM sublayer for long-range context and a Block Transformer sublayer for short-term sequence representation, improving language modeling perplexity and speed on parallelizable architectures.


Mirror Diffusion Models for Constrained and Watermarked Generation

Guan-Horng Liu, Tianrong Chen, Evangelos Theodorou, Molei Tao

https://openreview.net/forum?id=XPWEtXzlLy

Keywords: diffusion models, constrained generation, constrained manifold, mirror map, watermarked generation, generation privacy

Compressor summary: The paper proposes Mirror Diffusion Models, a new type of diffusion models that can generate data on constrained sets without losing tractability, by using a mirror map in a standard Euclidean space and embedding invisible information for safety and privacy purposes.


FreeMask: Synthetic Images with Dense Annotations Make Stronger Segmentation Models

Lihe Yang, Xiaogang Xu, Bingyi Kang, Yinghuan Shi, Hengshuang Zhao

https://openreview.net/forum?id=XOotfgPiUF

Keywords: learning from synthetic, semantic segmentation, generative models

Compressor summary: FreeMask uses synthetic images generated by generative models to ease the data collection and annotation process for semantic segmentation, achieving comparable or better performance than real images.


TD Convergence: An Optimization Perspective

Kavosh Asadi, Shoham Sabach, Yao Liu, Omer Gottesman, Rasool Fakoor

https://openreview.net/forum?id=XOCbdqxAR2

Keywords: Reinforcement Learning, Temporal Difference Learning, Value Function Optimization, Convergence

Compressor summary: The paper analyzes how the temporal-difference learning algorithm works as an iterative optimization process, identifies factors that affect its convergence or divergence, and proves its convergence in different settings, explaining its success in reinforcement learning.


CosNet: A Generalized Spectral Kernel Network

Yanfang Xue, Pengfei Fang, Jinyue Tian, Shipeng Zhu, hui xue

https://openreview.net/forum?id=XNBeTgYcAq

Keywords: Spectral kernel; complex-valued networks

Compressor summary: The CosNet is a generalized spectral kernel network that preserves the inherent complex-valued representation in time-sequential data and improves the analysis of time-varying statistical characteristics by combining complex-valued spectral kernels with neural networks.


DiffUTE: Universal Text Editing Diffusion Model

Haoxing Chen, Zhuoer Xu, Zhangxuan Gu, jun lan, 行 郑, Yaohui Li, Changhua Meng, Huijia Zhu, Weiqiang Wang

https://openreview.net/forum?id=XKeSauhUdJ

Keywords: Diffusion model, text editing, self-supervied learning

Compressor summary: The paper proposes DiffUTE, a self-supervised text editing diffusion model that can draw multilingual characters and edit images realistically using web data.


Incentives in Private Collaborative Machine Learning

Rachael Hwee Ling Sim, Yehong Zhang, Trong Nghia Hoang, Xinyi Xu, Bryan Kian Hsiang Low, Patrick Jaillet

https://openreview.net/forum?id=XKP3mAsNHd

Keywords: Incentives, Privacy, Shapley fairness, Collaborative machine learning, data valuation, reward, sufficient statistics

Compressor summary: The authors propose a method for collaborative machine learning that incentivizes data sharing while preserving privacy by using differential privacy and adjusting sufficient statistics.


Solving Linear Inverse Problems Provably via Posterior Sampling with Latent Diffusion Models

Litu Rout, Negin Raoof, Giannis Daras, Constantine Caramanis, Alex Dimakis, Sanjay Shakkottai

https://openreview.net/forum?id=XKBFdYwfRo

Keywords: Inverse Problems, Posterior Sampling, Latent Diffusion Model, Stable Diffusion, Sample Recovery

Compressor summary: The framework uses pre-trained latent diffusion models to solve linear inverse problems and shows superior performance in various tasks like image restoration and enhancement.


Cold Diffusion: Inverting Arbitrary Image Transforms Without Noise

Arpit Bansal, Eitan Borgnia, Hong-Min Chu, Jie S. Li, Hamid Kazemi, Furong Huang, Micah Goldblum, Jonas Geiping, Tom Goldstein

https://openreview.net/forum?id=XH3ArccntI

Keywords: Generative Models, Computer Vision, Diffusion Models

Compressor summary: Deterministic image degradations can be used to create a family of generative diffusion models without the need for added noise.


DreamWaltz: Make a Scene with Complex 3D Animatable Avatars

Yukun Huang, Jianan Wang, Ailing Zeng, He CAO, Xianbiao Qi, Yukai Shi, Zheng-Jun Zha, Lei Zhang

https://openreview.net/forum?id=XGXL1E8Yyo

Keywords: Avatar Generation, 3D Content Creation, NeRF, Diffusion Model

Compressor summary: DreamWaltz is a framework that generates and animates complex 3D avatars from text guidance using novel techniques like SDS, 3D-aware skeleton conditioning, and image priors.


Convergence analysis of ODE models for accelerated first-order methods via positive semidefinite kernels

Jungbin Kim, Insoon Yang

https://openreview.net/forum?id=XFE6zpevLc

Keywords: convex optimization, accelerated gradient methods

Compressor summary: The paper introduces a new way to analyze optimization methods using functional analysis tools that leads to faster convergence rates and reveals a link between function values and gradient norms.


Learning Regularized Monotone Graphon Mean-Field Games

Fengzhuo Zhang, Vincent Tan, Zhaoran Wang, Zhuoran Yang

https://openreview.net/forum?id=XF923QPCGw

Keywords: mean-field approximation, graphon games, multi-agent reinforcement learning

Compressor summary: The paper explores the existence and efficient learning of Nash Equilibria in regularized Graphon Mean-Field Games, extending previous results and proposing new algorithms for discrete-time learning under weak monotonicity conditions.


An Inductive Bias for Tabular Deep Learning

Ege Beyazit, Jonathan Kozaczuk, Bo Li, Vanessa Wallace, Bilal H Fadlallah

https://openreview.net/forum?id=XEUc1JegGt

Keywords: Tabular Deep Learning, Spectral Bias, Neural Networks

Compressor summary: The paper proposes a new neural network layer that helps improve deep learning performance on irregularly shaped tabular data by learning low-frequency representations of the input features.


GAIA: Delving into Gradient-based Attribution Abnormality for Out-of-distribution Detection

Jinggang Chen, Junjie Li, Xiaoyang Qu, Jianzong Wang, Jiguang Wan, Jing Xiao

https://openreview.net/forum?id=XEBzQP3e7B

Keywords: out-of-distribution detection, distribution shifts, attribution gradients

Compressor summary: The paper proposes a method called GAIA that uses gradient abnormality to detect out-of-distribution examples in deep neural networks, improving their reliability and safety.


Binarized Neural Machine Translation

Yichi Zhang, Ankush Garg, Yuan Cao, Lukasz Lew, Behrooz Ghorbani, Zhiru Zhang, Orhan Firat

https://openreview.net/forum?id=XAyPlfmWpu

Keywords: neural network quantization, binarized transformer, machine translation, scaling law

Compressor summary: The paper introduces BMT, a novel binarization technique for Transformers that achieves the same quality as float models while being much smaller, by using additional LayerNorms and residual connections to address one-bit weight variance issues.


Non-Convex Bilevel Optimization with Time-Varying Objective Functions

Sen Lin, Daouda Sow, Kaiyi Ji, Yingbin Liang, Ness Shroff

https://openreview.net/forum?id=X9Vjq9Fuhq

Keywords: Bilevel Optimization, Time-Varying Functions, Single-Loop, Sublinear Bilevel Local Regret

Compressor summary: SOBOW is an efficient online bilevel optimizer with window averaging that works well for streaming data and time-varying functions, achieving sublinear regret.


Self-Weighted Contrastive Learning among Multiple Views for Mitigating Representation Degeneration

Jie Xu, Shuo Chen, Yazhou Ren, Xiaoshuang Shi, Heng Tao Shen, Gang Niu, Xiaofeng Zhu

https://openreview.net/forum?id=X8dbFcAox2

Keywords: Multi-view learning, Contrastive learning, Representation degeneration, Self-supervised learning

Compressor summary: SEM is a new contrastive learning framework for multi-view scenarios that adaptively strengthens useful views and weakens unreliable ones, while improving representation quality with reconstruction regularization.


Stochastic Approximation Algorithms for Systems of Interacting Particles

Mohammad Reza Karimi Jaghargh, Ya-Ping Hsieh, Andreas Krause

https://openreview.net/forum?id=X6mwdEVYvc

Keywords: Stochastic Approximation, Mean-Field Dynamics, Dynamical Systems, Neural Networks, Sampling

Compressor summary: The paper proposes a new framework to connect discrete-time interacting particle systems and their mean-field limits, improving the analysis of these systems in machine learning tasks.


On the Planning Abilities of Large Language Models - A Critical Investigation

Karthik Valmeekam, Matthew Marquez, Sarath Sreedharan, Subbarao Kambhampati

https://openreview.net/forum?id=X6dEqXIsEW

Keywords: Large Language Models, Planning, LLMs for autonomous and heuristic planning guidance

Compressor summary: The paper investigates how well large language models (LLMs) can generate plans autonomously and as heuristic guidance for other planners, finding limited success in the former but more promise in the latter.


Explain Any Concept: Segment Anything Meets Concept-Based Explanation

Ao Sun, Pingchuan Ma, Yuanyuan Yuan, Shuai Wang

https://openreview.net/forum?id=X6TBBsz9qi

Keywords: EXplainable AI, Machine Learning, Computer Vision

Compressor summary: The paper introduces Explain Any Concept (EAC), a method that uses the Segment Anything Model (SAM) to explain deep neural network decisions with any concept, improving human understanding of computer vision tasks.


Latent SDEs on Homogeneous Spaces

Sebastian Zeng, Florian Graf, Roland Kwitt

https://openreview.net/forum?id=X6Eapo5paw

Keywords: Variational Bayesian inference, stochastic differential equation, homogeneous spaces, geometric Euler-Maruyama, time series

Compressor summary: The authors study variational Bayesian inference for latent stochastic differential equations (SDEs) on the unit sphere, which simplifies the learning process and achieves good results on time series tasks.


Towards Evaluating Transfer-based Attacks Systematically, Practically, and Fairly

Qizhang Li, Yiwen Guo, Wangmeng Zuo, Hao Chen

https://openreview.net/forum?id=X5MH7iut9K

Keywords: adversarial examples, adversarial transferability, black-box attack

Compressor summary: The paragraph introduces TA-Bench, a benchmark for transfer-based methods to attack black-box DNNs, which evaluates 30+ methods on 10 popular victim models using ImageNet data.


Performance-optimized deep neural networks are evolving into worse models of inferotemporal visual cortex

Drew Linsley, Ivan F Rodriguez Rodriguez, Thomas FEL, Michael Arcaro, Saloni Sharma, Margaret Livingstone, Thomas Serre

https://openreview.net/forum?id=X4mmXQ4Nxw

Keywords: neural system identification, behavioral alignment, neural object recognition

Compressor summary: The authors find that as deep neural networks (DNNs) improve at object recognition, they become worse models of inferotemporal (IT) neuron responses to images, and propose a method to align DNN representations with human vision using the neural harmonizer.


Causal Imitability Under Context-Specific Independence Relations

Fateme Jamshidi, Sina Akbari, Negar Kiyavash

https://openreview.net/forum?id=X3IeHRD0zf

Keywords: causal inference, conditional independence, context-specific independence relations, imitability

Compressor summary: The paper explores the benefits of using context-specific independence information in causal imitation learning and provides a necessary and sufficient graphical criterion for it, as well as an algorithm that considers both CSI and data.


Nearly Optimal Bounds for Cyclic Forgetting

William Joseph Swartworth, Deanna Needell, Rachel Ward, Mark Kong, Halyun Jeong

https://openreview.net/forum?id=X25L5AjHig

Keywords: catastrophic forgetting, linear systems

Compressor summary: The paper derives theoretical limits on how much information is forgotten in continual learning for linear tasks and introduces a new characterization of numerical ranges of products of projections.


Risk-Averse Model Uncertainty for Distributionally Robust Safe Reinforcement Learning

James Queeney, Mouhacine Benosman

https://openreview.net/forum?id=X0CIxqYc4Z

Keywords: deep reinforcement learning, model uncertainty, safety, risk-averse, distributionally robust

Compressor summary: The paper presents a deep reinforcement learning method for safe decision making in uncertain environments using risk-averse distortion risk measures, without minimax optimization, and shows its effectiveness on continuous control tasks with safety constraints.


STORM: Efficient Stochastic Transformer based World Models for Reinforcement Learning

Weipu Zhang, Gang Wang, Jian Sun, Yetian Yuan, Gao Huang

https://openreview.net/forum?id=WxnrX42rnS

Keywords: deep learning, reinforcement learning, model-based reinforcement learning, world model, learning in imagination, transformer, variational autoencoders, sequence modeling

Compressor summary: STORM is a world model architecture that combines Transformers and variational autoencoders to improve sequence modeling and generation in model-based reinforcement learning, achieving a record performance on the Atari 100k benchmark with less interaction experience and faster training.


Probabilistic Invariant Learning with Randomized Linear Classifiers

Leonardo Cotta, Gal Yehuda, Assaf Schuster, Chris J. Maddison

https://openreview.net/forum?id=WwP2JaXAtB

Keywords: Invariant Learning, Geometric Deep Learning, Set Representations, Graph Representations, Expressive Power, Randomized Algorithms

Compressor summary: The paper proposes Randomized Linear Classifiers, a probabilistic model that can approximate any smooth function and preserve invariances while using fewer resources than deterministic neural networks.


Weighted ROC Curve in Cost Space: Extending AUC to Cost-Sensitive Learning

Huiyang Shao, Qianqian Xu, Zhiyong Yang, Peisong Wen, Gao Peifeng, Qingming Huang

https://openreview.net/forum?id=WsmBcJarWW

Keywords: AUC, Cost Learning, Bilevel, machine learning

Compressor summary: The paper proposes a new cost-sensitive learning framework that integrates cost distribution into the AUC metric using a bilevel optimization method and shows its effectiveness in handling long-tail datasets.


Distributionally Robust Ensemble of Lottery Tickets Towards Calibrated Sparse Network Training

Hitesh Sapkota, Dingrong Wang, ZHIQIANG TAO, Qi Yu

https://openreview.net/forum?id=WrRG0C1Vo5

Keywords: sparse network training, model calibration

Compressor summary: The paper proposes a new method to improve the reliability of sparse network training by using Distributionally Robust Optimization (DRO) to create an ensemble of diverse and complementary sub-networks that capture different data distributions and reduce overconfidence.


MotionGPT: Human Motion as a Foreign Language

Biao Jiang, Xin Chen, Wen Liu, Jingyi Yu, Gang YU, Tao Chen

https://openreview.net/forum?id=WqiZJGNkjn

Keywords: 3d motion, motion generation, human motion synthesis, text-driven, text-to-motion

Compressor summary: The paper introduces MotionGPT, a unified model for language and human motion that can handle various motion-related tasks, by treating human motion as a specific language.


Regularizing Neural Networks with Meta-Learning Generative Models

Shin'ya Yamaguchi, Daiki Chijiwa, Sekitoshi Kanai, Atsutoshi Kumagai, Hisashi Kashima

https://openreview.net/forum?id=WpuBEtrn0t

Keywords: Deep Learning, Generative Models, Generative Data Augmentation, Regularization, Meta-Learning

Compressor summary: The paper proposes a method called meta generative regularization (MGR) to improve generative data augmentation for deep learning by using synthetic samples to regularize feature extractors instead of training classifiers, which can avoid performance degradation and boost accuracy, especially on smaller datasets.


Joint Training of Deep Ensembles Fails Due to Learner Collusion

Alan Jeffares, Tennison Liu, Jonathan Crabbé, Mihaela van der Schaar

https://openreview.net/forum?id=WpGLxnOWhn

Keywords: Deep Ensembles, Deep Learning

Compressor summary: Joint optimization of ensemble loss in deep ensembles leads to artificial diversity that fails to generalize, causing a larger performance gap.


Accelerating Value Iteration with Anchoring

Jongmin Lee, Ernest K. Ryu

https://openreview.net/forum?id=Wn82NbmvJy

Keywords: Value Iteration, Reinforcement Learning, Reinforcement Learning Theory, Dynamic Programming, Acceleration, Anchoring mechanism

Compressor summary: Anc-VI is an accelerated value iteration method that reduces Bellman error faster than standard VI by using an anchoring mechanism.


Contextual Bandits and Imitation Learning with Preference-Based Active Queries

Ayush Sekhari, Karthik Sridharan, Wen Sun, Runzhe Wu

https://openreview.net/forum?id=WmqYhqvz5i

Keywords: Contextual Bandit, Imitation Learning, Learning from Expert Feedback, Theory

Compressor summary: The paper proposes an algorithm for preference-based contextual bandits and imitation learning that minimizes regret and query complexity, and can outperform a sub-optimal expert with fewer queries.


Large Language Models as Commonsense Knowledge for Large-Scale Task Planning

Zirui Zhao, Wee Sun Lee, David Hsu

https://openreview.net/forum?id=Wjp1AYB8lH

Keywords: Embodied Task Planning, Large Language Models, Human-Robot Interaction

Compressor summary: The paper proposes an LLM-MCTS algorithm that combines a commonsense world model from LLMs with MCTS search, improving task planning efficiency and performance on complex tasks.


Action Inference by Maximising Evidence: Zero-Shot Imitation from Observation with World Models

Xingyuan Zhang, Philip Becker-Ehmck, Patrick van der Smagt, Maximilian Karl

https://openreview.net/forum?id=WjlCQxpuxU

Keywords: Imitation Learning, World Models, Latent Variable Model, Transfer Learning, Variational Inference

Compressor summary: The paper proposes AIME, a method for imitating expert behaviors using world models without needing additional environment interactions or training.


Repetition In Repetition Out: Towards Understanding Neural Text Degeneration from the Data Perspective

Huayang Li, Tian Lan, Zihao Fu, Deng Cai, Lemao Liu, Nigel Collier, Taro Watanabe, Yixuan Su

https://openreview.net/forum?id=WjgCRrOgip

Keywords: language modeling, text generation, natural language processing

Compressor summary: This paper proposes a simple explanation and solution for the neural text degeneration problem by showing that penalizing repetitions in training data reduces this issue.


Differentiable Registration of Images and LiDAR Point Clouds with VoxelPoint-to-Pixel Matching

Junsheng Zhou, Baorui Ma, Wenyuan Zhang, Yi Fang, Yu-Shen Liu, Zhizhong Han

https://openreview.net/forum?id=WjWifKqmcG

Keywords: LiDAR Point Clouds, 2D images, Cross-modality registration, Matching

Compressor summary: The paper proposes a method to improve cross-modality registration between 2D images and 3D point clouds using a structured latent space, a triplet network, and a differentiable probabilistic PnP solver, achieving better results on KITTI and nuScenes datasets.


Connected Superlevel Set in (Deep) Reinforcement Learning and its Application to Minimax Theorems

Sihan Zeng, Thinh T. Doan, Justin Romberg

https://openreview.net/forum?id=WjDj6W872v

Keywords: Reinforcement learning, superlevel sets, minimax optimization, robust reinforcement learning

Compressor summary: The paper explores new properties of policy optimization landscapes in reinforcement learning and uses them to derive minimax theorems for robust reinforcement learning problems under adversarial rewards.


Anonymous Learning via Look-Alike Clustering: A Precise Analysis of Model Generalization

Adel Javanmard, Vahab Mirrokni

https://openreview.net/forum?id=WfsWy59bX2

Keywords: high-dimensional regression, generalization error, asymptotic analysis, Convex Gaussian Minimax Theorem, regularization

Compressor summary: This paper explores how using anonymous cluster centers in personalized recommendation systems can protect user privacy and sometimes improve model performance, by analyzing the generalization error using the Convex Gaussian Minimax Theorem.


Towards Semi-Structured Automatic ICD Coding via Tree-based Contrastive Learning

Chang Lu, Chandan K. Reddy, Ping Wang, Yue Ning

https://openreview.net/forum?id=Wff6DWFY2W

Keywords: ICD Coding, Contrastive Learning, NLP, Healthcare, Text Categorization, Pre-training

Compressor summary: The authors propose an algorithm to segment clinical notes into sections and use contrastive pre-training and masked section training to improve ICD coding performance with limited data.


Low Tensor Rank Learning of Neural Dynamics

Arthur Pellegrino, N Alex Cayco Gajic, Angus Chadwick

https://openreview.net/forum?id=WcoX8eJJjI

Keywords: Recurrent Neural Networks, Computational Neuroscience, Neural Data Analysis, Tensor, Learning

Compressor summary: The study investigates how low rank structures in recurrent neural networks (RNNs) evolve during learning and their implications for population connectivity and learning dynamics in both biological and artificial neural networks.


Paraphrasing evades detectors of AI-generated text, but retrieval is an effective defense

Kalpesh Krishna, Yixiao Song, Marzena Karpinska, John Frederick Wieting, Mohit Iyyer

https://openreview.net/forum?id=WbFhFvjjKj

Keywords: AI-generated text detection, text detection, paraphrasing, attacks, retrieval, defenses, large language models, LLMs

Compressor summary: The authors develop a paraphrase generation model (DIPPER) that evades several AI text detection algorithms and propose a defense method based on searching for semantically-similar texts in a database.


DeWave: Discrete Encoding of EEG Waves for EEG to Text Translation

Yiqun Duan, Charles Zhou, Zhen Wang, Yu-Kai Wang, Chin-teng Lin

https://openreview.net/forum?id=WaLI8slhLw

Keywords: EEG; Neural Encoding; Brain Computer Interface

Compressor summary: DeWave is a novel framework that translates brain signals into natural language without relying on eye-tracking or event markers, improving EEG-to-text translation accuracy.


Online learning of long-range dependencies

Nicolas Zucchet, Robert Meier, Simon Schug, Asier Mujika, Joao Sacramento

https://openreview.net/forum?id=Wa1GGPqjUn

Keywords: online learning, linear recurrent units, temporal credit assignment, biologically-plausible learning, local learning rules, neuromorphic computing

Compressor summary: The paper introduces an efficient online learning algorithm for recurrent neural networks that leverages independent modules, enabling them to learn long-range dependencies and opening new possibilities in neuromorphic computing.


Necessary and Sufficient Conditions for Optimal Decision Trees using Dynamic Programming

Jacobus G.M. van der Linden, Mathijs de Weerdt, Emir Demirović

https://openreview.net/forum?id=WYYpxVsKpR

Keywords: optimal decision trees, dynamic programming, separability

Compressor summary: The authors propose a dynamic programming framework for global optimization of decision trees that can handle separable objectives and constraints, improving scalability and performance over general-purpose solvers.


Max-Margin Token Selection in Attention Mechanism

Davoud Ataee Tarzanagh, Yingcong Li, Xuechen Zhang, Samet Oymak

https://openreview.net/forum?id=WXc8O8ghLH

Keywords: attention mechanism, implicit bias, margin maximization, nonconvex optimization, prompt tuning

Compressor summary: The paper analyzes the attention mechanism in transformer models and proves that it converges to a max-margin solution for optimal token selection.


Mixture Weight Estimation and Model Prediction in Multi-source Multi-target Domain Adaptation

Yuyang Deng, Ilja Kuzborskij, Mehrdad Mahdavi

https://openreview.net/forum?id=WVmus8NWE8

Keywords: Multi-source domain adaptation; minimax optimization; learning theory

Compressor summary: The paper proposes algorithms to efficiently learn from heterogeneous data sources and adapt to different target distributions with theoretical guarantees.


Rank-DETR for High Quality Object Detection

Yifan Pu, Weicong Liang, Yiduo Hao, Yuhui Yuan, Yukang Yang, Chao Zhang, Han Hu, Gao Huang

https://openreview.net/forum?id=WUott1ZvRj

Keywords: Object Detection

Compressor summary: The Rank-DETR method improves DETR-based object detectors by enhancing ranking accuracy and reducing false positives, leading to better localization and detection performance.


Orthogonal Non-negative Tensor Factorization based Multi-view Clustering

Jing Li, Quanxue Gao, QIANQIAN WANG, Ming Yang, Wei Xia

https://openreview.net/forum?id=WRtlsxA5h7

Keywords: Multi-view clustering, tensor Schatten p-norm, non-negative matrix factorization.

Compressor summary: The paragraph describes a novel multi-view clustering method based on orthogonal non-negative tensor factorization, which considers both within-view and between-view information and achieves good performance on various datasets.


DYffusion: A Dynamics-informed Diffusion Model for Spatiotemporal Forecasting

Salva Rühling Cachay, Bo Zhao, Hailey James, Rose Yu

https://openreview.net/forum?id=WRGldGm5Hz

Keywords: AI for science, diffusion models, scientific machine learning, probabilistic forecasting

Compressor summary: The authors present a dynamics-informed diffusion model that leverages temporal data dynamics for forecasting complex systems and demonstrate its advantages over Gaussian noise-based models.


Full-Atom Protein Pocket Design via Iterative Refinement

ZAIXI ZHANG, Zepu Lu, Zhongkai Hao, Marinka Zitnik, Qi Liu

https://openreview.net/forum?id=WPdGRRJaPb

Keywords: Graph Representation Learning, AI for Science

Compressor summary: FAIR is a framework that co-designs protein pocket sequence and 3D structure for ligand binding by refining residue types, backbone coordinates, and sidechain atoms in an efficient full-shot manner.


A Robust Exact Algorithm for the Euclidean Bipartite Matching Problem

Akshaykumar G Gattani, Sharath Raghvendra, Pouyan Shirzadian

https://openreview.net/forum?id=WPbIAdB6aQ

Keywords: Euclidean bipartite matching, exact algorithms, primal dual method

Compressor summary: The paper presents a new algorithm for computing minimum-cost bipartite matching between two sets of points, which has faster execution time than previous methods, especially for stochastic point sets with real-valued coordinates and any dimension.


Projection-Free Methods for Solving Nonconvex-Concave Saddle Point Problems

Morteza Boroun, Erfan Yazdandoost Hamedani, Afrooz Jalilzadeh

https://openreview.net/forum?id=WO1kHC5Lfz

Keywords: Saddle Point Problem, Projection-free method

Compressor summary: The paper proposes efficient single-loop projection-free methods for constrained saddle point problems using regularization and nested approximation techniques, achieving convergence guarantees in limited iterations.


Unsupervised Anomaly Detection with Rejection

Lorenzo Perini, Jesse Davis

https://openreview.net/forum?id=WK8LQzzHwW

Keywords: Anomaly Detection, Learning with Rejection, Unsupervised Learning

Compressor summary: The paper proposes a method to improve anomaly detection by rejecting predictions with high uncertainty using a constant threshold on a stability metric, while providing theoretical guarantees and empirical results.


Contrastive Modules with Temporal Attention for Multi-Task Reinforcement Learning

Siming Lan, Rui Zhang, Qi Yi, Jiaming Guo, Shaohui Peng, Yunkai Gao, Fan Wu, Ruizhi Chen, Zidong Du, Xing Hu, Xishan Zhang, Ling Li, Yunji Chen

https://openreview.net/forum?id=WIrZh2XxLT

Keywords: reinforcement learning, multi-task learning, contrastive learning

Compressor summary: The paper proposes CMTA, a method for multi-task reinforcement learning that uses contrastive learning to prevent conflicts within tasks and temporal attention to combine shared modules at a finer level than tasks, improving performance on Meta-World benchmark tasks.


Rank-N-Contrast: Learning Continuous Representations for Regression

Kaiwen Zha, Peng Cao, Jeany Son, Yuzhe Yang, Dina Katabi

https://openreview.net/forum?id=WHedsAeatp

Keywords: regression, representation learning, continuity

Compressor summary: Rank-N-Contrast (RNC) is a framework that learns continuous representations for regression by contrasting samples based on their rankings in the target space, improving performance, robustness, efficiency, and generalization.


Top-Ambiguity Samples Matter: Understanding Why Deep Ensemble Works in Selective Classification

Qiang Ding, Yixuan Cao, Ping Luo

https://openreview.net/forum?id=WBq6Q4ml04

Keywords: selective classification, uncertainty estimation, ensemble learning

Compressor summary: The paper analyzes why ensemble methods improve prediction reliability, focusing on top-ambiguity samples where member models differ, and proves that ensembles have lower selective risk under certain conditions.


NCDL: A Framework for Deep Learning on non-Cartesian Lattices

Joshua John Horacsek, Usman Alim

https://openreview.net/forum?id=WBXYGBQXiB

Keywords: Computer Vision and Pattern Recognition

Compressor summary: The paper introduces the lattice tensor data structure to enable machine learning on non-Cartesian domains using standard algorithms and a software library.


Nonparametric Boundary Geometry in Physics Informed Deep Learning

Scott Alexander Cameron, Arnu Pretorius, Stephen J. Roberts

https://openreview.net/forum?id=WAd5ZRdFoc

Keywords: PINNs, physics informed neural networks, geometric deep learning, neural operator, PDEs

Compressor summary: The paper proposes a neural network that takes triangular meshes as input and solves partial differential equations quickly, without retraining or fixed geometry representation.


BadTrack: A Poison-Only Backdoor Attack on Visual Object Tracking

Bin Huang, Jiaqian Yu, Yiwei Chen, Siyang Pan, Qiang Wang, Zhi Wang

https://openreview.net/forum?id=W9pJx9sFCh

Keywords: Backdoor Attack, Visual Object Tracking, Deep Learning, Poison-Only

Compressor summary: This paper proposes a poison-only backdoor attack on visual object tracking that attaches a trigger pattern to the background of video frames, which degrades the performance of VOT trackers.


Black-box Backdoor Defense via Zero-shot Image Purification

Yucheng Shi, Mengnan Du, Xuansheng Wu, Zihan Guan, Jin Sun, Ninghao Liu

https://openreview.net/forum?id=W6U2xSbiE1

Keywords: backdoor defense, black-box defense, diffusion model

Compressor summary: The paper proposes a novel defense framework called Zero-shot Image Purification (ZIP) to defend against backdoor attacks on black-box models by applying a linear transformation and using a pre-trained diffusion model to recover the purified image.


Toward Understanding Generative Data Augmentation

Chenyu Zheng, Guoqiang Wu, Chongxuan Li

https://openreview.net/forum?id=W5Clq1bSrR

Keywords: generative data augmentation, algorithmic stability, non-i.i.d. learning

Compressor summary: Generative data augmentation can improve classification performance in various tasks by obtaining fake labeled examples from a trained conditional generative model, but its effect has not been well studied theoretically; this paper establishes a stability bound and shows how it helps learning in different settings.


Fair Canonical Correlation Analysis

Zhuoping Zhou, Davoud Ataee Tarzanagh, Bojian Hou, Boning Tong, Jia Xu, Yanbo Feng, Qi Long, Li Shen

https://openreview.net/forum?id=W3cDd5xlKZ

Keywords: Fairness, Canonical Correlation Analysis, Riemannian Optimization, Pareto Optimization

Compressor summary: The paper proposes a framework for minimizing unfairness in Canonical Correlation Analysis by balancing correlation levels across groups while maintaining accuracy and evaluating its effectiveness on various datasets.


PolyDiffuse: Polygonal Shape Reconstruction via Guided Set Diffusion Models

Jiacheng Chen, Ruizhi Deng, Yasutaka Furukawa

https://openreview.net/forum?id=W2ZBLdfa16

Keywords: Structured Reconstruction, Floorplan Reconstruction, HD Map Construction, Diffusion Models

Compressor summary: PolyDiffuse is a new algorithm that uses Diffusion Models to reconstruct structured data like polygonal shapes from sensor data, overcoming challenges related to denoising ambiguity and initial noise choice.


Are Vision Transformers More Data Hungry Than Newborn Visual Systems?

Lalit Pandey, Samantha Marie Waters Wood, Justin Newell Wood

https://openreview.net/forum?id=W23ZTdsabj

Keywords: vision transformer, newborn, controlled rearing, object recognition, data hungry

Compressor summary: ViTs and chicks both learn view-invariant object recognition in impoverished visual environments, challenging the assumption that ViTs are more data hungry than brains.


Exposing Attention Glitches with Flip-Flop Language Modeling

Bingbin Liu, Jordan T. Ash, Surbhi Goel, Akshay Krishnamurthy, Cyril Zhang

https://openreview.net/forum?id=VzmpXQAn6E

Keywords: Transformers, language models, hallucinations, long-range dependencies, generalization, extrapolation, out-of-distribution

Compressor summary: The text discusses attention glitches, a phenomenon where large language models sometimes make errors in reasoning, and introduces flip-flop language modeling as a way to study and potentially improve these models' extrapolative abilities.


Double Pessimism is Provably Efficient for Distributionally Robust Offline Reinforcement Learning: Generic Algorithm and Robust Partial Coverage

Jose Blanchet, Miao Lu, Tong Zhang, Han Zhong

https://openreview.net/forum?id=VzLBMkc7tB

Keywords: distributionally robust offline reinforcement learning, double pessimism, general function approximation

Compressor summary: The paper introduces Doubly Pessimistic Model-based Policy Optimization ($\texttt{P}^2\texttt{MPO}$), a novel algorithm framework for distributionally robust offline reinforcement learning that combines flexible model estimation and doubly pessimistic policy optimization to overcome distribution shift and perturbation.


VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks

Wenhai Wang, Zhe Chen, Xiaokang Chen, Jiannan Wu, Xizhou Zhu, Gang Zeng, Ping Luo, Tong Lu, Jie Zhou, Yu Qiao, Jifeng Dai

https://openreview.net/forum?id=Vx1JadlOIt

Keywords: Large Vision-Language Model, Detection, Image Caption

Compressor summary: The paper introduces VisionLLM, an LLM-based framework that treats images as foreign languages and uses language instructions to perform open-ended vision tasks with high task customization and competitive performance on COCO detection benchmark.


RevColV2: Exploring Disentangled Representations in Masked Image Modeling

Qi Han, Yuxuan Cai, Xiangyu Zhang

https://openreview.net/forum?id=VvnfMeC3gQ

Keywords: architecture design, representation learning, masked image modeling, self-supervised learning

Compressor summary: RevColV2 is a new architecture for masked image modeling that keeps the entire autoencoder during pre-training and fine-tuning, maintaining disentangled low-level and semantic information and achieving competitive performance on various vision tasks.


Multiplication-Free Transformer Training via Piecewise Affine Operations

Atli Kosson, Martin Jaggi

https://openreview.net/forum?id=Vtqymej1tA

Keywords: multiplication-free, neural architectures, piecewise linear networks, piecewise affine networks, efficient training, efficient arithmetics

Compressor summary: The authors propose a method to replace matrix multiplications and non-linearities with piecewise affine approximations, achieving fully multiplication-free training of neural networks without sacrificing performance.


Evaluating Cognitive Maps and Planning in Large Language Models with CogEval

Ida Momennejad, Hosein Hasanbeig, Felipe Vieira Frujeri, Hiteshi Sharma, Nebojsa Jojic, Hamid Palangi, Robert Ness, Jonathan Larson

https://openreview.net/forum?id=VtkGvGcGe3

Keywords: Large Language Models, LLM evaluation, model comparison, GPT-4, graph analysis, cognitive science, cognitive map, hippocampus, planning, multi-step planning, reasoning, community graph

Compressor summary: The authors propose CogEval, a systematic evaluation protocol for cognitive abilities in LLMs, and find that current LLMs have limited planning ability and fail to understand cognitive maps.


AdaptSSR: Pre-training User Model with Augmentation-Adaptive Self-Supervised Ranking

Yang Yu, Qi Liu, Kai Zhang, Yuren Zhang, Chao Song, Min Hou, Yuqing Yuan, ZHIhao Ye, ZAIXI ZHANG, Sanshi Lei Yu

https://openreview.net/forum?id=VsbrdJpwpT

Keywords: User Model Pre-training, Data Augmentation, Contrastive Learning

Compressor summary: AdaptSSR is a new pretext task for user modeling that improves performance by capturing similarity orders between implicitly and explicitly augmented views and other users' views, while adjusting the similarity constraint based on estimated similarities.


Autonomous Capability Assessment of Sequential Decision-Making Systems in Stochastic Settings

Pulkit Verma, Rushang Karia, Siddharth Srivastava

https://openreview.net/forum?id=VqclD6Nfaj

Keywords: Sequential Decision Making, Interpretable Models, Relational Model Learning, Black-Box Agents, Symbolic Descriptions

Compressor summary: The paper proposes an active-learning method to learn a probabilistic model of black-box AI systems with sequential decision-making capabilities, allowing users to understand and assess their safety in stochastic settings.


Does Graph Distillation See Like Vision Dataset Counterpart?

Beining Yang, Kai Wang, Qingyun Sun, Cheng Ji, Xingcheng Fu, Hao Tang, Yang You, Jianxin Li

https://openreview.net/forum?id=VqIWgUVsXc

Keywords: data-efficient learning, graph generation, graph neural networks

Compressor summary: The paper proposes a new method, SGDD, to preserve the original structure information of large-scale graphs when generating synthetic ones, improving performance in cross-architecture settings and specific tasks while reducing storage costs.


Patch n’ Pack: NaViT, a Vision Transformer for any Aspect Ratio and Resolution

Mostafa Dehghani, Basil Mustafa, Josip Djolonga, Jonathan Heek, Matthias Minderer, Mathilde Caron, Andreas Peter Steiner, Joan Puigcerver, Robert Geirhos, Ibrahim Alabdulmohsin, Avital Oliver, Piotr Padlewski, Alexey A. Gritsenko, Mario Lucic, Neil Houlsby

https://openreview.net/forum?id=VpGFHmI7e5

Keywords: Vision Transformer, variable aspect ratio, flexible inference, efficient training

Compressor summary: NaViT is a flexible and efficient computer vision model that can handle inputs of different resolutions and aspect ratios without resizing them, leading to improved performance and robustness.


Chanakya: Learning Runtime Decisions for Adaptive Real-Time Perception

Anurag Ghosh, Vaibhav Balloli, Akshay Nambi, Aditya Singh, Tanuja Ganu

https://openreview.net/forum?id=VpCjozUOM2

Keywords: approximate execution framework; real time perception; latency-accuracy tradeoffs

Compressor summary: The paper introduces Chanakya, a learned framework that optimizes real-time perception by automatically balancing accuracy and latency tradeoffs in decisions induced by intrinsic and extrinsic factors.


Fine-Tuning Language Models with Just Forward Passes

Sadhika Malladi, Tianyu Gao, Eshaan Nichani, Alex Damian, Jason D. Lee, Danqi Chen, Sanjeev Arora

https://openreview.net/forum?id=Vota6rFhBQ

Keywords: language models, fine-tuning, zeroth order optimization, memory efficiency

Compressor summary: MeZO is a memory-efficient zeroth-order optimizer that can fine-tune large language models with less memory and GPU resources than backpropagation, achieving comparable performance on various tasks.


Bounded rationality in structured density estimation

Tianyuan Teng, Li Kevin Wenliang, Hang Zhang

https://openreview.net/forum?id=VnfeOjR73Q

Keywords: human representation of uncertainty; Bayesian inference; bounded rationality; inductive bias; Chinese Restaurant Process

Compressor summary: This study examines how humans learn to represent environmental uncertainty, finding that they tend to overestimate the number of clusters in a distribution, possibly due to cognitive limitations.


Diffusion Hyperfeatures: Searching Through Time and Space for Semantic Correspondence

Grace Luo, Lisa Dunlap, Dong Huk Park, Aleksander Holynski, Trevor Darrell

https://openreview.net/forum?id=Vm1zeYqwdc

Keywords: semantic correspondence, hypercolumns, diffusion models, generative model representations

Compressor summary: The paper proposes Diffusion Hyperfeatures, a method to consolidate feature maps from diffusion models into per-pixel descriptors for semantic keypoint correspondence tasks on real and synthetic images.


Efficient Exploration in Continuous-time Model-based Reinforcement Learning

Lenart Treven, Jonas Hübotter, Bhavya Sukhija, Florian Dorfler, Andreas Krause

https://openreview.net/forum?id=VkhvDfY2dB

Keywords: Reinforcement Learning, Optimal Control, Continuous Time

Compressor summary: The paper presents a reinforcement learning algorithm that uses nonlinear ODEs to model continuous-time dynamics and shows how to choose when and how often to observe the system for optimal exploration.


Nonparametric Teaching for Multiple Learners

Chen Zhang, Xiaofeng Cao, Weiyang Liu, Ivor Tsang, James Kwok

https://openreview.net/forum?id=VkUNovXoxx

Keywords: Nonparametric machine teaching, Multiple learners

Compressor summary: The paper proposes a new framework called Multi-learner Nonparametric Teaching (MINT) that enables teachers to teach multiple students simultaneously more efficiently than traditional single-learner methods, especially when students can communicate with each other.


Efficient Activation Function Optimization through Surrogate Modeling

Garrett Bingham, Risto Miikkulainen

https://openreview.net/forum?id=ViFTWelHVZ

Keywords: automl, activation function, surrogate modeling, fisher information matrix, eigenvalues, optimization, umap, imagenet

Compressor summary: The paper introduces new datasets, methods, and findings to improve the design and optimization of activation functions in neural networks, including a surprising sigmoidal discovery.


DropPos: Pre-Training Vision Transformers by Reconstructing Dropped Positions

Haochen Wang, Junsong Fan, Yuxi Wang, Kaiyou Song, Tong Wang, Zhaoxiang Zhang

https://openreview.net/forum?id=VhcsIxVEd9

Keywords: Self-Supervised Learning, Vision Transformer, Visual Representation Learning

Compressor summary: DropPos is a self-supervised pretext task that improves location awareness in Vision Transformers by reconstructing dropped positions with visual appearance and attention.


Variance-Reduced Gradient Estimation via Noise-Reuse in Online Evolution Strategies

Oscar Li, James Harrison, Jascha Sohl-Dickstein, Virginia Smith, Luke Metz

https://openreview.net/forum?id=VhbV56AJNt

Keywords: Evolution Strategies, unrolled computation graph, online gradient estimation, variance reduction, stochastic gradient estimation

Compressor summary: The paper introduces Noise-Reuse Evolution Strategies (NRES), an unbiased online evolution strategies method for machine learning that has lower variance, faster convergence, and better parallelism than existing methods.


Estimating Riemannian Metric with Noise-Contaminated Intrinsic Distance

Jiaming Qiu, Xiongtao Dai

https://openreview.net/forum?id=VhLU3pStsl

Keywords: metric learning, manifold learning, local metric, dissimilarity, geometry

Compressor summary: The paper proposes a new local regression approach to learn the Riemannian metric tensor from similarity measures between data points, and provides theoretical convergence rates and empirical results on various datasets.


Uni-ControlNet: All-in-One Control to Text-to-Image Diffusion Models

Shihao Zhao, Dongdong Chen, Yen-Chun Chen, Jianmin Bao, Shaozhe Hao, Lu Yuan, Kwan-Yee K. Wong

https://openreview.net/forum?id=VgQw8zXrH8

Keywords: computer vision, diffusion model, text-to-image generation

Compressor summary: The paper introduces Uni-ControlNet, a framework that allows for flexible and composable control of text-to-image diffusion models using different local and global controls with minimal fine-tuning and model size costs.


Learning Large-Scale MTP$_2$ Gaussian Graphical Models via Bridge-Block Decomposition

Xiwen Wang, Jiaxi Ying, Daniel P. Palomar

https://openreview.net/forum?id=Vfp8sDST4g

Keywords: MTP2 Gaussian Graphical Model, High-dimensional precision matrix estimation, Bridge-block decomposition.

Compressor summary: The paper proposes a method to learn large Gaussian graphical models by decomposing the problem into smaller sub-problems and using bridge concepts, which reduces computational complexity and improves existing algorithms' performance.


RDumb: A simple approach that questions our progress in continual test-time adaptation

Ori Press, Steffen Schneider, Matthias Kuemmerer, Matthias Bethge

https://openreview.net/forum?id=VfP6VTVsHc

Keywords: test time adaptation, continual adaptation, benchmarking, imagenet-c, imagenet classification, robustness, continual learning, imagenet benchmark

Compressor summary: The Continually Changing Corruptions (CCC) benchmark reveals that most Test-Time Adaptation (TTA) methods perform poorly over long timescales and are beaten by a simple resetting strategy.


Conformal Prediction for Uncertainty-Aware Planning with Diffusion Dynamics Model

Jiankai Sun, Yiqi Jiang, Jianing Qiu, Parth Talpur Nobel, Mykel Kochenderfer, Mac Schwager

https://openreview.net/forum?id=VeO03T59Sh

Keywords: Uncertainty, Conformal Prediction, Dynamics Model

Compressor summary: PlanCP is a method that uses conformal prediction to quantify and reduce the uncertainty of diffusion dynamics models for trajectory prediction in robotic applications, improving performance on offline reinforcement learning and continuous planning tasks.


Efficient Algorithms for Generalized Linear Bandits with Heavy-tailed Rewards

Bo Xue, Yimu Wang, Yuanyu Wan, Jinfeng Yi, Lijun Zhang

https://openreview.net/forum?id=Vbm5UCaYeh

Keywords: linear bandits, heavy-tailed, truncated, mean of medians

Compressor summary: The paper proposes two new algorithms for generalized linear bandits with heavy-tailed rewards and shows they have better performance than existing methods in terms of regret bounds.


Efficient Adaptation of Large Vision Transformer via Adapter Re-Composing

Wei Dong, Dawei Yan, Zhijun Lin, Peng Wang

https://openreview.net/forum?id=VbYdaK8ek0

Keywords: computer vision; vision transformer; visual adapter; transfer learning

Compressor summary: The study proposes a new method called Adapter Re-Composing (ARC) to efficiently adapt pre-trained models for downstream tasks by reusing and sharing parameters in adapter design.


Correlation Aware Sparsified Mean Estimation Using Random Projection

Shuli Jiang, Pranay Sharma, Gauri Joshi

https://openreview.net/forum?id=VacSQpbI0U

Keywords: distributed vector mean estimation, communication efficiency, cross-client correlation

Compressor summary: Rand-Proj-Spatial is a new method for communication-efficient distributed vector mean estimation that uses projection to a random subspace and correlation information to outperform existing techniques like Rand-$k$-Spatial.


Score-based Data Assimilation

François Rozet, Gilles Louppe

https://openreview.net/forum?id=VUvLSnMZdX

Keywords: data assimilation, score-based, generative modeling, posterior inference, dynamical systems

Compressor summary: The paper introduces score-based data assimilation, a method for inferring state trajectories in stochastic dynamical systems by learning a generative model that can handle long time horizons and complex dynamics without relying on transition dynamics.


Understanding the Latent Space of Diffusion Models through the Lens of Riemannian Geometry

Yong-Hyun Park, Mingi Kwon, Jaewoong Choi, Junghyo Jo, Youngjung Uh

https://openreview.net/forum?id=VUlYp3jiEI

Keywords: diffusion models, semantic image editing, differential geometry

Compressor summary: This paper analyzes the geometric structure of diffusion models' latent space and shows how it enables image editing through traversal without additional training.


Batchnorm Allows Unsupervised Radial Attacks

Amur Ghose, Apurv Gupta, Yaoliang Yu, Pascal Poupart

https://openreview.net/forum?id=VQ1heZKSLQ

Keywords: Adversarial, Batch normalization, Robustness, Geometric, radial

Compressor summary: The text explains how intermediate latents from batch normalized deep image recognition models can be used to create adversarial examples without labels, and discusses the security implications of this finding.


Continuous-Time Functional Diffusion Processes

Giulio Franzese, Giulio Corallo, Simone Rossi, Markus Heinonen, Maurizio Filippone, Pietro Michiardi

https://openreview.net/forum?id=VPrir0p5b6

Keywords: Hilbert spaces, Diffusion models, Stochastic Partial Differential Equations

Compressor summary: Functional Diffusion Processes (FDPs) are a new type of generative model that works with continuous data in function spaces, requiring less complex network architectures and achieving high-quality image generation.


LogSpecT: Feasible Graph Learning Model from Stationary Signals with Recovery Guarantees

Shangyuan LIU, Linglingzhi Zhu, Anthony Man-Cho So

https://openreview.net/forum?id=VPTZVVP4tm

Keywords: Graph Signal Processing, Spectral Template, Network Inference, Optimization, Linearized ADMM

Compressor summary: The paper introduces LogSpecT, a novel graph learning model that is always feasible and has recovery guarantees, and proposes an efficient algorithm with convergence guarantees to learn graphs from stationary signals.


Hard Prompts Made Easy: Gradient-Based Discrete Optimization for Prompt Tuning and Discovery

Yuxin Wen, Neel Jain, John Kirchenbauer, Micah Goldblum, Jonas Geiping, Tom Goldstein

https://openreview.net/forum?id=VOstHxDdsN

Keywords: Diffusion Model, Generative AI, Prompt Discovery

Compressor summary: The paragraph discusses an optimization method for hard text prompts that can be used in various applications and allows users to easily generate and mix image concepts without prior knowledge.


Federated Learning via Meta-Variational Dropout

Insu Jeon, Minui Hong, Junhyeog Yun, Gunhee Kim

https://openreview.net/forum?id=VNyKBipt91

Keywords: Personalized Federated Learning, Variational Dropout, Meta-Learning, Bayesian Neural Network

Compressor summary: MetaVD is a novel Bayesian meta-learning approach that predicts dropout rates per client and improves Federated Learning in non-IID data settings by personalizing models and reducing communication costs.


Graph of Circuits with GNN for Exploring the Optimal Design Space

Aditya Hemant Shahane, Saripilli Venkata Swapna Manjiri, Ankesh Jain, Sandeep Kumar

https://openreview.net/forum?id=VNjJAWjuEU

Keywords: Analog design optimization, Analog synthesis, Graph Neural Networks, EDA, Graph learning, Optimization

Compressor summary: The paper introduces GCX, a framework that uses graph structure learning and graph neural networks to create a surrogate model for efficient exploration of the analog circuit design space in a semi-supervised learning framework.


Hyperbolic Space with Hierarchical Margin Boosts Fine-Grained Learning from Coarse Labels

ShuLin Xu, Yifan Sun, Faen Zhang, Anqi Xu, Xiu-Shen Wei, Yi Yang

https://openreview.net/forum?id=VMz5GhfxgV

Keywords: Fine-grained learning, Coarse-to-fine learning, Hyperbolic space, Hierarchical margin

Compressor summary: The paper proposes a novel hyperbolic space method for fine-grained recognition tasks that leverages hierarchical cosine margins to improve discriminative ability and achieves state-of-the-art results on five benchmark datasets.


UP-DP: Unsupervised Prompt Learning for Data Pre-Selection with Vision-Language Models

Xin Li, Sima Behpour, Thang Doan, Wenbin He, Liang Gou, Liu Ren

https://openreview.net/forum?id=VMAgvbBBts

Keywords: Unsupervised prompt learning, UP-DP, Data preselection

Compressor summary: This study proposes UP-DP, an unsupervised prompt learning method that improves data pre-selection by combining vision and text features from foundation models like BLIP-2, achieving up to 20% performance gain and showing generalizability across datasets.


Regret Minimization via Saddle Point Optimization

Johannes Kirschner, Alireza Bakhtiari, Kushagra Chandak, Volodymyr Tkachuk, Csaba Szepesvari

https://openreview.net/forum?id=VLnEFGu9V7

Keywords: sequential decision-making, decision-estimation coefficient, regret minimization, bandits, reinforcement learning, partial monitoring

Compressor summary: The paper presents a new algorithm (Anytime-E2D) that balances exploration and exploitation in sequential decision-making, based on re-parametrizing the existing DEC method, and demonstrates its effectiveness in high-dimensional linear bandits.


Test-Time Distribution Normalization for Contrastively Learned Visual-language Models

Yifei Zhou, Juntao Ren, Fengyu Li, Ramin Zabih, Ser-Nam Lim

https://openreview.net/forum?id=VKbEO2eh5w

Keywords: contrastive learning, pre-trained visual-language models, zero-shot learning, test-time augmentation

Compressor summary: The paper introduces Distribution Normalization (DN), which improves test-time performance in visual-language contrastive learning by approximating negative samples during inference without retraining or fine-tuning.


Interactive Multi-fidelity Learning for Cost-effective Adaptation of Language Model with Sparse Human Supervision

Jiaxin Zhang, Zhuohang Li, Kamalika Das, Sricharan Kumar

https://openreview.net/forum?id=VIaw1XHb4G

Keywords: multi-fidelity optimization, cost-effective learning, exploration-exploitation query, limited annotation budgets

Compressor summary: IMFL is a framework for developing small domain-specific language models using a mix of low-cost automatic and high-quality human annotations, achieving superior performance with limited budgets.


Delegated Classification

Eden Saig, Inbal Talgam-Cohen, Nir Rosenfeld

https://openreview.net/forum?id=VGLXjbTSYa

Keywords: Delegation, Algorithmic Contract Design, Moral Hazard, Learning Curves

Compressor summary: The paper proposes a framework for incentive-aware delegation of machine learning tasks using performance-based contracts that balance budget and accuracy, connecting classic statistical theory with modern learning curves and scaling laws.


Neural Polarizer: A Lightweight and Effective Backdoor Defense via Purifying Poisoned Features

Mingli Zhu, Shaokui Wei, Hongyuan Zha, Baoyuan Wu

https://openreview.net/forum?id=VFhN15Vlkj

Keywords: Backdoor Defense, Backdoor Learning, Trustworthy AI

Compressor summary: The proposed method filters trigger information from poisoned samples using a lightweight neural polarizer inserted as an intermediate layer in the backdoored model, requiring less clean data and being more efficient than other defense methods.


Mixed-Initiative Multiagent Apprenticeship Learning for Human Training of Robot Teams

Esmaeil Seraj, Jerry Yuyang Xiong, Mariah L Schrum, Matthew Gombolay

https://openreview.net/forum?id=VCOZaczCHg

Keywords: Learning from Demonstration, Multi-Robot Systems, Teaching Robot Teams

Compressor summary: MixTURE is a novel Multi-Agent Learning from Demonstration framework that learns both collaborative tasks and inter-agent communication from human expert data, reducing human workload and improving usability.


Training Neural Networks is NP-Hard in Fixed Dimension

Vincent Froese, Christoph Hertrich

https://openreview.net/forum?id=VAQp2EnZeW

Keywords: Computational Complexity, Neural Network, Rectified Linear Unit, Empirical Risk Minimization, Parameterized Complexity

Compressor summary: The paper investigates the complexity of training two-layer neural networks with different activation functions in various input dimensions and neuron counts, answering several open questions and proving some fixed-parameter tractability results.


Video-Mined Task Graphs for Keystep Recognition in Instructional Videos

Kumar Ashutosh, Santhosh Kumar Ramakrishnan, Triantafyllos Afouras, Kristen Grauman

https://openreview.net/forum?id=VAC7aB6qSG

Keywords: Instructional Videos, Task Graph, Keystep Recognition

Compressor summary: The paragraph describes a method for automatically discovering task graphs from how-to videos to improve keystep recognition in novel videos.


Temporally Disentangled Representation Learning under Unknown Nonstationarity

Xiangchen Song, Weiran Yao, Yewen Fan, Xinshuai Dong, Guangyi Chen, Juan Carlos Niebles, Eric Xing, Kun Zhang

https://openreview.net/forum?id=V8GHCGYLkf

Keywords: Unsupervised learning, Temporal disentanglement, Nonlinear ICA, Identifiability theory

Compressor summary: The paper proposes NCTRL, a method for identifying time-delayed causal influences in nonstationary sequential data without using auxiliary information or simplifying latent dynamics.


Nonparametric Identifiability of Causal Representations from Unknown Interventions

Julius von Kügelgen, Michel Besserve, Wendong Liang, Luigi Gresele, Armin Kekić, Elias Bareinboim, David Blei, Bernhard Schölkopf

https://openreview.net/forum?id=V87gZeSOL4

Keywords: Causal representation learning, identifiability, theory, nonparametric, interventions, multi-environment

Compressor summary: The paper proposes a method to learn latent causal variables and their relations from high-dimensional functions, using multiple datasets from unknown interventions and showing identifiability conditions without restrictive assumptions or partial knowledge.


SEEDS: Exponential SDE Solvers for Fast High-Quality Sampling from Diffusion Models

Martin Gonzalez, Nelson Fernandez, Thuy Vinh Dinh Tran, Elies Gherbi, Hatem Hajri, Nader Masmoudi

https://openreview.net/forum?id=V6IgkYKD8P

Keywords: Diffusion Probabilistic Models, Exponential SDE methods, Image Generation, Generative Models

Compressor summary: SEEDS are stochastic exponential derivative-free solvers that improve and generalize Exponential Integrator approaches for fast sampling from Diffusion Probabilistic Models without requiring derivatives or training, achieving optimal quality and speed.


Localized Symbolic Knowledge Distillation for Visual Commonsense Models

Jae Sung Park, Jack Hessel, Khyathi Chandu, Paul Pu Liang, Ximing Lu, Peter West, Youngjae Yu, Qiuyuan Huang, Jianfeng Gao, Ali Farhadi, Yejin Choi

https://openreview.net/forum?id=V5eG47pyVl

Keywords: multimodal, commonsense reasoning, instruction tuning, large language model

Compressor summary: The Localized Visual Commonsense model allows users to specify regions within images as input, enabling precise within-image reasoning and improving multimodal tasks with a reference-as-input interface.


Lockdown: Backdoor Defense for Federated Learning with Isolated Subspace Training

Tiansheng Huang, Sihao Hu, Ka-Ho Chow, Fatih Ilhan, Selim Furkan Tekin, Ling Liu

https://openreview.net/forum?id=V5cQH7JbGo

Keywords: Federated learning, backdoor defense, isolated subspace training.

Compressor summary: Lockdown is a novel isolated subspace training method that effectively defends against backdoor attacks in federated learning, while reducing communication and model complexity.


Causal Effect Regularization: Automated Detection and Removal of Spurious Correlations

Abhinav Kumar, Amit Deshpande, Amit Sharma

https://openreview.net/forum?id=V5Oh7Aqfft

Keywords: Spurious Correlation, Out of Distribution Generalization

Compressor summary: The paper proposes a method to automatically identify and remove spurious attributes in classification datasets, improving generalization and reducing dependence on noisy factors.


Strategic Behavior in Two-sided Matching Markets with Prediction-enhanced Preference-formation

Stefania Ionescu, Yuhao Du, Kenneth Joseph, Aniko Hannak

https://openreview.net/forum?id=V5FNSilWiC

Keywords: matching markets, strategic behaviour, ML-based forecasting, recommender systems, adversarial attacks, agent-based modelling

Compressor summary: The paper introduces adversarial interaction attacks, where agents return to a matching market and act non-optimally with their matches to disrupt future predictions, and shows how these attacks can benefit returning agents and increase inequality.


Perceptual Kalman Filters: Online State Estimation under a Perfect Perceptual-Quality Constraint

Dror Freirich, Tomer Michaeli, Ron Meir

https://openreview.net/forum?id=V4hqq2NGTW

Keywords: Kalman filter, estimation theory, causal filtering, signal processing, distortion-perception tradeoff

Compressor summary: The text discusses the challenge of achieving perfect perceptual-quality in reconstructing temporal signals from corrupted or missing data under a causal filtering constraint, which may require ignoring new information and increasing MSE.


Temporal Conditioning Spiking Latent Variable Models of the Neural Response to Natural Visual Scenes

Gehua Ma, Runhao Jiang, Rui Yan, Huajin Tang

https://openreview.net/forum?id=V4YeOvsQfu

Keywords: neuroscience, neural coding, sensory neuroscience, visual coding, SNN, spiking neural networks, generative model, latent variable model, cognitive computational neuroscience, computational neuroscience

Compressor summary: TeCoS-LVM is a new spiking neural network model that can capture temporal dependencies in natural stimuli and produce realistic spike activities, outperforming current methods.


Effective Human-AI Teams via Learned Natural Language Rules and Onboarding

Hussein Mozannar, Jimin J Lee, Dennis Wei, Prasanna Sattigeri, Subhro Das, David Sontag

https://openreview.net/forum?id=V2yFumwo5B

Keywords: human-ai, collaboration, onboarding, region-discovery, LLM, data description

Compressor summary: The paper proposes learning natural language rules for human-AI collaboration based on data regions discovered by an algorithm.


A Path to Simpler Models Starts With Noise

Lesia Semenova, Harry Chen, Ronald Parr, Cynthia Rudin

https://openreview.net/forum?id=Uzi22WryyX

Keywords: Rashomon Set, Simplicity, Interpretable Machine Learning, Model Selection, Model Multiplicity

Compressor summary: This paragraph discusses how data noise affects the performance of different models and introduces a measure called pattern diversity to quantify the difference in predictions between them.


$k$-Means Clustering with Distance-Based Privacy

Alessandro Epasto, Vahab Mirrokni, Shyam Narayanan, Peilin Zhong

https://openreview.net/forum?id=UzUhiKACmS

Keywords: differential Privacy, k-means, k-median, clustering, distance-based privacy

Compressor summary: The paper proposes constant-approximate algorithms for Euclidean clustering with Distance-based privacy, which protect exact locations and outperform existing methods.


Guiding Large Language Models via Directional Stimulus Prompting

Zekun Li, Baolin Peng, Pengcheng He, Michel Galley, Jianfeng Gao, Xifeng Yan

https://openreview.net/forum?id=UvIN8oQ4uI

Keywords: Black-box Large Language Models, Directional Stimulus Prompting, Hint, Reinforcement learning, Prompt optimization

Compressor summary: The paragraph introduces Directional Stimulus Prompting, a framework that uses a small policy model to generate hints for large language models, guiding them towards specific outputs and improving their performance on various tasks.


UP-NeRF: Unconstrained Pose Prior-Free Neural Radiance Field

Injae Kim, Minhyuk Choi, Hyunwoo J. Kim

https://openreview.net/forum?id=UvBwXdL95b

Keywords: neural radiance field, pose estimation

Compressor summary: UP-NeRF optimizes NeRF with unconstrained images without pose priors by using surrogate tasks, a separate module for transient occluders, and improved pose estimation and depth supervision.


Bayesian Optimisation of Functions on Graphs

Xingchen Wan, Pierre Osselin, Henry Kenlay, Binxin Ru, Michael A Osborne, Xiaowen Dong

https://openreview.net/forum?id=UuNd9A6noD

Keywords: graphs, Bayesian optimisation, scalability

Compressor summary: The paper proposes a new Bayesian optimization framework that adapts to graph-structured functions and improves sample efficiency for large-scale graphs using local modeling and suitable kernels.


Polynomially Over-Parameterized Convolutional Neural Networks Contain Structured Strong Winning Lottery Tickets

Arthur da Cunha, Francesco D'Amore, Emanuele Natale

https://openreview.net/forum?id=UqYrYB3dp5

Keywords: lottery ticket hypothesis, convolutional neural network, network pruning, structured pruning, random subset sum

Compressor summary: The paper introduces a new mathematical tool to prove that random neural networks have efficient subnetworks, which can help understand the role of over-parameterization in deep learning.


Language Is Not All You Need: Aligning Perception with Language Models

Shaohan Huang, Li Dong, Wenhui Wang, Yaru Hao, Saksham Singhal, Shuming Ma, Tengchao Lv, Lei Cui, Owais Khan Mohammed, Barun Patra, Qiang Liu, Kriti Aggarwal, Zewen Chi, Johan Bjorck, Vishrav Chaudhary, Subhojit Som, Xia Song, Furu Wei

https://openreview.net/forum?id=UpN2wfrLec

Keywords: multimodal large language model

Compressor summary: KOSMOS-1 is a Multimodal Large Language Model that can perceive various modalities, learn in context, and follow instructions, achieving impressive performance on language and perception tasks without gradient updates or finetuning.


Extending the Design Space of Graph Neural Networks by Rethinking Folklore Weisfeiler-Lehman

Jiarui Feng, Lecheng Kong, Hao Liu, Dacheng Tao, Fuhai Li, Muhan Zhang, Yixin Chen

https://openreview.net/forum?id=UlJcZoawgU

Keywords: Graph neural network, expressive power, Folklore Weisfeiler-Lehman test.

Compressor summary: The paper proposes $(k,t)$-FWL+ and Neighborhood$^2$-FWL (N$^2$-FWL), which are more powerful and flexible graph neural network frameworks than MPNNS, with better performance on various tasks.


Textually Pretrained Speech Language Models

Michael Hassid, Tal Remez, Tu Anh Nguyen, Itai Gat, Alexis Conneau, Felix Kreuk, Jade Copet, Alexandre Défossez, Gabriel Synnaeve, Emmanuel Dupoux, Roy Schwartz, Yossi Adi

https://openreview.net/forum?id=UlHueVjAKr

Keywords: LLM, speech, generative, GSLM

Compressor summary: TWIST trains SpeechLMs using a warm-start from pretrained textual models, leading to better performance than cold-start methods, and introduces new benchmarks for evaluation.


Approximate Heavy Tails in Offline (Multi-Pass) Stochastic Gradient Descent

Krunoslav Lehman Pavasovic, Alain Durmus, Umut Simsekli

https://openreview.net/forum?id=UkPeUXML7s

Keywords: SGD, heavy-tails, wasserstein convergence

Compressor summary: The paper explores why stochastic gradient descent (SGD) shows heavy-tailed behavior in practical settings and investigates the role of offline SGD's stationary distribution in achieving power-law tails as more data points are added.


Inner-Outer Aware Reconstruction Model for Monocular 3D Scene Reconstruction

Yu-Kun Qiu, Guohao Xu, Wei-Shi Zheng

https://openreview.net/forum?id=UkAGqeWTuL

Keywords: 3D reconstruction

Compressor summary: IOAR is a 3D scene reconstruction method that uses a coarse-to-fine strategy to distinguish outer-surface, inner-surface, and surface voxels, resulting in more precise mesh predictions.


LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning

Atsuyuki Miyai, Qing Yu, Go Irie, Kiyoharu Aizawa

https://openreview.net/forum?id=UjtiLdXGMC

Keywords: out-of-distribution detection, vision-language foundation model, prompt learning

Compressor summary: The authors propose LoCoOp, a novel prompt learning method for few-shot OOD detection that uses CLIP's local features to remove ID-irrelevant information and achieve superior performance over existing methods.


Multi-Agent First Order Constrained Optimization in Policy Space

Youpeng Zhao, Yaodong Yang, Zhenbo Lu, Wengang Zhou, Houqiang Li

https://openreview.net/forum?id=UgomCjCWjC

Keywords: Safe Multi-agent Reinforcement Learning, constrained policy optimisation, first-order optimisation

Compressor summary: MAFOCOPS is a novel safety-aware method for multi-agent systems in MARL, which effectively addresses high performance and enforcing safety constraints by solving a constrained optimization problem in the policy space.


DAMEX: Dataset-aware Mixture-of-Experts for visual understanding of mixture-of-datasets

Yash Jain, Harkirat Behl, Zsolt Kira, Vibhav Vineet

https://openreview.net/forum?id=UgSSOpqvPI

Keywords: mixture-of-experts, moe, object detection, mixture of datasets, multiple datasets

Compressor summary: The paper proposes a method called DAMEX that improves object detection by learning to route different datasets to their corresponding experts, achieving state-of-the-art results on various datasets and scenarios.


When can Regression-Adjusted Control Variate Help? Rare Events, Sobolev Embedding and Minimax Optimality

Jose Blanchet, Haoxuan Chen, Yiping Lu, Lexing Ying

https://openreview.net/forum?id=UdrybSp67L

Keywords: Information-theoretic Lower Bounds, Sobolev Embedding Theorem, Quadrature Rule

Compressor summary: The paper explores how machine learning-based control variates can reduce the variance of Monte Carlo sampling for estimating moments of a Sobolev function, and shows that they can improve the estimation rate under certain conditions and not in the presence of rare events.


Double Gumbel Q-Learning

David Yu-Tung Hui, Aaron Courville, Pierre-Luc Bacon

https://openreview.net/forum?id=UdaTyy0BNB

Keywords: deep reinforcement learning, Q-Learning, TD-Learning with function approximation, extreme value theory, maximum-likelihood estimation, moment-matching

Compressor summary: Double Gumbel Q-Learning is a new algorithm that handles heteroscedastic noise in Deep Q-Learning and performs well on various control tasks.


MoCa: Measuring Human-Language Model Alignment on Causal and Moral Judgment Tasks

Allen Nie, Yuhui Zhang, Atharva Amdekar, Christopher J Piech, Tatsunori Hashimoto, Tobias Gerstenberg

https://openreview.net/forum?id=UdByCgCNdr

Keywords: cognitive science, causal reasoning, moral reasoning, dataset, language models

Compressor summary: The paragraph discusses a study that examines how well large language models make causal and moral judgments about text-based scenarios compared to human participants, using a dataset of stories from cognitive science papers annotated with factors influencing judgments.


Generating Images with Multimodal Language Models

Jing Yu Koh, Daniel Fried, Ruslan Salakhutdinov

https://openreview.net/forum?id=Uczck6TlSZ

Keywords: multimodal, vision-and-language, language models

Compressor summary: The proposed method fuses large language models with image encoder and decoder models for multimodal capabilities including image retrieval, novel image generation, and multimodal dialogue.


Identification of Nonlinear Latent Hierarchical Models

Lingjing Kong, Biwei Huang, Feng Xie, Eric Xing, Yuejie Chi, Kun Zhang

https://openreview.net/forum?id=Uc5yyiytR1

Keywords: Causal discovery, causal representation learning, latent variable models, causal structure learning, causal identifiability.

Compressor summary: This paper presents a method for identifying causal structures and latent variables in nonlinear latent hierarchical models using novel identifiability guarantees and an estimation procedure.


Active Negative Loss Functions for Learning with Noisy Labels

Xichen Ye, Xiaoqiang Li, Songmin Dai, Tong Liu, Yan Sun, Weiqin Tong

https://openreview.net/forum?id=Uafbv4rfJc

Keywords: noisy label learning, robust loss function, multiclass classification, computer vision

Compressor summary: The authors propose a new robust passive loss function called Normalized Negative Loss Functions (NNLFs) that improves the Active Passive Loss (APL) framework by focusing more on memorized clean samples and outperform state-of-the-art methods.


Model-enhanced Vector Index

Hailin Zhang, Yujing Wang, Qi Chen, Ruiheng Chang, Ting Zhang, Ziming Miao, Yingyan Hou, Yang Ding, Xupeng Miao, Haonan Wang, Bochen Pang, Yuefeng Zhan, Hao Sun, Weiwei Deng, Qi Zhang, Fan Yang, Xing Xie, Mao Yang, Bin CUI

https://openreview.net/forum?id=UZlAjSnmvB

Keywords: document retrieval, model-based index, dense retrieval, residual quantization

Compressor summary: The paper proposes MEVI, a differentiable index enhanced by deep generative models, which improves vector index search efficiency while maintaining performance on academic benchmarks.


Decompose Novel into Known: Part Concept Learning For 3D Novel Class Discovery

Tingyu Weng, Jun Xiao, Haiyong Jiang

https://openreview.net/forum?id=UYl9IIsjq7

Keywords: 3D point clouds, 3D recognition, part-based representation, unsupervised class discovery

Compressor summary: The paper proposes DNIK, a method to discover novel 3D classes by decomposing them into known parts and using a part relation encoding module (PRE) for better recognition.


Intensity Profile Projection: A Framework for Continuous-Time Representation Learning for Dynamic Networks

Alexander Modell, Ian Gallagher, Emma Ceccherini, Nick Whiteley, Patrick Rubin-Delanchy

https://openreview.net/forum?id=UXtLrsG4Rf

Keywords: dynamic networks, representation learning, spectral methods

Compressor summary: The paragraph introduces a new method to represent network data over time using continuous trajectories that satisfy structural and temporal coherence, with estimation theory providing error control and smoothing trade-offs.


Expressive Sign Equivariant Networks for Spectral Geometric Learning

Derek Lim, Joshua Robinson, Stefanie Jegelka, Haggai Maron

https://openreview.net/forum?id=UWd4ysACo4

Keywords: Eigenvectors, spectral, geometry, universal approximation, graph, equivariance, invariance

Compressor summary: The authors propose novel sign equivariant neural network architectures for tasks such as building orthogonally equivariant models and learning node positional encodings, which they show have better performance than sign invariant models.


PAC-Bayes Generalization Certificates for Learned Inductive Conformal Prediction

Apoorva Sharma, Sushant Veer, Asher Hancock, Heng Yang, Marco Pavone, Anirudha Majumdar

https://openreview.net/forum?id=URrUpcp6Qh

Keywords: Conformal Prediction, PAC Bayes, Generalization Theory

Compressor summary: The authors propose a framework for optimizing the efficiency of set-valued predictions based on PAC-Bayes theory, which allows using the entire calibration dataset and provides test-time coverage and efficiency guarantees.


Horospherical Decision Boundaries for Large Margin Classification in Hyperbolic Space

Xiran Fan, Chun-Hao Yang, Baba C. Vemuri

https://openreview.net/forum?id=URAZeoIC1q

Keywords: Large-margin clssifier, Hyperbolic space, Horosphere, SVM, Geodesically convex, Global optimility, Busemann function

Compressor summary: The paper introduces a new large margin classifier with horospherical decision boundaries for hyperbolic spaces, which optimizes geodesically convex problems and outperforms existing methods.


Fast Asymptotically Optimal Algorithms for Non-Parametric Stochastic Bandits

Dorian Baudry, Fabien Pesquerel, Rémy Degenne, Odalric-Ambrym Maillard

https://openreview.net/forum?id=UPo8vlZ0wQ

Keywords: Multi-Armed Bandits

Compressor summary: The paper proposes methods to approximate the infimum KL in non-parametric stochastic bandits, reducing computational and memory costs while maintaining regret guaranties.


Brain encoding models based on multimodal transformers can transfer across language and vision

Jerry Tang, Meng Du, Vy A. Vo, Vasudev Lal, Alexander Huth

https://openreview.net/forum?id=UPefaFqjNQ

Keywords: fMRI, neuroscience, encoding models, multimodal transformers, language, vision

Compressor summary: The study used multimodal transformers to train encoding models that can transfer across fMRI responses to stories and movies, revealing shared semantic dimensions in language and vision processing.


Class-Distribution-Aware Pseudo-Labeling for Semi-Supervised Multi-Label Learning

Ming-Kun Xie, Jia-Hao Xiao, Hao-Zhe Liu, Gang Niu, Masashi Sugiyama, Sheng-Jun Huang

https://openreview.net/forum?id=UOB1UgPjuG

Keywords: Semi-supervised multi-label learning, pseudo labeling.

Compressor summary: The paper proposes Class-Aware Pseudo-Labeling (CAP), which uses class-aware thresholds to control pseudo-label assignment and improve semi-supervised multi-label learning by aligning pseudo-label distribution with the true one.


The s-value: evaluating stability with respect to distributional shifts

Suyash Gupta, Dominik Rothenhaeusler

https://openreview.net/forum?id=UKtjq3dIs0

Keywords: Distributional Stability, Distributional Robustness, Distributional Shifts, Generalizability

Compressor summary: The paper proposes a new measure of instability for statistical parameters that accounts for both distributional changes and directional shifts, and demonstrates its usefulness in transfer learning and improving estimation accuracy.


Learning to Tokenize for Generative Retrieval

Weiwei Sun, Lingyong Yan, Zheng Chen, Shuaiqiang Wang, Haichao Zhu, Pengjie Ren, Zhumin Chen, Dawei Yin, Maarten de Rijke, Zhaochun Ren

https://openreview.net/forum?id=UKd6dpVGdu

Keywords: Information Retrieval, Document Retrieval, Generative Retrieval

Compressor summary: The paper proposes a novel document tokenization learning method, GenRet, which learns to encode document semantics into docids for generative retrieval models.


Simple and Asymmetric Graph Contrastive Learning without Augmentations

Teng Xiao, Huaisheng Zhu, Zhengyu Chen, Suhang Wang

https://openreview.net/forum?id=UK8mA3DRnb

Keywords: Contrastive Learning, Graph Representation Learning

Compressor summary: GraphACL is a simple algorithm for contrastive learning on graph-structured data that works well on both homophilic and heterophilic graphs without relying on prefabricated augmentations or homophily assumptions.


Provably Safe Reinforcement Learning with Step-wise Violation Constraints

Nuoya Xiong, Yihan Du, Longbo Huang

https://openreview.net/forum?id=UJ9o8wbB5U

Keywords: safe reinforcement learning, step-wise violation, reinforcement learning theory

Compressor summary: The paper proposes two algorithms for safe reinforcement learning problems with step-wise violation constraints, one for when safe actions are known and another for when they are not, achieving near-optimal performance in both violation and regret.


Fast Attention Over Long Sequences With Dynamic Sparse Flash Attention

Matteo Pagliardini, Daniele Paliotta, Martin Jaggi, François Fleuret

https://openreview.net/forum?id=UINHuKeWUa

Keywords: self-attention, large language models, transformers

Compressor summary: The paper proposes FlashAttention improvements that allow for more dynamic sparse attention patterns with no computational overhead and significant runtime speedup, while maintaining perplexity and increasing training speed for longer sequences in transformer language models.


Attacks on Online Learners: a Teacher-Student Analysis

Riccardo Giuseppe Margiotta, Sebastian Goldt, Guido Sanguinetti

https://openreview.net/forum?id=UHwmoJYwSV

Keywords: Adversarial attacks, data poisoning, online learning, optimal control, teacher-student setup, solvable model

Compressor summary: The paper investigates how attackers can manipulate online learning dynamics by perturbing data labels and studies the effects on learners with different architectures using both theory and experiments.


Asynchrony-Robust Collaborative Perception via Bird's Eye View Flow

Sizhe Wei, Yuxi Wei, Yue Hu, Yifan Lu, Yiqi Zhong, Siheng Chen, Ya Zhang

https://openreview.net/forum?id=UHIDdtxmVS

Keywords: Collaborative Perception; BEV Flow; Time Asynchronization

Compressor summary: CoBEVFlow is a system to align asynchronous messages in collaborative perception using bird's eye view flow, handling irregular and continuous time stamps without discretization, and improving performance on real-world datasets.


Segment Everything Everywhere All at Once

Xueyan Zou, Jianwei Yang, Hao Zhang, Feng Li, Linjie Li, Jianfeng Wang, Lijuan Wang, Jianfeng Gao, Yong Jae Lee

https://openreview.net/forum?id=UHBrWeFWlL

Keywords: Generaic segmentation, interactive segmentation, referring segmentation, multi-modality prompting.

Compressor summary: SEEM is an interactive model for image segmentation that unifies different prompts, learns from segmentation history, and encodes text queries and mask labels semantically.


MEMTO: Memory-guided Transformer for Multivariate Time Series Anomaly Detection

Junho Song, Keonwoo Kim, Jeonglyul Oh, Sungzoon Cho

https://openreview.net/forum?id=UFW67uduJd

Keywords: Multivariate time series, Anomaly detection

Compressor summary: MEMTO is a memory-guided Transformer that uses a novel memory module and bi-dimensional deviation-based detection criterion to achieve high anomaly detection performance on real-world multivariate time series data.


Can Language Models Solve Graph Problems in Natural Language?

Heng Wang, Shangbin Feng, Tianxing He, Zhaoxuan Tan, Xiaochuang Han, Yulia Tsvetkov

https://openreview.net/forum?id=UDqHhbqYJV

Keywords: large language models, graph reasoning, structured reasoning

Compressor summary: The authors introduce NLGraph, a natural language benchmark for graph-based problem solving, and evaluate GPT-3/4 on it, finding that LLMs have preliminary graph reasoning abilities but are brittle in complex scenarios and improve with instruction-based prompting approaches.


Time Series Kernels based on Nonlinear Vector AutoRegressive Delay Embeddings

Giovanni De Felice, John Y Goulermas, Vladimir Gusev

https://openreview.net/forum?id=UBUWFEwn7p

Keywords: Time Series, Kernel methods, NVAR processes, Dynamical systems, Reservoir Computing

Compressor summary: The authors propose a new kernel for time series analysis that is based on Nonlinear Vector AutoRegressive processes, simplifies hyperparameter setting, and performs well in classification tasks.


Cross-modal Active Complementary Learning with Self-refining Correspondence

Yang Qin, Yuan Sun, Dezhong Peng, Joey Tianyi Zhou, Xi Peng, Peng Hu

https://openreview.net/forum?id=UBBeUjTja8

Keywords: Cross-modal learning, Image-text matching, Noisy correspondence.

Compressor summary: The text describes a new framework called CRCL that improves image-text matching by addressing challenges caused by noisy correspondence between visual and textual modalities using active complementary learning and self-refining correction.


A Unified Generalization Analysis of Re-Weighting and Logit-Adjustment for Imbalanced Learning

Zitai Wang, Qianqian Xu, Zhiyong Yang, Yuan He, Xiaochun Cao, Qingming Huang

https://openreview.net/forum?id=UAow2kPsYP

Keywords: Imbalanced Learning, Re-weighting, Logit Adjustment, Genralization Analysis

Compressor summary: The paragraph discusses imbalanced real-world datasets and proposes a novel technique called data-dependent contraction to analyze modified losses for such datasets, along with a fine-grained generalization bound, a learning algorithm, and empirical validation.


Unsupervised Graph Neural Architecture Search with Disentangled Self-Supervision

Zeyang Zhang, Xin Wang, Ziwei Zhang, Guangyao Shen, Shiqi Shen, Wenwu Zhu

https://openreview.net/forum?id=UAFa5ZhR85

Keywords: Graph Neural Architecture Search, Unsupervised Learning, Self-supervised Learning

Compressor summary: The paper proposes a novel model for unsupervised graph neural architecture search, which can discover optimal architectures by disentangling latent graph factors and using self-supervised training and contrastive search.


A Hierarchical Spatial Transformer for Massive Point Samples in Continuous Space

Wenchong He, Zhe Jiang, Tingsong Xiao, Zelin Xu, Shigang Chen, Ronald Fick, MILES D MEDINA, Christine Angelini

https://openreview.net/forum?id=U9zRgpgdFI

Keywords: Spatial representation learning, transformer, quadtree, efficiency

Compressor summary: The paper introduces a new transformer model for massive spatial point data, addressing challenges such as long-range dependencies, non-uniform distributions, and high computational costs, by using a hierarchical structure, coarse approximation, and uncertainty estimation.


Understanding Neural Network Binarization with Forward and Backward Proximal Quantizers

Yiwei Lu, Yaoliang Yu, Xinlin Li, Vahid Partovi Nia

https://openreview.net/forum?id=U6fp6IUBdr

Keywords: Neural network quantization, Model compression, Conditional gradient algorithm

Compressor summary: The paper introduces ProxConnect++, a generalization of BinaryConnect with forward-backward quantizers that enables principled binarization methods for neural networks without relying on heuristics or training tricks, and shows its effectiveness in image classification tasks.


Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs

Dongsheng Ding, Chen-Yu Wei, Kaiqing Zhang, Alejandro Ribeiro

https://openreview.net/forum?id=U6bhCLSPun

Keywords: Constrained Markov decision processes, policy gradient primal-dual methods, non-convex saddle-point problem, last-iterate convergence, entropy regularization, optimistic gradient

Compressor summary: The paper proposes two new methods for solving constrained Markov decision processes (MDPs) that have better convergence properties than existing Lagrangian-based policy search methods.


Masked Two-channel Decoupling Framework for Incomplete Multi-view Weak Multi-label Learning

Chengliang Liu, Jie Wen, Yabo Liu, Chao Huang, Zhihao Wu, Xiaoling Luo, Yong Xu

https://openreview.net/forum?id=U4pFV192JQ

Keywords: Incomplete Multi-view Weak Multi-label Learning, Multi-view learning, Multi-label Classification

Compressor summary: The paper proposes a new deep neural network framework for incomplete multi-view weak multi-label learning, which decouples single-channel view-level representation into shared and view-specific representations, and uses cross-channel contrastive loss and label-guided graph regularization to enhance the features and preserve geometric structure.


Towards Efficient and Accurate Winograd Convolution via Full Quantization

Chen Tianqi, Weixiang Xu, Weihan Chen, Peisong Wang, Jian Cheng

https://openreview.net/forum?id=U4WTG06Yu3

Keywords: Winograd Convolution, Quantization

Compressor summary: The paper introduces PAW, which optimizes transformation procedures for Winograd convolution with post-training quantization and FSQ, a hardware-friendly method that balances range differences in the domain, improving accuracy and efficiency.


Topological RANSAC for instance verification and retrieval without fine-tuning

Guoyuan An, Ju-hyeong Seon, Inkyu An, Yuchi Huo, Sung-eui Yoon

https://openreview.net/forum?id=U1Kr8FTyhQ

Keywords: Landmarks retrieval, non-fine-tuning, spatial verification, explainable AI, hypothesis and test

Compressor summary: The paper proposes a new image retrieval method that uses a topological model instead of a spatial one to improve performance and overcome limitations in recognizing features, while retaining high explainability and being lightweight.


Long Sequence Hopfield Memory

Hamza Tahir Chaudhry, Jacob A Zavatone-Veth, Dmitry Krotov, Cengiz Pehlevan

https://openreview.net/forum?id=Tz2uONpgpy

Keywords: Sequence Recall, Dense Associative Memory, Memory Capacity, Hopfield Networks, Biological Motor Control

Compressor summary: The authors propose a new sequence memory model that can store longer sequences by using nonlinear interactions and a novel recall rule, and discuss its biological implications for motor neuroscience.


An Efficient Doubly-Robust Test for the Kernel Treatment Effect

Diego Martinez-Taboada, Aaditya Ramdas, Edward Kennedy

https://openreview.net/forum?id=TyLjNSbSOe

Keywords: kernel treatment effect, causal inference, maximum mean discrepancy

Compressor summary: The authors propose a new kernel-based test for distributional effects of binary treatments that is computationally efficient and has valid type-I error.


Tools for Verifying Neural Models' Training Data

Dami Choi, Yonadav G Shavit, David Duvenaud

https://openreview.net/forum?id=TwLHB8sKme

Keywords: Large Scale Learning, ML Security, AI Governance

Compressor summary: The authors propose and explore efficient strategies for verifying the provenance of large neural models using "Proof-of-Training-Data" protocols that can detect various attacks, such as those based on harmful or beneficial data sources.


Optimal Parameter and Neuron Pruning for Out-of-Distribution Detection

Chao Chen, Zhihang Fu, Kai Liu, Ze Chen, Mingyuan Tao, Jieping Ye

https://openreview.net/forum?id=TtCPFN5fhO

Keywords: Out-of-Distribution Detection, Parameter Sensitivity, Parameter Pruning, Neuron Pruning

Compressor summary: The paper proposes an efficient method called OPNP to detect out-of-distribution samples in machine learning models without requiring extra training data or resources, and demonstrates its effectiveness on various tasks and architectures.


GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning

Haiteng Zhao, Shengchao Liu, Chang Ma, Hannan Xu, Jie Fu, Zhi-Hong Deng, Lingpeng Kong, Qi Liu

https://openreview.net/forum?id=Tt6DrRCgJV

Keywords: Instruction, Molecule, Zero Shot, Graph, Language Model

Compressor summary: GIMLET is a model that uses language instructions and graph structures to accomplish molecule-related tasks without expensive lab experiments, outperforming existing methods in instruction-based zero-shot learning.


Meta-Adapter: An Online Few-shot Learner for Vision-Language Model

Cheng Cheng, Lin Song, Ruoyi Xue, Hang Wang, Hongbin Sun, Yixiao Ge, Ying Shan

https://openreview.net/forum?id=Ts0d8PvTeB

Keywords: Few-shot Learning; Vision-Language Model Adaption

Compressor summary: CLIP is a powerful vision-language pre-training method for image recognition, but it needs offline fine-tuning and can overfit. The proposed Meta-Adapter improves CLIP's few-shot learning capabilities online, without extra fine-tuning, achieving high efficiency and performance.


SatLM: Satisfiability-Aided Language Models Using Declarative Prompting

Xi Ye, Qiaochu Chen, Isil Dillig, Greg Durrett

https://openreview.net/forum?id=TqW5PL1Poi

Keywords: Reasoning, Chain-of-thought, Logical Reasoning, Arithmetic Reasoning, Prompting, In-Context Learning, Large Language Model

Compressor summary: The paper proposes SatLM, which uses an LLM to generate a declarative task specification and an automated theorem prover to derive the answer, improving reasoning capabilities for constraint solving problems.


Predict-then-Calibrate: A New Perspective of Robust Contextual LP

Chunlin Sun, Linyu Liu, Xiaocheng Li

https://openreview.net/forum?id=TnTDiCppx5

Keywords: Uncertainty Quantification, Contextual LP, Robust Optimization, Distributionally Robust Optimization

Compressor summary: The paper proposes a method called predict-then-calibrate for solving optimization problems with contextual information, which separates the prediction and uncertainty quantification steps and allows for more flexible machine learning models and improved performance.


Kernel Stein Discrepancy thinning: a theoretical perspective of pathologies and a practical fix with regularization

Clement Benard, Brian Staber, Sébastien Da Veiga

https://openreview.net/forum?id=TjgG4UT62W

Keywords: Bayesian inference, Markov chain Monte Carlo, kernelized Stein discrepancy, Stein thinning, kernel methods

Compressor summary: Stein thinning is a post-processing method for MCMC outputs that minimizes kernelized Stein discrepancy, but has some drawbacks; the paper analyzes these issues and proposes an improved algorithm with theoretical guarantees and experimental results.


Exact recovery and Bregman hard clustering of node-attributed Stochastic Block Model

Maximilien Dreveton, Felipe Schreiber Fernandes, Daniel R. Figueiredo

https://openreview.net/forum?id=TjJJmcHw9p

Keywords: community detection, stochastic block model, bregman divergence

Compressor summary: The text describes a new approach to cluster nodes in networks based on their connections and attributes, using an information-theoretic criterion and an iterative algorithm that performs better than existing methods.


Loss Dynamics of Temporal Difference Reinforcement Learning

Blake Bordelon, Paul Masset, Henry Kuo, Cengiz Pehlevan

https://openreview.net/forum?id=Tj0eXVPnRX

Keywords: Reinforcement Learning, Statistical Mechanics, Stochastic Gradient Descent

Compressor summary: The paper uses statistical physics concepts to study the effects of parameter choice and feature representation on the learning dynamics of reinforcement learning models with linear function approximators and sparse feedback.


Bayesian target optimisation for high-precision holographic optogenetics

Marcus Triplett, Marta Agnieszka Gajowa, Hillel Adesnik, Liam Paninski

https://openreview.net/forum?id=TiFMYdQiqp

Keywords: Neuroscience, neural stimulation, optogenetics, calcium imaging

Compressor summary: Bayesian target optimization is a novel computational approach that reduces off-target stimulation in two-photon optogenetics by modeling neural responses and optimizing laser powers and locations for precise stimulation of desired activity patterns.


Sparse Modular Activation for Efficient Sequence Modeling

Liliang Ren, Yang Liu, Shuohang Wang, Yichong Xu, Chenguang Zhu, ChengXiang Zhai

https://openreview.net/forum?id=TfbzX6I14i

Keywords: Sequence Modeling, Modularity, Sparsity, Attention Mechanism, State Space Model, Mixture of Experts, Neural Network, Transformer

Compressor summary: Sparse Modular Activation (SMA) improves sequence modeling efficiency by allowing neural networks to dynamically activate sub-modules, enabling linear inference complexity and better quality-efficiency trade-offs.


4M: Massively Multimodal Masked Modeling

David Mizrahi, Roman Bachmann, Oguzhan Fatih Kar, Teresa Yeo, Mingfei Gao, Afshin Dehghan, Amir Zamir

https://openreview.net/forum?id=TegmlsD8oQ

Keywords: multimodal learning, multitask learning, representation learning, transfer learning, foundation models, generative models, computer vision

Compressor summary: The paper introduces 4M, a multimodal training scheme that trains a unified Transformer encoder-decoder on various input/output modalities, achieving versatile models for computer vision tasks.


The Rashomon Importance Distribution: Getting RID of Unstable, Single Model-based Variable Importance

Jon Donnelly, Srikar Katta, Cynthia Rudin, Edward P Browne

https://openreview.net/forum?id=TczT2jiPT5

Keywords: Rashomon Effect, Variable Importance, XAI, Stability, Interpretable Machine Learning

Compressor summary: The authors propose a new variable importance framework that quantifies the importance of a variable across all good models and is stable across data distribution, which can handle complex simulation setups and estimate true importance for real-world case studies like predicting HIV load.


L-CAD: Language-based Colorization with Any-level Descriptions using Diffusion Priors

Zheng Chang, Shuchen Weng, Peixuan Zhang, Yu Li, Si Li, Boxin Shi

https://openreview.net/forum?id=TcmjewOAd1

Keywords: Colorization, Language-based generation, Diffusion model

Compressor summary: The paper proposes a model that can colorize images using natural language descriptions of any level of detail, handling ambiguity and preserving local structures, while achieving better results than previous methods.


BERT Lost Patience Won't Be Robust to Adversarial Slowdown

Zachary Coalson, Gabriel Ritter, Rakesh B Bobba, Sanghyun Hong

https://openreview.net/forum?id=TcG8jhOPdv

Keywords: Efficient Methods for NLP; Multi-exit Language Models; Adversarial Slowdown

Compressor summary: The paper evaluates the robustness of multi-exit language models to an adversarial slowdown attack called WAFFLE, which generates natural text perturbations that reduce computational savings and require further research on efficient yet robust models.


Fast Trainable Projection for Robust Fine-tuning

Junjiao Tian, Yen-Cheng Liu, James Smith, Zsolt Kira

https://openreview.net/forum?id=Tb7np0MInj

Keywords: fine-tuning, transfer learning, regularization

Compressor summary: The paper introduces Fast Trainable Projection, a new projection-based fine-tuning algorithm that improves efficiency and scalability for robust fine-tuning of pre-trained models in various vision tasks.


MIM4DD: Mutual Information Maximization for Dataset Distillation

Yuzhang Shang, Zhihang Yuan, Yan Yan

https://openreview.net/forum?id=TZtw5YgxTE

Keywords: Dataset Distillation

Compressor summary: MIM4DD uses mutual information to measure and optimize the shared information between synthetic and real datasets in dataset distillation, improving their performance.


A Variational Perspective on High-Resolution ODEs

Hoomaan Maskan, Konstantinos C. Zygalakis, Alp Yurtsever

https://openreview.net/forum?id=TXq8PCRSoY

Keywords: Nesterov's accelerated gradient, gradient descent, Lyapunov function, gradient norm minimization, rate-matching, stochastic variance reduction, stochastic gradient descent, noisy gradient

Compressor summary: The text discusses a novel variational approach to unconstrained minimization that leads to faster convergence, interprets Nesterov's method as a high-resolution ODE discretization, and proposes a stochastic method for noisy gradients.


Improved Algorithms for Stochastic Linear Bandits Using Tail Bounds for Martingale Mixtures

Hamish Flynn, David Reeb, Melih Kandemir, Jan Peters

https://openreview.net/forum?id=TXoZiUZywf

Keywords: Linear bandits, confidence sequences, martingales, convex optimization, cumulative regret, regret analysis

Compressor summary: The paper proposes new algorithms with better regret guarantees for the stochastic linear bandit problem using tailored confidence sequences for reward function estimation and action selection.


Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced Datasets

Zhang-Wei Hong, Aviral Kumar, Sathwik Karnik, Abhishek Bhandwaldar, Akash Srivastava, Joni Pajarinen, Romain Laroche, Abhishek Gupta, Pulkit Agrawal

https://openreview.net/forum?id=TW99HrZCJU

Keywords: offline reinforcement learning, reinforcement learning, sampling, experience replay

Compressor summary: Our method improves offline RL by using importance sampling weights to emulate data from an optimal policy, addressing the challenge of distributional mismatch due to imbalanced datasets.


Fair Streaming Principal Component Analysis: Statistical and Algorithmic Viewpoint

Junghyun Lee, Hanseul Cho, Se-Young Yun, Chulhee Yun

https://openreview.net/forum?id=TW3ipYdDQG

Keywords: streaming, PCA, memory-limited, fair representation, online learning

Compressor summary: The paper proposes a new notion of fair principal component analysis (PCA) called PAFO-learnability and a memory-efficient algorithm for fair streaming PCA called fair noisy power method (FNPM).


Bounce: Reliable High-Dimensional Bayesian Optimization for Combinatorial and Mixed Spaces

Leonard Papenmeier, Luigi Nardi, Matthias Poloczek

https://openreview.net/forum?id=TVD3wNVH9A

Keywords: Bayesian optimization, global optimization, Gaussian process, combinatorial optimization, high-dimensional

Compressor summary: Bounce is a new algorithm for optimizing black-box functions with mixed and combinatorial input spaces, which outperforms existing methods in high-dimensional problems.


Correlative Information Maximization: A Biologically Plausible Approach to Supervised Deep Neural Networks without Weight Symmetry

Bariscan Bozkurt, Cengiz Pehlevan, Alper Tunga Erdogan

https://openreview.net/forum?id=TUGoUNkccV

Keywords: Correlative information maximization, Biologically-plausible learning, Multi-compartment neural model

Compressor summary: The authors propose a new approach for signal propagation in biological neural networks that addresses concerns about the plausibility of conventional artificial neural networks and backpropagation algorithm, and provides a natural resolution to the weight symmetry problem.


NAP: Neural 3D Articulated Object Prior

Jiahui Lei, Congyue Deng, Bokui Shen, Leonidas Guibas, Kostas Daniilidis

https://openreview.net/forum?id=TTkklyFv7e

Keywords: 3D articulated objects, diffusion models, generative models

Compressor summary: NAP is a 3D deep generative model that synthesizes articulated objects using an innovative articulation tree/graph parameterization and diffusion-denoising probabilistic model.


Scalarization for Multi-Task and Multi-Domain Learning at Scale

Amelie Royer, Tijmen Blankevoort, Babak Ehteshami Bejnordi

https://openreview.net/forum?id=TSuq3debnD

Keywords: multitask, multidomain, optimization, population based training

Compressor summary: The paragraph discusses how training a single model on multiple inputs and outputs improves efficiency and transfer learning, but optimizing such networks is challenging, and proposes a population-based training method to find optimal scalarization weights.


Context Shift Reduction for Offline Meta-Reinforcement Learning

Yunkai Gao, Rui Zhang, Jiaming Guo, Fan Wu, Qi Yi, Shaohui Peng, Siming Lan, Ruizhi Chen, Zidong Du, Xing Hu, Qi Guo, Ling Li, Yunji Chen

https://openreview.net/forum?id=TStMZH3Xqx

Keywords: offline meta-reinforcement learning, offline reinforcement learning, meta-reinforcement learning

Compressor summary: The paper proposes a new approach called CSRO to address the context shift problem in offline meta-reinforcement learning by reducing the influence of policy in context during both training and testing phases using mutual information representation learning and non-prior context collection strategy.


BayesTune: Bayesian Sparse Deep Model Fine-tuning

Minyoung Kim, Timothy Hospedales

https://openreview.net/forum?id=TRuqrVsmZK

Keywords: Parameter-efficient Foundation model fine-tuning, Bayesian methods, Stochastic-Gradient MCMC

Compressor summary: The paper proposes a new Bayesian sparse fine-tuning algorithm that uses a Laplace prior to select which parameters of foundation models need to be updated for specific tasks, achieving significant improvement over existing methods on NLP and vision benchmarks.


GLOBER: Coherent Non-autoregressive Video Generation via GLOBal Guided Video DecodER

Mingzhen Sun, Weining Wang, Zihan Qin, Jiahui Sun, Sihan Chen, Jing Liu

https://openreview.net/forum?id=TRbklCR2ZW

Keywords: Video Generation, Video Autoencoder, Diffusion Probabilistic Model

Compressor summary: The GLOBER method generates coherent videos using a novel non-autoregressive approach that first creates global features and then synthesizes video frames based on them.


Neural Injective Functions for Multisets, Measures and Graphs via a Finite Witness Theorem

Tal Amir, Steven J. Gortler, Ilai Avni, Ravina Ravina, Nadav Dym

https://openreview.net/forum?id=TQlpqmCeMe

Keywords: Equivariant Neural Networks, Universal approximation, Geometric deep learning, multiset learning, injective multiset functions, learning on measures. WL test

Compressor summary: This paper shows that using non-polynomial activation in neural networks creates injective multiset functions, filling a gap between theory and practice, and also provides new results for graph neural networks and multiset functions approximation.


Parameter and Computation Efficient Transfer Learning for Vision-Language Pre-trained Models

Qiong Wu, Wei Yu, Yiyi Zhou, Shubin Huang, Xiaoshuai Sun, Rongrong Ji

https://openreview.net/forum?id=TPeAmxwPK2

Keywords: vision and language, parameter and computation efficient transfer learning

Compressor summary: The paper proposes a dynamic architecture skipping approach to achieve efficient transfer learning for vision-language models by observing module significance with reinforcement learning and reducing redundant parameters and computation.


Multi-task Graph Neural Architecture Search with Task-aware Collaboration and Curriculum

Yijian Qin, Xin Wang, Ziwei Zhang, Hong Chen, Wenwu Zhu

https://openreview.net/forum?id=TOxpAwp0VE

Keywords: graph neural network, neural architecture search, multi-task learning

Compressor summary: The paper proposes a novel method called MTGC3 that can automatically design optimal graph neural network architectures for multiple tasks simultaneously by learning their collaborative relationships and using a task-wise curriculum training strategy.


Information Geometry of the Retinal Representation Manifold

Xuehao Ding, Dongsoo Lee, Joshua Brendan Melander, George Sivulka, Surya Ganguli, Stephen Baccus

https://openreview.net/forum?id=TNLO8KNFFZ

Keywords: neural coding, theoretical neuroscience, stochastic methods, neural networks

Compressor summary: The study uses a neural network model of salamander retinal cells to show that retinal noise correlations limit information transmission in natural scenes and that population coding benefits from complementary coding.


Learning Rate Free Bayesian Inference in Constrained Domains

Louis Sharrock, Lester Mackey, Christopher Nemeth

https://openreview.net/forum?id=TNAGFUcSP7

Keywords: Bayesian Inference, Particle Based Variational Inference, Sampling, Wasserstein Gradient Descent, Coin Betting, Constrained Domains

Compressor summary: The paper presents new particle-based algorithms for sampling on constrained domains that don't require learning rate adjustments and are based on convex optimization ideas and probability measure space perspective.


CoDet: Co-occurrence Guided Region-Word Alignment for Open-Vocabulary Object Detection

Chuofan Ma, Yi Jiang, Xin Wen, Zehuan Yuan, XIAOJUAN QI

https://openreview.net/forum?id=TKjX41IP7n

Keywords: Open-vocabulary Object Detection; Object-level Vision-Language Pretraining

Compressor summary: The paper introduces CoDet, a new method for open-vocabulary object detection that improves region-word alignment by grouping images with shared caption concepts and leveraging visual similarities.


White-Box Transformers via Sparse Rate Reduction

Yaodong Yu, Sam Buchanan, Druv Pai, Tianzhe Chu, Ziyang Wu, Shengbang Tong, Benjamin David Haeffele, Yi Ma

https://openreview.net/forum?id=THfl8hdVxH

Keywords: white-box deep neural networks, representation learning, transformer, sparse coding

Compressor summary: The paper argues that representation learning aims to compress and transform data distributions using low-dimensional Gaussian mixtures, and shows how transformer blocks achieve this using sparse rate reduction as a unified objective.


Uniform Convergence with Square-Root Lipschitz Loss

Lijia Zhou, Zhen Dai, Frederic Koehler, Nathan Srebro

https://openreview.net/forum?id=TEpRn67828

Keywords: Uniform Convergence, Square-Root Lipschitz, Benign Overfitting, Minimal Norm Interpolation, Phase Retrieval, ReLU Regression, Matrix Sensing

Compressor summary: The paper provides uniform convergence guarantees for Gaussian data using Radamacher complexity and square-root-Lipschitz losses, generalizing previous smoothness-based results and covering non-smooth loss functions like phase retrieval and ReLU regression.


REx: Data-Free Residual Quantization Error Expansion

Edouard YVINEC, Arnaud Dapogny, Matthieu Cord, Kevin Bailly

https://openreview.net/forum?id=TDS3kqRteY

Keywords: deep learning, quantization, compression, acceleration, data-free

Compressor summary: REx is a data-free quantization method for DNNs that adapts to different devices and bit widths, improving accuracy-speed trade-offs and solving the outlier problem in large language models.


GEQ: Gaussian Kernel Inspired Equilibrium Models

Mingjie Li, Yisen Wang, Zhouchen Lin

https://openreview.net/forum?id=TBOfDCX4Gz

Keywords: equilibirum models, neural networks

Compressor summary: The paper proposes GEQ, a new equilibrium model that uses Gaussian kernels to capture nonlinear feature dependencies in input data, improving performance over traditional OptEqs models.


Curve Your Enthusiasm: Concurvity Regularization in Differentiable Generalized Additive Models

Julien Niklas Siems, Konstantin Ditschuneit, Winfried Ripken, Alma Lindborg, Maximilian Schambach, Johannes Otterbach, Martin Genzel

https://openreview.net/forum?id=TAIYBdRb3C

Keywords: Interpretable Machine Learning, Generalized Additive Models, Concurvity, Multicollinearity, Regularization, Time-Series Forecasting, Interpretability

Compressor summary: The paper introduces a regularizer to improve interpretability of generalized additive models by reducing feature dependencies, and demonstrates its effectiveness on synthetic and real datasets.


Mass-Producing Failures of Multimodal Systems with Language Models

Shengbang Tong, Erik Jones, Jacob Steinhardt

https://openreview.net/forum?id=T6iiOqsGOh

Keywords: safety, red-teaming, robustness, explainability, failures, multimodal models, vision-language, natural-language explanations

Compressor summary: MultiMon is a system that automatically finds and describes systematic failures in multimodal models using natural language.


UNSSOR: Unsupervised Neural Speech Separation by Leveraging Over-determined Training Mixtures

Zhong-Qiu Wang, Shinji Watanabe

https://openreview.net/forum?id=T5h69frFF7

Keywords: Speech separation, microphone array processing, deep learning

Compressor summary: UNSSOR is an algorithm that uses over-determined mixtures to separate speakers in noisy recordings with deep neural networks and linear filters.


PUCA: Patch-Unshuffle and Channel Attention for Enhanced Self-Supervised Image Denoising

Hyemi Jang, Junsung Park, Dahuin Jung, Jaihyun Lew, Ho Bae, Sungroh Yoon

https://openreview.net/forum?id=T3SstRu5fq

Keywords: self-supervised image denoising, low-level vision

Compressor summary: PUCA is a novel J-invariant U-Net architecture for self-supervised image denoising that leverages patch-unshuffle/shuffle and dilated attention blocks to achieve state-of-the-art performance.


Connecting Certified and Adversarial Training

Yuhao Mao, Mark Niklas Mueller, Marc Fischer, Martin Vechev

https://openreview.net/forum?id=T2lM4ohRwb

Keywords: Certified Training, Certified Robustness, Adversarial Robustness, Robustness Verification

Compressor summary: TAPS is a new certified training method that combines IBP and PGD to optimize worst-case loss approximations, improving both certification and standard accuracies.


Precision-Recall Divergence Optimization for Generative Modeling with GANs and Normalizing Flows

Alexandre Verine, benjamin negrevergne, Muni Sreenivas Pydi, Yann Chevaleyre

https://openreview.net/forum?id=SzYHu7EIwZ

Keywords: Generative Models, Precision, Recall, Optimization, f-Divergeces

Compressor summary: The paper proposes a new method for training generative models that optimizes a user-defined trade-off between image quality (precision) and diversity (recall), based on a novel family of $f$-divergences called PR-divergences.


Implicit Variational Inference for High-Dimensional Posteriors

Anshuk Uppal, Kristoffer Stensbo-Smidt, Wouter Boomsma, Jes Frellsen

https://openreview.net/forum?id=Sxu7xlUJGx

Keywords: Implicit models, Variational Inference, Bayesian Deep Learning, Bayesian Inference, Generative Modelling

Compressor summary: The paper proposes a new method for Bayesian models using neural samplers with implicit distributions to capture complex posteriors in high-dimensional spaces, enabling better performance in large Bayesian neural networks.


Offline Reinforcement Learning for Mixture-of-Expert Dialogue Management

Dhawal Gupta, Yinlam Chow, Azamat Tulepbergenov, Mohammad Ghavamzadeh, Craig Boutilier

https://openreview.net/forum?id=SxXN3kNTsV

Keywords: Reinforcement Learning, Mixture of Experts, Dialogue Management

Compressor summary: The text discusses challenges and solutions for using reinforcement learning in developing conversational chatbots, focusing on combining mixture-of-expert language models to reduce action space and improve dialogue management.


DeepSimHO: Stable Pose Estimation for Hand-Object Interaction via Physics Simulation

Rong Wang, Wei Mao, Hongdong Li

https://openreview.net/forum?id=SxVHyYavHy

Keywords: hand-object pose estimation, physics simulation

Compressor summary: The paper proposes DeepSimHO, a method that combines deep learning with physics simulation to improve 3D pose estimation of hands interacting with objects in a single image.


The Distortion of Binomial Voting Defies Expectation

Yannai Gonczarowski, Gregory Kehne, Ariel D. Procaccia, Ben Schiffer, Shirley Zhang

https://openreview.net/forum?id=Sv5bo2StIx

Keywords: computational social choice, statistics, distortion

Compressor summary: The paper explores how well voting rules perform on average with different preference distributions, proposing a new rule called binomial voting that works well across various situations.


Scaling Up Differentially Private LASSO Regularized Logistic Regression via Faster Frank-Wolfe Iterations

Edward Raff, Amol Ashish Khanna, Fred Lu

https://openreview.net/forum?id=SuvDnzrKCo

Keywords: Sparsity, Differential Privacy, Regression

Compressor summary: The paper proposes an algorithm for training private regression models on sparse data, which can greatly reduce the runtime compared to existing methods.


Topological Parallax: A Geometric Specification for Deep Perception Models

Abraham David Smith, Michael J. Catanzaro, Gabrielle Angeloro, Nirav Patel, Paul Bendich

https://openreview.net/forum?id=SthlUe5xDP

Keywords: topological data analysis, persistent homology, convexity, AI safety, interpolation

Compressor summary: Topological parallax is a tool that compares AI models and datasets to ensure their geometric similarity, which is important for safety and robustness.


Dataset Diffusion: Diffusion-based Synthetic Data Generation for Pixel-Level Semantic Segmentation

Quang Ho Nguyen, Truong Tuan Vu, Anh Tuan Tran, Khoi Nguyen

https://openreview.net/forum?id=StD4J5ZlI5

Keywords: Deep learning; Diffusion Models; Semantic Segmentation; Text-to-Image

Compressor summary: The paragraph describes a novel method for generating pixel-level semantic segmentation labels using the text-to-image generative model Stable Diffusion, which reduces the need for labor-intensive pixel-wise annotation and outperforms current approaches.


Cappy: Outperforming and Boosting Large Multi-Task LMs with a Small Scorer

Bowen Tan, Yun Zhu, Lijuan Liu, Eric Xing, Zhiting Hu, Jindong Chen

https://openreview.net/forum?id=Srt1hhQgqa

Keywords: multi-task, large language models, pretrain model

Compressor summary: Cappy is a small pretrained scorer that improves the performance and efficiency of multi-task large language models without requiring their finetuning or access to their parameters.


Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language Model

Zirui Liu, Guanchu Wang, Shaochen Zhong, Zhaozhuo Xu, Daochen Zha, Ruixiang Tang, Zhimeng Jiang, Kaixiong Zhou, Vipin Chaudhary, Shuai Xu, Xia Hu

https://openreview.net/forum?id=SquMNyrk1O

Keywords: memory-efficient tuning, language model, transformers

Compressor summary: The authors propose a new unbiased estimator called \sas\ for reducing memory usage in large pre-trained language models, which can achieve significant memory reduction and improve downstream task performance.


Persuading Farsighted Receivers in MDPs: the Power of Honesty

Martino Bernasconi, Matteo Castiglioni, Alberto Marchesi, Mirco Mutti

https://openreview.net/forum?id=SqTUGq0R7j

Keywords: Bayesian Persuasion, MPD, information design, signaling

Compressor summary: The paper proposes a new class of optimal signaling schemes for Bayesian persuasion that consider the receiver's future rewards and can be computed efficiently using a promise-form representation.


Solving a Class of Non-Convex Minimax Optimization in Federated Learning

Xidong Wu, Jianhui Sun, Zhengmian Hu, Aidong Zhang, Heng Huang

https://openreview.net/forum?id=SpStmVboGy

Keywords: Federated Learning, Non-Convex Optimization, Minimax Optimization

Compressor summary: The paper proposes FL algorithms for federated nonconvex minimax optimization problems, reducing communication complexity and achieving similar performance to centralized methods in some settings.


On Generalization Bounds for Projective Clustering

Maria Sofia Bucarelli, Matilde Fjeldsø Larsen, Chris Schwiegelshohn, Mads Toftrup

https://openreview.net/forum?id=Sp0yOBfelp

Keywords: Subspace Clustering, Learning Theory, Clustering, Error bounds

Compressor summary: The paper studies learning bounds for clustering problems with center-based and subspace-based objectives and provides near-optimal convergence rates for both cases.


Robust and Actively Secure Serverless Collaborative Learning

Nicholas Franzese, Adam Dziedzic, Christopher A. Choquette-Choo, Mark R. Thomas, Muhammad Ahmad Kaleem, Stephan Rabanser, Congyu Fang, Somesh Jha, Nicolas Papernot, Xiao Wang

https://openreview.net/forum?id=SouroWC5Un

Keywords: collaborative learning, robust aggregation, secure machine learning

Compressor summary: The paper proposes a peer-to-peer learning scheme that is secure from malicious servers and robust to malicious clients, and transforms any compatible algorithm for aggregation of model updates in such settings.


ARTree: A Deep Autoregressive Model for Phylogenetic Inference

Tianyu Xie, Cheng Zhang

https://openreview.net/forum?id=SoLebIqHgZ

Keywords: phylogenetic inference, autoregressive model, graph neural network, density estimation, variational inference

Compressor summary: The paper introduces ARTree, a deep autoregressive model for phylogenetic inference using graph neural networks, which can efficiently learn probabilistic models over tree topologies without relying on hand-engineered features.


Bayesian Risk-Averse Q-Learning with Streaming Observations

Yuhao Wang, Enlu Zhou

https://openreview.net/forum?id=SjiLtmZETc

Keywords: Q-learning, risk-averse reinforcement learning, off-policy learning, Bayesian risk Markov decision process, distributionally robust Markov decision process

Compressor summary: The paper proposes a robust reinforcement learning framework using BRMDP and a multi-stage Q-learning algorithm that learns a risk-averse optimal policy with real environment observations and strong convergence.


Errors-in-variables Fr\'echet Regression with Low-rank Covariate Approximation

Dogyoon Song, Kyunghee Han

https://openreview.net/forum?id=Sg3aCpWUQP

Keywords: Frechet regression, principal component regression, non-Euclidean, low-rank matrix, errors-in-variables analysis

Compressor summary: The paper proposes a new method that combines global Fr\'echet regression and principal component regression to improve efficiency and accuracy in non-Euclidean regression analysis by leveraging low-rank structure in the covariate matrix.


An Optimal Structured Zeroth-order Algorithm for Non-smooth Optimization

Marco Rando, Cesare Molinari, Lorenzo Rosasco, Silvia Villa

https://openreview.net/forum?id=SfdkS6tt81

Keywords: nonsmooth optimization;zeroth order optimization;nonsmooth zeroth-order

Compressor summary: O-ZD is a new structured finite-difference algorithm for non-smooth black-box optimization that uses smooth approximations of the target function and orthogonal directions to approximate its gradient efficiently.


Sampling from Gaussian Process Posteriors using Stochastic Gradient Descent

Jihao Andreas Lin, Javier Antoran, Shreyas Padhy, David Janz, José Miguel Hernández-Lobato, Alexander Terenin

https://openreview.net/forum?id=Sf9goJtTCE

Keywords: Gaussian processes, scalable learning, posterior sampling, Bayesian optimization

Compressor summary: Stochastic gradient descent is a computationally efficient method for solving linear systems in Gaussian processes, achieving state-of-the-art performance and accurate predictions even when not converging quickly to the optimum.


Enhancing Sharpness-Aware Optimization Through Variance Suppression

Bingcong Li, Georgios B. Giannakis

https://openreview.net/forum?id=Sf3t6Bth4P

Keywords: generalization, optimization, neural networks

Compressor summary: VaSSO improves SAM by making adversaries more aggressive, leading to better generalization, stability, and robustness in neural network training.


Optimistic Exploration in Reinforcement Learning Using Symbolic Model Estimates

Sarath Sreedharan, Michael Katz

https://openreview.net/forum?id=Sf17j2pkCU

Keywords: Planning, Reinforcement Learning, Exploration

Compressor summary: The paragraph discusses a new method for learning optimistic symbolic models that guide reinforcement learning agents in sparse reward settings, improving exploration speed through action generalization.


DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation

Shentong Mo, Enze Xie, Ruihang Chu, Lanqing HONG, Matthias Nießner, Zhenguo Li

https://openreview.net/forum?id=Se71ks7Mfz

Keywords: Diffusion Models, Transformers, 3D Shape Generation

Compressor summary: The paper proposes a novel Diffusion Transformer (DiT-3D) for generating high-quality 3D shapes from voxelized point clouds, which improves scalability and quality over existing U-Net methods and leverages pre-trained DiT-2D on ImageNet.


SpecTr: Fast Speculative Decoding via Optimal Transport

Ziteng Sun, Ananda Theertha Suresh, Jae Hun Ro, Ahmad Beirami, Himanshu Jain, Felix Yu

https://openreview.net/forum?id=SdYHLTCC5J

Keywords: autoregressive sampling; computation efficiency; optimal transport

Compressor summary: The authors propose SpecTr, a new autoregressive sampling algorithm that uses optimal transport with membership cost to improve the speed and quality of decoding from large language models.


ReMaX: Relaxing for Better Training on Efficient Panoptic Segmentation

Shuyang Sun, Weijun Wang, Andrew G. Howard, Qihang Yu, Philip Torr, Liang-Chieh Chen

https://openreview.net/forum?id=SaMrN9tnxE

Keywords: Panoptic segmentation, efficient models

Compressor summary: The paper introduces ReMaX, a technique that improves the training of mask transformers for efficient panoptic segmentation by adding relaxations to predictions.


Learning to Parameterize Visual Attributes for Open-set Fine-grained Retrieval

Shijie Wang, Jianlong Chang, Haojie Li, Zhihui Wang, Wanli Ouyang, Qi Tian

https://openreview.net/forum?id=SaII5qMgKH

Keywords: Open-set Fine-grained Retrieval, Visual Attribute, Unknown Categories

Compressor summary: VAPNet is a novel network that learns visual attributes from known categories and integrates them into a retrieval model for unknown categories, without manual annotations.


Neural Sampling in Hierarchical Exponential-family Energy-based Models

Xingsi Dong, Si Wu

https://openreview.net/forum?id=SWU8YLlFVH

Keywords: Baysian brain, sampling-based inference, energy-based models, local learning, exponential-family

Compressor summary: The Hierarchical Exponential-family Energy-based (HEE) model is a new method for inferring and learning in the brain, which uses neural adaptation and allows fast computation and representation similar to biological vision systems.


Improving CLIP Training with Language Rewrites

Lijie Fan, Dilip Krishnan, Phillip Isola, Dina Katabi, Yonglong Tian

https://openreview.net/forum?id=SVjDiiVySh

Keywords: contrastive learning; CLIP; large language model

Compressor summary: LaCLIP is a method that enhances CLIP training by rewriting image text descriptions with diverse sentences while preserving original meanings, improving transfer performance without extra computation or memory.


DiffKendall: A Novel Approach for Few-Shot Learning with Differentiable Kendall's Rank Correlation

Kaipeng Zheng, Huishuai Zhang, Weiran Huang

https://openreview.net/forum?id=SVUQX1W7RL

Keywords: Few-shot learning

Compressor summary: This paper shows that using Kendall's rank correlation instead of geometric similarity metrics improves few-shot learning performance across different methods and datasets.


Language Models Meet World Models: Embodied Experiences Enhance Language Models

Jiannan Xiang, Tianhua Tao, Yi Gu, Tianmin Shu, Zirui Wang, Zichao Yang, Zhiting Hu

https://openreview.net/forum?id=SVBR6xBaMl

Keywords: Language Model, World Model, Embodied Experience

Compressor summary: The paper proposes a method to improve large language models' abilities in reasoning and acting in the physical world by finetuning them with embodied experiences from a simulated environment and using techniques like EWC and LoRA to preserve generality.


Learning Robust Statistics for Simulation-based Inference under Model Misspecification

Daolang Huang, Ayush Bharti, Amauri H Souza, Luigi Acerbi, Samuel Kaski

https://openreview.net/forum?id=STrXsSIEiq

Keywords: Simulation-based inference, model misspecification, likelihood-free inference, approximate Bayesian computation, neural posterior estimation

Compressor summary: The authors propose a general approach to handle model misspecification in simulation-based inference methods and demonstrate its superior performance on artificially and realistically misspecified models.


Plug-and-Play Stability for Intracortical Brain-Computer Interfaces: A One-Year Demonstration of Seamless Brain-to-Text Communication

Chaofei Fan, Nick Hahn, Foram Kamdar, Donald Avansino, Guy H Wilson, Leigh Hochberg, Krishna V. Shenoy, Jaimie M. Henderson, Francis R Willett

https://openreview.net/forum?id=STqaMqhtDi

Keywords: brain-computer interface, self-training, continual online learning

Compressor summary: The paper proposes a method to improve intracortical brain-computer interfaces (iBCIs) for communication by enabling self-recalibration using large language models, achieving high performance and long-term stability in a human participant.


MomentDiff: Generative Video Moment Retrieval from Random to Real

Pandeng Li, Chen-Wei Xie, Hongtao Xie, Liming Zhao, Lei Zhang, Yun Zheng, Deli Zhao, Yongdong Zhang

https://openreview.net/forum?id=SQouRKRIXY

Keywords: Video Moment Retrieval, Diffusion Model

Compressor summary: MomentDiff is a generative diffusion-based framework that learns to retrieve specific temporal segments of videos based on language descriptions, overcoming temporal location biases in existing methods.


Leveraging Locality and Robustness to Achieve Massively Scalable Gaussian Process Regression

Robert F Allison, Anthony Stephenson, Samuel F, Edward Pyzer-Knapp

https://openreview.net/forum?id=SQP1H9Jy8W

Keywords: Gaussian Processes, Bayesian Inference, Regression, Bayesian Nonparametrics, Kernel Methods

Compressor summary: GP nearest-neighbour prediction is accurate and computationally efficient even when data size is large, as parameter estimation and model assumptions become less important.


Cognitive Model Discovery via Disentangled RNNs

Kevin J Miller, Maria K Eckstein, Matthew Botvinick, Zeb Kurth-Nelson

https://openreview.net/forum?id=SOEF0i0G1z

Keywords: Cognitive modeling, neural networks, interpretability, disentangling, neuroscience, rodent behavior

Compressor summary: The authors propose an alternative approach to construct cognitive models from behavioral data using sparse and interpretable recurrent neural networks.


Variational Annealing on Graphs for Combinatorial Optimization

Sebastian Sanokowski, Wilhelm Franz Berghammer, Sepp Hochreiter, Sebastian Lehner

https://openreview.net/forum?id=SLx7paoaTU

Keywords: Combinatorial Optimization, Entropy Regularization, Graph Neural Networks, Statistical Mechanics

Compressor summary: Unsupervised learning methods using probabilistic approaches for combinatorial optimization may not perform well on difficult problems due to the assumption of independent solution variables; author propose Subgraph Tokenization, which represents a set of solution variables by one token, and annealed entropy regularization, which improves efficiency and stability.


PLANNER: Generating Diversified Paragraph via Latent Language Diffusion Model

Yizhe Zhang, Jiatao Gu, Zhuofeng Wu, Shuangfei Zhai, Joshua M. Susskind, Navdeep Jaitly

https://openreview.net/forum?id=SLwy8UVS8Y

Keywords: Text generation, diffusion model, NLP

Compressor summary: PLANNER combines latent semantic diffusion with autoregressive generation to produce fluent and controllable long-form text for various tasks.


CoLA: Exploiting Compositional Structure for Automatic and Efficient Numerical Linear Algebra

Andres Potapczynski, Marc Anton Finzi, Geoff Pleiss, Andrew Gordon Wilson

https://openreview.net/forum?id=SLtNFERsHo

Keywords: Machine Learning, Numerical Linear Algebra, partial differential equations, Gaussian processes, equivariance, graph learning, spectral analysis

Compressor summary: CoLA is a framework for large-scale linear algebra problems in machine learning that combines a linear operator abstraction with compositional dispatch rules to automatically construct memory and runtime efficient numerical algorithms.


Distributional Model Equivalence for Risk-Sensitive Reinforcement Learning

Tyler Kastner, Murat A Erdogdu, Amir-massoud Farahmand

https://openreview.net/forum?id=SLTQluG80x

Keywords: Reinforcement learning, Risk-Sensitive Reinforcement Learning, Model-Based Reinforcement Learning, Distributional Reinforcement Learning

Compressor summary: The paper explores learning optimal models for risk-sensitive reinforcement learning using distributional reinforcement learning and introduces two new notions of model equivalence.


Adversarial Robustness through Random Weight Sampling

Yanxiang Ma, Minjing Dong, Chang Xu

https://openreview.net/forum?id=SIE9N5nnHg

Keywords: Adversarial robustness; Randomized defense; Random parameters optimization

Compressor summary: The authors propose a new randomized defense mechanism for deep neural networks, called Constrained Trainable Random Weight (CTRW), which optimizes randomness parameters to balance natural performance and adversarial robustness, achieving better results on several datasets and benchmarks.


Unlocking Deterministic Robustness Certification on ImageNet

Kai Hu, Andy Zou, Zifan Wang, Klas Leino, Matt Fredrikson

https://openreview.net/forum?id=SHyVaWGTO4

Keywords: adversarial robustness, ImageNet, Lipschitz-based certification, ResNet, adversarial examples, ML security

Compressor summary: The paper proposes LiResNet, a new architecture and EMMA, a loss function, that improve the efficiency of calculating Lipschitz bounds and stabilize robust training for deep networks on large-scale datasets like ImageNet.


Hyperbolic Graph Neural Networks at Scale: A Meta Learning Approach

Nurendra Choudhary, Nikhil Rao, Chandan K. Reddy

https://openreview.net/forum?id=SHVwG9yOEk

Keywords: meta learning, hyperbolic networks, scalability, graph neural networks

Compressor summary: H-GRAM is a novel meta-learning method for hyperbolic neural networks that learns inductive biases from local subgraphs and transfers them to new tasks, improving generalization and scalability in few-shot settings.


Contextual Stochastic Bilevel Optimization

Yifan Hu, Jie Wang, Yao Xie, Andreas Krause, Daniel Kuhn

https://openreview.net/forum?id=SHBksHKutP

Keywords: stochastic optimization, bilevel optimization, contextual stochastic optimization, Multilevel Monte Carlo

Compressor summary: CSBO is a new framework for bilevel optimization with contextual information that requires a double-loop gradient method based on MLMC to handle its challenges and has applications in various areas like meta-learning and federated learning.


Concept Algebra for (Score-Based) Text-Controlled Generative Models

Zihao Wang, Lin Gui, Jeffrey Negrea, Victor Veitch

https://openreview.net/forum?id=SGlrCuwdsB

Keywords: disentanglement; representation learning; text-controlled generative models; diffusion models

Compressor summary: The paper explores how text-guided generative models like Stable Diffusion represent and manipulate disentangled concepts in a subspace of their representation space.


Tracking Most Significant Shifts in Nonparametric Contextual Bandits

Joe Suk, Samory Kpotufe

https://openreview.net/forum?id=SGerL9HMrp

Keywords: multi-armed bandits, non-stationary, contextual bandits, nonparametric, Lipschitz

Compressor summary: The paper studies nonparametric contextual bandits with changing reward functions and proposes a new concept of change that accounts for locality and significance, achieving adaptivity in the minimax regret rate.


Collaborative Learning via Prediction Consensus

Dongyang Fan, Celestine Mendler-Dünner, Martin Jaggi

https://openreview.net/forum?id=SGKbHXoLCI

Keywords: Collaborative training, decentralized learning, consensus reaching

Compressor summary: Our method helps agents improve their models by sharing unlabeled data and using trust weights to reach consensus on pseudo-labels, which enhances performance and reduces bad influence.


Graph Convolutional Kernel Machine versus Graph Convolutional Networks

Zhihao Wu, Zhao Zhang, Jicong Fan

https://openreview.net/forum?id=SFfOt1oDsX

Keywords: graph neural network, kernel method

Compressor summary: Graph convolutional kernel machine (GCKM) is a framework for graph-based machine learning that uses kernel functions with graph convolution and has advantages over deep graph convolutional networks (GCN).


Nearly Optimal VC-Dimension and Pseudo-Dimension Bounds for Deep Neural Network Derivatives

Yahong Yang, Haizhao Yang, Yang Xiang

https://openreview.net/forum?id=SE73LzWNjr

Keywords: VC-dimension, pseudo-dimension, Sobolev space, generalization error, nearly optimal approximation

Compressor summary: The paper studies how to accurately estimate the complexity of derivative functions in deep neural networks and applies these estimations to various physics-informed machine learning tasks.


Structure Learning with Adaptive Random Neighborhood Informed MCMC

Xitong Liang, Alberto Caron, Samuel Livingstone, Jim Griffin

https://openreview.net/forum?id=SCsJFNcSHQ

Keywords: Bayesian Networks, structure MCMC on graphs, Structure Learning, Random neighborhood samplers, Locally informed Metropolis-Hastings schemes

Compressor summary: PARNI-DAG is a new MCMC sampler that efficiently learns Directed Acyclic Graphs (DAGs) from observational data under causal sufficiency by using adaptive random neighborhood proposals and pre-tuning parameters based on a skeleton graph.


Towards Self-Interpretable Graph-Level Anomaly Detection

Yixin Liu, Kaize Ding, Qinghua Lu, Fuyi Li, Leo Yu Zhang, Shirui Pan

https://openreview.net/forum?id=SAzaC8f3cM

Keywords: Anomaly Detection, Graph Neural Networks, Explanation, Self-Interpretation

Compressor summary: SIGNET is a model that detects abnormal graphs and explains why they are abnormal using subgraphs from the input graph and its dual hypergraph.


Category-Extensible Out-of-Distribution Detection via Hierarchical Context Descriptions

Kai Liu, Zhihang Fu, Chao Chen, Sheng Jin, Ze Chen, Mingyuan Tao, Rongxin Jiang, Jieping Ye

https://openreview.net/forum?id=SA2KrosYjY

Keywords: out-of-distribution detection, vision-language models, category-extendable classification

Compressor summary: This paper introduces two contexts, perceptual and spurious, for detecting out-of-distribution samples in vision-language models using automatic prompt tuning and applies them to a novel detection method called CATEX.


$p$-Poisson surface reconstruction in curl-free flow from point clouds

Yesom Park, Taekyung Lee, Jooyoung Hahn, Myungjoo Kang

https://openreview.net/forum?id=S8hg5LpFvz

Keywords: Surface reconstruction, Signed distance function, Implicit neural representations, Point cloud

Compressor summary: The paper proposes a new surface reconstruction method using implicit neural representations, which learns a signed distance function from a point cloud and imposes physical constraints for better results.


Learning Nonparametric Latent Causal Graphs with Unknown Interventions

Yibo Jiang, Bryon Aragam

https://openreview.net/forum?id=S8DFqgmEbe

Keywords: graphical models, directed acyclic graphs, causality, identifiability, causal representation learning, unknown interventions

Compressor summary: The paper presents methods for identifying latent causal graphs from interventions without making parametric assumptions or knowing the number of hidden variables, introducing new graphical concepts and characterizing the limits of edge orientations in a general setting.


Posterior Sampling for Competitive RL: Function Approximation and Partial Observation

Shuang Qiu, Ziyu Dai, Han Zhong, Zhaoran Wang, Zhuoran Yang, Tong Zhang

https://openreview.net/forum?id=S75ccNdOYG

Keywords: Markov game, Partial observation, Function approximation, Posterior sampling, Reinforcement Learning

Compressor summary: The paper presents new algorithms for competitive reinforcement learning in different game settings, using complexity measures and posterior sampling methods that handle partial observability and trade-off exploration and exploitation.


A3FL: Adversarially Adaptive Backdoor Attacks to Federated Learning

Hangfan Zhang, Jinyuan Jia, Jinghui Chen, Lu Lin, Dinghao Wu

https://openreview.net/forum?id=S6ajVZy6FA

Keywords: Backdoor Attack, Federated Learning

Compressor summary: The paper introduces a new backdoor attack, A3FL, that adapts the trigger to survive global training dynamics in federated learning.


The Clock and the Pizza: Two Stories in Mechanistic Explanation of Neural Networks

Ziqian Zhong, Ziming Liu, Max Tegmark, Jacob Andreas

https://openreview.net/forum?id=S5wmbQc1We

Keywords: mechanistic interpretability, algorithmic phase transitions, arithmetic learning, neural network, transformer, ensemble

Compressor summary: The study shows that neural networks can learn different algorithms, including some novel ones, when trained on modular addition tasks, indicating the complexity of algorithm discovery in neural networks.


Efficient Equivariant Transfer Learning from Pretrained Models

Sourya Basu, Pulkit Katdare, Prasanna Sattigeri, Vijil Chenthamarakshan, Katherine Rose Driggs-Campbell, Payel Das, Lav R. Varshney

https://openreview.net/forum?id=S4NN3OOiwP

Keywords: zero-shot learning, equivariant machine learning, equivariant fine-tuning, pretrained models

Compressor summary: λ-equitune is a method to improve equivariant outputs from non-equivariant neural networks by averaging features with importance weights learned from data, leading to better zero-shot and finetuned results than equitune on diverse applications and models.


The Benefits of Being Distributional: Small-Loss Bounds for Reinforcement Learning

Kaiwen Wang, Kevin Zhou, Runzhe Wu, Nathan Kallus, Wen Sun

https://openreview.net/forum?id=S3Y0VvegGm

Keywords: Reinforcement Learning Theory, Distributional Reinforcement Learning, Small-Loss Bounds, First-order regret

Compressor summary: The paper explains why distributional reinforcement learning (DistRL) is better than non-distributional RL by using small-loss bounds, proposes a DistCB algorithm with empirical results, and shows that pessimistic DistRL has novel small-loss PAC bounds in offline RL.


Self-Refine: Iterative Refinement with Self-Feedback

Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, Shashank Gupta, Bodhisattwa Prasad Majumder, Katherine Hermann, Sean Welleck, Amir Yazdanbakhsh, Peter Clark

https://openreview.net/forum?id=S37hOerQLB

Keywords: LLMs, Iterative Refinement, Feedback-driven Generation

Compressor summary: The Self-Refine approach improves the output of large language models by iteratively refining their own outputs based on feedback, without needing additional data or training.


Transportability for Bandits with Data from Different Environments

Alexis Bellot, Alan Malek, Silvia Chiappa

https://openreview.net/forum?id=S2k5dBb91q

Keywords: Transportability, transfer learning, bandits

Compressor summary: The paper proposes a new bandit algorithm that uses causal models to learn from batch data and exploit invariances across related environments, leading to lower regret and faster learning.


ClusterFomer: Clustering As A Universal Visual Learner

James Chenhao Liang, Yiming Cui, Qifan Wang, Tong Geng, Wenguan Wang, Dongfang Liu

https://openreview.net/forum?id=S1KGaTSOTS

Keywords: Universal Model, Clustering

Compressor summary: ClusterFormer is a novel vision model that uses clustering with TransFormers, enabling explainable and transferable image processing for various tasks and outperforming specialized architectures.


Dense-Exponential Random Features: Sharp Positive Estimators of the Gaussian Kernel

Valerii Likhosherstov, Krzysztof Marcin Choromanski, Kumar Avinava Dubey, Frederick Liu, Tamas Sarlos, Adrian Weller

https://openreview.net/forum?id=S0xrBMFihS

Keywords: Gaussian kernel, softmax kernel

Compressor summary: The authors propose new parameterized random features to approximate Gaussian and softmax kernels, which reduce variance and improve performance in kernel methods and Transformers.


SwiftSage: A Generative Agent with Fast and Slow Thinking for Complex Interactive Tasks

Bill Yuchen Lin, Yicheng Fu, Karina Yang, Faeze Brahman, Shiyu Huang, Chandra Bhagavatula, Prithviraj Ammanabrolu, Yejin Choi, Xiang Ren

https://openreview.net/forum?id=Rzk3GP1HN7

Keywords: interactive reasoning, text game, agents, action planning, large language models

Compressor summary: SwiftSage is a new agent framework that combines behavior cloning and large language models to improve action planning for complex reasoning tasks by using two modules: one for fast thinking and another for deliberate thought.


Similarity, Compression and Local Steps: Three Pillars of Efficient Communications for Distributed Variational Inequalities

Aleksandr Beznosikov, Martin Takáč, Alexander Gasnikov

https://openreview.net/forum?id=Rvk1wdwz1L

Keywords: convex optimization, variational inequalities, similarity, local methods, compression, partial participation

Compressor summary: The paper proposes a novel triple synergy technique to reduce communication costs for solving large-scale variational inequality problems in machine learning, achieving the best theoretical guarantees and outperforming existing methods.


Diversified Outlier Exposure for Out-of-Distribution Detection via Informative Extrapolation

Jianing Zhu, Geng Yu, Jiangchao Yao, Tongliang Liu, Gang Niu, Masashi Sugiyama, Bo Han

https://openreview.net/forum?id=RuxBLfiEqI

Keywords: out-of-distribution detection, outlier exposure

Compressor summary: DivOE is a novel framework for OOD detection that diversifies auxiliary outliers and extrapolates them during training using a multi-step optimization method.


Dynamic Personalized Federated Learning with Adaptive Differential Privacy

Xiyuan Yang, Wenke Huang, Mang Ye

https://openreview.net/forum?id=RteNLuc8D9

Keywords: federated learning, differential privacy, personalization

Compressor summary: The paper proposes FedDPA, a novel federated learning method that uses layer-wise Fisher information to flexibly personalize models while mitigating convergence difficulties caused by clipping operations in differential privacy.


A Partially-Supervised Reinforcement Learning Framework for Visual Active Search

Anindya Sarkar, Nathan Jacobs, Yevgeniy Vorobeychik

https://openreview.net/forum?id=Rs6pzz21U4

Keywords: Visual Active Search, Reinforcement Learning

Compressor summary: Visual active search (VAS) has been proposed as a modeling framework in which visual cues are used to guide exploration, with the goal of identifying regions of interest in a large geospatial area. Its potential applications include identifying hot spots of rare wildlife poaching activity, search-and-rescue scenarios, identifying illegal trafficking of weapons, drugs, or people, and many others. State of the art approaches to VAS include applications of deep reinforcement learning (DRL), which yield end-to-end search policies, and traditional active search, which combines predictions with custom algorithmic approaches. While the DRL framework has been shown to greatly outperform traditional active search in such domains, its end-to-end nature does not make full use of supervised information attained either during training, or during actual search, a significant limitation if search tasks differ significantly from those in the training distribution. We propose an approach that combines the strength of both DRL and conventional active search approaches by decomposing the search policy into a prediction module, which produces a geospatial distribution of regions of interest based on task embedding and search history, and a search module, which takes the predictions and search history as input and outputs the search distribution. In addition, we develop a novel meta-learning approach for jointly learning the resulting combined policy that can make effective use of supervised information obtained both at training and decision time. Our extensive experiments demonstrate that the proposed representation and meta-learning frameworks significantly outperform state of the art in visual active search on several problem domains.


Cluster-aware Semi-supervised Learning: Relational Knowledge Distillation Provably Learns Clustering

Yijun Dong, Kevin Miller, Qi Lei, Rachel Ward

https://openreview.net/forum?id=RrdBNXBUIF

Keywords: Relational knowledge distillation, Semi-supervised learning, Spectral clustering, Sample complexity

Compressor summary: The paper provides a theoretical understanding and analysis of relational knowledge distillation in semi-supervised classification problems using spectral clustering and cluster-aware frameworks.


Spectral Co-Distillation for Personalized Federated Learning

Zihan Chen, Howard Hao Yang, Tony Quek, Kai Fong Ernest Chong

https://openreview.net/forum?id=RqjQL08UFc

Keywords: Personalized federated learning, spectral bias, co-distillation, communication efficiency

Compressor summary: The text introduces spectral distillation, a new personalized federated learning method that leverages model spectrum information to improve performance and efficiency on diverse data settings.


Unsupervised Video Domain Adaptation for Action Recognition: A Disentanglement Perspective

Pengfei Wei, Lingdong Kong, Xinghua Qu, Yi Ren, zhiqiang xu, Jing Jiang, Xiang Yin

https://openreview.net/forum?id=Rp4PA0ez0m

Keywords: action recognition, unsupervised domain adaptation, video analysis

Compressor summary: This paper introduces a new method called TranSVAE to generate videos across different domains by disentangling static and dynamic information and removing spatial and temporal domain differences using constraints and adversarial learning.


Agnostic Multi-Group Active Learning

Nicholas Rittler, Kamalika Chaudhuri

https://openreview.net/forum?id=RmxP5ZcQhC

Keywords: learning theory, active learning, multi-group learning

Compressor summary: The paper presents a novel active learning algorithm for learning from multiple groups with minimal label queries, and shows its advantages in some scenarios, as well as providing results for special cases of realizable groups and approximations for the general case.


H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models

Zhenyu Zhang, Ying Sheng, Tianyi Zhou, Tianlong Chen, Lianmin Zheng, Ruisi Cai, Zhao Song, Yuandong Tian, Christopher Re, Clark Barrett, Zhangyang Wang, Beidi Chen

https://openreview.net/forum?id=RkRrPp7GKO

Keywords: Large Language Models; Efficient Generative Inference

Compressor summary: The paper proposes a new memory management technique for large language models that reduces their cost by focusing on the most important tokens and evicting less relevant ones.


Posterior Sampling with Delayed Feedback for Reinforcement Learning with Linear Function Approximation

Nikki Lijing Kuang, Ming Yin, Mengdi Wang, Yu-Xiang Wang, Yian Ma

https://openreview.net/forum?id=RiyH3z7oIF

Keywords: Posterior Sampling, Reinforcement Learning Theory, Linear Markov Decision Processes, Delayed Feedback, Langevin Monte Carlo

Compressor summary: This paper proposes two optimistic value-based algorithms for reinforcement learning with delayed feedback, and analyzes their performance in terms of regret.


Active representation learning for general task space with applications in robotics

Yifang Chen, Yingbing Huang, Simon Shaolei Du, Kevin Jamieson, Guanya Shi

https://openreview.net/forum?id=RiwPYAMLur

Keywords: active learning, representation learning, robotics, theory

Compressor summary: The paper proposes a general framework for active representation learning that optimally chooses source tasks to sample from and shows its effectiveness in various domains with theoretical and empirical results.


SimMMDG: A Simple and Effective Framework for Multi-modal Domain Generalization

Hao Dong, Ismail Nejjar, Han Sun, Eleni Chatzi, Olga Fink

https://openreview.net/forum?id=RiSMijlsLT

Keywords: Domain Generalization, Multi-modal Learning, Distribution Shift, Out-of-distribution Generalization

Compressor summary: SimMMDG is a multi-modal domain generalization framework that splits features into modality-specific and modality-shared components, uses supervised contrastive learning to encourage joint properties, imposes distance constraints for diversity, and employs cross-modal translation for regularization and missing-modality generalization.


Feature Selection in the Contrastive Analysis Setting

Ethan Weinberger, Ian Connick Covert, Su-In Lee

https://openreview.net/forum?id=RhE01dqo8u

Keywords: Feature selection, contrastive analysis, computational biology, representation learning, information theory

Compressor summary: The paper introduces contrastive feature selection (CFS), a method for selecting features in the contrastive analysis setting, and shows that it outperforms existing methods on biomedical datasets.


Reusing Pretrained Models by Multi-linear Operators for Efficient Training

Yu Pan, Ye Yuan, Yichun Yin, Zenglin Xu, Lifeng Shang, Xin Jiang, Qun Liu

https://openreview.net/forum?id=RgNXKIrWyU

Keywords: Model Growth, Efficient Training, Pretrained Model, Multi-linearity

Compressor summary: The paper proposes a method to accelerate training large models by linearly correlating each weight of the target model with all weights of the pretrained model, reducing computational and spatial complexity.


DeepPCR: Parallelizing Sequential Operations in Neural Networks

Federico Danieli, Miguel Sarabia, Xavier Suau, Pau Rodriguez, Luca Zappella

https://openreview.net/forum?id=RgD92idA32

Keywords: Acceleration, layer-parallelization, diffusion, Parallel Cyclic Reduction

Compressor summary: DeepPCR is a novel algorithm that parallelizes typically sequential operations in deep neural networks, such as forward and backward passes, to achieve significant speedups.


Learning Environment-Aware Affordance for 3D Articulated Object Manipulation under Occlusions

Ruihai Wu, Kai Cheng, Yan Zhao, Chuanruo Ning, Guanqi Zhan, Hao Dong

https://openreview.net/forum?id=Re2NHYoZ5l

Keywords: Visual Affordance for Robotics, Articulated Object Manipulation, Occlusion Handling

Compressor summary: The paper proposes an environment-aware affordance framework for home-assistant robots that learns from object-level actionable priors and various occlusions using a contrastive learning approach.


GenS: Generalizable Neural Surface Reconstruction from Multi-View Images

Rui Peng, Xiaodong Gu, Luyang Tang, Shihe Shen, Fanqi Yu, Ronggang Wang

https://openreview.net/forum?id=Rcit6V3vus

Keywords: Generalizable Neural Surface, Volume Rendering, Signed Distance Function

Compressor summary: The paper introduces GenS, a neural surface reconstruction model that generalizes well to new scenes and outperforms existing methods without 3D supervision or long-time optimizations.


Modality-Agnostic Self-Supervised Learning with Meta-Learned Masked Auto-Encoder

Huiwon Jang, Jihoon Tack, Daewon Choi, Jongheon Jeong, Jinwoo Shin

https://openreview.net/forum?id=RZGtK2nDDJ

Keywords: Self-Supervised Learning, Modality-Agnostic Self-Supervised Learning, Meta-Learning, Masked Auto-Encoder

Compressor summary: The paper proposes MetaMAE, a meta-learning enhanced version of Masked Auto-Encoder, to improve self-supervised learning across diverse modalities.


BanditPAM++: Faster $k$-medoids Clustering

Mo Tiwari, Ryan Kang, Donghyun Lee, Sebastian Thrun, Ilan Shomorony, Martin Jinye Zhang

https://openreview.net/forum?id=RWcfpmjlYm

Keywords: multi-armed bandits, clustering, k-medoids, best-arm identification

Compressor summary: BanditPAM++ is a faster and improved version of BanditPAM, which is a randomized $k$-medoids algorithm with state-of-the-art complexity and accuracy.


Federated Spectral Clustering via Secure Similarity Reconstruction

Dong Qiao, Chris Ding, Jicong Fan

https://openreview.net/forum?id=RW7rZ8Y3Bp

Keywords: clustering, federated learning, privacy

Compressor summary: Federated learning has a significant advantage in protecting information privacy. Many scholars proposed various secure learning methods within the framework of federated learning but the study on secure federated unsupervised learning especially clustering is limited. We in this work propose a secure kernelized factorization method for federated spectral clustering on distributed dataset. The method is non-trivial because the kernel or similarity matrix for spectral clustering is computed by data pairs, which violates the principle of privacy protection. Our method implicitly constructs an approximation for the kernel matrix on distributed data such that we can perform spectral clustering under the constraint of privacy protection. We provide a convergence guarantee of the optimization algorithm, reconstruction error bounds of the Gaussian kernel matrix, and the sufficient condition of correct clustering of our method. We also present some results of differential privacy. Numerical results on synthetic and real datasets demonstrate that the proposed method is efficient and accurate in comparison to the baselines.


Promises and Pitfalls of Threshold-based Auto-labeling

Harit Vishwakarma, Heguang Lin, Frederic Sala, Ramya Korlakai Vinayak

https://openreview.net/forum?id=RUCFAKNDb2

Keywords: Auto Labeling, Active Learning, Selective Classification

Compressor summary: The paragraph discusses threshold-based auto-labeling (TBAL), a technique to reduce manual annotation in machine learning, and analyzes its sample complexity, quality, and potential pitfalls using theoretical and empirical methods.


ANPL: Towards Natural Programming with Interactive Decomposition

Di Huang, Ziyuan Nan, Xing Hu, Pengwei Jin, Shaohui Peng, Yuanbo Wen, Rui Zhang, Zidong Du, Qi Guo, Yewen Pu, Yunji Chen

https://openreview.net/forum?id=RTRS3ZTsSj

Keywords: programming language, large language models, program synthesis, code generation, human-ai interaction

Compressor summary: ANPL is an interactive programming system that allows users to refine code generated by LLMs using structured decompositions and natural language specifications.


How to Turn Your Knowledge Graph Embeddings into Generative Models

Lorenzo Loconte, Nicola Di Mauro, Robert Peharz, Antonio Vergari

https://openreview.net/forum?id=RSGNGiB1q4

Keywords: knowledge graph, knowledge graph embeddings, probabilistic circuits, probabilistic reasoning, tractable inference

Compressor summary: The authors propose generative circuit models for knowledge graph embedding that enable exact maximum-likelihood estimation, efficient sampling, and logical constraint satisfaction while maintaining performance for link prediction.


Spectral Invariant Learning for Dynamic Graphs under Distribution Shifts

Zeyang Zhang, Xin Wang, Ziwei Zhang, Zhou Qin, Weigao Wen, Hui Xue', Haoyang Li, Wenwu Zhu

https://openreview.net/forum?id=RRUVZygUtr

Keywords: Dynamic Graph Neural Networks, Out-of-Distribution Generalization

Compressor summary: The paper proposes a new method, SILD, to handle distribution shifts in dynamic graphs by capturing and utilizing invariant and variant spectral patterns.


IMPRESS: Evaluating the Resilience of Imperceptible Perturbations Against Unauthorized Data Usage in Diffusion-Based Generative AI

Bochuan Cao, Changjiang Li, Ting Wang, Jinyuan Jia, Bo Li, Jinghui Chen

https://openreview.net/forum?id=RRSltzPc7w

Keywords: Image Generation Godels, Latent Diffusion Models, Image Purifying

Compressor summary: IMPRESS is a platform that evaluates the effectiveness of imperceptible perturbations in protecting original images from unauthorized data usage by diffusion-based image generation models.


On Imitation in Mean-field Games

Giorgia Ramponi, Pavel Kolev, Olivier Pietquin, Niao He, Mathieu Lauriere, Matthieu Geist

https://openreview.net/forum?id=RPFd3D3P3L

Keywords: Mean-field games, Imitation Learning

Compressor summary: The paper studies imitation learning in mean-field games and introduces a new solution concept called the Nash imitation gap, showing that it is harder than single-agent imitation learning when the dynamics depend on the population distribution.


Towards Consistent Video Editing with Text-to-Image Diffusion Models

Zicheng Zhang, Bonan Li, Xuecheng Nie, Congying Han, Tiande Guo, Luoqi Liu

https://openreview.net/forum?id=RNVwm4BzXO

Keywords: diffusion model, video editing, text-to-video diffusion model

Compressor summary: The paper proposes a new EI$^2$ model to improve Text-to-Image diffusion models for video editing by addressing inconsistency issues caused by modules learning temporal information and using two attention mechanisms to enhance spatial and temporal features.


Elastic Decision Transformer

Yueh-Hua Wu, Xiaolong Wang, Masashi Hamaya

https://openreview.net/forum?id=RMeQjexaRj

Keywords: Offline Reinforcement Learning, Trajectory Stitching, Decision Transformer

Compressor summary: The Elastic Decision Transformer (EDT) improves upon Decision Transformer (DT) by better handling trajectory stitching and adapting history length for optimal performance, leading to competitive results with Q Learning-based methods.


Symbol-LLM: Leverage Language Models for Symbolic System in Visual Human Activity Reasoning

Xiaoqian Wu, Yong-Lu Li, Jianhua Sun, Cewu Lu

https://openreview.net/forum?id=RJq9bVEf6N

Keywords: neuro-symbolic, visual reasoning, human activity understanding

Compressor summary: The paragraph discusses a new symbolic system for visual activity understanding that uses large language models to improve explainability, generalization, and data efficiency.


Cheaply Estimating Inference Efficiency Metrics for Autoregressive Transformer Models

Deepak Narayanan, Keshav Santhanam, Peter Henderson, Rishi Bommasani, Tony Lee, Percy Liang

https://openreview.net/forum?id=RJpAz15D0S

Keywords: Systems for Machine Learning, Inference efficiency, Transformer models, Text generation APIs, Capability-efficiency tradeoffs

Compressor summary: The authors propose idealized runtime and cost models to measure inference efficiency for large language models, and compare ten LLMs from 2022 using these metrics.


Relax, it doesn’t matter how you get there: A new self-supervised approach for multi-timescale behavior analysis

Mehdi Azabou, Michael Jacob Mendelson, Nauman Ahad, Maks Sorokin, Shantanu Thakoor, Carolina Urzay, Eva L Dyer

https://openreview.net/forum?id=RInTOCEL3l

Keywords: animal behavior, behavioral neuroscience, self-supervised learning, multi-timescale

Compressor summary: The paper presents a multi-task representation learning model for animal behavior that predicts actions over future timesteps and incorporates short- and long-term dynamics, achieving success in various environments and ranking first in the MABe 2022 Multi-Agent Behavior challenge.


Learning Rule-Induced Subgraph Representations for Inductive Relation Prediction

Tianyu Liu, Qitan Lv, Jie Wang, Shuling Yang, Hanzhu Chen

https://openreview.net/forum?id=RHDXkRPNQa

Keywords: inductive relation prediction, knowledge graph completion, knowledge graph reasoning

Compressor summary: REST is a novel GNN model for inductive relation prediction that initializes edge features only for the target link and uses RNN-based functions for edge-wise message passing to learn rule-induced subgraph representations.


On-the-Fly Adapting Code Summarization on Trainable Cost-Effective Language Models

Yufan Cai, Yun Lin, Chenyan Liu, Jinglian Wu, Yifan Zhang, Yiming Liu, Yeyun Gong, Jin Song Dong

https://openreview.net/forum?id=RFgv7cfMUy

Keywords: Code Summarization, Adaptation, Language Model

Compressor summary: Adacom is a novel approach to improve code comment generation by adapting deep learning models on the fly using contradictory training samples.


Public Opinion Field Effect Fusion in Representation Learning for Trending Topics Diffusion

Junliang Li, Yajun Yang, Qinghua Hu, Xin Wang, Hong Gao

https://openreview.net/forum?id=RFE1eI0zNZ

Keywords: public opinion field effect, heterogeneous networks, representation learning, trending topic diffusion

Compressor summary: The paper proposes a new representation learning framework that considers public opinion field and social circle influence effects for trending topic diffusion analysis.


Neural Combinatorial Optimization with Heavy Decoder: Toward Large Scale Generalization

Fu Luo, Xi Lin, Fei Liu, Qingfu Zhang, Zhenkun Wang

https://openreview.net/forum?id=RBI4oAbdpm

Keywords: Neural Combinatorial Optimization, Generalization, Large scale problem, Heavy decoder

Compressor summary: The LEHD model is a novel neural combinatorial optimization method that can learn to generalize across problem sizes and scales, enabling it to solve large-scale real-world problems more effectively than previous methods.


Beyond Black-Box Advice: Learning-Augmented Algorithms for MDPs with Q-Value Predictions

Tongxin Li, Yiheng Lin, Shaolei Ren, Adam Wierman

https://openreview.net/forum?id=RACcp8Zbr9

Keywords: Time-varying MDP, Learning-augmented online algorithm, consistency and robustness tradeoff

Compressor summary: The paper investigates how using information about the advice generation process affects the tradeoff between consistency and robustness in single-trajectory time-varying MDPs with machine-learned Q-value advice, showing that it can lead to near-optimal performance.


Segment Anything in High Quality

Lei Ke, Mingqiao Ye, Martin Danelljan, Yifan liu, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu

https://openreview.net/forum?id=RA7ND878XP

Keywords: segment anything, zero-shot segmentation, high-quality segmentation

Compressor summary: The authors propose HQ-SAM, an improved version of the Segment Anything Model (SAM) that can accurately segment any object with minimal additional parameters and computation, while maintaining SAM's efficiency and zero-shot generalizability.


PROTES: Probabilistic Optimization with Tensor Sampling

Anastasia Batsheva, Andrei Chertkov, Gleb Ryzhakov, Ivan Oseledets

https://openreview.net/forum?id=R9R7YDOar1

Keywords: Tensor Train, Black Box Optimization, Sampling, Optimal Control

Compressor summary: PROTES is a novel black-box optimization method that uses probabilistic sampling from a low-parametric tensor train format and performs better than popular discrete methods on complex problems.


Projection-Free Methods for Stochastic Simple Bilevel Optimization with Convex Lower-level Problem

Jincheng Cao, Ruichen Jiang, Nazanin Abolfazli, Erfan Yazdandoost Hamedani, Aryan Mokhtari

https://openreview.net/forum?id=R8GF0EsNsI

Keywords: Bilevel optimization, stochastic optimization

Compressor summary: The paper proposes new methods for solving stochastic bilevel optimization problems that improve the complexity of existing methods by using stochastic cutting planes and conditional gradient updates with variance reduction.


The Utility of “Even if” Semifactual Explanation to Optimise Positive Outcomes

Eoin M. Kenny, Weipeng Fuzzy Huang

https://openreview.net/forum?id=R6wXP7txer

Keywords: Semifactual Explanation, Counterfactual Explanation, Explainable AI, Recourse, User Study

Compressor summary: The paper introduces semifactuals, a new type of XAI that optimizes positive outcomes without crossing decision boundaries, and shows they are more useful than counterfactuals for users who receive positive outcomes like loan approvals.


Dynamo-Depth: Fixing Unsupervised Depth Estimation for Dynamical Scenes

Yihong Sun, Bharath Hariharan

https://openreview.net/forum?id=R6qMmdl4qP

Keywords: monocular, depth estimation, dynamical scenes, motion segmentation, self-supervised

Compressor summary: Dynamo-Depth is a method that estimates depth, motion, and segmentation from unlabeled videos by first separating moving objects from static scenes.


Additive Decoders for Latent Variables Identification and Cartesian-Product Extrapolation

Sebastien Lachapelle, Divyat Mahajan, Ioannis Mitliagkas, Simon Lacoste-Julien

https://openreview.net/forum?id=R6KJN1AUAR

Keywords: identifiability, nonlinear ICA, causal representation learning, disentanglement, object-centric representation learning, extrapolation

Compressor summary: The paragraph discusses how additive decoders can identify latent variables and generate "out-of-support" images for representation learning, with theoretical guarantees and applications in nonlinear ICA and OCRL methods.


Simplifying and Empowering Transformers for Large-Graph Representations

Qitian Wu, Wentao Zhao, Chenxiao Yang, Hengrui Zhang, Fan Nie, Haitian Jiang, Yatao Bian, Junchi Yan

https://openreview.net/forum?id=R4xpvDTWkV

Keywords: graph transformers, large graphs

Compressor summary: The authors propose Simplified Graph Transformers (SGFormer), which uses one-layer attention to efficiently predict node properties on large graphs without complex models or pre-processing, achieving significant inference acceleration and scalability.


Scaling MLPs: A Tale of Inductive Bias

Gregor Bachmann, Sotiris Anagnostidis, Thomas Hofmann

https://openreview.net/forum?id=R45A8eKcax

Keywords: MLP, scaling-laws, inductive bias, DL theory

Compressor summary: This paper investigates the limits and performance of multi-layer perceptrons (MLPs) in vision tasks, showing that they improve with scale and can mimic modern models.


Two Sides of The Same Coin: Bridging Deep Equilibrium Models and Neural ODEs via Homotopy Continuation

Shutong Ding, Tianyu Cui, Jingya Wang, Ye Shi

https://openreview.net/forum?id=R2rJq5OHdr

Keywords: Deep Equilibrium Models, Neural Ordinary Differential Equations, Homotopy Continuation

Compressor summary: The authors propose a new implicit model called HomoODE, which combines the advantages of DEQs and Neural ODEs, and show its effectiveness on image classification tasks using homotopy continuation.


Emergence of Shape Bias in Convolutional Neural Networks through Activation Sparsity

Tianqin Li, Ziqi Wen, Yangfan Li, Tai Sing Lee

https://openreview.net/forum?id=QzcZb3fWmW

Keywords: neuroscience, computer vision, shape & texture bias

Compressor summary: The paper proposes that sparse coding can introduce shape bias into deep learning models, improving their robustness and structure generation, and provides code at a specified GitHub repository.


Equal Opportunity of Coverage in Fair Regression

Fangxin Wang, Lu Cheng, Ruocheng Guo, Kay Liu, Philip S. Yu

https://openreview.net/forum?id=QxYzmYmQQe

Keywords: Equal Opportunity; Fair Machine Learning; Conformal Prediction; Uncertainty Quantification

Compressor summary: The authors propose Equal Opportunity of Coverage (EOC), a fair machine learning approach that ensures equal coverage rates for different groups and maintains population-level coverage, while using Binned Fair Quantile Regression to improve prediction interval width.


Beyond Myopia: Learning from Positive and Unlabeled Data through Holistic Predictive Trends

Wang Xinrui, wan wenhai, Chuanxing Geng, Shao-Yuan Li, Songcan Chen

https://openreview.net/forum?id=QwvaqV48fB

Keywords: positive and unlabeled learning, machine learning, deep learning, temporal point process, data imbalance

Compressor summary: The paper proposes a novel balanced resampling technique for PUL methods and uses temporal point processes to detect trends in scores, improving performance in real-world settings.


Is Distance Matrix Enough for Geometric Deep Learning?

Zian Li, Xiyuan Wang, Yinan Huang, Muhan Zhang

https://openreview.net/forum?id=QwQ5HhhSNo

Keywords: geometric deep learning, expressiveness, equivariant neural networks, universality

Compressor summary: $k$-DisGNNs improve geometric deep learning by capturing high-order information, unifying existing models, and being universal function approximators.


SOAR: Improved Indexing for Approximate Nearest Neighbor Search

Philip Sun, David Simcha, Dave Dopson, Ruiqi Guo, Sanjiv Kumar

https://openreview.net/forum?id=QvIvWMaQdX

Keywords: ann, quantization, mips, nearest neighbor search, retrieval

Compressor summary: SOAR is a new data indexing technique that uses multiple redundant representations and an orthogonality-amplified residual loss to improve approximate nearest neighbor search performance, speed, and memory efficiency.


Sorting with Predictions

Xingjian Bai, Christian Coester

https://openreview.net/forum?id=Qv7rWR9JWa

Keywords: sorting, learning-augmented algorithms, algorithms with predictions, adaptive sorting

Compressor summary: The paper explores how learning-augmented algorithms can use predictions to improve sorting efficiency, and designs new algorithms that adapt to prediction quality and achieve optimal comparison complexity.


PDE-Refiner: Achieving Accurate Long Rollouts with Neural PDE Solvers

Phillip Lippe, Bastiaan S. Veeling, Paris Perdikaris, Richard E Turner, Johannes Brandstetter

https://openreview.net/forum?id=Qv6468llWS

Keywords: Neural PDE Solvers, Neural Operators, Temporal Stability, Long-Horizon Modeling, Autoregressive Forecasting

Compressor summary: This paper introduces a new neural network model called PDE-Refiner that improves the accuracy and stability of solving time-dependent partial differential equations by using a multistep refinement process and spectral data augmentation, outperforming existing methods.


Streaming Factor Trajectory Learning for Temporal Tensor Decomposition

Shikai Fang, Xin Yu, Shibo Li, Zheng Wang, Robert Kirby, Shandian Zhe

https://openreview.net/forum?id=Qu6Ln7d9df

Keywords: Tensor Decomposition, streaming method, Bayesian model

Compressor summary: SFTL is a method that uses Gaussian processes to model the temporal evolution of factors in tensor decomposition for streaming data.


Towards a Unified Framework of Contrastive Learning for Disentangled Representations

Stefan Matthes, Zhiwei Han, Hao Shen

https://openreview.net/forum?id=QrB38MAAEP

Keywords: Disentanglement, Contrastive Learning, Identifiability, Representation Learning, Nonlinear ICA

Compressor summary: This paper extends disentanglement guarantees and identifiability of latents to a broader family of contrastive learning methods and evaluates them on benchmark datasets.


Advice Querying under Budget Constraint for Online Algorithms

Ziyad Benomar, Vianney Perchet

https://openreview.net/forum?id=QpZubU4yD9

Keywords: online algorithms, competitive ratio, learning augmented algorithms, scheduling, ski-rental, secretary

Compressor summary: The paper explores algorithms that can request a limited number of predictions at any time while solving classical problems in competitive analysis.


Selective Sampling and Imitation Learning via Online Regression

Ayush Sekhari, Karthik Sridharan, Wen Sun, Runzhe Wu

https://openreview.net/forum?id=QoeOVgayLp

Keywords: Selective Sampling, Imitation Learning, Learning from Expert Feedback, Theory, General purpose algorithms

Compressor summary: This paper proposes an interactive algorithm for Imitation Learning that uses selective sampling to query a noisy expert for feedback and achieves the best-known bounds for regret and queries in general function classes and multiple actions, with tight theoretical results.


Bifurcations and loss jumps in RNN training

Lukas Eisenmann, Zahra Monfared, Niclas Alexander Göring, Daniel Durstewitz

https://openreview.net/forum?id=QmPf29EHyI

Keywords: dynamical systems, bifurcations, Recurrent Neural Networks, attractors, training algorithm, BPTT, exploding and vanishing gradient problem, nonlinear dynamics, time series

Compressor summary: The authors propose a novel heuristic algorithm to detect bifurcations and fixed points in ReLU-based RNNs, which can help understand their dynamical behavior and the effects of parameter variations during training.


Convergence of Actor-Critic with Multi-Layer Neural Networks

Haoxing Tian, Alex Olshevsky, Ioannis Paschalidis

https://openreview.net/forum?id=QlfGOVD5PO

Keywords: Reinforcement Learning, Actor-Critic, gradient splitting, neural network

Compressor summary: This paper demonstrates that actor-critic methods can converge using deep neural networks with any number of hidden layers, improving the connection between theory and practice in this area.


Goal-conditioned Offline Planning from Curious Exploration

Marco Bagatella, Georg Martius

https://openreview.net/forum?id=QlbZabgMdK

Keywords: deep reinforcement learning, unsupervised reinforcement learning, goal-conditioned reinforcement learning, model-based planning

Compressor summary: Curiosity-driven exploration helps learn robust dynamics models, but extracting goal-conditioned behavior from unsupervised data is challenging; combining model-based planning and graph-based aggregation improves zero-shot goal-reaching performance.


Weakly-Supervised Concealed Object Segmentation with SAM-based Pseudo Labeling and Multi-scale Feature Grouping

Chunming He, Kai Li, Yachao Zhang, Guoxia Xu, Longxiang Tang, Yulun Zhang, Zhenhua Guo, Xiu Li

https://openreview.net/forum?id=QlHosp050r

Keywords: Concealed Object Segmentation, Weakly-Supervised Learning, Segment Anything Model

Compressor summary: The paper proposes a new WSCOS method that tackles intrinsic similarity and weak supervision challenges by using multi-scale feature grouping and SAM-based segmentation with strategies to improve supervision quality, achieving state-of-the-art performance.


ProPILE: Probing Privacy Leakage in Large Language Models

Siwon Kim, Sangdoo Yun, Hwaran Lee, Martin Gubri, Sungroh Yoon, Seong Joon Oh

https://openreview.net/forum?id=QkLpGxUboF

Keywords: Personal identifiable information, Private data leakage, Large language model

Compressor summary: ProPILE is a tool that helps people check how much of their personal information might be leaked by large language models like OPT-1.3B trained on web data.


Neural-Logic Human-Object Interaction Detection

Liulei Li, Jianan Wei, Wenguan Wang, Yi Yang

https://openreview.net/forum?id=QjI36zxjbW

Keywords: Human-Object Interaction, Neuro-Symbolic Computing, Compositional Generalization

Compressor summary: LogicHOI is a new HOI detector that uses neural-logic reasoning and Transformer to infer feasible interactions between entities, taking into account affordances and proxemics, and achieves significant improvements over existing methods.


Adaptive Privacy Composition for Accuracy-first Mechanisms

Ryan Rogers, Gennady Samorodnitsky, Steven Wu, Aaditya Ramdas

https://openreview.net/forum?id=QezJbfW01r

Keywords: differential privacy, brownian motion, composition, martingale

Compressor summary: The authors develop privacy filters that combine differentially private mechanisms with ex-post private mechanisms, such as noise reduction, for better performance in tasks requiring accuracy and privacy.


Adaptive Data Analysis in a Balanced Adversarial Model

Kobbi Nissim, Uri Stemmer, Eliad Tsfadia

https://openreview.net/forum?id=QatZNssk7T

Keywords: Adaptive Data Analysis, Differential Privacy, Statistical Queries

Compressor summary: The paper investigates adaptive data analysis with a balanced adversary model, where the analyst does not know the underlying distribution and shows that it is hard to answer more than n^2 queries under standard cryptography assumptions.


Counterfactually Fair Representation

Zhiqun Zuo, Mohammad Mahdi Khalili, Xueru Zhang

https://openreview.net/forum?id=QZo1cge4Tc

Keywords: Counterfactual fairness, Representation learning

Compressor summary: This paper introduces a new algorithm for training machine learning models that are fair under Counterfactual Fairness, a notion that requires individuals to be treated equally regardless of their social group, by using all available features instead of excluding those related to sensitive attributes.


CoDA: Collaborative Novel Box Discovery and Cross-modal Alignment for Open-vocabulary 3D Object Detection

Yang Cao, Yihan Zeng, Hang Xu, Dan Xu

https://openreview.net/forum?id=QW5ouyyIgG

Keywords: 3D vision, open-vocabulary perception, multi-modal learning, point cloud, 3D object detection

Compressor summary: The paper proposes CoDA, a unified framework for simultaneous novel object localization and classification in 3D scenes using both geometric and semantic priors, as well as cross-modal alignment between point cloud and image/text modalities.


Bottleneck Structure in Learned Features: Low-Dimension vs Regularity Tradeoff

Arthur Jacot

https://openreview.net/forum?id=QVpfk2C3Dm

Keywords: Feature Learning, Symmetry Learning, Theory of Deep Learning, Weight Decay

Compressor summary: The paper investigates how deep neural networks learn low-dimensional representations and balance complexity/irregularity, proving a conjectured bottleneck structure in the limit of infinite depth.


FaceDNeRF: Semantics-Driven Face Reconstruction, Prompt Editing and Relighting with Diffusion Models

Hao ZHANG, Tianyuan DAI, Yanbo Xu, Yu-Wing Tai, Chi-Keung Tang

https://openreview.net/forum?id=QUkYZNhfc6

Keywords: NeRF Editing, NeRF Relighting, Face, Diffusion model, 3d synthesis, GAN inversion

Compressor summary: The paper introduces FaceDNeRF, a method to create high-quality 3D faces from single images with semantic editing and relighting capabilities, outperforming existing 2D editing approaches.


Networks are Slacking Off: Understanding Generalization Problem in Image Deraining

Jinjin Gu, Xianzheng Ma, Xiangtao Kong, Yu Qiao, Chao Dong

https://openreview.net/forum?id=QRWA5nTWuM

Keywords: Image Deraining, Generalization, Interpretation

Compressor summary: Despite being successful in labs, deep deraining networks struggle with real-world applications due to overfitting; simplifying training background images can improve their generalization performance.


CADet: Fully Self-Supervised Out-Of-Distribution Detection With Contrastive Learning

Charles Guille-Escuret, Pau Rodriguez, David Vazquez, Ioannis Mitliagkas, Joao Monteiro

https://openreview.net/forum?id=QRAS5wSgEy

Keywords: Contrastive learning, OOD detection, adversarial detection, MMD, ImageNet-O, Anomaly detection, CIFAR-10.1

Compressor summary: The authors propose a novel self-supervised contrastive learning method (CADet) that can detect both unseen classes and adversarial perturbations in machine learning systems without needing labels or additional OOD samples.


Fractal Landscapes in Policy Optimization

Tao Wang, Sylvia Lee Herbert, Sicun Gao

https://openreview.net/forum?id=QQidjdmyPp

Keywords: Reinforcement learning, policy gradient, non-smooth landscape

Compressor summary: The paper proposes a framework to understand and handle non-smooth or fractal optimization landscapes in deep reinforcement learning, which can cause failures in policy gradient methods.


Adversarial Self-Training Improves Robustness and Generalization for Gradual Domain Adaptation

Lianghe Shi, Weiwei Liu

https://openreview.net/forum?id=QNUs3Ramad

Keywords: learning theory

Compressor summary: Gradual self-training with adversarial training improves both clean and adversarial accuracies in domains with intermediate adaptations, as it outperforms standard training when dealing with incorrect pseudo-labels.


Where2Explore: Few-shot Affordance Learning for Unseen Novel Categories of Articulated Objects

Chuanruo Ning, Ruihai Wu, Haoran Lu, Kaichun Mo, Hao Dong

https://openreview.net/forum?id=QLllDwizVd

Keywords: articulated object manipulation, few-shot learning, visual affordance for robotics

Compressor summary: Where2Explore is a framework that helps robots learn how to manipulate novel object categories by efficiently exploring local geometries shared across different objects.


An Optimal and Scalable Matrix Mechanism for Noisy Marginals under Convex Loss Functions

Yingtai Xiao, Guanlin He, Danfeng Zhang, Daniel Kifer

https://openreview.net/forum?id=QKSejqE8Vp

Keywords: differential privacy, marginals, matrix mechanism, scalability

Compressor summary: ResidualPlanner is an optimal and scalable matrix mechanism for confidentiality-protecting data release that can optimize various loss functions and handle large scale settings efficiently.


Causal normalizing flows: from theory to practice

Adrián Javaloy, Pablo Sanchez Martin, Isabel Valera

https://openreview.net/forum?id=QIFoCI7ca1

Keywords: causality, causal inference, normalizing flows, identifiability, interventions, counterfactuals

Compressor summary: The authors present a method to identify and learn causal models from observational data using autoregressive normalizing flows, and demonstrate its applicability on real-world problems with mixed discrete-continuous data.


Lossy Image Compression with Conditional Diffusion Models

Ruihan Yang, Stephan Mandt

https://openreview.net/forum?id=QIBpzaDCAv

Keywords: generative model, diffusion model, image compression, computer vision

Compressor summary: The paper presents a lossy image compression framework using diffusion generative models that maps images to and from latent space, achieving good performance on various metrics.


RADAR: Robust AI-Text Detection via Adversarial Learning

Xiaomeng Hu, Pin-Yu Chen, Tsung-Yi Ho

https://openreview.net/forum?id=QGrkbaan79

Keywords: Large Language Models, Text Detection, Adversarial Learning, Paraphrase

Compressor summary: RADAR is a framework that uses adversarial learning to train a robust AI-text detector capable of detecting paraphrased texts generated by large language models, outperforming existing methods.


Learning Large-scale Neural Fields via Context Pruned Meta-Learning

Jihoon Tack, Subin Kim, Sihyun Yu, Jaeho Lee, Jinwoo Shin, Jonathan Richard Schwarz

https://openreview.net/forum?id=QGmNMtK3pQ

Keywords: Meta-Learning, Efficient Meta-Learning, Neural Fields, Implicit Neural Representations, Data Pruning

Compressor summary: The paragraph describes an efficient optimization-based meta-learning technique for large-scale neural field training that reduces memory usage, improves model quality, and allows fast learning of high-quality neural fields across multiple modalities.


Neural Oscillators are Universal

Samuel Lanthaler, T. Konstantin Rusch, Siddhartha Mishra

https://openreview.net/forum?id=QGQsOZcQ2H

Keywords: neural ODE, universal approximation, oscillator

Compressor summary: The paper introduces neural oscillators as a universal class for machine learning architectures and proves their ability to approximate continuous and causal operator mappings using forced harmonic oscillators with a nonlinear read-out.


CAT-Walk: Inductive Hypergraph Learning via Set Walks

Ali Behrouz, Farnoosh Hashemi, Sadaf Sadeghian, Margo Seltzer

https://openreview.net/forum?id=QG4nJBNEar

Keywords: Hypergraph Learning, Temporal Networks, Higher-order Temporal Motifs, Inductive Representation Learning

Compressor summary: CAT-Walk is a method to learn dynamic laws of temporal hypergraphs using a higher-order walk, SetWalk, and an adaptive pooling strategy, SetMixer, that achieves high performance in various tasks.


Nearly Tight Bounds For Differentially Private Multiway Cut

Mina Dalirrooyfard, Slobodan Mitrovic, Yuriy Nevmyvaka

https://openreview.net/forum?id=QDByreuQyk

Keywords: Differential Privacy, clustering, multiway cut, min cut, graph partitioning

Compressor summary: The authors study the complexity of maintaining differential privacy while finding minimal edge removals to disconnect two or more nodes in a graph, and develop efficient algorithms for this problem with near optimal trade-offs between privacy and efficiency.


DiffAttack: Evasion Attacks Against Diffusion-Based Adversarial Purification

Mintong Kang, Dawn Song, Bo Li

https://openreview.net/forum?id=QB7ot7p6j7

Keywords: adversarial attack, adversarial purification, adversarial robustness, diffusion model

Compressor summary: The paper proposes DiffAttack, an efficient method to break diffusion-based purification defenses in adversarial examples, improving on existing attacks by over 10% on ImageNet and over 20% on CIFAR-10.


SegRefiner: Towards Model-Agnostic Segmentation Refinement with Discrete Diffusion Process

Mengyu Wang, Henghui Ding, Jun Hao Liew, Jiajun Liu, Yao Zhao, Yunchao Wei

https://openreview.net/forum?id=Q9CNA7B7v2

Keywords: Refinement, Segmentation, Discrete Diffusion

Compressor summary: The paper introduces SegRefiner, a model-agnostic method that improves object masks using denoising diffusion steps and shows its superior performance on various segmentation tasks.


Unified 3D Segmenter As Prototypical Classifiers

Zheyun Qin, Cheng Han, Qifan Wang, Xiushan Nie, Yilong Yin, Xiankai Lu

https://openreview.net/forum?id=Q6zd1hr7sD

Keywords: Point Cloud Segmentation, Prototypical Classifier, Unified Framework

Compressor summary: ProtoSEG is a prototype-based model that unifies semantic, instance, and panoptic segmentation tasks for point clouds using a Transformer architecture and achieves competitive results on several benchmarks.


Refining Diffusion Planner for Reliable Behavior Synthesis by Automatic Detection of Infeasible Plans

Kyowoon Lee, Seongun Kim, Jaesik Choi

https://openreview.net/forum?id=Q5tuGgqJwt

Keywords: Offline Reinforcement Learning, Trajectory Optimization, Diffusion Models, Sequential Decision Making

Compressor summary: The paper proposes a method to improve the reliability of long-horizon plans generated by diffusion models using a restoration gap metric and an attribution map regularizer, which enables refinement of infeasible plans and provides explainability.


VanillaNet: the Power of Minimalism in Deep Learning

Hanting Chen, Yunhe Wang, Jianyuan Guo, Dacheng Tao

https://openreview.net/forum?id=Q5Eb6qIKux

Keywords: computer vision, foundation models.

Compressor summary: VanillaNet is a simple neural network architecture that achieves high performance in computer vision tasks while being efficient and easy to deploy.


Fast and Simple Spectral Clustering in Theory and Practice

Peter Macgregor

https://openreview.net/forum?id=Q3FXnCPZ1X

Keywords: spectral clustering, power method, spectral graph theory, graph algorithms

Compressor summary: The paper proposes a fast and accurate spectral clustering algorithm using fewer eigenvectors and evaluates its performance on various datasets.


Approximate Allocation Matching for Structural Causal Bandits with Unobserved Confounders

Lai Wei, Muhammad Qasim Elahi, Mahsa Ghasemi, Murat Kocaoglu

https://openreview.net/forum?id=Q3CRHnttxW

Keywords: multi-armed bandits, causal Inference, sequential decision-making

Compressor summary: Structural causal bandit is an online decision-making framework using causal models to learn from unknown environments, balancing exploration and exploitation for maximizing rewards and minimizing regret.


OneNet: Enhancing Time Series Forecasting Models under Concept Drift by Online Ensembling

YiFan Zhang, Qingsong Wen, Xue Wang, Weiqi Chen, Liang Sun, Zhang Zhang, Liang Wang, Rong Jin, Tieniu Tan

https://openreview.net/forum?id=Q25wMXsaeZ

Keywords: Time series forecasting, concept drift, online learning, online convex programming

Compressor summary: OneNet is an online time series forecasting algorithm that combines two models, one for temporal dependency and one for cross-variable dependency, using reinforcement learning to adjust their weights dynamically and improve performance against concept drift.


Near-optimal learning with average Hölder smoothness

Guy Kornowski, Steve Hanneke, Aryeh Kontorovich

https://openreview.net/forum?id=Q0ntwxVtcy

Keywords: Hölder smoothness, average smoothness, bracketing numbers, generalization, risk bounds, metric space

Compressor summary: The paper generalizes a smoothness measure for functions and shows that using an average smoothness instead of a worst-case one leads to better learning rates in different settings.


Does Visual Pretraining Help End-to-End Reasoning?

Chen Sun, Calvin Luo, Xingyi Zhou, Anurag Arnab, Cordelia Schmid

https://openreview.net/forum?id=PzYAMXmIT3

Keywords: visual reasoning, self-supervised learning

Compressor summary: The authors propose a self-supervised method to compress video frames into tokens using a transformer network, and show that visual pretraining is crucial for end-to-end visual reasoning.


Vulnerabilities in Video Quality Assessment Models: The Challenge of Adversarial Attacks

Aoxiang Zhang, Yu Ran, Weixuan Tang, Yuan-Gen Wang

https://openreview.net/forum?id=Pz8xvVCLNJ

Keywords: video quality assessment, adversarial attack, black-box, just noticeable difference

Compressor summary: The paper introduces a method to evaluate the robustness of NR-VQA models against adversarial attacks and proposes a patch-based random search technique for black-box attacks.


Private estimation algorithms for stochastic block models and mixture models

Hongjie Chen, Vincent Cohen-Addad, Tommaso d'Orsi, Alessandro Epasto, Jacob Imola, David Steurer, Stefan Tiegel

https://openreview.net/forum?id=Pya0kCEpDk

Keywords: differential privacy, stochastic block model, Gaussian mixture model, sum of squares

Compressor summary: The authors propose efficient private estimation algorithms that match non-private counterparts in high-dimensional settings and apply them to stochastic block models and mixtures of spherical Gaussians problems.


Learning Exponential Families from Truncated Samples

Jane Lee, Andre Wibisono, Manolis Zampetakis

https://openreview.net/forum?id=PxcWJqO3qj

Keywords: truncated statistics, robustness, exponential families, extrapolation

Compressor summary: The paper presents an estimation algorithm that can extrapolate from truncated data for log-concave exponential families, using Projected Stochastic Gradient Descent, which is simpler and more efficient than previous methods.


Understanding Few-Shot Learning: Measuring Task Relatedness and Adaptation Difficulty via Attributes

Minyang Hu, Hong Chang, Zong Guo, Bingpeng Ma, Shiguang Shan, Xilin CHEN

https://openreview.net/forum?id=Pvgxecj5aS

Keywords: Few-shot Learning, Meta-Learning, Task Relatedness, Task Adaptation Difficulty

Compressor summary: The paper proposes a metric called Task Attribute Distance (TAD) to measure the similarity between training and novel tasks in few-shot learning, and shows how it relates to adaptation difficulty.


Interpretable Graph Networks Formulate Universal Algebra Conjectures

Francesco Giannini, Stefano Fioravanti, Oguzhan Keskin, Alisia Maria Lupidi, Lucie Charlotte Magister, Pietro Lio, Pietro Barbiero

https://openreview.net/forum?id=Psnph85KYc

Keywords: universal algebra, interpretability, graph neural networks, concept-based models

Compressor summary: This paper uses AI to explore Universal Algebra's conjectures, creating datasets and a new neural layer for interpretable graph networks that can validate and generate conjectures.


Contextually Affinitive Neighborhood Refinery for Deep Clustering

Chunlin Yu, Ye Shi, Jingya Wang

https://openreview.net/forum?id=Psj0jHocm1

Keywords: Deep Clustering, Self-supervised learning, re-ranking

Compressor summary: The paragraph discusses a new method for deep clustering that uses online re-ranking to find more informative neighbors, cross-view consistency, and boundary filtering to handle noisy data and improve performance on benchmark datasets.


The Shaped Transformer: Attention Models in the Infinite Depth-and-Width Limit

Lorenzo Noci, Chuning Li, Mufan Bill Li, Bobby He, Thomas Hofmann, Chris J. Maddison, Daniel M. Roy

https://openreview.net/forum?id=PqfPjS9JRX

Keywords: Deep Learning Theory, Covariance SDE, Attention Mechanism, Infinite-Depth-and-Width, Scaling Limit

Compressor summary: The paper studies how modifying a Softmax-based attention model with skip connections and width-dependent temperature can improve trainability and stability in deep learning models, leading to a new architecture called shaped Transformer.


Metis: Understanding and Enhancing In-Network Regular Expressions

Zhengxin Zhang, Yucheng Huang, Guanglin Duan, Qing Li, Dan Zhao, Yong Jiang, Lianbo Ma, Xi Xiao, Hengyang Xu

https://openreview.net/forum?id=Pplq1TRnma

Keywords: Network Security; Regular Expression; Knowledge Distillation; Machine Learning; Programmable Switch

Compressor summary: Metis converts regular expressions to neural networks for better accuracy and device deployment in networking tasks.


A Spectral Algorithm for List-Decodable Covariance Estimation in Relative Frobenius Norm

Ilias Diakonikolas, Daniel Kane, Jasper C.H. Lee, Ankit Pensia, Thanasis Pittas

https://openreview.net/forum?id=PpI7XvOXkF

Keywords: robust statistics, covariance estimation, list-decodable learning

Compressor summary: The paper presents a spectral algorithm for list-decodable Gaussian covariance estimation that works with unknown fractions of Gaussian samples and can also learn robust partial clusters of GMMs efficiently.


What Knowledge Gets Distilled in Knowledge Distillation?

Utkarsh Ojha, Yuheng Li, Anirudh Sundara Rajan, Yingyu Liang, Yong Jae Lee

https://openreview.net/forum?id=Poj71ASubN

Keywords: knowledge distillation

Compressor summary: The paragraph discusses the need to understand what kind of knowledge is transferred in knowledge distillation techniques and how it affects a student network's properties beyond task performance.


Goal Driven Discovery of Distributional Differences via Language Descriptions

Ruiqi Zhong, Peter Zhang, Steve Li, Jinwoo Ahn, Dan Klein, Jacob Steinhardt

https://openreview.net/forum?id=PnbCA4ylIc

Keywords: large language model, prompting, exploratory text analysis

Compressor summary: The authors introduce a new task (D5) that uses language models to automatically discover differences between two large corpora based on user-specified goals, and evaluate its performance on synthetic and real datasets.


FABind: Fast and Accurate Protein-Ligand Binding

Qizhi Pei, Kaiyuan Gao, Lijun Wu, Jinhua Zhu, Yingce Xia, Shufang Xie, Tao Qin, Kun He, Tie-Yan Liu, Rui Yan

https://openreview.net/forum?id=PnWakgg1RL

Keywords: protein-ligand docking

Compressor summary: The authors propose FABind, a fast and accurate model for predicting protein-ligand binding structures using a pocket prediction and docking approach that leverages ligand information.


Trajectory Alignment: Understanding the Edge of Stability Phenomenon via Bifurcation Theory

Minhak Song, Chulhee Yun

https://openreview.net/forum?id=PnJaA0A8Lr

Keywords: non-convex optimization, trajectory alignment of GD, edge of stability, progressive sharpening, bifurcation theory

Compressor summary: The paper investigates how the sharpness of neural networks changes during training, and shows that different gradient descent trajectories align on a bifurcation diagram when the Edge of Stability occurs.


Adaptive Principal Component Regression with Applications to Panel Data

Anish Agarwal, Keegan Harris, Justin Whitehouse, Steven Wu

https://openreview.net/forum?id=PmqBJ02V1p

Keywords: adaptive data collection, principal component regression, error-in-variables regression, panel data, synthetic controls, synthetic interventions, causal inference

Compressor summary: The paper introduces time-uniform finite sample guarantees for online principal component regression in error-in-variables settings and uses it to estimate counterfactual treatment effects in panel data with adaptive interventions.


Neural Processes with Stability

Huafeng Liu, Liping Jing, Jian Yu

https://openreview.net/forum?id=PmlNxZoXr4

Keywords: Neural processes, stability

Compressor summary: The paper introduces a method to improve neural processes by adding algorithmic stability, which leads to better accuracy and robustness.


Taking the neural sampling code very seriously: A data-driven approach for evaluating generative models of the visual system

Suhas Shrinivasan, Konstantin-Klemens Lurz, Kelli Restivo, George Denfield, Andreas S. Tolias, Edgar Y. Walker, Fabian H. Sinz

https://openreview.net/forum?id=Pl416tPkNv

Keywords: Neural Sampling Code, Probabilistic Inference, Bayesian Brain, Macaque V1, Natural Images, Population Recordings, Normalizing Flows, Probabilistic Models, Computational Neuroscience, Theoretical Neuroscience

Compressor summary: The authors propose a novel formalization of the Neural Sampling Code theory that allows fitting more flexible generative models to recorded neuronal activity, enabling quantitative evaluation and comparison of different models on natural images.


Truncating Trajectories in Monte Carlo Policy Evaluation: an Adaptive Approach

Riccardo Poiani, Nicole Nobili, Alberto Maria Metelli, Marcello Restelli

https://openreview.net/forum?id=PkKpTK7hJ6

Keywords: Reinforcement Learning, Policy Evaluation, Budget Optimization, Monte Carlo

Compressor summary: The paper proposes RIDO, an adaptive algorithm that splits the interaction budget into mini-batches and adjusts the length of trajectories to minimize the variance of policy return estimators in Monte Carlo Reinforcement Learning.


Optimal Preconditioning and Fisher Adaptive Langevin Sampling

Michalis Titsias

https://openreview.net/forum?id=Pk9CdOZYRA

Keywords: MCMC, Langevin diffusion, preconditioning, Fisher information, adaptive MCMC, score function

Compressor summary: The paper proposes an optimal preconditioning for the Langevin diffusion that improves the performance of the Metropolis adjusted Langevin algorithm (MALA) in high dimensions, and compares it with other methods.


ViCA-NeRF: View-Consistency-Aware 3D Editing of Neural Radiance Fields

Jiahua Dong, Yu-Xiong Wang

https://openreview.net/forum?id=Pk49a9snPe

Keywords: neural radiance field, diffusion model, editing

Compressor summary: ViCA-NeRF is a method for 3D editing with text instructions that uses view-consistency and two types of regularization to ensure multi-view consistency and produce detailed results.


One Less Reason for Filter Pruning: Gaining Free Adversarial Robustness with Structured Grouped Kernel Pruning

Shaochen Zhong, Zaichuan You, Jiamu Zhang, Sebastian Zhao, Zachary LeClaire, Zirui Liu, Daochen Zha, Vipin Chaudhary, Shuai Xu, Xia Hu

https://openreview.net/forum?id=Pjky9XG8zP

Keywords: pruning, structured pruning, adversarial robustness, grouped kernel pruning, CNN, one-shot

Compressor summary: The authors investigate the vulnerability of modern structured pruning methods to adversarial attacks and propose a new method that improves robustness while maintaining compression and acceleration benefits.


Implicit Bias of (Stochastic) Gradient Descent for Rank-1 Linear Neural Network

Bochen Lyu, Zhanxing Zhu

https://openreview.net/forum?id=PjBEUTVzoe

Keywords: implicit bias, gradient descent, stochastic gradient descent, linear networks

Compressor summary: The paper investigates the implicit bias of rank-1 linear networks and finds new insights on how gradient descent and stochastic gradient descent behave in over-parameterized regression problems, which could help understand deep learning better.


Learning Fine-grained View-Invariant Representations from Unpaired Ego-Exo Videos via Temporal Alignment

Zihui Xue, Kristen Grauman

https://openreview.net/forum?id=Pj6X6GqNy8

Keywords: fine-grained video understanding, egocentric video, self-supervised learning, temporal alignment

Compressor summary: The paragraph describes a new method called AE2 that learns action features invariant to viewpoints by aligning egocentric and exocentric videos, even when not captured at the same time or place, and shows its effectiveness on four datasets.


Provable Guarantees for Generative Behavior Cloning: Bridging Low-Level Stability and High-Level Behavior

Adam Block, Ali Jadbabaie, Daniel Pfrommer, Max Simchowitz, Russ Tedrake

https://openreview.net/forum?id=PhFVF0gwid

Keywords: Imitation Learning, Control, Diffusion Models, Optimal Transport

Compressor summary: The paper presents a theory for imitating expert actions using generative models and low-level controllers to ensure stability and match the distribution of the expert trajectories.


Overcoming Recency Bias of Normalization Statistics in Continual Learning: Balance and Adaptation

Yilin Lyu, Liyuan Wang, Xingxing Zhang, Zicheng Sun, Hang Su, Jun Zhu, Liping Jing

https://openreview.net/forum?id=Ph65E1bE6A

Keywords: Continual Learning, Batch Normalization, Recency Bias, Catastrophic Forgetting

Compressor summary: The paper analyzes the limitations of Batch Normalization in continual learning and proposes Adaptive Balance of BN, which adapts to task-wise contributions and balances BN statistics, achieving significant performance gains on benchmarks.


Behavior Alignment via Reward Function Optimization

Dhawal Gupta, Yash Chandak, Scott M. Jordan, Philip S. Thomas, Bruno Castro da Silva

https://openreview.net/forum?id=PfpAQuyZCB

Keywords: Reinforcement Learning, Behavior Alignment, Implicit Gradient, Bi-level Optimization

Compressor summary: The paper introduces a new framework that uses a bi-level objective to learn behavior alignment reward functions that integrate designer's heuristics with environment's primary rewards for guiding reinforcement learning agents efficiently and robustly.


Learning Transformer Programs

Dan Friedman, Alexander Wettig, Danqi Chen

https://openreview.net/forum?id=Pe9WxkN8Ff

Keywords: mechanistic interpretability, transformers

Compressor summary: The authors propose a method for training mechanistically interpretable Transformers by converting them into human-readable programs and show their effectiveness on various tasks, including NLP problems.


An Information-Theoretic Evaluation of Generative Models in Learning Multi-modal Distributions

Mohammad Jalali, Cheuk Ting Li, Farzan Farnia

https://openreview.net/forum?id=PdZhf6PiAb

Keywords: Generative Models; Evaluation in Learning; Information Measures

Compressor summary: The paper proposes an information-theoretic method to measure the number of modes in multi-modal image datasets using R\'enyi Kernel Entropy and evaluates state-of-the-art generative models with it.


Responsible AI (RAI) Games and Ensembles

Yash Gupta, Runtian Zhai, Arun Suggala, Pradeep Kumar Ravikumar

https://openreview.net/forum?id=PcNpL9Q39p

Keywords: Responsible AI, fairness, DRO, robustness

Compressor summary: The authors propose a general framework called Responsible AI (RAI) games to study societal effects of AI, and present two classes of algorithms for solving them.


On the Statistical Consistency of Risk-Sensitive Bayesian Decision-Making

Prateek Jaiswal, Harsha Honnappa, Vinayak Rao

https://openreview.net/forum?id=PcKHQFsvel

Keywords: Variational Bayes, Loss Calibration, Bayesian Statistics, Variational Inference, Statistical Theory

Compressor summary: The authors propose a new variational Bayesian method for risk-sensitive decision-making when the posterior distribution is difficult to compute, and study its theoretical and practical properties.


Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threat Models

Naman Deep Singh, Francesco Croce, Matthias Hein

https://openreview.net/forum?id=Pbpk9jUzAi

Keywords: adversarial robustness, deep learning, vision transformers, convnext

Compressor summary: The paper compares adversarial training on ImageNet for ViTs and ConvNeXts using different architectural modifications and shows that these changes affect robustness in different ways depending on the threat model.


Interpretable and Explainable Logical Policies via Neurally Guided Symbolic Abstraction

Quentin Delfosse, Hikaru Shindo, Devendra Singh Dhami, Kristian Kersting

https://openreview.net/forum?id=PbMBfRpVgU

Keywords: Reinforcement Learning, First-Order-Logic, Symbolic Abstraction, Interpretable Reinforcement Learning, Logic Reinforcement Learning

Compressor summary: NUDGE is a neuro-symbolic RL method that combines neural networks and differentiable logic to create interpretable and explainable policies, performing better than pure neural methods in various environments.


Performance Bounds for Policy-Based Average Reward Reinforcement Learning Algorithms

Yashaswini Murthy, Mehrdad Moharrami, R. Srikant

https://openreview.net/forum?id=PaSpImjKm2

Keywords: Average Reward MDPs, Reinforcement Learning Theory, Approximate Policy Iteration, Policy Based Methods, Performance Bounds

Compressor summary: The authors propose a solution to the open problem of obtaining performance bounds for approximate policy iteration and reinforcement learning algorithms in the average-reward setting, by deriving non-trivial finite time error bounds that converge to zero as policy evaluation and improvement errors decrease.


ALIM: Adjusting Label Importance Mechanism for Noisy Partial Label Learning

Mingyu Xu, Zheng Lian, Lei Feng, Bin Liu, Jianhua Tao

https://openreview.net/forum?id=PYSfn5xXEe

Keywords: Partial label learning; Noisy label learning

Compressor summary: The paragraph introduces a new method called "Adjusting Label Importance Mechanism" (ALIM) for handling noisy partial label learning, which reduces the impact of detection errors by adjusting the candidate set and model outputs.


Feature Learning for Interpretable, Performant Decision Trees

Jack Henry Good, Torin Kovach, Kyle Miller, Artur Dubrawski

https://openreview.net/forum?id=PYEgC56flW

Keywords: explainability, interpretability, decision tree, feature learning

Compressor summary: The text discusses a new system that combines sparse feature learning and differentiable decision tree construction to create small, interpretable trees with good performance.


Eliciting User Preferences for Personalized Multi-Objective Decision Making through Comparative Feedback

Han Shao, Lee Cohen, Avrim Blum, Yishay Mansour, Aadirupa Saha, Matthew Walter

https://openreview.net/forum?id=PYASzxr2OP

Keywords: preference learning, algorithms, linear model, Markov decision processes, learning theory, multi-objective decision making, preference elicitation

Compressor summary: The paper presents a method to learn user preferences over multiple objectives and find near-optimal policies based on feedback from comparisons between policies or trajectories.


SHOT: Suppressing the Hessian along the Optimization Trajectory for Gradient-Based Meta-Learning

JunHoo Lee, Jayeon Yoo, Nojun Kwak

https://openreview.net/forum?id=PXsqbAjpQd

Keywords: meta learning, Hessian, Gradient-Based meta learning, Feature Reuse, Implicit Prior

Compressor summary: SHOT is a new algorithm that suppresses the Hessian in gradient-based meta-learning, improving performance on few-shot learning tasks.


Universality laws for Gaussian mixtures in generalized linear models

Yatin Dandi, Ludovic Stephan, Florent Krzakala, Bruno Loureiro, Lenka Zdeborova

https://openreview.net/forum?id=PU3deePP2S

Keywords: theoretical analysis, high-dimensional statistics, Universality, weak convergence, mixture models, sampling, statistical physics

Compressor summary: The paragraph discusses results in high-dimensional statistics under the Gaussian mixture hypothesis and provides rigorous proofs for applying these results to a general class of datasets using generalized linear models and investigating their asymptotic joint statistics.


Simplicity Bias in 1-Hidden Layer Neural Networks

Depen Morwani, jatin batra, Prateek Jain, Praneeth Netrapalli

https://openreview.net/forum?id=PTvxck0QDE

Keywords: Simplicity Bias, Gradient Descent, Implicit Bias, Neural Networks

Compressor summary: The paper investigates simplicity bias in neural networks with one hidden layer, showing that they learn only the simplest features even when more complex ones exist, and propose an ensemble method to improve robustness.


Beyond Exponential Graph: Communication-Efficient Topologies for Decentralized Learning via Finite-time Convergence

Yuki Takezawa, Ryoma Sato, Han Bao, Kenta Niwa, Makoto Yamada

https://openreview.net/forum?id=PSngfm5B9q

Keywords: decentralized learning, distributed optimization, network topology, consensus rate

Compressor summary: The study proposes a new graph topology for decentralized learning that combines fast consensus rate and small maximum degree, improving accuracy and communication efficiency over existing topologies.


Policy Space Diversity for Non-Transitive Games

Jian Yao, Weiming Liu, Haobo Fu, Yaodong Yang, Stephen Marcus McAleer, QIANG FU, Yang Wei

https://openreview.net/forum?id=PRgvdEbhdH

Keywords: Policy Diversity, Policy-Space Response Oracles, Nash Equilibrium, Multi-agent Reinforcement Learning

Compressor summary: The paper proposes a new diversity metric for Policy Space Response Oracles (PSRO) that improves the approximation of Nash Equilibrium and presents PSD-PSRO, a variant that converges and performs better in experiments.


Efficient Beam Tree Recursion

Jishnu Ray Chowdhury, Cornelia Caragea

https://openreview.net/forum?id=PR5znB6BZ2

Keywords: Recursive Models, Recursive Neural Networks, RvNNs, Length Generalization, Structured Encoding, Representation Learning

Compressor summary: The paper proposes strategies to reduce memory usage of BT-RvNN, a neural network for encoding sentences, and to use it as a token contextualizer.


Fair Adaptive Experiments

Waverly Wei, Xinwei Ma, Jingshen Wang

https://openreview.net/forum?id=PMvudWa53L

Keywords: Adaptive Randomized Experiment; Adaptive Design; Causal Inference

Compressor summary: The paragraph describes a fair adaptive experiment strategy that balances data efficiency, equity, and participant welfare in randomized experiments without assuming parametric models on outcomes.


How Re-sampling Helps for Long-Tail Learning?

Jiang-Xin Shi, Tong Wei, Yuke Xiang, Yu-Feng Li

https://openreview.net/forum?id=PLzCXefcpE

Keywords: long-tail learning, class-imbalanced learning, re-sampling

Compressor summary: This paper investigates how re-sampling affects long-tail learning, finding that it can improve generalization if the training data has relevant contexts but may introduce spurious correlations otherwise; they propose a new context shift augmentation module to address this issue.


Dynamics Generalisation in Reinforcement Learning via Adaptive Context-Aware Policies

Michael Beukman, Devon Jarvis, Richard Klein, Steven James, Benjamin Rosman

https://openreview.net/forum?id=PJhjkSFlbG

Keywords: Deep Reinforment Learning, Contextual Markov Decision Process, Neural Network Architecture

Compressor summary: The paper proposes the Decision Adapter, a neural network architecture that incorporates context information into behaviour learning and improves generalisation performance and robustness.


User-Level Differential Privacy With Few Examples Per User

Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Raghu Meka, Chiyuan Zhang

https://openreview.net/forum?id=PITeSdYQkv

Keywords: differential privacy, user-level privacy, PAC learning

Compressor summary: This paper presents algorithms for user-level differential privacy in the example-scarce regime, improving existing results and providing new bounds for various learning tasks.


Domain Watermark: Effective and Harmless Dataset Copyright Protection is Closed at Hand

Junfeng Guo, Yiming Li, Lixu Wang, Shu-Tao Xia, Heng Huang, Cong Liu, Bo Li

https://openreview.net/forum?id=PIDNxRRJ8w

Keywords: Ownership Verification, Dataset Protection, Copyright Protection, Backdoor Attack, AI Security

Compressor summary: The paper proposes a new approach for dataset ownership verification using domain watermarks that make watermarked DNNs classify hard samples correctly while maintaining stealthiness.


Contrastive Moments: Unsupervised Halfspace Learning in Polynomial Time

Xinyuan Cao, Santosh Vempala

https://openreview.net/forum?id=PHbqznMa1i

Keywords: Unsupervised Learning, Learning Halfspaces, Non-Gaussian Component analysis

Compressor summary: The paper presents a polynomial-time algorithm for learning high-dimensional halfspaces with margins from an unknown affine transformation and a symmetric one-dimensional logconcave distribution, using only the first two moments of suitable re-weightings and achieving TV distance guarantees.


An Iterative Self-Learning Framework for Medical Domain Generalization

Zhenbang Wu, Huaxiu Yao, David Liebovitz, Jimeng Sun

https://openreview.net/forum?id=PHKkBbuJWM

Keywords: healthcare, clinical predictive model, domain generalization

Compressor summary: SLGD is a self-learning framework that trains personalized classifiers for decoupled domains, improving deep learning models' domain generalization performance on EHR datasets.


ATMAN: Understanding Transformer Predictions Through Memory Efficient Attention Manipulation

Björn Deiseroth, Mayukh Deb, Samuel Weinbach, Manuel Brack, Patrick Schramowski, Kristian Kersting

https://openreview.net/forum?id=PBpEb86bj7

Keywords: explainability, attention manipulation, perturbation, large language model, multi-modality, generative decoder, efficiency, transformer

Compressor summary: AtMan explains generative transformer models efficiently by manipulating their attention mechanisms using cosine similarity neighborhood in the embedding space instead of backpropagation.


Effective Robustness against Natural Distribution Shifts for Models with Different Training Data

Zhouxing Shi, Nicholas Carlini, Ananth Balashankar, Ludwig Schmidt, Cho-Jui Hsieh, Alex Beutel, Yao Qin

https://openreview.net/forum?id=PAYXfIUKWY

Keywords: Effective robustness, natural distribution shifts, out-of-distribution robustness

Compressor summary: The paper introduces a new metric to evaluate and compare the effective robustness of models trained on different data distributions by controlling for accuracy on multiple in-distribution test sets.


Type-to-Track: Retrieve Any Object via Prompt-based Tracking

Pha Nguyen, Kha Gia Quach, Kris M. Kitani, Khoa Luu

https://openreview.net/forum?id=PARMyW6xX0

Keywords: Grounded Object Tracking, Multiple Object Tracking, Vision Language

Compressor summary: The paper proposes a new method for tracking objects in videos using natural language captions and introduces a new dataset, GroOT, as well as evaluation protocols and metrics for this task. It also presents a fast and accurate model called MENDER based on tensor decomposition.


Cause-Effect Inference in Location-Scale Noise Models: Maximum Likelihood vs. Independence Testing

Xiangyu Sun, Oliver Schulte

https://openreview.net/forum?id=P9I2VQv1uC

Keywords: Causal Discovery, Cause-Effect Inference, Location-Scale Noise Models

Compressor summary: The paper discusses a problem of inferring cause-effect relationships between random variables and proposes a more robust method for selecting causal models when the noise distribution assumptions are incorrect.


Temporal Robustness against Data poisoning

Wenxiao Wang, Soheil Feizi

https://openreview.net/forum?id=P5vzRpoOj2

Keywords: Robustness, Data Poisoning, Security, Machine Learning, Backdoor, Adversarial

Compressor summary: The paper proposes a temporal threat model of data poisoning that considers how long an attack started and lasted, and evaluates a baseline defense called temporal aggregation.


Window-Based Distribution Shift Detection for Deep Neural Networks

Guy Bar-Shalom, Yonatan Geifman, Ran El-Yaniv

https://openreview.net/forum?id=P3n4wFJGs5

Keywords: Distribution shift detection, Window-based detection

Compressor summary: The authors propose a method to detect distribution deviations in deep neural networks, which is faster and more efficient than existing methods, and can handle large datasets.


EvoFed: Leveraging Evolutionary Strategies for Communication-Efficient Federated Learning

Mohammad Mahdi Rahimi, Hasnain Irshad Bhatti, Younghyun Park, Humaira Kousar, Do-Yeon Kim, Jaekyun Moon

https://openreview.net/forum?id=P3Z59Okb5I

Keywords: evolutionary strategies, federated learning, gradient compression, distributed learning

Compressor summary: EvoFed is a novel approach for federated learning that reduces communication costs by using evolutionary strategies and fitness-based information sharing instead of exchanging model parameters.


Hierarchically Gated Recurrent Neural Network for Sequence Modeling

Zhen Qin, Songlin Yang, Yiran Zhong

https://openreview.net/forum?id=P1TCHxJwLB

Keywords: RNN, Sequence Modeling, NLP

Compressor summary: The paper introduces HGRN, a gated linear RNN model that uses forget gates with a learnable lower bound to enable efficient long-term dependency modeling in different tasks.


Differentially Private Image Classification by Learning Priors from Random Processes

Xinyu Tang, Ashwinee Panda, Vikash Sehwag, Prateek Mittal

https://openreview.net/forum?id=P0Avuii9iI

Keywords: Differential privacy, image classification, deep learning

Compressor summary: DP-RandP improves privacy-utility tradeoff of DP-SGD by learning priors from random image generation and transferring them to private data, achieving state-of-the-art accuracy on various datasets.


Rethinking the Backward Propagation for Adversarial Transferability

Xiaosen Wang, Kangheng Tong, Kun He

https://openreview.net/forum?id=OzpTd2EsH1

Keywords: Adversarial examples, Convolutional neural networks, Adversarial transferability, Backward propagation

Compressor summary: The paper proposes Backward Propagation Attack (BPA), a novel method that mitigates information loss in non-linear layers to improve the transferability of adversarial examples.


Achieving $\mathcal{O}(\epsilon^{-1.5})$ Complexity in Hessian/Jacobian-free Stochastic Bilevel Optimization

Yifan Yang, Peiyao Xiao, Kaiyi Ji

https://openreview.net/forum?id=OzjBohmLvE

Keywords: Stochastic bilevel optimization, Hessian-free algorithms, near-optimal complexity

Compressor summary: The paper introduces FdeHBO, a novel method for solving nonconvex-strongly-convex bilevel optimization without second-order derivative computation, achieving an $\mathcal{O}(\epsilon^{-1.5})$ sample complexity and iteration complexity.


VCC: Scaling Transformers to 128K Tokens or More by Prioritizing Important Tokens

Zhanpeng Zeng, Cole Hawkins, Mingyi Hong, Aston Zhang, Nikolaos Pappas, Vikas Singh, Shuai Zheng

https://openreview.net/forum?id=Ozc8XVzwd4

Keywords: efficient, transformer, roberta, T5, language modeling, question answering, summarization

Compressor summary: The paper proposes a method to improve the efficiency and performance of Transformers for ultra long sequences by compressing them based on VIP-tokens that are most relevant to the final prediction.


Jigsaw: Learning to Assemble Multiple Fractured Objects

Jiaxin Lu, Yifan Sun, Qixing Huang

https://openreview.net/forum?id=OwpaO4w6K7

Keywords: Shape Matching; Reassembly; Shape Segmentation;

Compressor summary: Jigsaw is a novel framework that uses hierarchical features to assemble physically broken 3D objects from multiple pieces and outperforms existing methods.


Towards Unbounded Machine Unlearning

Meghdad Kurmanji, Peter Triantafillou, Jamie Hayes, Eleni Triantafillou

https://openreview.net/forum?id=OveBaTtUAT

Keywords: machine unlearning, deep learning

Compressor summary: The paper introduces SCRUB, a novel unlearning algorithm for neural networks that performs well across various applications and metrics, including privacy protection.


Importance-aware Co-teaching for Offline Model-based Optimization

Ye Yuan, Can Chen, Zixuan Liu, Willie Neiswanger, Xue Liu

https://openreview.net/forum?id=OvPnc5kVsb

Keywords: offline model-based optimization, co-teaching, meta-learning, sample reweighting

Compressor summary: ICT is a method that uses three symmetric proxies and pseudo-labeling to improve offline model-based optimization, addressing the out-of-distribution issue in gradient ascent.


Learning Multi-agent Behaviors from Distributed and Streaming Demonstrations

Shicheng Liu, Minghui Zhu

https://openreview.net/forum?id=Ou1VRZ4j4y

Keywords: inverse reinforcement learning; distributed online bi-level optimization

Compressor summary: The paper presents MA-BIRDS, an algorithm that infers expert behaviors from sequential demonstrations using a distributed bi-level optimization framework and achieves consensus, low regret, and sub-linear constraint violation.


Learning to Augment Distributions for Out-of-distribution Detection

Qizhou Wang, Zhen Fang, Yonggang Zhang, Feng Liu, Yixuan Li, Bo Han

https://openreview.net/forum?id=OtU6VvXJue

Keywords: OOD Detection

Compressor summary: The text discusses a problem in open-world classification systems and proposes a solution called Distributional-Augmented OOD Learning (DAOL) that improves performance by reducing distribution discrepancy between auxiliary and unseen out-of-distribution data.


Weakly Supervised 3D Open-vocabulary Segmentation

Kunhao Liu, Fangneng Zhan, Jiahui Zhang, MUYU XU, Yingchen Yu, Abdulmotaleb El Saddik, Christian Theobalt, Eric Xing, Shijian Lu

https://openreview.net/forum?id=Orp1K2dZvY

Keywords: 3D, open-vocabulary segmentation, neural radiance field

Compressor summary: The paper proposes a method to learn 3D open-vocabulary segmentation from 2D images and text using pre-trained models CLIP and DINO without any manual annotations.


Modulated Neural ODEs

Ilze Amanda Auzina, Cagatay Yildiz, Sara Magliacane, Matthias Bethge, Efstratios Gavves

https://openreview.net/forum?id=Op9z2QfXbC

Keywords: Neural ODEs, Modulator Variables, Dynamical Systems, Disentanglment

Compressor summary: The paper introduces MoNODEs, a novel framework for neural ordinary differential equations that improves generalization and far-horizon forecasting by learning time-invariant modulator variables from data.


Semantic Image Synthesis with Unconditional Generator

JungWoo Chae, Hyunin Cho, Sooyeon Go, Kyungmook Choi, Youngjung Uh

https://openreview.net/forum?id=OoPLRGBKjM

Keywords: Generative model

Compressor summary: The paper presents a novel method for semantic image synthesis that uses a proxy mask derived from intermediate feature maps to guide a pretrained unconditional generator without heavy annotation.


Reading Relevant Feature from Global Representation Memory for Visual Object Tracking

Xinyu Zhou, Pinxue Guo, Lingyi Hong, Jinglun Li, Wei Zhang, Weifeng Ge, Wenqiang Zhang

https://openreview.net/forum?id=On0IDMYKw2

Keywords: object tracking;global representation memory;transformer

Compressor summary: The text introduces a new tracking paradigm that uses a relevance attention mechanism and a global representation memory to select and read the most relevant historical information for visual object tracking, improving performance and reducing redundancy.


MAViL: Masked Audio-Video Learners

Po-Yao Huang, Vasu Sharma, Hu Xu, Chaitanya Ryali, Haoqi Fan, Yanghao Li, Shang-Wen Li, Gargi Ghosh, Jitendra Malik, Christoph Feichtenhofer

https://openreview.net/forum?id=OmTMaTbjac

Keywords: self-supervised learning, audio representation learning, audio classification

Compressor summary: MAViL learns audio-visual representations using three forms of self-supervision, achieves state-of-the-art performance in audio-video classification and improves the quality of both modalities individually.


Federated Multi-Objective Learning

Haibo Yang, Zhuqing Liu, Jia Liu, Chaosheng Dong, Michinari Momma

https://openreview.net/forum?id=OlSTwlz96r

Keywords: Multi-Objective Learning, Federated Learning

Compressor summary: The text introduces a new federated multi-objective learning framework that allows multiple clients to collaboratively solve a multi-objective optimization problem while keeping their data private and proposes two new algorithms for this framework.


KD-Zero: Evolving Knowledge Distiller for Any Teacher-Student Pairs

Lujun Li, Peijie Dong, Anggeng Li, Zimian Wei, Yang Ya

https://openreview.net/forum?id=OlMKa5YZ8e

Keywords: Knowledge distillation

Compressor summary: The paper introduces KD-Zero, a novel framework that uses evolutionary search to automatically discover optimal knowledge distillation designs for any teacher-student model pairs, achieving superior performance across various tasks and architectures.


Percentile Criterion Optimization in Offline Reinforcement Learning

Cyrus Cousins, Elita Lobo, Marek Petrik, Yair Zick

https://openreview.net/forum?id=OjlZqQzw51

Keywords: Reinforcement Learning, Bayesian Uncertainty, Robustness

Compressor summary: The paper proposes a new dynamic programming algorithm for reinforcement learning that optimizes the percentile criterion without explicitly creating uncertainty sets, leading to smaller sets and more efficient policies.


Eliminating Catastrophic Overfitting Via Abnormal Adversarial Examples Regularization

Runqi Lin, Chaojian Yu, Tongliang Liu

https://openreview.net/forum?id=Oj7Mrb4009

Keywords: adversarial training, catastrophic overfitting

Compressor summary: The paper proposes AAER, a method to prevent catastrophic overfitting in SSAT by regularizing the variation of abnormal adversarial examples, which are associated with classifier distortion.


Transformers are uninterpretable with myopic methods: a case study with bounded Dyck grammars

Kaiyue Wen, Yuchen Li, Bingbin Liu, Andrej Risteski

https://openreview.net/forum?id=OitmaxSAUu

Keywords: Transformer, Self Attention, Dyck Language, Context Free Grammar, Formal Language, Theory, Interpretability

Compressor summary: This paper critically examines transformer interpretability methods that focus only on parts of the model instead of considering it holistically, using theoretical results and synthetic data experiments.


Augmentation-Aware Self-Supervision for Data-Efficient GAN Training

Liang Hou, Qi Cao, Yige Yuan, Songtao Zhao, Chongyang Ma, Siyuan Pan, Pengfei Wan, Zhongyuan Wang, Huawei Shen, Xueqi Cheng

https://openreview.net/forum?id=OiivS2mqQf

Keywords: generative adversarial networks, limited data, self-supervised learning

Compressor summary: The paper proposes a novel self-supervised discriminator that predicts the augmentation parameter and encourages the generator to produce realistic and augmentation-predictable data, improving data efficiency for training GANs with limited data.


Quantum speedups for stochastic optimization

Aaron Sidford, Chenyi Zhang

https://openreview.net/forum?id=OiatK9W6tR

Keywords: continuous optimization, quantum algorithms, stochastic optimization, gradient oracle

Compressor summary: The text describes new quantum methods for minimizing a Lipschitz convex function, achieving unachievable trade-offs and optimal rates in low dimensions, as well as quantum algorithms for solving smooth non-convex functions using quantum multivariate mean estimation and variance reduction techniques.


EgoDistill: Egocentric Head Motion Distillation for Efficient Video Understanding

Shuhan Tan, Tushar Nagarajan, Kristen Grauman

https://openreview.net/forum?id=OfjVAKx44G

Keywords: Egocentric Video; IMU; Efficient Video Understanding

Compressor summary: EgoDistill is a distillation-based approach that improves efficiency of ego-centric video understanding by combining semantics from sparse video frames and head motion from IMU readings, achieving state-of-the-art results on Ego4D and EPIC-Kitchens datasets.


The Transient Nature of Emergent In-Context Learning in Transformers

Aaditya K Singh, Stephanie C.Y. Chan, Ted Moskovitz, Erin Grant, Andrew M Saxe, Felix Hill

https://openreview.net/forum?id=Of0GBzow8P

Keywords: in-context learning, transformers, emergence, transience

Compressor summary: The paragraph discusses how in-context learning in transformer neural networks is often transient and not persistent, raising questions about overtraining and suggesting L2 regularization as a way to achieve more lasting ICL.


Practical Differentially Private Hyperparameter Tuning with Subsampling

Antti Koskela, Tejas Kulkarni

https://openreview.net/forum?id=OeLInnFKUK

Keywords: differential privacy, hyperparameter tuning, Rényi differential privacy, computational efficiency, DP-SGD

Compressor summary: The paper proposes a method to lower the privacy and compute cost of differentially private hyperparameter tuning algorithms by using a random subset of sensitive data and extrapolating optimal values.


Red Teaming Deep Neural Networks with Feature Synthesis Tools

Stephen Casper, Tong Bu, Yuxiao Li, Jiawei Li, Kevin Zhang, Kaivalya Hariharan, Dylan Hadfield-Menell

https://openreview.net/forum?id=Od6CHhPM7I

Keywords: interpretability, benchmarking, auditing, diagnostics, debugging, adversarial attacks, feature synthesis

Compressor summary: The paper introduces a benchmark with human-interpretable trojans to test interpretability tools for model debugging, showing that current methods often fail to discover them.


Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors

Thomas Hartvigsen, Swami Sankaranarayanan, Hamid Palangi, Yoon Kim, Marzyeh Ghassemi

https://openreview.net/forum?id=Oc1SIKxwdV

Keywords: Model Editing, Continual Learning, Model Repair

Compressor summary: GRACE is a method that allows targeted edits on deployed language models without retraining or degrading performance, using streaming errors to update a latent codebook.


The Gain from Ordering in Online Learning

Vasilis Kontonis, Mingchen Ma, Christos Tzamos

https://openreview.net/forum?id=OaUT4hX40s

Keywords: Online Learning, Self-directed Learning, Hardness of Approximation

Compressor summary: The paper explores how to order datapoints in online linear regression to minimize regret and shows that efficient algorithms are hard for arbitrary datasets but can achieve a log(d)-approximation for structured ones.


SyncDiffusion: Coherent Montage via Synchronized Joint Diffusions

Yuseung Lee, Kunho Kim, Hyunjin Kim, Minhyuk Sung

https://openreview.net/forum?id=OZEfMD7axv

Keywords: Diffusion model, Text-to-image generation, Panorama generation

Compressor summary: SyncDiffusion is a module that helps generate coherent image montages by synchronizing multiple diffusions using perceptual similarity loss, and it can be applied to various image generation tasks.


Scale Alone Does not Improve Mechanistic Interpretability in Vision Models

Roland S. Zimmermann, Thomas Klein, Wieland Brendel

https://openreview.net/forum?id=OZ7aImD4uQ

Keywords: feature visualization, interpretability, explainability, deep learning, neural networks, analysis, activation maximization, psychophysics

Compressor summary: The study finds that larger neural networks do not improve interpretability and suggests the need for more interpretable models and methods.


Laughing Hyena Distillery: Extracting Compact Recurrences From Convolutions

Stefano Massaroli, Michael Poli, Daniel Y Fu, Hermann Kumbong, Rom Nishijima Parnichkun, David W. Romero, Aman Timalsina, Quinn McIntyre, Beidi Chen, Atri Rudra, Ce Zhang, Christopher Re, Stefano Ermon, Yoshua Bengio

https://openreview.net/forum?id=OWELckerm6

Keywords: Long convolutions, recurrence, attention, language models, signal processing, throughput, auto-regressive generation

Compressor summary: The paper proposes methods to improve the efficiency of attention-free sequence models using linear state-space models and architectural improvements like Hyena, achieving better performance than Transformers and Hyena.


QLoRA: Efficient Finetuning of Quantized LLMs

Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, Luke Zettlemoyer

https://openreview.net/forum?id=OUIFPHEgJU

Keywords: finetuning, llama, instructions, quantization

Compressor summary: QLora is a memory-efficient finetuning method that achieves near-ChatGPT performance with the Guanaco model family, while introducing new innovations to save memory without sacrificing performance.


Rigorous Runtime Analysis of MOEA/D for Solving Multi-Objective Minimum Weight Base Problems

Anh Viet Do, Aneta Neumann, Frank Neumann, Andrew M. Sutton

https://openreview.net/forum?id=ORmVvN94B9

Keywords: minimum weight base problem, multi-objective optimization, approximation, evolutionary algorithm

Compressor summary: The paper studies the multi-objective minimum weight base problem, proves properties of its convex hull, and analyzes the performance of the MOEA/D algorithm on this problem compared to another algorithm called GSEMO.


Are aligned neural networks adversarially aligned?

Nicholas Carlini, Milad Nasr, Christopher A. Choquette-Choo, Matthew Jagielski, Irena Gao, Pang Wei Koh, Daphne Ippolito, Florian Tramèr, Ludwig Schmidt

https://openreview.net/forum?id=OQQoD8Vc3B

Keywords: Adversarial examples, large language models, alignment

Compressor summary: The paper investigates how well large language models can resist harmful inputs from adversarial users, and finds that multimodal models are more vulnerable than text-only models.


Robust Contrastive Language-Image Pretraining against Data Poisoning and Backdoor Attacks

Wenhan Yang, Jingdong Gao, Baharan Mirzasoleiman

https://openreview.net/forum?id=ONwL9ucoYG

Keywords: Contrastive Learning, Adversarial Learning, Model Robustness

Compressor summary: RoCLIP is a method for pre-training multimodal vision-language models that defends against targeted data poisoning and backdoor attacks by using random captions and image/text augmentations.


EICIL: Joint Excitatory Inhibitory Cycle Iteration Learning for Deep Spiking Neural Networks

Zihang Shao, Xuanye Fang, Yaxin Li, Chaoran Feng, Jiangrong Shen, Qi Xu

https://openreview.net/forum?id=OMDgOjdqoZ

Keywords: spiking neural networks cycle learning spike encoding

Compressor summary: EICIL is a novel learning method for spiking neural networks that mimics biological neurons by integrating excitatory and inhibitory behaviors, improving bio-mimicry and adaptability, and expanding the representation space.


An $\varepsilon$-Best-Arm Identification Algorithm for Fixed-Confidence and Beyond

Marc Jourdan, Rémy Degenne, Emilie Kaufmann

https://openreview.net/forum?id=OLk3F64eSg

Keywords: multi-armed bandits, pure-exploration, epsilon best arm identification, Top Two algorithm, anytime

Compressor summary: EB-TCε is a new sampling method for finding the best arm in stochastic bandits with theoretical guarantees on its performance and simulation results showing its effectiveness.


LLMScore: Unveiling the Power of Large Language Models in Text-to-Image Synthesis Evaluation

Yujie Lu, Xianjun Yang, Xiujun Li, Xin Eric Wang, William Yang Wang

https://openreview.net/forum?id=OJ0c6um1An

Keywords: Text-to-Image Evaluation, Visio-linguistic Compositionality, Large Language Models

Compressor summary: LLMScore is a new framework that uses large language models to evaluate text-to-image synthesis by considering object-level compositionality, achieving high correlation with human judgments and outperforming existing metrics.


RePo: Resilient Model-Based Reinforcement Learning by Regularizing Posterior Predictability

Chuning Zhu, Max Simchowitz, Siri Gadipudi, Abhishek Gupta

https://openreview.net/forum?id=OIJ3VXDy6s

Keywords: Model-Based Reinforcement Learning, Deep Reinforcement Learning

Compressor summary: The paper proposes a visual model-based RL method that learns a latent representation resilient to spurious variations and shows how it can adapt to different environments without relearning the dynamics and policy.


Inserting Anybody in Diffusion Models via Celeb Basis

Ge Yuan, Xiaodong Cun, Yong Zhang, Maomao Li, Chenyang Qi, Xintao Wang, Ying Shan, Huicheng Zheng

https://openreview.net/forum?id=OGQWZ3p0Zn

Keywords: Text-to-Image Synthesis, Personalized Synthesis, Face Embedding

Compressor summary: The paper proposes a new personalization method for text-to-image models that uses one facial photo and 1024 learnable parameters to integrate a unique individual into the model, enabling generation of stunning images featuring the person in various scenarios.


Jaccard Metric Losses: Optimizing the Jaccard Index with Soft Labels

Zifu Wang, Xuefei Ning, Matthew B. Blaschko

https://openreview.net/forum?id=OFMPrCAMKi

Keywords: Semantic Segmentation

Compressor summary: JMLs are compatible with soft labels and improve semantic segmentation performance in various scenarios, such as label smoothing, knowledge distillation, and semi-supervised learning.


Experimental Designs for Heteroskedastic Variance

Justin David Naggar Weltz, Tanner Fiez, Alexander Volfovsky, Eric Laber, Blake Mason, houssam nassif, Lalit K Jain

https://openreview.net/forum?id=OFDApY678F

Keywords: Heteroskedastic Variance, Linear Bandits, Experimental design

Compressor summary: The paragraph discusses a novel design for adaptive experimental design problems under heteroskedastic noise, which improves sample complexity and estimation accuracy by bounding the error of variance parameters.


A Guide Through the Zoo of Biased SGD

Yury Demidovich, Grigory Malinovsky, Igor Sokolov, Peter Richtárik

https://openreview.net/forum?id=OCtv4NyahI

Keywords: Stochastic optimization, biased SGD, Non-convex analysis

Compressor summary: The paper studies SGD with biased gradient estimators, clarifies existing assumptions and relationships among them, introduces a new set of weaker assumptions, and shows advantages of biased estimators over unbiased ones in various settings using theory and experiments.


Scalable Primal-Dual Actor-Critic Method for Safe Multi-Agent RL with General Utilities

Donghao Ying, YUNKAI ZHANG, Yuhao Ding, Alec Koppel, Javad Lavaei

https://openreview.net/forum?id=O63qgtebjH

Keywords: Reinforcement Learning Theory, Safe reinforcement learning, Multi-agent reinforcement learning

Compressor summary: The paper presents a safe multi-agent reinforcement learning method with general utilities and safety constraints, and proves its convergence and sample efficiency. It also demonstrates its effectiveness in simulations.


Training biologically plausible recurrent neural networks on cognitive tasks with long-term dependencies

Wayne WM Soo, Vishwa Goudar, Xiao-Jing Wang

https://openreview.net/forum?id=O453PHSthc

Keywords: neuroscience, recurrent neural network, neural circuits, cortical circuits, cognitive tasks, working memory

Compressor summary: The authors propose a method to train recurrent neural networks (RNNs) for cognitive tasks with long temporal dependencies by adding specialized skip-connections through time and revert to the original architecture, enabling RNNs to learn tasks that are difficult or impossible using conventional methods.


S-CLIP: Semi-supervised Vision-Language Learning using Few Specialist Captions

Sangwoo Mo, Minkyu Kim, Kyungmin Lee, Jinwoo Shin

https://openreview.net/forum?id=O1lYncfVOO

Keywords: vision-language model, semi-supervised learning, specialist domain

Compressor summary: S-CLIP is a semi-supervised learning method that uses unpaired images and pseudo-labeling strategies to improve vision-language models like CLIP in specialized domains with limited training data.


A U-turn on Double Descent: Rethinking Parameter Counting in Statistical Learning

Alicia Curth, Alan Jeffares, Mihaela van der Schaar

https://openreview.net/forum?id=O0Lz8XZT2b

Keywords: Double Descent, Statistical Machine Learning, Interpolation Regime, Effective Parameters

Compressor summary: The paragraph discusses how recent work on the relationship between model complexity and prediction error has suggested a new phenomenon called double descent, but argues that this is not incompatible with conventional statistical wisdom if one considers multiple axes of complexity.


Evaluating the Moral Beliefs Encoded in LLMs

Nino Scherrer, Claudia Shi, Amir Feder, David Blei

https://openreview.net/forum?id=O06z2G18me

Keywords: Language Models, Moral Decision Making, Social Aspects of Machine Learning, Ethics

Compressor summary: The paper investigates how LLMs make moral choices and express uncertainty in ambiguous situations using a large survey with moral dilemmas.


Information Design in Multi-Agent Reinforcement Learning

Yue Lin, Wenhao Li, Hongyuan Zha, Baoxiang Wang

https://openreview.net/forum?id=NyQwBttTnG

Keywords: multi-agent reinforcement learning, multi-agent communication, information design, signaling gradient, obedience constraints

Compressor summary: The paragraph discusses how reinforcement learning agents can influence each other through information design, addressing challenges like non-stationarity and information obedience in a Markov signaling game framework.


Efficient Policy Adaptation with Contrastive Prompt Ensemble for Embodied Agents

Wonje Choi, Woo Kyung Kim, SeungHyun Kim, Honguk Woo

https://openreview.net/forum?id=Ny3GcHLyzj

Keywords: Prompt Learining, Domain Adaptation, Embodied AI

Compressor summary: ConPE is a novel framework that uses visual prompts and a pretrained vision-language model to enable efficient policy learning and adaptation for embodied agents in various environments.


Sheaf Hypergraph Networks

Iulia Duta, Giulia Cassarà, Fabrizio Silvestri, Pietro Lio

https://openreview.net/forum?id=NvcVXzJvhX

Keywords: hypergraph neural networks, hypergraph, sheaf, higher-order

Compressor summary: The authors propose cellular sheaves for hypergraphs to enhance representation of complex interactions and develop two types of models that improve hypergraph node classification performance.


Label-Only Model Inversion Attacks via Knowledge Transfer

Ngoc-Bao Nguyen, Keshigeyan Chandrasegaran, Milad Abdollahzadeh, Ngai-man Cheung

https://openreview.net/forum?id=NuoIThPPag

Keywords: Model Inversion attacks, Generative models, Surrogate models, Knowledge transfer

Compressor summary: The authors propose LOKT, a method to perform label-only MI attacks using surrogate models and knowledge transfer from the target model, achieving significant improvements over existing methods.


Blocked Collaborative Bandits: Online Collaborative Filtering with Per-Item Budget Constraints

Soumyabrata Pal, Arun Suggala, Karthikeyan Shanmugam, Prateek Jain

https://openreview.net/forum?id=Ntd6X7uWYF

Keywords: Blocked Bandits, Collaborative Filtering, Clustering

Compressor summary: B-LATTICE is a novel algorithm that efficiently solves the blocked collaborative bandit problem by clustering users with similar rewards, collaborating across clusters, and respecting the arm sampling budget.


Reverse Engineering Self-Supervised Learning

Ido Ben-Shaul, Ravid Shwartz-Ziv, Tomer Galanti, Shai Dekel, Yann LeCun

https://openreview.net/forum?id=NsVEjx6YPd

Keywords: Self-Supervised Learning, Deep Learning, Representation Learning

Compressor summary: The paper analyzes self-supervised learning models to reveal that they naturally form semantic label-based clusters, which improve downstream tasks and information compression.


On the Complexity of Differentially Private Best-Arm Identification with Fixed Confidence

Achraf Azize, Marc Jourdan, Aymen Al Marjani, Debabrota Basu

https://openreview.net/forum?id=NsPbMwyxRl

Keywords: Differential Privacy, Multi-armed Bandits, Best Arm Identification, Fixed Confidence

Compressor summary: The authors study best arm identification problems under differential privacy and propose a new algorithm (AdaP-TT) with a good trade-off between privacy and utility.


On the Convergence to a Global Solution of Shuffling-Type Gradient Algorithms

Lam M. Nguyen, Trang H. Tran

https://openreview.net/forum?id=Nr1XSeDzpn

Keywords: stochastic gradient, shuffling type gradient method, global convergence

Compressor summary: Shuffling SGD, a scalable and efficient machine learning method, can converge to global solutions for certain non-convex functions even when over-parameterized.


Distributionally Robust Skeleton Learning of Discrete Bayesian Networks

Yeshu Li, Brian D Ziebart

https://openreview.net/forum?id=NpyZkaEEun

Keywords: structure learning, Bayesian network, robustness

Compressor summary: The paper proposes a method to learn the structure of discrete Bayesian networks from potentially corrupted data using distributionally robust optimization and regression, with guarantees for successful learning and efficient algorithms.


IPMix: Label-Preserving Data Augmentation Method for Training Robust Classifiers

Zhenglin Huang, Xiaoan Bao, Na Zhang, Qingqi Zhang, Xiao mei Tu, Biao Wu, Xi Yang

https://openreview.net/forum?id=No52399wXA

Keywords: data augmentation, robustness, safety

Compressor summary: IPMix is a data augmentation technique that improves robustness of deep neural networks without sacrificing clean accuracy by applying image-level, patch-level, and pixel-level augmentations with structural complexity and random mixing.


Human spatiotemporal pattern learning as probabilistic program synthesis

Tracey Mills, Joshua B. Tenenbaum, Samuel J Cheyette

https://openreview.net/forum?id=NnXznLurw5

Keywords: pattern learning; probabilistic programs; program synthesis; gaussian process; human learning

Compressor summary: The paragraph discusses a study on how humans learn structured patterns from small amounts of data and suggests that people might "learn by programming" using probabilistic models, with the best fit being a structured "Language of Thought" model.


Understanding Diffusion Objectives as the ELBO with Simple Data Augmentation

Diederik P Kingma, Ruiqi Gao

https://openreview.net/forum?id=NnMEadcdyD

Keywords: Diffusion Model, Evidence Lower Bound, Maximum Likelihood

Compressor summary: This paper shows that diffusion model objectives are closely related to the ELBO and explores new monotonic weightings for better performance.


Mix-of-Show: Decentralized Low-Rank Adaptation for Multi-Concept Customization of Diffusion Models

Yuchao Gu, Xintao Wang, Jay Zhangjie Wu, Yujun Shi, Yunpeng Chen, Zihan Fan, WUYOU XIAO, Rui Zhao, Shuning Chang, Weijia Wu, Yixiao Ge, Ying Shan, Mike Zheng Shou

https://openreview.net/forum?id=NnIaEaBfXD

Keywords: Text-to-Image Diffusion Models, Concept Customization

Compressor summary: The paper proposes Mix-of-Show, a framework for decentralized multi-concept customization in large-scale text-to-image diffusion models, which uses embedding-decomposed LoRAs and regionally controllable sampling to preserve concept identity and handle attribute binding.


Policy Optimization in a Noisy Neighborhood: On Return Landscapes in Continuous Control

Nathan Rahn, Pierluca D'Oro, Harley Wiltzer, Pierre-Luc Bacon, Marc G Bellemare

https://openreview.net/forum?id=Nn0daSf6CW

Keywords: deep reinforcement learning, continuous control, return landscape, stability

Compressor summary: The authors study how deep reinforcement learning agents for continuous control behave over time and propose a method to improve their stability by navigating the return landscape.


Towards Generic Semi-Supervised Framework for Volumetric Medical Image Segmentation

Haonan Wang, Xiaomeng Li

https://openreview.net/forum?id=NibgkUin5n

Keywords: Volumetric Medical Image Segmentation, Semi-supervised Learning, Unsupervised Domain Adaptation, Semi-supervised Domain Generalization

Compressor summary: The paper proposes a new semi-supervised learning framework that can handle different settings, such as unsupervised domain adaptation and semi-supervised domain generalization, by addressing the issues of capturing distribution-invariant features and avoiding over-fitting to labeled data.


Lexinvariant Language Models

Qian Huang, Eric Zelikman, Sarah Li Chen, Yuhuai Wu, Gregory Valiant, Percy Liang

https://openreview.net/forum?id=NiQTy0NW1L

Keywords: Large Language Model, in-context learning, pretraining

Compressor summary: The paper explores lexinvariant language models that do not use fixed token embeddings but rely on context to determine token meanings, and shows their performance, properties, and potential applications.


Improving neural network representations using human similarity judgments

Lukas Muttenthaler, Lorenz Linhardt, Jonas Dippel, Robert A. Vandermeulen, Katherine Hermann, Andrew Kyle Lampinen, Simon Kornblith

https://openreview.net/forum?id=Nh5dp6Uuvx

Keywords: representational alignment; human similarity judgments; neural networks; representation learning; few-shot learning; anomaly detection

Compressor summary: The authors explore how supervising the global structure of deep neural networks' representations with human similarity judgments can improve their performance on computer vision tasks such as few-shot learning and anomaly detection.


Bucks for Buckets (B4B): Active Defenses Against Stealing Encoders

Jan Dubiński, Stanisław Pawlak, Franziska Boenisch, Tomasz Trzcinski, Adam Dziedzic

https://openreview.net/forum?id=NfpYgGZC3B

Keywords: model stealing, model defenses, self-supervised learning

Compressor summary: B4B is a novel defense for MLaaS APIs that protects against model stealing attacks by dynamically changing the quality and uniqueness of vector representations for each user.


GNeSF: Generalizable Neural Semantic Fields

Hanlin Chen, Chen Li, Mengqi Guo, Zhiwen Yan, Gim Hee Lee

https://openreview.net/forum?id=NemifGnD2E

Keywords: NeRF; Semantic Segmentation; 3D vision; Scene understanding; Generalizable

Compressor summary: The authors propose a generalizable 3D segmentation framework based on neural implicit representation that uses multi-view image features and semantic maps as inputs, and achieves comparable or better performance than existing methods with only 2D supervision.


Reward-agnostic Fine-tuning: Provable Statistical Benefits of Hybrid Reinforcement Learning

Gen Li, Wenhao Zhan, Jason D. Lee, Yuejie Chi, Yuxin Chen

https://openreview.net/forum?id=Nd3FennRJZ

Keywords: reward-agnostic reinforcement learning, policy finetuning, offline reinforcement learning, online reinforcement learning

Compressor summary: The paper proposes a hybrid RL algorithm that combines tabular methods with reward-agnostic exploration and offline RL, achieving better sample complexity than pure offline or online RL methods without needing reward information during data collection.


Minimum Description Length and Generalization Guarantees for Representation Learning

Milad Sefidgaran, Abdellatif Zaidi, Piotr Krasnowski

https://openreview.net/forum?id=Ncb0MvVqRV

Keywords: Information Bottleneck, Representation Learning, Generalization Error, Minimum Description Length

Compressor summary: The paper proposes an information-theoretic framework to derive upper bounds on generalization error for representation learning algorithms, using the Minimum Description Length of labels or latent variables.


Provably Robust Temporal Difference Learning for Heavy-Tailed Rewards

Semih Cayci, Atilla Eryilmaz

https://openreview.net/forum?id=NapL36HSBT

Keywords: temporal difference learning, natural actor-critic, reinforcement learning, policy evaluation, policy gradient, markov decision processes

Compressor summary: This paper proposes a dynamic gradient clipping mechanism for robustifying temporal difference and natural actor-critic methods against heavy-tailed rewards in reinforcement learning, and proves their sample complexities under certain assumptions.


Geometric Neural Diffusion Processes

Emile Mathieu, Vincent Dutordoir, Michael John Hutchinson, Valentin De Bortoli, Yee Whye Teh, Richard E Turner

https://openreview.net/forum?id=NaYAsbv2jF

Keywords: diffusion model, functional space, stochastic process, time-series, neural processes, Gaussian processes, random fields, invariance, equivariance, symmetries, stationarity

Compressor summary: The authors propose a geometric diffusion model that incorporates symmetries in infinite-dimensional modelling and show its applicability to weather data.


Disambiguated Attention Embedding for Multi-Instance Partial-Label Learning

Wei Tang, Weijia Zhang, Min-Ling Zhang

https://openreview.net/forum?id=NYwbmCrrni

Keywords: Machine Learning, Multi-Instance Partial-Label Learning, Multi-Instance Learning, Partial-Label Learning

Compressor summary: DEMIPL is a novel algorithm that uses disambiguation attention to embed multi-instance bags into single vectors, improving performance in Multi-Instance Partial-Label Learning tasks for colorectal cancer classification.


Intra-Modal Proxy Learning for Zero-Shot Visual Categorization with CLIP

Qi Qian, Yuanhong Xu, Juhua Hu

https://openreview.net/forum?id=NXLjaYdgaL

Keywords: zero-shot; clip; proxy learning

Compressor summary: The paper proposes InMaP, a method to learn vision proxies from unlabeled data using text proxies and refining pseudo labels for zero-shot transfer tasks.


Moment Matching Denoising Gibbs Sampling

Mingtian Zhang, Alex Hawkins-Hooker, Brooks Paige, David Barber

https://openreview.net/forum?id=NWrN6cMG2x

Keywords: denoising score-matching, gibbs sampling, diffusion model

Compressor summary: The authors propose a framework for efficient sampling from EBMs trained with DSM, which addresses the inconsistency issues of DSM and allows for scalability to high-dimensional data.


Prefix-Tree Decoding for Predicting Mass Spectra from Molecules

Samuel Goldman, John Bradshaw, Jiayi Xin, Connor W. Coley

https://openreview.net/forum?id=NWEbeI2HNQ

Keywords: molecules, prefix tree, mass spectra, mass spectrum prediction, metabolomics, GNNs, chemistry, biology

Compressor summary: The paper presents a new method for predicting mass spectra from molecules by treating them as sets of subformulae and using a prefix tree structure to decode them, improving on existing limitations in computational tools.


SceneScape: Text-Driven Consistent Scene Generation

Rafail Fridman, Amit Abecasis, Yoni Kasten, Tali Dekel

https://openreview.net/forum?id=NU2kGsA4TT

Keywords: Computer Vision, Image & Video Editing, Video Generation, Perpetual View Generation, Texture Synthesis & Inpainting

Compressor summary: The paragraph describes a method for generating realistic long-term videos of various scenes based on text prompts and camera poses, using a combination of text-to-image and depth prediction models with online test-time training to ensure 3D consistency.


First- and Second-Order Bounds for Adversarial Linear Contextual Bandits

Julia Olkhovskaya, Jack Mayo, Tim van Erven, Gergely Neu, Chen-Yu Wei

https://openreview.net/forum?id=NTSbj2otOA

Keywords: contextual bandits, bandits, sequential learning, regret bounds

Compressor summary: The paper studies an online learning problem where the loss functions for each arm change over time and gives improved bounds on regret using a modified continuous exponential weights algorithm.


Geometric Analysis of Matrix Sensing over Graphs

Haixiang Zhang, Ying Chen, Javad Lavaei

https://openreview.net/forum?id=NRnm5xO8Hz

Keywords: Low-rank matrix optimization, non-convex optimization

Compressor summary: This paper introduces a new condition called the $\Omega$-RIP condition for matrix sensing over graphs and proves that it leads to polynomial-time global convergence of saddle-avoiding methods.


Fairly Recommending with Social Attributes: A Flexible and Controllable Optimization Approach

Jinqiu Jin, Haoxuan Li, Fuli Feng, Sihao Ding, Peng Wu, Xiangnan He

https://openreview.net/forum?id=NP5xb00Y6a

Keywords: Recommender System, Fairness

Compressor summary: This paper proposes social attribute-aware item group fairness metrics for recommender systems, and develops a gradient-based optimization algorithm to balance direct and social utility in training such models.


Efficient Low-rank Backpropagation for Vision Transformer Adaptation

Yuedong Yang, Hung-Yueh Chiang, Guihong Li, Diana Marculescu, Radu Marculescu

https://openreview.net/forum?id=NNtsO5L27J

Keywords: Low-rank backpropagation, model adaptation, transfer learning, vision transformer, Edge AI

Compressor summary: LBP-WHT is a new method that projects gradients into a low-rank space, reducing the computation needed for adapting large vision transformers (ViT) models and improving their accuracy on various datasets.


To Stay or Not to Stay in the Pre-train Basin: Insights on Ensembling in Transfer Learning

Ildus Sadrtdinov, Dmitrii Pozdeev, Dmitry P. Vetrov, Ekaterina Lobacheva

https://openreview.net/forum?id=NNooZoQpP4

Keywords: ensembles, transfer learning, loss landscape basins, model soups

Compressor summary: StarSSE is a modified Snapshot Ensembles method that improves the performance and diversity of neural network ensembles trained from a single pre-trained checkpoint by better exploring the pre-train basin without losing the benefits of transfer learning.


Extracting Reward Functions from Diffusion Models

Felipe Pinto Coelho Nuti, Tim Franzmeyer, Joao F. Henriques

https://openreview.net/forum?id=NN60HKTur2

Keywords: Diffusion models, sequential decision making, inverse reinforcement learning

Compressor summary: The authors propose a method to extract a reward function from two diffusion models with different behavior, and show its effectiveness in navigation and image generation tasks.


Policy Gradient for Rectangular Robust Markov Decision Processes

Navdeep Kumar, Esther Derman, Matthieu Geist, Kfir Yehuda Levy, Shie Mannor

https://openreview.net/forum?id=NLpXRrjpa6

Keywords: robust Markov decision process, policy gradient

Compressor summary: The paper introduces robust policy gradient (RPG), a method to train reinforcement learning agents that account for transition uncertainty and is efficient in computation.


Understanding the Limitations of Deep Models for Molecular property prediction: Insights and Solutions

Jun Xia, Lecheng Zhang, Xiao Zhu, Yue Liu, Zhangyang Gao, Bozhen Hu, Cheng Tan, Jiangbin Zheng, Siyuan Li, Stan Z. Li

https://openreview.net/forum?id=NLFqlDeuzt

Keywords: Graph Neural Networks

Compressor summary: The study compares 12 models on MPP tasks, finding that deep models generally underperform non-deep ones, and proposes a feature mapping method to improve their performance.


Latent Diffusion for Language Generation

Justin Lovelace, Varsha Kishore, Chao Wan, Eliot Seo Shekhtman, Kilian Q Weinberger

https://openreview.net/forum?id=NKdtztladR

Keywords: diffusion, language generation

Compressor summary: The paragraph discusses a new method for language generation using encoder-decoder language models and continuous diffusion models in the latent space, which improves over previous diffusion language models on various data sets.


Robust low-rank training via approximate orthonormal constraints

Dayana Savostianova, Emanuele Zangrando, Gianluca Ceruti, Francesco Tudisco

https://openreview.net/forum?id=NJPSvv0u3R

Keywords: low-rank neural networks, Stiefel manifold, orthogonal neural networks, pruning, adversarial robustness, neural network condition number, neural network singular values

Compressor summary: The authors propose a new method to prune neural networks while maintaining accuracy, robustness, and efficiency in the face of adversarial attacks.


Variational Gaussian processes for linear inverse problems

Thibault Christophe RANDRIANARISOA, Botond Szabo

https://openreview.net/forum?id=NJK3aSB0z4

Keywords: Linear inverse problems, Gaussian processes, Variational inference, Inducing variables, Asymptotics, Contraction rates

Compressor summary: The paper analyzes variational Bayesian methods for Gaussian process priors to solve linear inverse problems with different levels of ill-posedness and shows that inducing variable methods can attain the minimax estimation rate.


Boundary Guided Learning-Free Semantic Control with Diffusion Models

Ye Zhu, Yu Wu, Zhiwei Deng, Olga Russakovsky, Yan Yan

https://openreview.net/forum?id=NIrTSCiIZ7

Keywords: Diffusion probabilistic models, learning-free applications, high-dimensional semantic boundary, markov mixing

Compressor summary: BoundaryDiffusion is a learning-free method for efficient and effective image semantic editing using pre-trained generative denoising diffusion models without fine-tuning or extra networks.


Zero-sum Polymatrix Markov Games: Equilibrium Collapse and Efficient Computation of Nash Equilibria

Fivos Kalogiannis, Ioannis Panageas

https://openreview.net/forum?id=NGiq8qCQNk

Keywords: network games, Nash equilibrium, equilibrium, game theory, learning

Compressor summary: The paper proposes a class of zero-sum multi-agent Markov games with pairwise interactions and shows how to efficiently find an approximate Nash equilibrium using generalized techniques from previous work on coarse-correlated equilibria.


Monte Carlo Tree Search with Boltzmann Exploration

Michael Painter, Mohamed Baioumy, Nick Hawes, Bruno Lacerda

https://openreview.net/forum?id=NG4DaApavi

Keywords: Monte Carlo Tree Search, Planning, Entropy, Reinforcement Learning

Compressor summary: The paper introduces two improved MCTS methods (BTS and DENTS) that balance exploration and exploitation better than previous methods (MENTS and UCT) in automated planning problems like the game of Go.


Frequency Domain-Based Dataset Distillation

DongHyeok Shin, Seungjae Shin, Il-chul Moon

https://openreview.net/forum?id=NEawU0TgKG

Keywords: Dataset distillation, Frequency domain, Dataset condensation

Compressor summary: FreD is a new method to create smaller synthetic datasets from large ones by optimizing frequency representations and selecting relevant dimensions, improving distillation performance and preserving original data information.


Alternation makes the adversary weaker in two-player games

Volkan Cevher, Ashok Cutkosky, Ali Kavis, Georgios Piliouras, Stratis Skoulakis, Luca Viano

https://openreview.net/forum?id=NBMIsOS6B7

Keywords: Online Learning, Regret Minimization, Game Theory

Compressor summary: The paper proposes and analyzes two algorithms for alternating online linear optimization that achieve sub-linear regrets, depending on the problem domain and the number of actions.


SLaM: Student-Label Mixing for Distillation with Unlabeled Examples

Vasilis Kontonis, Fotis Iliopoulos, Khoa Trinh, Cenk Baykal, Gaurav Menghani, Erik Vee

https://openreview.net/forum?id=N7tw0QXx3z

Keywords: Distillation, teacher, student

Compressor summary: The paper proposes SLaM, a method to improve knowledge distillation with unlabeled data by mixing student and teacher labels, and provides theoretical guarantees and empirical results.


Online Learning under Adversarial Nonlinear Constraints

Pavel Kolev, Georg Martius, Michael Muehlebach

https://openreview.net/forum?id=N6YNe4KxDc

Keywords: online learning, online convex optimization, constrained optimization, adversarial nonlinear constraints, constraint violation oracle

Compressor summary: The CVV-Pro algorithm can learn from continuous non-stationary data streams with adversarial and time-varying constraints by using local sparse linear approximations of the feasible set.


Label Robust and Differentially Private Linear Regression: Computational and Statistical Efficiency

Xiyang Liu, Prateek Jain, Weihao Kong, Sewoong Oh, Arun Suggala

https://openreview.net/forum?id=N6FhEMnxCU

Keywords: Differential Privacy; Private Estimation

Compressor summary: The paper proposes an efficient linear regression method under differential privacy that works well even when some data points are corrupted or in the no-corruption setting.


Understanding and Improving Ensemble Adversarial Defense

Yian Deng, Tingting Mu

https://openreview.net/forum?id=N5uUTWLz0E

Keywords: adversarial defense, ensemble diversity, robustness, curvature

Compressor summary: The paragraph describes a new error theory for ensemble adversarial defense, which improves robustness by selectively allocating challenging examples to base classifiers and addressing their weaknesses.


PackQViT: Faster Sub-8-bit Vision Transformers via Full and Packed Quantization on the Mobile

Peiyan Dong, LEI LU, Chao Wu, Cheng Lyu, Geng Yuan, Hao Tang, Yanzhi Wang

https://openreview.net/forum?id=N56hAiQvot

Keywords: Vision Transformers, Quantization, Real-time on mobile, Sub-8-bit

Compressor summary: The paper proposes PackQViT, a framework for efficiently and accurately accelerating ViTs on mobile devices using sub-8-bit quantization and activation-aware training techniques.


Neural Multi-Objective Combinatorial Optimization with Diversity Enhancement

Jinbiao Chen, Zizhen Zhang, Zhiguang Cao, Yaoxin Wu, Yining Ma, Te Ye, Jiahai Wang

https://openreview.net/forum?id=N4JkStI1fe

Keywords: neural heuristic, diversity enhancement, deep reinforcement learning, multi-objective combinatorial optimization

Compressor summary: NHDE is a novel neural heuristic method that uses deep reinforcement learning and graph attention to generate diverse Pareto solutions for multi-objective combinatorial optimization problems, outperforming existing decomposition-based approaches.


Proportional Response: Contextual Bandits for Simple and Cumulative Regret Minimization

Sanath Kumar Krishnamurthy, Ruohan Zhan, Susan Athey, Emma Brunskill

https://openreview.net/forum?id=N1feehMSG9

Keywords: Contextual Bandits; Adaptive Experimentation; Simple Regret; Reinforcement Learning

Compressor summary: The authors propose efficient bandit algorithms for learning optimal treatment policies in stochastic settings, using conformal arm sets to balance simple and cumulative regret guarantees.


Active Learning-Based Species Range Estimation

Christian Lange, Elijah Cole, Grant Van Horn, Oisin Mac Aodha

https://openreview.net/forum?id=N0m9c0FqUV

Keywords: species range estimation, active learning, implicit networks

Compressor summary: The paper proposes an active learning method that uses transfer learned spatial representations to estimate the geographic range of unmapped species from limited observations and crowdsourced data.


Theoretical Analysis of the Inductive Biases in Deep Convolutional Networks

Zihao Wang, Lei Wu

https://openreview.net/forum?id=N0KwVdaaaJ

Keywords: Convolutional neural network, Inductive bias, Universality, Sparse function, Equivariance group

Compressor summary: This paper analyzes CNNs' inductive biases, shows they can approximate any continuous function with depth $\mathcal{O}(\log d)$, and reveals the importance of weight sharing and locality for efficient learning.


A Holistic Approach to Unifying Automatic Concept Extraction and Concept Importance Estimation

Thomas FEL, Victor Boutin, Louis Béthune, Remi Cadene, Mazda Moayeri, Léo Andéol, Mathieu Chalvidal, Thomas Serre

https://openreview.net/forum?id=MziFFGjpkb

Keywords: Explainable AI, Concept-based explainability, Interpretability, Concept extraction, Concept importance, Attribution methods

Compressor summary: The paragraph introduces a unifying framework for concept-based explainability methods in ANNs, which connects concept extraction and importance estimation as special cases of dictionary learning and attribution methods, respectively, and provides tools to evaluate and visualize them.


Mitigating Test-Time Bias for Fair Image Retrieval

Fanjie Kong, Shuai Yuan, Weituo Hao, Ricardo Henao

https://openreview.net/forum?id=Mxhb2lCOKL

Keywords: Vision-language, Fairness, Text-based Image Retrieval, Deep Learning, Application

Compressor summary: The paper proposes a post-processing technique, Post-hoc Bias Mitigation (PBM), to generate fair and unbiased image retrieval results from neutral textual queries while maintaining the performance of the vision-language model.


State Sequences Prediction via Fourier Transform for Representation Learning

Mingxuan Ye, Yufei Kuang, Jie Wang, Rui Yang, Wengang Zhou, Houqiang Li, Feng Wu

https://openreview.net/forum?id=MvoMDD6emT

Keywords: Reinforcement learning, Representation learning, State sequences prediction, Fourier transform

Compressor summary: SPF uses Fourier transform of state sequences to learn efficient representations for data-efficient RL, improving sample efficiency and long-term decision-making.


Mitigating the Popularity Bias of Graph Collaborative Filtering: A Dimensional Collapse Perspective

Yifei Zhang, Hao Zhu, yankai Chen, Zixing Song, Piotr Koniusz, Irwin King

https://openreview.net/forum?id=MvCq52yt9Y

Keywords: Graph, Collaborative Filtering, Recommendation

Compressor summary: The paper proposes a decorrelation-enhanced graph-based collaborative filtering (GCF) method using non-Euclidean geometry to improve feature diversity and reduce popularity bias in personalized recommendation systems.


Near-Linear Time Algorithm for the Chamfer Distance

Ainesh Bakshi, Piotr Indyk, Rajesh Jayaram, Sandeep Silwal, Erik Waingarten

https://openreview.net/forum?id=Mv96iC6TMX

Keywords: chamfer distance, earth mover distance, high dimensional data analysis, nearest neighbor search, high dimensional data, high-dimensional geometry, sublinear algorithms, point clouds, theory

Compressor summary: The paper proposes an efficient algorithm for approximating the Chamfer distance between large point sets in high-dimensional spaces, which can enable new applications and analyses of such data.


An Efficient Dataset Condensation Plugin and Its Application to Continual Learning

Enneng Yang, Li Shen, Zhenyi Wang, Tongliang Liu, Guibing Guo

https://openreview.net/forum?id=Murj6wcjRw

Keywords: Data Condensation, Continual Learning, Few-shot Learning

Compressor summary: The authors propose a new dataset condensation method that uses low-rank matrices to represent raw images, improving performance and reducing privacy risks in training networks from synthetic data.


Convolutional Neural Operators for robust and accurate learning of PDEs

Bogdan Raonic, Roberto Molinaro, Tim De Ryck, Tobias Rohner, Francesca Bartolucci, Rima Alaifari, Siddhartha Mishra, Emmanuel de Bezenac

https://openreview.net/forum?id=MtekhXRP4h

Keywords: PDEs, Neural Operators, Scientific Machine Learning, Convolutional Neural Networks

Compressor summary: The authors propose convolutional neural operators (CNOs), which are adaptations of convolutional neural networks that can learn operators for partial differential equations (PDEs) while preserving their continuous nature, achieving better performance than existing methods.


Task-Robust Pre-Training for Worst-Case Downstream Adaptation

Jianghui Wang, Yang Chen, Xingyu Xie, Cong Fang, Zhouchen Lin

https://openreview.net/forum?id=Mr4OpbZEiB

Keywords: Pre-training, Robustness, Multi-task learning

Compressor summary: The paper proposes a pre-training method that ensures good performance on downstream tasks by using a minimax loss and shows improved results on natural language processing and computer vision datasets.


Trade-off Between Efficiency and Consistency for Removal-based Explanations

Yifan Zhang, Haowei He, Zhiquan Tan, Yang Yuan

https://openreview.net/forum?id=MmCtXvW6GO

Keywords: AI interpretability, explainable AI, deep learning theory

Compressor summary: The paper shows that popular feature importance methods have a trade-off between interpretability, efficiency, and consistency, and proposes two new algorithms to minimize this trade-off using interpretation error as a metric.


Minimum norm interpolation by perceptra: Explicit regularization and implicit bias

Jiyoung Park, Ian Pelakh, Stephan Wojtowytsch

https://openreview.net/forum?id=MlrFYNo1yc

Keywords: Artificial neural network, interpolation, explicit regularization, implicit bias, weight decay, Barron class

Compressor summary: The paper explores how deep ReLU networks approach the minimal error solution when given enough data and relaxed regularization.


Birder: Communication-Efficient 1-bit Adaptive Optimizer for Practical Distributed DNN Training

Hanyang Peng, Shuang Qin, Yue Yu, Jin Wang, Hui Wang, Ge Li

https://openreview.net/forum?id=Mlo2kM11ZB

Keywords: optimizer, 1-bit optimizer, distributed learning, optimization for deep networks, communication efficiency

Compressor summary: The paper proposes Birder, a 1-bit adaptive optimizer that offers fast distributed DNN training with similar convergence rate and inference performance as uncompressed SGDM/Adam, while reducing communication volume.


Diverse Conventions for Human-AI Collaboration

Bidipta Sarkar, Andy Shih, Dorsa Sadigh

https://openreview.net/forum?id=MljeRycu9s

Keywords: Multi-Agent RL, Multi-Agent Coordination, Human-AI Coordination

Compressor summary: The paper proposes a method for creating diverse and effective cooperative strategies in multi-agent games by combining self-play and cross-play optimization techniques.


Delayed Algorithms for Distributed Stochastic Weakly Convex Optimization

Wenzhi Gao, Qi Deng

https://openreview.net/forum?id=MirclT6zpv

Keywords: Stochastic optimization, Distributed optimization, Prox-linear method, Stochastic gradient method

Compressor summary: The paper presents new delayed stochastic optimization algorithms with improved convergence rates and robustness against network delays, and shows their effectiveness in distributed weakly convex problems.


Optimize Planning Heuristics to Rank, not to Estimate Cost-to-Goal

Leah Chrestien, Stefan Edelkamp, Antonin Komenda, Tomáš Pevný

https://openreview.net/forum?id=Mgy6sgslPY

Keywords: Learning heuristic functions, deep learning, Immitation learning, planning, A*, best first search

Compressor summary: The paper studies optimal heuristics for forward search algorithms and proposes ranking-based loss functions for imitation learning, with experiments showing good results.


Protein Design with Guided Discrete Diffusion

Nate Gruver, Samuel Don Stanton, Nathan C. Frey, Tim G. J. Rudner, Isidro Hotzel, Julien Lafrance-Vanasse, Arvind Rajpal, Kyunghyun Cho, Andrew Gordon Wilson

https://openreview.net/forum?id=MfiK69Ga6p

Keywords: protein design, diffusion model, classifier guidance

Compressor summary: The authors propose NOS, a guidance method for discrete diffusion models that uses gradients from the denoising network to optimize protein design directly in sequence space, improving performance and addressing limitations of structure-based methods.


Optimization of Inter-group criteria for clustering with minimum size constraints

Eduardo Sany Laber, Lucas Murtinho

https://openreview.net/forum?id=MdJX5wwKwx

Keywords: Single Link, clustering, approximation algorithms, complexity, inter-group criterion

Compressor summary: The authors propose algorithms for optimizing inter-group criteria in clustering and show provable guarantees for two natural measures, with results for both unrestricted and constrained cases, and an empirical study on 10 real datasets.


Energy Transformer

Benjamin Hoover, Yuchen Liang, Bao Pham, Rameswar Panda, Hendrik Strobelt, Duen Horng Chau, Mohammed J Zaki, Dmitry Krotov

https://openreview.net/forum?id=MbwVNEx9KS

Keywords: Hopfield Network, Dense Associative Memory, Energy-based models, Attention Mechanism

Compressor summary: The Energy Transformer combines attention mechanism, energy-based models, and associative memory to improve machine learning tasks such as image completion, graph anomaly detection, and graph classification.


Approximate inference of marginals using the IBIA framework

Shivani Bathla, Vinita Vasudevan

https://openreview.net/forum?id=MamHShmHiX

Keywords: Bayesian inference, posterior marginals, probabilistic graphical models

Compressor summary: The paper introduces a new algorithm for inferring marginals in probabilistic graphical models using a sequence of linked clique tree forests and a heuristic belief update method, which is faster and more accurate than existing methods.


Architecture Matters: Uncovering Implicit Mechanisms in Graph Contrastive Learning

Xiaojun Guo, Yifei Wang, Zeming Wei, Yisen Wang

https://openreview.net/forum?id=MYfqIVcQrp

Keywords: graph contrastive learning

Compressor summary: The paragraph discusses common phenomena in graph contrastive learning (GCL) methods that differ from visual contrastive learning (VCL), and suggests paying attention to the unique architecture of graph learning when designing GCL methods.


Efficient Training of Energy-Based Models Using Jarzynski Equality

Davide Carbone, Mengjian Hua, Simon Coste, Eric Vanden-Eijnden

https://openreview.net/forum?id=MXxZ0Z5MNz

Keywords: Energy-Based Model, contrastive learning, generative models, Jarzynski identity, ULA

Compressor summary: The text describes a new method to train energy-based models more efficiently by using tools from nonequilibrium thermodynamics and sequential Monte-Carlo sampling, which improves performance over existing methods.


Explainable and Efficient Randomized Voting Rules

Soroush Ebadian, Aris Filos-Ratsikas, Mohamad Latifian, Nisarg Shah

https://openreview.net/forum?id=MWxsYPVmLS

Keywords: explainability, efficiency, voting, distortion, randomized decision-making

Compressor summary: The text discusses how adding simple randomization to deterministic voting rules can improve efficiency without losing explainability, and focuses on two families of such rules.


MCUFormer: Deploying Vision Tranformers on Microcontrollers with Limited Memory

Yinan Liang, Ziwei Wang, Xiuwei Xu, Yansong Tang, Jie Zhou, Jiwen Lu

https://openreview.net/forum?id=MWp3SwoHmH

Keywords: Vision transformer, microcontroller, network architecture search

Compressor summary: The paper introduces MCUFormer, a method to deploy vision transformers on IoT devices with limited memory by jointly designing architecture and constructing an inference operator library for efficient memory usage.


Restless Bandits with Average Reward: Breaking the Uniform Global Attractor Assumption

Yige Hong, Qiaomin Xie, Yudong Chen, Weina Wang

https://openreview.net/forum?id=MWQjqtV1z4

Keywords: restless bandits, average reward MDP, simulation-based method, asymptotic optimality

Compressor summary: The paper proposes a simulation-based method to design near-optimal policies for multi-armed bandit problems without relying on complex assumptions.


Tanimoto Random Features for Scalable Molecular Machine Learning

Austin Tripp, Sergio Bacallado, Sukriti Singh, José Miguel Hernández-Lobato

https://openreview.net/forum?id=MV0INFAKGq

Keywords: Tanimoto, Kernel, MinMax, Gaussian process, molecule, chemistry, interpretability

Compressor summary: The paper introduces new random features that enable faster computation of the Tanimoto kernel for large molecular datasets, and extend it to real-valued vectors with theoretical and empirical analysis.


Improvements on Uncertainty Quantification for Node Classification via Distance Based Regularization

Russell Alan Hart, Linlin Yu, Yifei Lou, Feng Chen

https://openreview.net/forum?id=MUzdCW2hC6

Keywords: Uncertainty Quantification, Graph Posterior Network, Bayesian

Compressor summary: The paragraph discusses uncertainty quantification for interdependent node-level classification using graph posterior networks, highlighting the limitations of the uncertainty cross-entropy loss function and proposing a distance-based regularization to improve out-of-distribution and misclassification detection.


HEDNet: A Hierarchical Encoder-Decoder Network for 3D Object Detection in Point Clouds

Gang Zhang, Chen Junnan, Guohuan Gao, Jianmin Li, Xiaolin Hu

https://openreview.net/forum?id=MUwr2YVJfN

Keywords: 3D object detection; encoder-decoder structure

Compressor summary: HEDNet is a hierarchical encoder-decoder network for 3D object detection in point clouds that captures long-range dependencies among features and outperforms previous methods on Waymo Open and nuScenes datasets.


On the Exploration of Local Significant Differences For Two-Sample Test

Zhijian Zhou, Jie Ni, Jia-He Yao, Wei Gao

https://openreview.net/forum?id=MRiitgpcUy

Keywords: two-sample test, local significant difference, directional information

Compressor summary: The paper proposes ME$_\text{MaBiD}$, a new method for two-sample testing that uses multiple Mahalanobis kernels, partitions the embedding space into regions, and tests for local significant differences with a bi-directional approach.


Enhancing Adversarial Robustness via Score-Based Optimization

Boya Zhang, Weijian Luo, Zhihua Zhang

https://openreview.net/forum?id=MOAHXRzHhm

Keywords: Adversarial Defense, Adversarial Attack, Score-based Models, Diffusion Models

Compressor summary: The paper proposes ScoreOpt, a novel defense against adversarial attacks for deep neural networks that optimizes adversarial samples towards clean data, achieving better robustness and faster inference than existing methods.


Diff-Instruct: A Universal Approach for Transferring Knowledge From Pre-trained Diffusion Models

Weijian Luo, Tianyang Hu, Shifeng Zhang, Jiacheng Sun, Zhenguo Li, Zhihua Zhang

https://openreview.net/forum?id=MLIs5iRq4w

Keywords: diffusion model, data-free distillation, implicit generator, knowledge transfer

Compressor summary: Diff-Instruct is a framework for instructing the training of different generative models using pre-trained diffusion models, based on a novel divergence called Integral Kullback-Leibler (IKL) divergence.


Bridging the Domain Gap: Self-Supervised 3D Scene Understanding with Foundation Models

Zhimin Chen, Longlong Jing, Yingwei Li, Bing Li

https://openreview.net/forum?id=MJbDy2155j

Keywords: 3D self-supervised learning, Multi-modal Representation Learning, Masked autoencoders, Knowledge distillation

Compressor summary: The authors propose Bridge3D, a method that uses features, semantic masks, and captions from foundation models to pre-train 3D models for better 3D scene representation learning and improve 3D object detection and semantic segmentation tasks.


Sequential Preference Ranking for Efficient Reinforcement Learning from Human Feedback

Minyoung Hwang, Gunmin Lee, Hogun Kee, Chan Woo Kim, Kyungjae Lee, Songhwai Oh

https://openreview.net/forum?id=MIYBTjCVjR

Keywords: Reinforcement Learning; Reinforcement Learning from Human Feedback; Preference-based Reinforcement Learning; Human-Robot Interaction

Compressor summary: The proposed SeqRank framework uses sequential preference ranking to enhance feedback efficiency in reinforcement learning from human feedback, leading to higher average feedback efficiency and improved task performance.


Debiasing Conditional Stochastic Optimization

Lie He, Shiva Kasiviswanathan

https://openreview.net/forum?id=MH7E7AME1r

Keywords: Optimization, Bilevel Optimization, Stochastic Optimization

Compressor summary: The paper proposes a general stochastic extrapolation technique to reduce bias and improve sample complexity for conditional stochastic optimization (CSO) problems, which have applications in various fields such as portfolio selection and reinforcement learning.


Learning Efficient Surrogate Dynamic Models with Graph Spline Networks

Chuanbo Hua, Federico Berto, Michael Poli, Stefano Massaroli, Jinkyoo Park

https://openreview.net/forum?id=MGPST5I9DO

Keywords: Graph, Spline Collocation Method, Graph Neural Networks, Simulation, Partial Differential Equations, PDEs, Physics, Scientific Computing, Surrogate Models, Weather Forecasting

Compressor summary: GraphSplineNets is a deep learning method that uses spline collocation to forecast physical systems faster by reducing grid size and iteration steps, with an adaptive strategy for sampling important regions.


Should Under-parameterized Student Networks Copy or Average Teacher Weights?

Berfin Simsek, Amire Bendjeddou, Wulfram Gerstner, Johanni Brea

https://openreview.net/forum?id=MG0mYskXN2

Keywords: shallow neural networks, non-convex optimization, approximation error, loss landscape

Compressor summary: The paragraph discusses how an under-parameterized neural network (student) can be fitted to an over-parameterized one (teacher) and shows that "copy-average" configurations are critical points for shallow networks with specific activation functions.


Random Cuts are Optimal for Explainable k-Medians

Konstantin Makarychev, Liren Shan

https://openreview.net/forum?id=MFWgLCWgUB

Keywords: Clustering, k-medians, Decision Tree, Explainability

Compressor summary: The RandomCoordinateCut algorithm achieves the best possible performance for solving the explainable $k$-medians problem in the $\ell_1$ norm, with a competitive ratio of at most $2\ln k+2$.


Imbalanced Mixed Linear Regression

Pini Zilber, Boaz Nadler

https://openreview.net/forum?id=MDxZYFR5Me

Keywords: Mixture regression model, Mixture of linear models, Iteratively reweighted least squares

Compressor summary: Mix-IRLS is a novel algorithm that efficiently solves mixed linear regression problems with imbalanced and challenging data settings, outperforming existing methods.


Nash Regret Guarantees for Linear Bandits

Ayush Sawarni, Soumyabrata Pal, Siddharth Barman

https://openreview.net/forum?id=MCkUS1P3Sh

Keywords: Sub-Poisson Distribution, Nash Social Welfare, Fairness Quantification, John Ellipsoid, Kiefer-Wolfowitz Optimal Design, Algorithmic Game Theory, Online Learning

Compressor summary: The paper presents a new notion of regret called Nash regret, which measures the performance of a bandit algorithm as collective welfare, and develops an algorithm that achieves tight upper bounds on it for stochastic linear bandits with sub-Poisson rewards.


Adversarial Attacks on Online Learning to Rank with Click Feedback

Jinhang Zuo, Zhiyao Zhang, Zhiyong Wang, Shuai Li, Mohammad Hajiesmaili, Adam Wierman

https://openreview.net/forum?id=MCj7DLkYqS

Keywords: online learning to rank, adversarial attack, click model

Compressor summary: This paper investigates various attack strategies against online learning to rank algorithms and proposes a general attack strategy that works under different feedback models.


Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples

Abulhair Saparov, Richard Yuanzhe Pang, Vishakh Padmakumar, Nitish Joshi, Mehran Kazemi, Najoung Kim, He He

https://openreview.net/forum?id=MCVfX7HgPO

Keywords: large language models, reasoning, out-of-distribution generalization, chain-of-thought, in-context learning

Compressor summary: The paragraph discusses testing the general deductive reasoning ability of large language models on various deduction rules and complex proofs, using a new synthetic dataset.


Fast Rank-1 Lattice Targeted Sampling for Black-box Optimization

Yueming Lyu

https://openreview.net/forum?id=M8CYKLHoEN

Keywords: Black-box Optimization, Derivate-free Optimization, Kernel methods

Compressor summary: RLTS is a novel technique that uses rank-1 lattices to improve query efficiency in high-dimensional optimization problems, enabling faster training and inference of Gaussian processes.


Geometric Algebra Transformer

Johann Brehmer, Pim De Haan, Sönke Behrends, Taco Cohen

https://openreview.net/forum?id=M7r2CO4tJC

Keywords: Geometry, geometric algebra, equivariance, transformer

Compressor summary: The Geometric Algebra Transformer (GATr) is a new architecture for geometric data that leverages projective geometric algebra and symmetry to achieve better performance in various applications.


Feature Dropout: Revisiting the Role of Augmentations in Contrastive Learning

Alex Tamkin, Margalit Glasgow, Xiluo He, Noah Goodman

https://openreview.net/forum?id=M7hijAPA4B

Keywords: self-supervised learning, contrastive learning

Compressor summary: Augmentations that destroy labels can be beneficial in foundation model settings, where diverse, general-purpose representations are needed for multiple downstream tasks.


Enhancing Minority Classes by Mixing: An Adaptative Optimal Transport Approach for Long-tailed Classification

Jintong Gao, He Zhao, Zhuo Li, Dan dan Guo

https://openreview.net/forum?id=M7FQpIdo0X

Keywords: Long-tailed Classification, Optimal Transport, Image-mixing, Semantic Similarity

Compressor summary: The authors propose an adaptive image-mixing method based on optimal transport to generate semantically reasonable mixed images for minority classes in long-tailed classification tasks, improving performance and serving as a general data augmentation method.


Multiply Robust Federated Estimation of Targeted Average Treatment Effects

Larry Han, Zhu Shen, Jose R Zubizarreta

https://openreview.net/forum?id=M6UccKMFGl

Keywords: Causal inference, Covariate mismatch, Federated learning, Multiple robustness, Transportation

Compressor summary: The paper proposes a new method to make valid causal inferences from multi-site studies while preserving privacy and accounting for differences between sites.


Language Models can Solve Computer Tasks

Geunwoo Kim, Pierre Baldi, Stephen Marcus McAleer

https://openreview.net/forum?id=M6OmjAZ4CX

Keywords: Large Language models, Web Navigation, Foundation Models, Decision Making

Compressor summary: The paper presents a new method called RCI, which uses a pre-trained language model to automate computer tasks via natural language commands with fewer demonstrations and without task-specific rewards, outperforming existing methods on benchmarks.


H-nobs: Achieving Certified Fairness and Robustness in Distributed Learning on Heterogeneous Datasets

Guanqiang Zhou, Ping Xu, Yue Wang, Zhi Tian

https://openreview.net/forum?id=M4h1UAxI3b

Keywords: distributed learning, federated learning, fairness, robustness, Byzantine attack, norm-based screening, q-FFL, optimization, convergence analysis

Compressor summary: The paper proposes a framework called H-nobs that combines fairness and robustness in distributed learning systems, addressing challenges like data heterogeneity and trade-offs between the two properties.


Transformers learn to implement preconditioned gradient descent for in-context learning

Kwangjun Ahn, Xiang Cheng, Hadi Daneshmand, Suvrit Sra

https://openreview.net/forum?id=LziniAXEI9

Keywords: Transformers, In-context learning, adaptive gradient methods

Compressor summary: The authors explore if transformers can learn to implement algorithms like gradient descent by training over random problem instances and provide theoretical analysis of the loss landscape for linear transformers in this setting.


Diffused Redundancy in Pre-trained Representations

Vedant Nanda, Till Speicher, John P Dickerson, Krishna P. Gummadi, Soheil Feizi, Adrian Weller

https://openreview.net/forum?id=LyAuNoZkGP

Keywords: representation learning, redundancy, transfer learning, fairness

Compressor summary: This study analyzes how pre-trained neural network features encode diffuse redundancy, where a subset of neurons can perform similar tasks as the whole layer, depending on the downstream task and pre-training parameters.


Sharp Spectral Rates for Koopman Operator Learning

Vladimir R Kostic, Karim Lounici, Pietro Novelli, massimiliano pontil

https://openreview.net/forum?id=Lt3jqxsbVO

Keywords: Statistical Learning Theory, Dynamical Systems

Compressor summary: The paper presents non-asymptotic learning bounds for Koopman eigenvalues and eigenfunctions in time-reversal-invariant stochastic dynamical systems, comparing two estimators (EDMD and RRR) and discussing their bias and variance trade-off.


Prompt-augmented Temporal Point Process for Streaming Event Sequence

Siqiao Xue, Yan Wang, Zhixuan Chu, Xiaoming Shi, Caigao JIANG, Hongyan Hao, Gangwei Jiang, Xiaoyun Feng, James Y. Zhang, JUN ZHOU

https://openreview.net/forum?id=LswqtKU9op

Keywords: prompt, point process, event sequence, continual learning.

Compressor summary: The text introduces PromptTPP, a framework that uses Continual Learning to monitor Neural Temporal Point Processes and learn event sequences without forgetting or buffering past data.


Bridging RL Theory and Practice with the Effective Horizon

Cassidy Laidlaw, Stuart Russell, Anca Dragan

https://openreview.net/forum?id=Lr2swAfwff

Keywords: reinforcement learning, RL theory, theory of reinforcement learning, instance-dependent bounds, empirical validation of theory

Compressor summary: The authors introduce BRIDGE, a new dataset for comparing deep RL algorithms and bounds, and propose the effective horizon as a better measure of MDP complexity that correlates with empirical performance and exploration techniques.


Stabilizing the Optimization of Neural Signed Distance Functions and Finer Shape Representation

Huizong Yang, Yuxin Sun, Ganesh Sundaramoorthi, Anthony Yezzi

https://openreview.net/forum?id=Lqv7VS1iBF

Keywords: Implicit Neural Representation, Surface Reconstruction

Compressor summary: The authors propose a novel approach to learning implicit neural representations of shapes, which addresses the instability issues in existing methods and enables capturing finer shape details.


Compositional Generalization from First Principles

Thaddäus Wiedemer, Prasanna Mayilvahanan, Matthias Bethge, Wieland Brendel

https://openreview.net/forum?id=LqOQ1uJmSx

Keywords: compositional generalization, compositionality, generalization, combinatorial generalization, out-of-distribution, out-of-domain, identifiability, disentanglement, object-centric learning, DSprites

Compressor summary: The text discusses an approach to improve machine learning's compositional generalization by considering it as a property of the data-generating process, rather than the data itself, and provides conditions and empirical validation for this approach.


Flow Factorized Representation Learning

Yue Song, T. Anderson Keller, Nicu Sebe, Max Welling

https://openreview.net/forum?id=LnySNEJAQt

Keywords: Generative modelling, latent disentanglement, variational autoencoders

Compressor summary: The authors propose Flow Factorized Representation Learning, a novel approach to structured representation learning that uses latent flow paths generated by learned potentials to achieve more efficient and better structured representations than existing methods.


From Tempered to Benign Overfitting in ReLU Neural Networks

Guy Kornowski, Gilad Yehudai, Ohad Shamir

https://openreview.net/forum?id=LnZuxp3Tx7

Keywords: benign overfitting, implicit bias, interpolating predictors, neural networks, theory

Compressor summary: The text discusses how the type of overfitting in neural networks depends on input dimension, sample size, architecture, and training algorithm, and provides theoretical and empirical evidence for this claim.


Individualized Dosing Dynamics via Neural Eigen Decomposition

Stav Belogolovsky, Ido Greenberg, Danny Eytan, Shie Mannor

https://openreview.net/forum?id=Lmxo0RVNx2

Keywords: personalized medicine, dosing dynamics, sequential prediction, stochastic differential equations, Kalman filter, recurrent neural networks, medical drug control

Compressor summary: NESDE is a novel algorithm that uses neural differential equations to model biological dynamics in individual patients, generalize to new treatment policies, and handle different noise levels while providing fast and continuous predictions.


What functions can Graph Neural Networks compute on random graphs? The role of Positional Encoding

Nicolas Keriven, Samuel Vaiter

https://openreview.net/forum?id=LmmjiTwYm0

Keywords: graph neural network; random graph; positional encoding

Compressor summary: This paper investigates the expressive power of equivariant Graph Neural Networks (GNNs) for node-tasks on large graphs, considering input features and positional encodings, and provides theoretical and empirical insights.


Non-autoregressive Machine Translation with Probabilistic Context-free Grammar

Shangtong Gui, Chenze Shao, Zhengrui Ma, Xishan Zhang, Yunji Chen, Yang Feng

https://openreview.net/forum?id=LloZFVwWvj

Keywords: Machine translation, Non-autoregressive generation, Probabilistic Context-free Grammar

Compressor summary: PCFG-NAT improves non-autoregressive neural machine translation by using a probabilistic context-free grammar to capture complex dependencies among output tokens and enhance explainability.


Training shallow ReLU networks on noisy data using hinge loss: when do we overfit and is it benign?

Erin George, Michael Murray, William Joseph Swartworth, Deanna Needell

https://openreview.net/forum?id=LlERoXEKjh

Keywords: benign overfitting, neural networks, relu, hinge loss

Compressor summary: The paper investigates how two-layer ReLU networks with gradient descent and hinge loss handle linearly separable data with some noisy labels, and identifies three possible outcomes depending on the margin conditions of the clean data.


CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation

Yexiong Lin, Yu Yao, Xiaolong Shi, Mingming Gong, Xu Shen, Dong Xu, Tongliang Liu

https://openreview.net/forum?id=Lkc0KjsDFv

Keywords: learning with label errors

Compressor summary: The paper proposes a method to infer content factors from noisy labels by using data augmentation and regularization, improving the learning of hard examples for image classification.


When Can We Track Significant Preference Shifts in Dueling Bandits?

Joe Suk, Arpit Agarwal

https://openreview.net/forum?id=LjWJLkSpjh

Keywords: non-stationary, multi-armed bandits, dueling bandits, preference-based learning

Compressor summary: The paper investigates how preferences can change over time in dueling bandits problems and explores the feasibility of adaptive algorithms with low dynamic regret under different preference distributions.


AlberDICE: Addressing Out-Of-Distribution Joint Actions in Offline Multi-Agent RL via Alternating Stationary Distribution Correction Estimation

Daiki E. Matsunaga, Jongmin Lee, Jaeseok Yoon, Stefanos Leonardos, Pieter Abbeel, Kee-Eung Kim

https://openreview.net/forum?id=LhVJdq4cZm

Keywords: Offline Reinforcement Learning, Multi-Agent Reinforcement Learning

Compressor summary: AlberDICE is an offline multi-agent reinforcement learning algorithm that trains agents individually based on stationary distribution optimization and avoids out-of-distribution joint actions, achieving better performance than baselines.


Deep Stochastic Processes via Functional Markov Transition Operators

Jin Xu, Emilien Dupont, Kaspar Märtens, Tom Rainforth, Yee Whye Teh

https://openreview.net/forum?id=Lg1ODJGGiI

Keywords: Neural Processes, Bayesian Nonparammetric Models

Compressor summary: Markov Neural Processes combine neural networks with stochastic processes, preserving their properties and improving performance on various tasks.


Hybrid Policy Optimization from Imperfect Demonstrations

Hanlin Yang, Chao Yu, peng sun, Siji Chen

https://openreview.net/forum?id=LftAvFt54C

Keywords: reinforcement learning, sparse reward, exploration, learning from demonstrations

Compressor summary: The paper proposes a novel RL algorithm called HYPO that uses imperfect demonstrations to train an offline guider policy to instruct an online agent policy for efficient exploration in challenging tasks with sparse rewards.


On Dynamic Programming Decompositions of Static Risk Measures in Markov Decision Processes

Jia Lin Hau, Erick Delage, Mohammad Ghavamzadeh, Marek Petrik

https://openreview.net/forum?id=LelK6Mfoey

Keywords: reinforcement learning, markov decision processes, monetary risk measures

Compressor summary: The paragraph discusses the limitations of common risk-averse optimization methods in Markov decision processes and proposes a new decomposition for Value-at-Risk that avoids suboptimality.


ODE-based Recurrent Model-free Reinforcement Learning for POMDPs

Xuanle Zhao, Duzhen Zhang, Liyuan Han, Tielin Zhang, Bo XU

https://openreview.net/forum?id=LdvVd0bNyO

Keywords: neural ode, POMDPs, reinforcement learning

Compressor summary: The paragraph discusses a novel ODE-based recurrent model combined with model-free RL for solving POMDPs and demonstrates its effectiveness in various tasks, including robustness against irregular observations.


Residual Q-Learning: Offline and Online Policy Customization without Value

Chenran Li, Chen Tang, Haruki Nishimura, Jean Mercat, Masayoshi Tomizuka, Wei Zhan

https://openreview.net/forum?id=LaNeRwDrTk

Keywords: reinforcement learning, imitation learning

Compressor summary: The paragraph discusses imitation learning and introduces a new problem setting called policy customization, which involves training a policy that inherits the characteristics of a prior policy while satisfying additional requirements from a downstream task, using a novel framework called Residual Q-learning.


Echoes Beyond Points: Unleashing the Power of Raw Radar Data in Multi-modality Fusion

Yang Liu, Feng Wang, Naiyan Wang, Zhaoxiang Zhang

https://openreview.net/forum?id=LZzsn51DPr

Keywords: 4D Radar; Transformer; Multi-modality

Compressor summary: The paper introduces EchoFusion, a novel radar signal processing method that fuses raw radar data with other sensors to improve autonomous driving performance and approach LiDAR accuracy.


High-dimensional Contextual Bandit Problem without Sparsity

Junpei Komiyama, Masaaki Imaizumi

https://openreview.net/forum?id=LZ4WgwmrUJ

Keywords: multi-armed bandits, linear bandits, contextual bandits, overparameterized models, high-dimensional models, online learning

Compressor summary: The paper studies how to solve high-dimensional linear contextual bandit problems using explore-then-commit and adaptive explore-then-commit algorithms, without imposing sparsity on regression coefficients.


Large Language Models can Implement Policy Iteration

Ethan Brooks, Logan A Walls, Richard Lewis, Satinder Singh

https://openreview.net/forum?id=LWxjWoBTsr

Keywords: Reinforcement Learning, In-Context Learning, Foundation Models

Compressor summary: The authors propose a method to use a large language model for policy iteration without expert demonstrations or gradients, by iteratively updating the prompt and interacting with an RL environment.


Diffusion Models and Semi-Supervised Learners Benefit Mutually with Few Labels

Zebin You, Yong Zhong, Fan Bao, Jiacheng Sun, Chongxuan Li, Jun Zhu

https://openreview.net/forum?id=LVHEcVgEGm

Keywords: diffusion models, semi-supervised generation, semi-supervised diffusion models, semi-supervised classification, image generation.

Compressor summary: The authors propose dual pseudo training, a method that uses partial labels to train a classifier, then a generative model, and finally retrains the classifier with real and generated images, achieving state-of-the-art results in semi-supervised image generation and classification.


Exploring Question Decomposition for Zero-Shot VQA

Zaid Khan, Vijay Kumar b g, Samuel Schulter, Manmohan Chandraker, Yun Fu

https://openreview.net/forum?id=LV83JEihHu

Keywords: visual question answering, in-context learning, vision-language

Compressor summary: The authors propose a question decomposition strategy for visual question answering, which improves performance by selectively decomposing questions and using human-written decompositions to guide large vision-language models.


Stochastic Optimal Control for Collective Variable Free Sampling of Molecular Transition Paths

Lars Holdijk, Yuanqi Du, Ferry Hooft, Priyank Jaini, Bernd Ensing, Max Welling

https://openreview.net/forum?id=LUVqEs90mq

Keywords: Transition Path Sampling, Stochastic Optimal Control

Compressor summary: The paragraph describes a problem of finding transition paths between metastable states in molecular systems, proposes a machine learning method called PIPS that does not rely on collective variables, and demonstrates its effectiveness on small proteins.


Learning Visual Prior via Generative Pre-Training

Jinheng Xie, Kai Ye, Yudong Li, Yuexiang Li, Kevin Qinghong Lin, Yefeng Zheng, Linlin Shen, Mike Zheng Shou

https://openreview.net/forum?id=LUT4b9gOtS

Keywords: Visual Prior, Generative Pre-Training, Conditional Image Synthesis

Compressor summary: The paper proposes VisorGPT, a method to learn and customize visual prior using generative pre-training and prompt engineering for conditional image synthesis tasks.


Feature-Learning Networks Are Consistent Across Widths At Realistic Scales

Nikhil Vyas, Alexander Atanasov, Blake Bordelon, Depen Morwani, Sabarish Sainathan, Cengiz Pehlevan

https://openreview.net/forum?id=LTdfYIvbHc

Keywords: mean-field, muP, feature learning, infinite width, deep ensembles

Compressor summary: The dynamics of feature-learning neural networks differ across various architectures and datasets, but their behavior can be captured by infinite-width limits for simple tasks, while deviations grow systematically for harder tasks due to initialization-dependent output variance scaling and the bias of narrower width.


Sample Efficient Reinforcement Learning in Mixed Systems through Augmented Samples and Its Applications to Queueing Networks

Honghao Wei, Xin Liu, Weina Wang, Lei Ying

https://openreview.net/forum?id=LTbIUkN95h

Keywords: Reinforcement Learning, Mixed Systems, Queueing Network, Sample Efficient

Compressor summary: The paper proposes a sample-efficient reinforcement learning method for mixed systems with stochastic and pseudo-stochastic states, which improves learning by generating augmented data samples and reducing the sample complexity.


Three Towers: Flexible Contrastive Learning with Pretrained Image Models

Jannik Kossen, Mark Collier, Basil Mustafa, Xiao Wang, Xiaohua Zhai, Lucas Beyer, Andreas Peter Steiner, Jesse Berent, Rodolphe Jenatton, Effrosyni Kokiopoulou

https://openreview.net/forum?id=LSYQB4CwD3

Keywords: three towers, contrastive learning, transformers, vision transformers, pretrained models, representation learning, finetuning, CLIP, ALIGN, classification, zero-shot, few-shot, retrieval

Compressor summary: Three Towers (3T) is a method that enhances contrastive learning of vision-language models by combining pretrained image classifiers with contrastive training, improving performance on retrieval tasks and classification for some pretraining datasets.


Asymptotically Optimal Quantile Pure Exploration for Infinite-Armed Bandits

Xiao-Yue Gong, Mark Sellke

https://openreview.net/forum?id=LROEcjVkv5

Keywords: pure exploration, multi-armed bandits, Fisher information

Compressor summary: The paper studies how to efficiently select one high quality arm in a setting with infinitely many arms, considering both fixed confidence and fixed budget scenarios.


An Inverse Scaling Law for CLIP Training

Xianhang Li, Zeyu Wang, Cihang Xie

https://openreview.net/forum?id=LMU2RNwdh2

Keywords: CLIP, inverse scaling, efficient training

Compressor summary: The paper finds an inverse scaling law for CLIP training, reducing its computational cost and enabling faster training with limited resources.


Disentanglement via Latent Quantization

Kyle Hsu, Will Dorrell, James C. R. Whittington, Jiajun Wu, Chelsea Finn

https://openreview.net/forum?id=LLETO26Ga2

Keywords: disentanglement, unsupervised learning, quantization

Compressor summary: The authors propose an inductive bias for latent space encoding and decoding using quantization and regularization, which improves disentangled representation learning in various generative models, and introduce new metrics called InfoMEC to evaluate disentanglement reliably.


Variational Weighting for Kernel Density Ratios

Sangwoong Yoon, Frank C. Park, Gunsu S YUN, Iljung Kim, Yung-Kyun Noh

https://openreview.net/forum?id=LIsJHQHi4z

Keywords: Kernel Density Estimation, KL-divergence, Density Ratio

Compressor summary: Kernel density estimation (KDE) improves predictions and information-theoretic measures by using an optimal weight function derived from multidimensional calculus of variations for density ratios.


Dynamic Tensor Decomposition via Neural Diffusion-Reaction Processes

Zheng Wang, Shikai Fang, Shibo Li, Shandian Zhe

https://openreview.net/forum?id=LGqIAn2OaZ

Keywords: Tensor Decomposition, Representation Learning

Compressor summary: DEMOTE is a method that uses neural networks to capture temporal structure in sparse multiway data, by estimating dynamic embeddings for entities in each mode of a tensor.


Handling Data Heterogeneity via Architectural Design for Federated Visual Recognition

Sara Pieri, Jose Renato Restom, Samuel Horváth, Hisham Cholakkal

https://openreview.net/forum?id=LGKxz9clGG

Keywords: Computer Vision, Federated Learning, Image Classification, Neural Network Architectures, Transformer, CNN, Data Hetereogenity, non-IID

Compressor summary: The paper reviews federated learning for visual recognition, highlighting the crucial role of architecture choices to achieve optimal performance and narrow the gap with centralized learning.


Certified Robustness via Dynamic Margin Maximization and Improved Lipschitz Regularization

Mahyar Fazlyab, Taha Entesari, Aniket Roy, Rama Chellappa

https://openreview.net/forum?id=LDhhi8HBO3

Keywords: Deep Learning, Adversarial Robustness, Certified Radius, Lipschitz Constants

Compressor summary: The paper proposes a differentiable regularizer that uses Lipschitz bounds on neural networks to improve robustness against adversarial perturbations, and shows better results than existing methods on image classification tasks.


Hardness of Low Rank Approximation of Entrywise Transformed Matrix Products

Tamas Sarlos, Xingyou Song, David Woodruff, Qiuyi Zhang

https://openreview.net/forum?id=LCwToX315b

Keywords: Low rank approximation, kernel methods, fine-grained complexity

Compressor summary: The authors study low rank approximation of a function involving two matrices, and show that certain conditions are necessary for fast algorithms, while also giving new hardness results and efficient algorithms for some cases.


Normalization-Equivariant Neural Networks with Application to Image Denoising

Sébastien Herbreteau, Emmanuel Moebel, Charles Kervrann

https://openreview.net/forum?id=LCnjG1IEfm

Keywords: equivariance, normalization, image denoising, activation functions, ReLU, interpretability, robustness, deep learning, analysis of neural networks

Compressor summary: The text proposes a method to adapt neural networks for normalization-equivariance, which is beneficial for many applications, by replacing existing layers and functions with new ones.


Cascading Bandits: Optimizing Recommendation Frequency in Delayed Feedback Environments

Dairui Wang, Junyu Cao, Yan Zhang, Wei Qi

https://openreview.net/forum?id=LClyG4vZmS

Keywords: delayed feedback, recommender system, frequency control

Compressor summary: This paper proposes a new bandit problem in dynamic recommender systems and presents several algorithms to optimize recommendation frequency and user engagement based on delayed feedback.


SNAP: Self-Supervised Neural Maps for Visual Positioning and Semantic Understanding

Paul-Edouard Sarlin, Eduard Trulls, Marc Pollefeys, Jan Hosang, Simon Lynen

https://openreview.net/forum?id=LCHmP68Gtj

Keywords: neural maps, visual positioning, semantic mapping

Compressor summary: SNAP is a deep network that learns 2D neural maps from images and can create better, more detailed, and semantically richer maps than traditional methods.


Trial matching: capturing variability with data-constrained spiking neural networks

Christos Sourmpis, Carl C. H. Petersen, Wulfram Gerstner, Guillaume Bellec

https://openreview.net/forum?id=LAbxkhkjbD

Keywords: neuroscience, spiking networks, data-constrained modeling, electrophysiological recordings, optimal transport, trial variability, RNN, interpretable machine learning

Compressor summary: The authors use a neural network model fitted to mouse behavioral and neural data to simulate sensory-motor processes and identify different modes of variability in the data.


Explaining the Uncertain: Stochastic Shapley Values for Gaussian Process Models

Siu Lun Chau, Krikamol Muandet, Dino Sejdinovic

https://openreview.net/forum?id=LAGxc2ybuH

Keywords: Gaussian Processes, Shapley values, Uncertainty Modelling

Compressor summary: The authors introduce a method to explain Gaussian processes using Shapley values from stochastic cooperative games, which allows for uncertainty quantification and predictive explanations.


GradOrth: A Simple yet Efficient Out-of-Distribution Detection with Orthogonal Projection of Gradients

Sima Behpour, Thang Doan, Xin Li, Wenbin He, Liang Gou, Liu Ren

https://openreview.net/forum?id=L9nTuSbAws

Keywords: out-of-distribution detection, OOD, uncertainty estimation, gradient projection

Compressor summary: GradOrth is a novel approach for detecting out-of-distribution data that uses important features from lower-rank subspaces of in-distribution data, achieving better performance than existing methods.


Latent Field Discovery in Interacting Dynamical Systems with Neural Fields

Miltiadis Kofinas, Erik J Bekkers, Naveen Shankar Nagaraja, Efstratios Gavves

https://openreview.net/forum?id=L9ZTvJ5jVx

Keywords: Graph Neural Networks, Neural Fields, Field Discovery, Equivariance, Interacting Dynamical Systems, Geometric Graphs

Compressor summary: The authors propose a method to infer hidden force fields from observed dynamics using equivariant graph networks and neural fields, which can improve prediction of system evolution.


Efficient Batched Algorithm for Contextual Linear Bandits with Large Action Space via Soft Elimination

Osama Hanna, Lin Yang, Christina Fragouli

https://openreview.net/forum?id=L7Whl9pXd0

Keywords: efficient bandit algorithms, contextual linear bandits

Compressor summary: The paper presents an efficient batched algorithm for contextual linear bandits with large action spaces using a novel soft elimination approach and achieving low regret bounds.


PrObeD: Proactive Object Detection Wrapper

Vishal Asnani, Abhinav Kumar, Suya You, Xiaoming Liu

https://openreview.net/forum?id=L74NTrzH1O

Keywords: Object detection, proactive, Camouflage, 2D

Compressor summary: PrObeD is a proactive scheme that enhances 2D object detection performance by learning an image-dependent template that acts as a mask for the input images, improving detection for generic and camouflaged objects.


Truly Scale-Equivariant Deep Nets with Fourier Layers

Md Ashiqur Rahman, Raymond A. Yeh

https://openreview.net/forum?id=L0QwnevT0F

Keywords: Scale Equivariance, Fourier Neural Network

Compressor summary: The paper presents a new scale-equivariant convolutional neural network architecture that considers anti-aliasing in down-scaling and performs well on image classification tasks.


Learning Descriptive Image Captioning via Semipermeable Maximum Likelihood Estimation

Zihao Yue, Anwen Hu, Liang Zhang, Qin Jin

https://openreview.net/forum?id=Kvaa3DhvlZ

Keywords: Image Captioning, Learning Objective, Natural Language Processing

Compressor summary: SMILE is a method for image captioning that encourages richer and more detailed descriptions by preventing conciseness optimization.


Spatial-frequency channels, shape bias, and adversarial robustness

Ajay Subramanian, Elena Sizikova, Najib J. Majaj, Denis G. Pelli

https://openreview.net/forum?id=KvPwXVcslY

Keywords: object recognition, critical band masking, spatial-frequency channels, shape bias, adversarial robustness

Compressor summary: Critical band masking shows humans and neural networks use different frequency filters for object recognition, with networks having broader filters that make them less robust to noise.


Tester-Learners for Halfspaces: Universal Algorithms

Aravind Gollakota, Adam Klivans, Konstantinos Stavropoulos, Arsen Vasilyan

https://openreview.net/forum?id=Kv8GJkV19S

Keywords: testable learning, pac learning, agnostic learning, Massart label noise, adversarial label noise, distribution testing

Compressor summary: The paragraph describes a universal tester-learner for halfspaces that works on many structured distributions, has low error, and uses sum-of-squares programs to check hypercontractivity of unknown distributions.


Describe, Explain, Plan and Select: Interactive Planning with LLMs Enables Open-World Multi-Task Agents

Zihao Wang, Shaofei Cai, Guanzhou Chen, Anji Liu, Xiaojian Ma, Yitao Liang

https://openreview.net/forum?id=KtvPdGb31Z

Keywords: open-ended learning, multi task, large language models, zero-shot planning

Compressor summary: The paper presents DESPs, a novel interactive planning method for Minecraft agents using LLMs, which enables a zero-shot multi-task agent to achieve better results in various domains.


IDRNet: Intervention-Driven Relation Network for Semantic Segmentation

Zhenchao Jin, Xiaowei Hu, Lingting Zhu, Luchuan Song, Li Yuan, Lequan Yu

https://openreview.net/forum?id=KtHquQuyA5

Keywords: semantic segmentation, relation modeling, object detection

Compressor summary: IDRNet is a novel context modeling paradigm that uses deletion diagnostics to improve dense prediction tasks by enhancing pixel-level representations with semantic-level interactions.


Circuit as Set of Points

Jialv Zou, Xinggang Wang, JiaHao Guo, Wenyu Liu, Qian Zhang, Chang Huang

https://openreview.net/forum?id=KsICioDlYs

Keywords: EDA, Circuit Design, Congestion prediction, DRC violation prediction

Compressor summary: The authors propose a new method for circuit design using Transformer-based point cloud perception to extract features directly from raw data, achieving high performance in various tasks.


Time-Independent Information-Theoretic Generalization Bounds for SGLD

Futoshi Futami, Masahiro Fujisawa

https://openreview.net/forum?id=Ks0RSFNxPO

Keywords: SGLD, Langevin dynamics, Generalization, Information theoretic analysis

Compressor summary: The paper proposes new generalization bounds for SGLD that are time-independent, decay with sample size, and account for stability of datasets, improving upon previous work.


StyleDrop: Text-to-Image Synthesis of Any Style

Kihyuk Sohn, Lu Jiang, Jarred Barber, Kimin Lee, Nataniel Ruiz, Dilip Krishnan, Huiwen Chang, Yuanzhen Li, Irfan Essa, Michael Rubinstein, Yuan Hao, Glenn Entis, Irina Blok, Daniel Castro Chin

https://openreview.net/forum?id=KoaFh16uOc

Keywords: text-to-image synthesis, fine-tuning, stylization

Compressor summary: StyleDrop is a versatile method that allows synthesizing images with specific styles using text-to-image models, even with just one image as input, and outperforms other methods in style tuning.


Distributed Personalized Empirical Risk Minimization

Yuyang Deng, Mohammad Mahdi Kamani, Pouria Mahdavinia, Mehrdad Mahdavi

https://openreview.net/forum?id=KoQgA0coZ9

Keywords: distributed learning, heterogeneous data, heterogeneous system, convergence analysis

Compressor summary: The paper introduces PERM, a new learning method for heterogeneous data that adapts models to each client's needs and constraints, using a distributed algorithm that optimizes PERM objectives and allows different model architectures.


Disentangling Voice and Content with Self-Supervision for Speaker Recognition

TIANCHI LIU, Kong Aik Lee, Qiongqiong Wang, Haizhou Li

https://openreview.net/forum?id=KoFYzuwjCA

Keywords: speaker recognition, disentanglement learning, self-supervision

Compressor summary: The paper introduces a framework that uses three Gaussian inference layers to model speaker traits and content variability in speech, disentangling them without labels other than speaker identities, and shows its effectiveness on VoxCeleb and SITW datasets.


Generalizing Importance Weighting to A Universal Solver for Distribution Shift Problems

Tongtong Fang, Nan Lu, Gang Niu, Masashi Sugiyama

https://openreview.net/forum?id=KmdlUP23qh

Keywords: importance weighting, distribution shift, deep learning

Compressor summary: The paper proposes a generalized importance weighting method to handle distribution shift problems in machine learning, addressing the limitations of existing methods.


An Information Theory Perspective on Variance-Invariance-Covariance Regularization

Ravid Shwartz-Ziv, Randall Balestriero, Kenji Kawaguchi, Tim G. J. Rudner, Yann LeCun

https://openreview.net/forum?id=KipjqOPaZ0

Keywords: Self-Supervised Learning, Generalization Bounds, Information-Theory, Deep Neural Networks

Compressor summary: The paper explores the mechanisms behind VICReg, an SSL method, and introduces new SSL methods based on information-theoretic principles that perform better than existing ones.


Transferable Adversarial Robustness for Categorical Data via Universal Robust Embeddings

Klim Kireev, Maksym Andriushchenko, Carmela Troncoso, Nicolas Flammarion

https://openreview.net/forum?id=Kig2YJVYfq

Keywords: Tabular data, Categorical data, Robust ML, Adversarial Robustness

Compressor summary: The paper proposes a method to train robust deep networks and transfer robustness to other classifiers using categorical feature embeddings for tabular data, which improves performance under adversarial attacks compared to existing methods.


Online Label Shift: Optimal Dynamic Regret meets Practical Algorithms

Dheeraj Baby, Saurabh Garg, Tzu-Ching Yen, Sivaraman Balakrishnan, Zachary Chase Lipton, Yu-Xiang Wang

https://openreview.net/forum?id=Ki6DqBXss4

Keywords: online learning, label shift, distribution shift, unsupervised domain adaptation

Compressor summary: The paper presents algorithms for adapting to online label shift in both supervised and unsupervised settings, using bootstrapped online regression oracles to track drifting class marginals and achieve optimal dynamic regret with high accuracy.


VoxDet: Voxel Learning for Novel Instance Detection

Bowen Li, Jiashun Wang, Yaoyu Hu, Chen Wang, Sebastian Scherer

https://openreview.net/forum?id=KgqucdSwIe

Keywords: Unseen object detection, instance perception, voxel representation

Compressor summary: The text introduces VoxDet, a 3D geometry-aware framework that uses voxel representation and matching techniques to improve object detection in challenging scenarios involving occlusion and pose variations.


Demystifying Oversmoothing in Attention-Based Graph Neural Networks

Xinyi Wu, Amir Ajorlou, Zihui Wu, Ali Jadbabaie

https://openreview.net/forum?id=Kg65qieiuB

Keywords: graph neural networks, attention mechanisms, oversmoothing, dynamical systems, theory

Compressor summary: The paper analyzes the effect of graph attention on oversmoothing in different types of Graph Neural Networks (GNNs) using mathematical tools and shows that it does not prevent oversmoothing and loses expressive power exponentially.


Towards Optimal Effective Resistance Estimation

Rajat Vadiraj Dwaraknath, Ishani Karmarkar, Aaron Sidford

https://openreview.net/forum?id=KffE8iXAw7

Keywords: effective resistances, spectral sketch, fine-grained complexity, triangle detection, numerical linear algebra

Compressor summary: The paper presents new algorithms for estimating effective resistances in undirected expander graphs with faster runtimes than previous methods, as well as a conditional lower bound showing that these algorithms are optimal up to a factor of $\epsilon$.


Wasserstein distributional robustness of neural networks

Xingjian Bai, Guangyi He, Yifan Jiang, Jan Obloj

https://openreview.net/forum?id=KfOUAlraMP

Keywords: adversarial attack, adversarial robustness of DNN, adversarial training, Wasserstein distance, distributionally robust optimization, sensitivity analysis, asymptotic bounds

Compressor summary: The authors propose a method for evaluating the robustness of deep neural networks against adversarial attacks using Wasserstein distributionally robust optimization and show its effectiveness on various image recognition tasks.


Layer-Neighbor Sampling --- Defusing Neighborhood Explosion in GNNs

Muhammed Fatih Balin, Umit Catalyurek

https://openreview.net/forum?id=Kd5W4JRsfV

Keywords: Graph Neural Networks, Graph Sampling, GNN, Layer Sampling, Minibatch Training

Compressor summary: LABOR is a new sampling algorithm for Graph Neural Networks that samples fewer vertices with better performance and convergence than existing methods.


Blockwise Parallel Transformers for Large Context Models

Hao Liu, Pieter Abbeel

https://openreview.net/forum?id=KbqQMoqfLQ

Keywords: Language Model, Long Context Modeling, Reinforcement Learning

Compressor summary: The Blockwise Parallel Transformer (BPT) is a new method that reduces memory costs and enables longer input sequences for natural language processing models by processing them in blocks, outperforming previous methods on language modeling and reinforcement learning tasks.


Prediction and Control in Continual Reinforcement Learning

Nishanth Anand, Doina Precup

https://openreview.net/forum?id=KakzVASqul

Keywords: reinforcement learning, continual reinforcement learning, lifelong learning, never-ending learning, prediction, control, multi-task learning, complementary learning systems

Compressor summary: The paper proposes a value function decomposition into permanent and transient components for continual reinforcement learning, showing theoretical and empirical benefits.


ReContrast: Domain-Specific Anomaly Detection via Contrastive Reconstruction

Jia Guo, shuai lu, LIze JIa, Weihang Zhang, Huiqi Li

https://openreview.net/forum?id=KYxD9YCQBH

Keywords: Unsupervised Anomaly Detection, Contrastive Learning, Medical Anomaly Detection, Transfer Learning

Compressor summary: The paper introduces ReContrast, a new unsupervised anomaly detection method that adapts to different domains by optimizing the entire network for feature reconstruction and contrastive learning.


Faster Relative Entropy Coding with Greedy Rejection Coding

Gergely Flamich, Stratis Markou, José Miguel Hernández-Lobato

https://openreview.net/forum?id=KXbAgvLi2l

Keywords: Compression, Learnt Compression, Relative Entropy Coding, Information Theory

Compressor summary: The paper introduces Greedy Rejection Coding (GRC), a new algorithm for relative entropy coding that improves runtime and codelength for continuous distributions, and evaluates it on a compression pipeline with variational autoencoders.


Trading-off price for data quality to achieve fair online allocation

Mathieu Molina, Nicolas Gast, Patrick Loiseau, Vianney Perchet

https://openreview.net/forum?id=KUBFYAPdqN

Keywords: Fairness, Online allocation, Bandits algorithms

Compressor summary: The paper proposes an algorithm that balances fairness and efficiency in online allocation problems by choosing data sources and allocating resources based on multi-armed bandit theory.


Reinforcement Learning with Fast and Forgetful Memory

Steven Morad, Ryan Kortvelesy, Stephan Liwicki, Amanda Prorok

https://openreview.net/forum?id=KTfAtro6vP

Keywords: reinforcement learning, partially observable, POMDP, memory, rnn, transformer

Compressor summary: Fast and Forgetful Memory is a memory model for Reinforcement Learning that improves reward and training speed compared to recurrent neural networks, without changing hyperparameters or adding more resources.


On the Constrained Time-Series Generation Problem

Andrea Coletta, Sriram Gopalakrishnan, Daniel Borrajo, Svitlana Vyetrenko

https://openreview.net/forum?id=KTZttLZekH

Keywords: time-series, generative models, constrained optimization, machine learning

Compressor summary: The paper introduces novel methods for generating realistic constrained time series using a constrained optimization framework and a guided diffusion model, achieving better performance and efficiency than existing approaches.


Conformal Prediction for Time Series with Modern Hopfield Networks

Andreas Auer, Martin Gauch, Daniel Klotz, Sepp Hochreiter

https://openreview.net/forum?id=KTRwpWCMsC

Keywords: time series, uncertainty, prediction interval, conformal prediction, modern hopfield networks

Compressor summary: HopCPT is a novel conformal prediction method for time series that handles temporal dependencies and performs better than existing methods on various datasets.


Aligning Optimization Trajectories with Diffusion Models for Constrained Design Generation

Giorgio Giannone, Akash Srivastava, Ole Winther, Faez Ahmed

https://openreview.net/forum?id=KTR33hMnMX

Keywords: diffusion models, engineering design, generative optimization, trajectory matching

Compressor summary: Diffusion Optimization Models (DOM) align diffusion model sampling with physics-based optimization for efficient, feasible, and high-performance designs in constrained settings with limited data.


DAW: Exploring the Better Weighting Function for Semi-supervised Semantic Segmentation

Rui Sun, Huayu Mai, Tianzhu Zhang, Feng Wu

https://openreview.net/forum?id=KRlG7NJUCD

Keywords: semi-supervised semantic segmentation

Compressor summary: The paper analyzes a trade-off in semi-supervised semantic segmentation methods when handling pseudo-labels and proposes Distribution-Aware Weighting (DAW) to improve the model's generalization performance.


SOC: Semantic-Assisted Object Cluster for Referring Video Object Segmentation

Zhuoyan Luo, Yicheng Xiao, Yong Liu, Shuyan Li, Yitong Wang, Yansong Tang, Xiu Li, Yujiu Yang

https://openreview.net/forum?id=KQyXyIAfK8

Keywords: Referring Video Object Segmentation, Video-Level Multi-Modal Understanding, Object Cluster, Visual-Linguistic Contrastive Learning

Compressor summary: The paper proposes a new approach, Semantic-assisted Object Cluster (SOC), that improves referring video object segmentation by aggregating video content and textual guidance for unified temporal modeling and cross-modal alignment.


Fairness Continual Learning Approach to Semantic Scene Understanding in Open-World Environments

Thanh-Dat Truong, Hoang-Quan Nguyen, Bhiksha Raj, Khoa Luu

https://openreview.net/forum?id=KQ25VgEvOJ

Keywords: Fairness Continual Learning; Semantic Segmentation; Contrastive Clustering;

Compressor summary: The paper proposes a new fairness continual learning framework for semantic segmentation that addresses challenges like catastrophic forgetting and background shift using novel losses and achieves state-of-the-art results on three benchmarks.


Fast Attention Requires Bounded Entries

Josh Alman, Zhao Song

https://openreview.net/forum?id=KOVWXcrFIK

Keywords: fast attention computation, algorithm, hardness

Compressor summary: The paper investigates faster algorithms for inner product attention computation using implicit use of the attention matrix A, and shows a sharp transition at matrix dimensions B = Θ(√log n).


One-Pass Distribution Sketch for Measuring Data Heterogeneity in Federated Learning

Zichang Liu, Zhaozhuo Xu, Benjamin Coleman, Anshumali Shrivastava

https://openreview.net/forum?id=KMxRQO7P98

Keywords: Distribution sketch, federated learning

Compressor summary: The paper proposes a one-pass distribution sketch that measures data heterogeneity in federated learning, improving client selection and handling cold start problems.


First Order Methods with Markovian Noise: from Acceleration to Variational Inequalities

Aleksandr Beznosikov, Sergey Samsonov, Marina Sheshukova, Alexander Gasnikov, Alexey Naumov, Eric Moulines

https://openreview.net/forum?id=KMeFZopsqP

Keywords: convex optimization, stochastic optimization, Markovian noise, acceleration, variational inequalities, lower bounds

Compressor summary: The paper proposes a unified approach for analyzing first-order gradient methods for stochastic optimization and variational inequalions with Markovian noise, eliminating some limiting assumptions and achieving optimal dependence on mixing time.


SE(3) Equivariant Augmented Coupling Flows

Laurence Illing Midgley, Vincent Stimper, Javier Antoran, Emile Mathieu, Bernhard Schölkopf, José Miguel Hernández-Lobato

https://openreview.net/forum?id=KKxO6wwx8p

Keywords: Boltzmann generator, normalizing flow

Compressor summary: Coupling normalizing flows with SE(3) and permutation equivariance are proposed for fast probabilistic modeling of physical systems, enabling efficient sampling and estimations.


SEGA: Instructing Text-to-Image Models using Semantic Guidance

Manuel Brack, Felix Friedrich, Dominik Hintersdorf, Lukas Struppek, Patrick Schramowski, Kristian Kersting

https://openreview.net/forum?id=KIPAIy329j

Keywords: diffusion, text-to-image, generation, semantics

Compressor summary: The paper introduces a method to interact with text-to-image diffusion models for semantic guidance, enabling users to control the image generation process more effectively.


MeCo: Zero-Shot NAS with One Data and Single Forward Pass via Minimum Eigenvalue of Correlation

Tangyu Jiang, Haodi Wang, Rongfang Bie

https://openreview.net/forum?id=KFm2lZiI7n

Keywords: Neural Architecture Search, Zero-Cost Proxy, Evaluation Strategy, Feature Map

Compressor summary: The paper proposes a novel zero-cost proxy called $\mathsf{MeCo}$ for neural architecture search, which uses the Pearson correlation matrix of feature maps and requires only one random data for a single forward pass.


Multi-Step Generalized Policy Improvement by Leveraging Approximate Models

Lucas Nunes Alegre, Ana L. C. Bazzan, Ann Nowe, Bruno Castro da Silva

https://openreview.net/forum?id=KFj0Q1EXvU

Keywords: generalized policy improvement, successor features, transfer learning, model-based reinforcement learning

Compressor summary: The paper presents a new method, h-GPI, for zero-shot transfer in reinforcement learning that combines model-free and model-based approaches and improves performance over existing methods.


On the Generalization Error of Stochastic Mirror Descent for Quadratically-Bounded Losses: an Improved Analysis

Ta Duy Nguyen, Alina Ene, Huy Nguyen

https://openreview.net/forum?id=KF4LCXz8Np

Keywords: high probability, generalization, convex optimization, nonconvex optimization

Compressor summary: The authors propose a new method to analyze the generalization error of stochastic mirror descent for quadratically bounded losses and obtain improved bounds in various settings.


The probability flow ODE is provably fast

Sitan Chen, Sinho Chewi, Holden Lee, Yuanzhi Li, Jianfeng Lu, Adil Salim

https://openreview.net/forum?id=KD6MFeWSAd

Keywords: DDIM, deterministic samplers, diffusion models, predictor-corrector, probability flow ODE, score-based generative modeling

Compressor summary: The text describes a probabilistic flow ODE method for generative modeling that has polynomial-time convergence guarantees and outperforms DDPM in terms of dimension dependence.


Towards Free Data Selection with General-Purpose Models

Yichen Xie, Mingyu Ding, Masayoshi Tomizuka, Wei Zhan

https://openreview.net/forum?id=KBXcDAaZE7

Keywords: data selection, unsupervised learning

Compressor summary: The paper proposes FreeSel, an efficient data selection method that uses existing models to select data without training or supervision, achieving significant speedups compared to current active learning methods.


LIMA: Less Is More for Alignment

Chunting Zhou, Pengfei Liu, Puxin Xu, Srini Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, LILI YU, Susan Zhang, Gargi Ghosh, Mike Lewis, Luke Zettlemoyer, Omer Levy

https://openreview.net/forum?id=KBMOKmX2he

Keywords: large language models, supervised instruction fine-tuning, chat assistant

Compressor summary: The paragraph discusses LIMA, a language model fine-tuned with supervised loss on a small number of prompts and responses, which shows strong performance and generalization, suggesting that pretraining is the most important stage for large language models.


PromptIR: Prompting for All-in-One Image Restoration

Vaishnav Potlapalli, Syed Waqas Zamir, Salman Khan, Fahad Khan

https://openreview.net/forum?id=KAlSIL4tXU

Keywords: Image Restoration

Compressor summary: The paper introduces PromptIR, a method that uses prompts to guide a restoration network and achieve high-quality image restoration from various degradation types and levels without prior knowledge of the corruptions.


DropCompute: simple and more robust distributed synchronous training via compute variance reduction

Niv Giladi, Shahar Gottlieb, Moran Shkolnik, Asaf Karnieli, Ron Banner, Elad Hoffer, Kfir Yehuda Levy, Daniel Soudry

https://openreview.net/forum?id=KAWaeKOEkx

Keywords: distributed optimization, large-scale parallel SGD, synchronous training

Compressor summary: The paper analyzes how worker variability affects the scalability and robustness of synchronous distributed DNN training, and proposes a decentralized method to reduce this variation.


Graph Mixture of Experts: Learning on Large-Scale Graphs with Explicit Diversity Modeling

Haotao Wang, Ziyu Jiang, Yuning You, Yan Han, Gaowen Liu, Jayanth Srinivasa, Ramana Rao Kompella, Zhangyang Wang

https://openreview.net/forum?id=K9xHDD6mic

Keywords: graph neural networks

Compressor summary: The GMoE model improves GNNs by allowing nodes to select expert aggregators that capture different graph structures and information distances, leading to better performance on various graph tasks.


A Recurrent Neural Circuit Mechanism of Temporal-scaling Equivariant Representation

Junfeng Zuo, Xiao Liu, Ying Nian Wu, Si Wu, Wenhao Zhang

https://openreview.net/forum?id=K9dmkfZcMu

Keywords: temporal-scaling group, equivariant representation, disentangled representation, motor timing, continuous attractor networks

Compressor summary: The study proposes a recurrent circuit model that explains how the brain generates temporal sequences at different speeds, using control inputs and Lie group operators.


Corruption-Robust Offline Reinforcement Learning with General Function Approximation

Chenlu Ye, Rui Yang, Quanquan Gu, Tong Zhang

https://openreview.net/forum?id=K9M7XNS9BX

Keywords: offline RL; adversarial corruption; general function approximation

Compressor summary: The paper proposes a new algorithm to solve offline RL problems under corruption and provides a suboptimality bound that depends on the corruption level and other factors, showing tightness for linear MDPs.


Regularization properties of adversarially-trained linear regression

Antonio H. Ribeiro, Dave Zachariah, Francis Bach, Thomas B. Schön

https://openreview.net/forum?id=K8gLHZIgVW

Keywords: adversarial training; regularization; linear models

Compressor summary: Adversarial training for linear models can lead to minimum-norm interpolators, equivalent parameter shrinking methods, or independent of noise variance in some cases.


Neural Modulation for Flash Memory: An Unsupervised Learning Framework for Improved Reliability

Jonathan Zedaka, Elisha Halperin, Evgeny Blaichman, Amit Berman

https://openreview.net/forum?id=K7u3RkoBP9

Keywords: WGAN, GAN, Autoencoder, Unsupervised Learning, Generative models, Flash Memory, NAND, Modulation, Reliability, Flash, Communication system

Compressor summary: The authors propose a machine learning approach that reduces errors and data degradation in NAND flash memory by using a neural modulator that adapts programming operations for each memory cell.


Invariant Learning via Probability of Sufficient and Necessary Causes

Mengyue Yang, Yonggang Zhang, Zhen Fang, Yali Du, Furui Liu, Jean-Francois Ton, Jianhong Wang, Jun Wang

https://openreview.net/forum?id=K5e5tFZuur

Keywords: OOD Generalization, Invariant Representation Learning

Compressor summary: The paper proposes a method to improve out-of-distribution generalization by using probability of sufficiency and necessary causes (PNS) to learn representations with high PNS value, which is theoretically analyzed and experimentally validated.


MAG-GNN: Reinforcement Learning Boosted Graph Neural Network

Lecheng Kong, Jiarui Feng, Hao Liu, Dacheng Tao, Yixin Chen, Muhan Zhang

https://openreview.net/forum?id=K4FK7I8Jnl

Keywords: Graph Neural Network; Reinforcement Learning

Compressor summary: MAG-GNN is a reinforcement learning boosted graph neural network that uses combinatorial optimization to find the optimal subset of subgraphs for efficient and expressive graph learning tasks.


Fast Optimal Locally Private Mean Estimation via Random Projections

Hilal Asi, Vitaly Feldman, Jelani Nelson, Huy Nguyen, Kunal Talwar

https://openreview.net/forum?id=K3JgUvDSYX

Keywords: Differential Privacy, mean estimation, private federated learning, communication complexity

Compressor summary: The paper proposes ProjUnit, a simple and efficient algorithmic framework for private mean estimation of high-dimensional vectors with low communication and run-time complexity.


Controlling Text-to-Image Diffusion by Orthogonal Finetuning

Zeju Qiu, Weiyang Liu, Haiwen Feng, Yuxuan Xue, Yao Feng, Zhen Liu, Dan Zhang, Adrian Weller, Bernhard Schölkopf

https://openreview.net/forum?id=K30wTdIIYc

Keywords: Text-to-image, diffusion models, finetuning, generative models, orthogonality

Compressor summary: The Orthogonal Finetuning (OFT) method helps adapt text-to-image diffusion models for different tasks by preserving their semantic generation ability and improving finetuning stability.


Learning Mask-aware CLIP Representations for Zero-Shot Segmentation

Siyu Jiao, Yunchao Wei, Yaowei Wang, Yao Zhao, Humphrey Shi

https://openreview.net/forum?id=K1Uzj8tuwd

Keywords: Zero-Shot Segmentation, Open-Vocabulary Segmentation, Fine-tuning

Compressor summary: The paper proposes Mask-aware Fine-tuning (MAFT), a method to improve zero-shot segmentation by making CLIP more responsive to different mask proposals using a new encoder and two types of loss functions.


Block Low-Rank Preconditioner with Shared Basis for Stochastic Optimization

Jui-Nan Yen, Sai Surya Duvvuri, Inderjit S Dhillon, Cho-Jui Hsieh

https://openreview.net/forum?id=JzQlGqBm8d

Keywords: Second Order Optimization, Optimization for deep networks

Compressor summary: The paper proposes a low-rank approximation method for preconditioning in second-order optimization that reduces time and memory complexity while maintaining performance on deep neural networks.


Opening the Vocabulary of Egocentric Actions

Dibyadip Chatterjee, Fadime Sener, Shugao Ma, Angela Yao

https://openreview.net/forum?id=JzQ7QClAdF

Keywords: video understanding, egocentric videos, open vocabulary

Compressor summary: The paper introduces a new action recognition task that can handle verbs and objects not seen during training using an object-agnostic verb encoder and a prompt-based object encoder based on CLIP representations.


Provably (More) Sample-Efficient Offline RL with Options

Xiaoyan Hu, Ho-fung Leung

https://openreview.net/forum?id=JwNXeBdkeo

Keywords: Learning with Options, Offline RL, Provably Efficient RL

Compressor summary: The paper studies how using options in offline reinforcement learning can improve sample efficiency and provide a novel information-theoretic lower bound for this scenario. It also proposes an algorithm (PEVIO) and analyzes two data-collection methods.


CODA: Generalizing to Open and Unseen Domains with Compaction and Disambiguation

Chaoqi Chen, Luyao Tang, Yue Huang, Xiaoguang Han, Yizhou Yu

https://openreview.net/forum?id=Jw0KRTjsGA

Keywords: Domain generalization, domain shift, open class, source compaction, target disambiguation

Compressor summary: CODA is a novel framework for learning compact representations and adapting to open classes in machine learning, addressing domain shift and unseen categories by introducing virtual unknown classes and optimizing a new training objective.


DASpeech: Directed Acyclic Transformer for Fast and High-quality Speech-to-Speech Translation

Qingkai Fang, Yan Zhou, Yang Feng

https://openreview.net/forum?id=JvYSSPtQyk

Keywords: speech-to-speech translation, non-autoregressive translation, speech translation, directed acyclic transformer

Compressor summary: DASpeech is a non-autoregressive direct speech-to-speech translation model that uses a two-pass architecture, achieving fast and high-quality translations while preserving the source speaker's voice.


Train Hard, Fight Easy: Robust Meta Reinforcement Learning

Ido Greenberg, Shie Mannor, Gal Chechik, Eli Meirom

https://openreview.net/forum?id=JvOZ4IIjwP

Keywords: meta reinforcement learning, robust reinforcement learning, safe reinforcement learning, risk sensitive reinforcement learning

Compressor summary: The paper proposes Robust Meta RL (RoML), a meta-algorithm that improves the performance of standard Meta RL methods in high-risk or difficult tasks by over-sampling them, thus increasing system reliability.


Penalising the biases in norm regularisation enforces sparsity

Etienne Boursier, Nicolas Flammarion

https://openreview.net/forum?id=JtIqG47DAQ

Keywords: Neural networks, Min norm interpolators, Sparsity, Representational cost

Compressor summary: This paper investigates how controlling the norm of neural network parameters affects generalization and shows that using a specific weighting factor ensures unique and sparse solutions for one hidden ReLU layer networks with unidimensional data.


From Distribution Learning in Training to Gradient Search in Testing for Combinatorial Optimization

Yang Li, Jinpei Guo, Runzhong Wang, Junchi Yan

https://openreview.net/forum?id=JtF0ugNMv2

Keywords: Machine Learning, Combinatorial Optimization, Generative Modeling, Diffusion Model

Compressor summary: The paragraph describes a new framework called T2TCO that improves Combinatorial Optimization problem solving by using generative modeling and gradient-based search, achieving significant performance gains on TSP and MIS problems.


Attentive Transfer Entropy to Exploit Transient Emergence of Coupling Effect

Xiaolei Ru, Xin-Ya Zhang, Zijia Liu, Jack Murdoch Moore, Gang Yan

https://openreview.net/forum?id=JpU5YmMKx7

Keywords: Directed coupled network reconstruction; Neuronal dynamics; Mutual information estimator; Attention mechanism; Transfer entropy.

Compressor summary: The text describes a novel method to reconstruct coupled biological neural networks from time series data using attention mechanism and Attentive Transfer Entropy, which can identify critical regions of strong coupling-drive.


ALGO: Synthesizing Algorithmic Programs with Generated Oracle Verifiers

Kexun Zhang, Danqing Wang, Jingtao Xia, William Yang Wang, Lei Li

https://openreview.net/forum?id=JolrEmMim6

Keywords: Large Language Models, Code Generation, Code Intelligence, Automatic Verification

Compressor summary: The proposed ALGO framework synthesizes algorithms with LLM-generated oracles to guide code generation and verify correctness, achieving significantly improved one-submission pass rates on algorithmic problems.


On Single-Index Models beyond Gaussian Data

Aaron Zweig, Loucas Pillaud-Vivien, Joan Bruna

https://openreview.net/forum?id=JkmvrheMe7

Keywords: gradient descent, shallow neural networks

Compressor summary: This paper studies how gradient descent methods can learn non-linear features from data using single-index models, and shows that it works well for non-Gaussian data as well.


Towards Symmetry-Aware Generation of Periodic Materials

Youzhi Luo, Chengkai Liu, Shuiwang Ji

https://openreview.net/forum?id=Jkc74vn1aZ

Keywords: material generation, symmetries, variational auto-encoder, score-based diffusion model

Compressor summary: SyMat is a new method for generating periodic materials with different symmetries using deep models, which can create atom types, lattices, and coordinates that are invariant to various symmetry transformations.


Module-wise Adaptive Distillation for Multimodality Foundation Models

Chen Liang, Jiahui Yu, Ming-Hsuan Yang, Matthew Brown, Yin Cui, Tuo Zhao, Boqing Gong, Tianyi Zhou

https://openreview.net/forum?id=JhQP33aMx2

Keywords: Multimodality foundation models, knowledge distillation

Compressor summary: The proposed method uses a modified-Thompson sampling algorithm to selectively distill different modules of a multimodal foundation model based on their contributions to reduce its size and maintain performance.


Minimax-Optimal Location Estimation

Shivam Gupta, Jasper C.H. Lee, Eric Price, Paul Valiant

https://openreview.net/forum?id=JeKXmYb4kd

Keywords: location estimation, minimax estimation

Compressor summary: The paper proposes two optimal location estimators for parametric statistics with different criteria, one minimizing estimation error and success probability, and another minimizing expected squared interval width.


Neural Functional Transformers

Allan Zhou, Kaien Yang, Yiding Jiang, Kaylee Burns, Winnie Xu, Samuel Sokota, J Zico Kolter, Chelsea Finn

https://openreview.net/forum?id=JdhyIa0azI

Keywords: equivariance, permutation, implicit neural representation, generalization, transformers, attention

Compressor summary: The paper proposes neural functional Transformers (NFTs) that use attention mechanisms to process neural networks' weights and achieve better performance in weight-space methods, enabling improved classification of implicit neural representations (INRs).


Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time

Zichang Liu, Aditya Desai, Fangshuo Liao, Weitao Wang, Victor Xie, Zhaozhuo Xu, Anastasios Kyrillidis, Anshumali Shrivastava

https://openreview.net/forum?id=JZfg6wGi6g

Keywords: Large language model; KV Cache Compression

Compressor summary: Scissorhands is a system that reduces memory usage and inference time for large language models by storing pivotal tokens more frequently.


Joint Attribute and Model Generalization Learning for Privacy-Preserving Action Recognition

Duo Peng, Li Xu, Qiuhong Ke, Ping Hu, Jun Liu

https://openreview.net/forum?id=JYUN0vYjh9

Keywords: Privacy Preservation, Action Recognition, Meta-Learning

Compressor summary: The paper proposes MPPAR, a meta-learning framework for privacy-preserving action recognition that can adapt to novel privacy attributes and attack models.


Switching Temporary Teachers for Semi-Supervised Semantic Segmentation

Jaemin Na, Jung-Woo Ha, Hyung Jin Chang, Dongyoon Han, Wonjun Hwang

https://openreview.net/forum?id=JXvszuOqY3

Keywords: Semi-supervised Learning, Semantic Segmentation

Compressor summary: The paper proposes Dual Teacher, a method to improve semi-supervised semantic segmentation by using two temporary teachers that shift roles and prevent the student model from losing its unique features.


Parameterizing Non-Parametric Meta-Reinforcement Learning Tasks via Subtask Decomposition

Suyoung Lee, Myungsik Cho, Youngchul Sung

https://openreview.net/forum?id=JX6UloWrmE

Keywords: Deep reinforcement learning, Meta-reinforcement learning, Subtask decomposition

Compressor summary: SDVT is a novel meta-RL approach that decomposes non-parametric tasks into subtasks, uses a Gaussian mixture VAE to meta-learn the decomposition, and employs virtual training for better generalization.


ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation

Jiazheng Xu, Xiao Liu, Yuchen Wu, Yuxuan Tong, Qinkai Li, Ming Ding, Jie Tang, Yuxiao Dong

https://openreview.net/forum?id=JVzeOYEx6d

Keywords: Generative Models, Text-to-Image, Learning from Human Feedback, Multimodality, Evaluation

Compressor summary: The authors introduce ImageReward, a new model that encodes human preferences for text-to-image synthesis, and ReFL, an algorithm to optimize diffusion models using ImageReward.


DIFUSCO: Graph-based Diffusion Solvers for Combinatorial Optimization

Zhiqing Sun, Yiming Yang

https://openreview.net/forum?id=JV8Ff0lgVV

Keywords: neural-symbolic reasoning, combinatorial optimization, diffusion models

Compressor summary: The paper introduces DIFUSCO, a graph-based diffusion framework that generates high-quality solutions for NP-complete problems using neural networks and outperforms previous state-of-the-art methods.


MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers

LILI YU, Daniel Simig, Colin Flaherty, Armen Aghajanyan, Luke Zettlemoyer, Mike Lewis

https://openreview.net/forum?id=JTmO2V9Xpz

Keywords: byte level language model, model architecture, efficient pretraining

Compressor summary: Megabyte is a multi-scale decoder architecture that enables efficient and large-scale tokenization-free autoregressive sequence modeling for various domains such as language, image, and audio.


Randomized Sparse Neural Galerkin Schemes for Solving Evolution Equations with Deep Networks

Jules Berman, Benjamin Peherstorfer

https://openreview.net/forum?id=JTKd7zYROf

Keywords: numerical methods, deep networks, evolution equations, scientific computing, partial differential equations, model reduction

Compressor summary: The paper proposes Neural Galerkin schemes that use randomized sparse updates to train neural networks for time-dependent partial differential equations, improving accuracy and speed over dense updates.


Latent exploration for Reinforcement Learning

Alberto Silvio Chiappa, Alessandro Marin Vargas, Ann Huang, Alexander Mathis

https://openreview.net/forum?id=JSVXZKqfLU

Keywords: Reinforcement learning, efficient exploration, curse of dimensionality, motor control, musculoskeletal control

Compressor summary: Lattice introduces temporally-correlated noise into policy networks to improve exploration and control in overactuated systems, achieving state of the art results on locomotion and musculoskeletal tasks.


MMD-Fuse: Learning and Combining Kernels for Two-Sample Testing Without Data Splitting

Felix Biggs, Antonin Schrab, Arthur Gretton

https://openreview.net/forum?id=JOkgEY9os2

Keywords: Testing, MMD, Kernel Methods, Two-sample testing

Compressor summary: The paper introduces new statistics that improve the ability to test if two samples come from the same distribution using a modified version of the Maximum Mean Discrepancy (MMD) test, and shows their effectiveness on both synthetic and real data.


Towards Combinatorial Generalization for Catalysts: A Kohn-Sham Charge-Density Approach

Phil Pope, David Jacobs

https://openreview.net/forum?id=JOHp5SmckS

Keywords: graph neural networks, equivariance, materials science, chemistry, density functional theory, combinatorial generalization, catalysts

Compressor summary: The authors propose using pointwise learning of the Kohn-Sham charge-density to model catalysts, which shows improved convergence and combinatorial generalization compared to energy prediction.


On Slicing Optimality for Mutual Information

Ammar Fayad, Majd Ibrahim

https://openreview.net/forum?id=JMuKfZx2xU

Keywords: Mutual information, Information Theory

Compressor summary: The paper proposes a framework to find the optimal distribution of slices for measuring mutual information between high-dimensional variables, improving on existing methods that use uniform distributions and discard informative features.


Generalised f-Mean Aggregation for Graph Neural Networks

Ryan Kortvelesy, Steven Morad, Amanda Prorok

https://openreview.net/forum?id=JMrIeKjTAe

Keywords: Aggregation, Graph Neural Networks

Compressor summary: The paper introduces GenAgg, a generalised aggregation operator for graph neural networks that improves performance by parametrising a function space including all standard aggregators.


Vocabulary-free Image Classification

Alessandro Conti, Enrico Fini, Massimiliano Mancini, Paolo Rota, Yiming Wang, Elisa Ricci

https://openreview.net/forum?id=JKhyQHpx7B

Keywords: language and vision, zero-shot classification, image classification

Compressor summary: The paragraph introduces a new task called Vocabulary-free Image Classification (VIC), which aims to classify images without knowing the categories beforehand, and proposes CaSED, a method that uses a pre-trained vision-language model and an external database to achieve this goal.


Global Identifiability of $\ell_1$-based Dictionary Learning via Matrix Volume Optimization

Jingzhou Hu, Kejun Huang

https://openreview.net/forum?id=JK2oPrP8B3

Keywords: dictionary learning, matrix volume, nonconvex optimization

Compressor summary: The authors propose a novel dictionary learning method that ensures global identifiability of the groundtruth matrices by minimizing the volume of the dictionary matrix while maintaining unit $\ell_1$ norm for each row of the sparse coefficient matrix, without requiring mutual incoherence of dictionary atoms.


Simple, Scalable and Effective Clustering via One-Dimensional Projections

Moses Charikar, Monika Henzinger, Lunjia Hu, Maximilian Vötsch, Erik Waingarten

https://openreview.net/forum?id=JIYdbHDonF

Keywords: clustering, k-means, random projection, massive datasets

Compressor summary: A simple randomized clustering algorithm with improved running time and cluster quality compared to popular methods is introduced for data analysis tasks such as k-means clustering and coreset construction.


DatasetDM: Synthesizing Data with Perception Annotations Using Diffusion Models

Weijia Wu, Yuzhong Zhao, Hao Chen, Yuchao Gu, Rui Zhao, Yefei He, Hong Zhou, Mike Zheng Shou, Chunhua Shen

https://openreview.net/forum?id=JIKM2vS8XU

Keywords: Diffusion Model; Text-guided dataset generation

Compressor summary: DatasetDM is a model that generates synthetic images and perception annotations using a diffusion model, enabling efficient training of various perception models on downstream tasks with minimal effort and cost.


Partial Multi-Label Learning with Probabilistic Graphical Disambiguation

Jun-Yi Hang, Min-Ling Zhang

https://openreview.net/forum?id=JDw50IX4TY

Keywords: Machine learning, multi-label learning, partial multi-label learning, label disambiguation

Compressor summary: The paper proposes a probabilistic graphical model for partial multi-label learning, which uses stochastic gradient variational Bayes to infer ground-truth labeling information from inaccurate annotations and performs better than existing methods.


Adversarial Examples Might be Avoidable: The Role of Data Concentration in Adversarial Robustness

Ambar Pal, Jeremias Sulam, Rene Vidal

https://openreview.net/forum?id=JDoA6admhv

Keywords: Adversarial Robustness, Geometry in Data, Low Dimensional Modeling

Compressor summary: The authors explore whether adversarial examples are unavoidable by analyzing how data distribution concentration affects the existence of robust machine learning classifiers.


Improving Few-Shot Generalization by Exploring and Exploiting Auxiliary Data

Alon Albalak, Colin Raffel, William Yang Wang

https://openreview.net/forum?id=JDnLXc4NOn

Keywords: Few-shot learning, natural language processing, few shot learning, NLP, multi-armed bandit, multi armed bandit

Compressor summary: FLAD is a method to improve few-shot learning with auxiliary data, and this paper introduces two algorithms that explore and exploit different datasets, achieving better results than previous methods and enabling larger language models.


Deep Non-line-of-sight Imaging from Under-scanning Measurements

Yue Li, Yueyi Zhang, Juntian Ye, Feihu Xu, Zhiwei Xiong

https://openreview.net/forum?id=JCN9YsZiwB

Keywords: Non-line-of-sight imaging, Transient Recovery, Volume Reconstruction

Compressor summary: The paragraph describes a new deep learning approach to non-line-of-sight imaging that uses under-scanning measurements to reconstruct hidden volumes with high resolution and robustness, outperforming traditional methods.


Grounded Decoding: Guiding Text Generation with Grounded Models for Embodied Agents

Wenlong Huang, Fei Xia, Dhruv Shah, Danny Driess, Andy Zeng, Yao Lu, Pete Florence, Igor Mordatch, Sergey Levine, Karol Hausman, brian ichter

https://openreview.net/forum?id=JCCi58IUsh

Keywords: robotics, language models, embodied agents

Compressor summary: The paragraph discusses the challenge of combining language models' semantic knowledge with grounded robots' real-world understanding and proposes a solution that decodes action sequences using both types of models.


Make Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning

Baohao Liao, Shaomu Tan, Christof Monz

https://openreview.net/forum?id=J8McuwS3zY

Keywords: large language model, parameter-efficient learning, memory-efficient learning, reversible neural network

Compressor summary: MEFT is a memory-efficient fine-tuning method that uses adapters to insert reversibility into PLMs, preserving their starting point and achieving comparable performance to full fine-tuning with much less activation memory.


LLM-Pruner: On the Structural Pruning of Large Language Models

Xinyin Ma, Gongfan Fang, Xinchao Wang

https://openreview.net/forum?id=J8Ajf9WfXP

Keywords: model compression, structural pruning, large language model

Compressor summary: The paper introduces LLM-pruner, a method to compress large language models without relying on their original training data, by selectively removing non-critical structures based on gradient information.


Unlocking Feature Visualization for Deep Network with MAgnitude Constrained Optimization

Thomas FEL, Thibaut Boissin, Victor Boutin, Agustin Martin Picard, Paul Novello, Julien Colin, Drew Linsley, Tom ROUSSEAU, Remi Cadene, Lore Goetschalckx, Laurent Gardes, Thomas Serre

https://openreview.net/forum?id=J7VoDuzuKs

Keywords: explainable AI, feature visualization, interpretability, optimization

Compressor summary: MACO is a method for improving feature visualization in neural networks by optimizing image phase spectrum and introducing three metrics for evaluation.


Glance and Focus: Memory Prompting for Multi-Event Video Question Answering

Ziyi Bai, Ruiping Wang, Xilin CHEN

https://openreview.net/forum?id=J6Niv3yrMq

Keywords: Video Question Answering; Multi-Event Reasoning; Spatial-Temporal Reasoning

Compressor summary: The Glance- Focus model uses dynamic event memories to improve Video Question Answering by mimicking human reasoning strategies, achieving state-of-the-art results on four benchmarks.


Kernel Quadrature with Randomly Pivoted Cholesky

Ethan Nicholas Epperly, Elvira Moreno Ferreira

https://openreview.net/forum?id=J66ptjMkAG

Keywords: kernel quadrature, Nyström approximation, reproducing kernel Hilbert space, randomly pivoted Cholesky

Compressor summary: The paper introduces a fast and accurate kernel quadrature method using randomly pivoted Choleskey nodes that works well with complex geometries and kernels.


Stability-penalty-adaptive follow-the-regularized-leader: Sparsity, game-dependency, and best-of-both-worlds

Taira Tsuchiya, Shinji Ito, Junya Honda

https://openreview.net/forum?id=J3taqrzyyA

Keywords: follow-the-regularized-leader, adaptive learning rate, multi-armed bandits, partial monitoring, data-dependent bound, sparsity, game-dependency, best-of-both-worlds

Compressor summary: The paper proposes a new adaptive learning rate for FTRL called SPA, which enables algorithms with three types of adaptivity in sequential decision-making problems, including sparsity, game-dependency, and best-of-both-worlds.


DesCo: Learning Object Recognition with Rich Language Descriptions

Liunian Harold Li, Zi-Yi Dou, Nanyun Peng, Kai-Wei Chang

https://openreview.net/forum?id=J2Cso0wWZX

Keywords: Vision language, fine-grained recognition, object detection

Compressor summary: The text describes a new vision-language model for object detection that uses rich language descriptions and context-sensitive queries to improve accuracy and adaptability.


Quantifying & Modeling Multimodal Interactions: An Information Decomposition Framework

Paul Pu Liang, Yun Cheng, Xiang Fan, Chun Kai Ling, Suzanne Nie, Richard J. Chen, Zihao Deng, Nicholas Allen, Randy Auerbach, Faisal Mahmood, Ruslan Salakhutdinov, Louis-Philippe Morency

https://openreview.net/forum?id=J1gBijopla

Keywords: multimodal learning, feature interactions, partial information decomposition, information theory, quantification, model selection

Compressor summary: The paragraph discusses an information-theoretic approach called PID to measure interactions between input modalities and output tasks in multimodal applications, and its usefulness in various aspects of multimodal modeling.


A graphon-signal analysis of graph neural networks

Ron Levie

https://openreview.net/forum?id=J0RD92Tmfc

Keywords: graph neural network, graphon, generalization, stability, sampling, Szemerédi regularity lemma

Compressor summary: The paper proposes a new similarity measure for graph-signals in message passing graph neural networks (MPNNs) and shows how it leads to Lipschitz continuity and generalization bounds for MPNNs.


SUBP: Soft Uniform Block Pruning for 1$\times$N Sparse CNNs Multithreading Acceleration

Jingyang Xiang, Siqi Li, Jun Chen, Guang Dai, Shipeng Bai, Yukai Ma, Yong Liu

https://openreview.net/forum?id=J0Pvvxspmz

Keywords: Soft Uniform Block Pruning, Block Angular Redundancy, Hardware Acceleration

Compressor summary: The paper proposes a new method to train sparse neural networks that reduces storage, improves performance, and balances workload across threads.


Smoothed Online Learning for Prediction in Piecewise Affine Systems

Adam Block, Max Simchowitz, Russ Tedrake

https://openreview.net/forum?id=Izt7rDD7jN

Keywords: Smoothed Online Learning, Piecewise Affine Prediction, Learning Dynamics

Compressor summary: The paper presents polynomial-regret algorithms for prediction and simulation in piecewise affine systems using smoothed online learning and introduces new technical tools.


Swap Agnostic Learning, or Characterizing Omniprediction via Multicalibration

Parikshit Gopalan, Michael P. Kim, Omer Reingold

https://openreview.net/forum?id=IzlRh5qwmG

Keywords: Agnostic Learning, Omniprediction, Multicalibration

Compressor summary: Swap Agnostic Learning is a game where a predictor selects a hypothesis and an adversary tries to minimize the loss, and the paper shows how it relates to other concepts like Omniprediction and Multicalibration.


Conditional Adapters: Parameter-efficient Transfer Learning with Fast Inference

Tao Lei, Junwen Bai, Siddhartha Brahma, Joshua Ainslie, Kenton Lee, Yanqi Zhou, Nan Du, Vincent Y Zhao, Yuexin Wu, Bo Li, Yu Zhang, Ming-Wei Chang

https://openreview.net/forum?id=IyYyKov0Aj

Keywords: conditional computation, inference efficiency, parameter efficiency, large models

Compressor summary: CoDA is a transfer learning method that boosts inference efficiency while maintaining accuracy and parameter efficiency on various tasks.


Block Coordinate Plug-and-Play Methods for Blind Inverse Problems

Weijie Gan, Shirin Shoushtari, Yuyang Hu, Jiaming Liu, Hongyu An, Ulugbek Kamilov

https://openreview.net/forum?id=IyWpP2e0bF

Keywords: inverse problems, plug-and-play priors, computational imaging, nonconvex optimization

Compressor summary: BC-PnP is a new method that uses learned denoisers to solve blind imaging inverse problems by combining physical measurements and image estimation.


Assumption violations in causal discovery and the robustness of score matching

Francesco Montagna, Atalanti A. Mastakouri, Elias Eulig, Nicoletta Noceti, Lorenzo Rosasco, Dominik Janzing, Bryon Aragam, Francesco Locatello

https://openreview.net/forum?id=IyTArtpuCK

Keywords: Causal discovery; empirical study; robust inference; benchmark

Compressor summary: The paper benchmarks various causal discovery algorithms on observational data with different conditions, finding that score matching-based methods perform well and providing theoretical explanations.


A Unified Approach to Count-Based Weakly Supervised Learning

Vinay Shukla, Zhe Zeng, Kareem Ahmed, Guy Van den Broeck

https://openreview.net/forum?id=IyAHCbMq3a

Keywords: weakly supervised learning, constraint, label proportion, learning from positive and unlabeled data, multiple instance learning

Compressor summary: The paper proposes a method called count-based weakly-supervised learning that learns from data with inferred weak labels by computing and penalizing deviations from expected label counts.


Perceptual adjustment queries and an inverted measurement paradigm for low-rank metric learning

Austin Xu, Andrew McRae, Jingyan Wang, Mark A. Davenport, Ashwin Pananjady

https://openreview.net/forum?id=IwyymRXfzL

Keywords: human querying, high dimensional low rank matrix estimation, metric learning

Compressor summary: The paper introduces PAQ, a query mechanism that combines cardinal and ordinal feedback to learn an unknown Mahalanobis distance using a two-stage estimator with sample complexity guarantees.


Conformal Meta-learners for Predictive Inference of Individual Treatment Effects

Ahmed Alaa, Zaid Ahmad, Mark van der Laan

https://openreview.net/forum?id=IwnINorSZ5

Keywords: Heterogeneous treatment effects, conformal prediction

Compressor summary: The paper proposes a method to estimate uncertainty in individual treatment effects using machine learning and conformal prediction techniques, showing that it is valid under certain conditions and performs well in simulations.


Accelerating Exploration with Unlabeled Prior Data

Qiyang Li, Jason Zhang, Dibya Ghosh, Amy Zhang, Sergey Levine

https://openreview.net/forum?id=Itorzn4Kwf

Keywords: Reinforcement Learning, Exploration

Compressor summary: The paper proposes a method to improve exploration in sparse reward tasks by using prior data without reward labels to guide an agent, leading to better performance in challenging domains.


Hidden Poison: Machine Unlearning Enables Camouflaged Poisoning Attacks

Jimmy Z. Di, Jack Douglas, Jayadev Acharya, Gautam Kamath, Ayush Sekhari

https://openreview.net/forum?id=Isy7gl1Hqc

Keywords: Machine unlearning, new attack vector, Camouflaging poisoning attacks

Compressor summary: The paper introduces a new data poisoning attack that uses camouflaged points to affect model predictions when some points are removed during model retraining or unlearning.


Model-free Posterior Sampling via Learning Rate Randomization

Daniil Tiapkin, Denis Belomestny, Daniele Calandriello, Eric Moulines, Remi Munos, Alexey Naumov, pierre perrault, Michal Valko, Pierre MENARD

https://openreview.net/forum?id=IrjXmIKFyx

Keywords: reinforcement learning, exploration, q-learning

Compressor summary: Randomized Q-learning (RandQL) is a new algorithm for minimizing regret in episodic MDPs that uses random learning rates and achieves better performance than existing methods.


Why Did This Model Forecast This Future? Information-Theoretic Saliency for Counterfactual Explanations of Probabilistic Regression Models

Chirag Raman, Alec Nonnemaker, Amelia Villegas-Morcillo, Hayley Hung, Marco Loog

https://openreview.net/forum?id=IrEYkhuxup

Keywords: Probabilistic Forecasting, Saliency, Explainability, XAI, Probabilistic Regression

Compressor summary: The authors propose a method to explain probabilistic time-series forecasts by identifying salient timesteps based on human visual cognition and test it on synthetic and real data.


Debiasing Pretrained Generative Models by Uniformly Sampling Semantic Attributes

Walter Gerych, Kevin Hickey, Luke Buquicchio, Kavin Chandrasekaran, Abdulaziz Alajaji, Elke Rundensteiner, Emmanuel Agu

https://openreview.net/forum?id=Iq7v0sZw2H

Keywords: generative models, generative modeling, bias, GANs, debiasing

Compressor summary: The distribution mapping module proposes a way to debias generative models without retraining them by using fair noise distribution, and it works better than current methods.


Emergent and Predictable Memorization in Large Language Models

Stella Biderman, USVSN Sai Prashanth, Lintang Sutawika, Hailey Schoelkopf, Quentin Gregory Anthony, Shivanshu Purohit, Edward Raff

https://openreview.net/forum?id=Iq0DvhB4Kf

Keywords: large language model, emergent properties, memorization

Compressor summary: The paper proposes a method to predict memorization in language models using lower-compute trials, and provides insights on memorization behavior and recommendations for reliability.


Improving the Privacy and Practicality of Objective Perturbation for Differentially Private Linear Learners

Rachel Emily Redberg, Antti Koskela, Yu-Xiang Wang

https://openreview.net/forum?id=IpUJd3KG3c

Keywords: differential privacy, empirical risk minimization, objective perturbation

Compressor summary: The paper improves the objective perturbation mechanism for privacy-preserving machine learning, making it more efficient and competitive with differentially private stochastic gradient descent.


Skill-it! A data-driven skills framework for understanding and training language models

Mayee F Chen, Nicholas Roberts, Kush Bhatia, Jue WANG, Ce Zhang, Frederic Sala, Christopher Re

https://openreview.net/forum?id=IoizwO1NLf

Keywords: language models, data selection

Compressor summary: The authors propose a framework for selecting and ordering training data based on language model skills, which improves performance on downstream tasks with fewer tokens.


SLM: A Smoothed First-Order Lagrangian Method for Structured Constrained Nonconvex Optimization

Songtao Lu

https://openreview.net/forum?id=IobxuwPnWt

Keywords: Functional constrained optimization, bilevel optimization, primal dual method, Lagrangian method

Compressor summary: The paper proposes a smoothed first-order Lagrangian method for solving nonconvex functional constrained optimization problems with nonconvex constraints, and shows its effectiveness through theoretical convergence guarantees and numerical results.


Utilitarian Algorithm Configuration

Devon R. Graham, Kevin Leyton-Brown, Tim Roughgarden

https://openreview.net/forum?id=InB9Loet1u

Keywords: algorithm configuration, algorithm selection, data-driven algorithm design, utility of runtime

Compressor summary: The authors propose a new method for configuring heuristic algorithms that maximizes utility and offers theoretical guarantees about performance, unlike existing methods that only minimize expected runtime.


Improved Bayes Risk Can Yield Reduced Social Welfare Under Competition

Meena Jagadeesan, Michael Jordan, Jacob Steinhardt, Nika Haghtalab

https://openreview.net/forum?id=IltQ87ZdT6

Keywords: competition, equilibria, inverse scaling, digital marketplaces

Compressor summary: Increasing the scale of machine learning models can hurt overall predictive accuracy when multiple providers compete for users, as improving data representation quality may reduce social welfare.


Perturbation Towards Easy Samples Improves Targeted Adversarial Transferability

Junqi Gao, Biqing Qi, Yao Li, Zhichang Guo, Dong Li, Yuming Xing, Dazhi Zhang

https://openreview.net/forum?id=IkD1EWFF8c

Keywords: Adversarial Attacks; Generative Attack; Transferable Targeted Attack

Compressor summary: The paper proposes ESMA, a generative targeted attack strategy that improves transferability by adding perturbations to easy samples in the target class, reducing storage space and computation time compared to current methods.


StableFDG: Style and Attention Based Learning for Federated Domain Generalization

Jungwuk Park, Dong-Jun Han, Jinho Kim, Shiqiang Wang, Christopher Brinton, Jaekyun Moon

https://openreview.net/forum?id=IjZa2fQ8tL

Keywords: Federated Learning, Domain Generalization

Compressor summary: The paper proposes StableFDG, a learning strategy for federated domain generalization, which uses style and attention-based methods to improve domain diversity and learn domain-invariant characteristics in data-poor scenarios.


On the Adversarial Robustness of Out-of-distribution Generalization Models

Xin Zou, Weiwei Liu

https://openreview.net/forum?id=IiwTFcGGTq

Keywords: Adversarial Robustness, Out-of-distribution Generalization

Compressor summary: The text discusses out-of-distribution generalization, its vulnerability to attacks, and proposes methods to improve it.


Calibration by Distribution Matching: Trainable Kernel Calibration Metrics

Charles Thomas Marx, Sofian Zalouk, Stefano Ermon

https://openreview.net/forum?id=IhxD94i5ra

Keywords: Uncertainty Quantification, Calibration, Decision Making, Probabilistic Forecasting

Compressor summary: The authors propose kernel-based calibration metrics for both classification and regression that improve predictive uncertainty, sharpness, and decision-making by incorporating them into empirical risk minimization.


Bayes beats Cross Validation: Efficient and Accurate Ridge Regression via Expectation Maximization

Shu Tew, Mario Boley, Daniel F. Schmidt

https://openreview.net/forum?id=Ih2yL7o2Gq

Keywords: Ridge Regression, Cross validation, Expectation Maximisation, Bayesian methods

Compressor summary: A faster and better method for tuning ridge regression hyper-parameter $\lambda$ using Bayesian formulation and iterative EM procedure is proposed.


Efficient Model-Free Exploration in Low-Rank MDPs

Zakaria Mhammedi, Adam Block, Dylan J Foster, Alexander Rakhlin

https://openreview.net/forum?id=IgDa5Ynm9l

Keywords: Reinforcement learning, Representation Learning, Low-rank MDPs, Model-Free Learning

Compressor summary: The paper proposes SpanRL, a sample-efficient exploration algorithm for high-dimensional reinforcement learning that works without structural assumptions and uses barycentric spanners for efficient exploration.


Conditional independence testing under misspecified inductive biases

Felipe Maia Polo, Yuekai Sun, Moulinath Banerjee

https://openreview.net/forum?id=Ifq8GMdqJK

Keywords: conditional independence, hypothesis testing, misspecification

Compressor summary: The paper studies how regression-based tests for conditional independence can fail due to misspecified models or algorithms, and proposes a new test (RBPT) that is robust against such failures.


Online Performative Gradient Descent for Learning Nash Equilibria in Decision-Dependent Games

Zihan Zhu, Ethan X Fang, Zhuoran Yang

https://openreview.net/forum?id=IdF7VT6eEs

Keywords: Performative Prediction, Nash Equilibrium, Reproducing Kernel Hilbert Space, Online Learning, Stochastic Gradient Methods

Compressor summary: The paper proposes a novel online algorithm (OPGD) to find the Nash equilibrium of decision-dependent games in the bandit feedback setting, where agents' actions affect population data and traditional gradient-based methods are infefficient without a gradient oracle.


Stochastic Approximation Approaches to Group Distributionally Robust Optimization

Lijun Zhang, Peng Zhao, Zhenhua Zhuang, Tianbao Yang, Zhi-Hua Zhou

https://openreview.net/forum?id=IcIQbCWoFj

Keywords: Group distributionally robust optimization, Stochastic mirror descent, Non-oblivious online learning, Sample complexity, Stochastic mirror-prox algorithm, Mini-batch

Compressor summary: The paper studies group distributionally robust optimization (GDRO) and proposes methods to reduce sample complexity using techniques from online learning and non-uniform sampling.


Can Language Models Teach? Teacher Explanations Improve Student Performance via Personalization

Swarnadeep Saha, Peter Hase, Mohit Bansal

https://openreview.net/forum?id=IacxcFpvWQ

Keywords: Language Models, Reasoning, Explanations

Compressor summary: The study explores how to make LLMs good teachers for weaker agents by using natural language explanations and a communication budget. It shows that teacher LLMs can improve student performance by intervening at the right time, personalizing explanations, and avoiding misinformation.


ResoNet: Noise-Trained Physics-Informed MRI Off-Resonance Correction

Alfredo De Goyeneche, Shreya Ramachandran, Ke Wang, Ekin Karasan, Joseph Yitan Cheng, Stella X. Yu, Michael Lustig

https://openreview.net/forum?id=Ia4dmqst0Z

Keywords: Inverse problem, MRI, Medical Imaging, Computational Imaging, Deep Learning, Off-Resonance

Compressor summary: The paper proposes a physics-informed deep learning framework for correcting off-resonance artifacts in MRI using synthetic data, which could enable faster and more accurate scans with non-Cartesian sampling trajectories.


Efficient Test-Time Adaptation for Super-Resolution with Second-Order Degradation and Reconstruction

Zeshuai Deng, Zhuokun Chen, Shuaicheng Niu, Thomas H. Li, Bohan Zhuang, Mingkui Tan

https://openreview.net/forum?id=IZRlMABK4l

Keywords: Image Super-resolution, Test-time Adaptation, Self-supervised Learning, Second-Order Degradation

Compressor summary: The authors propose a test-time adaptation framework for image super-resolution, called SRTTA, that can handle different degradation types by predicting the degradation type and adapting the model using feature-level reconstruction learning.


Intriguing Properties of Quantization at Scale

Arash Ahmadian, Saurabh Dash, Hongyu Chen, Bharat Venkitesh, Zhen Stephen Gou, Phil Blunsom, Ahmet Üstün, Sara Hooker

https://openreview.net/forum?id=IYe8j7Gy8f

Keywords: Quantization, optimization, language modelling, efficiency

Compressor summary: This paper investigates if quantization cliffs, which cause performance drops in large models, are solely a factor of scale and proposes a method to optimize for efficient quantization by suppressing outlier dimensions.


Optimization or Architecture: How to Hack Kalman Filtering

Ido Greenberg, Netanel Yannay, Shie Mannor

https://openreview.net/forum?id=IXWaWPkGke

Keywords: non-linear filtering, Kalman filter, noise estimation, optimization, Cholesky parameterization

Compressor summary: The Optimized Kalman Filter (OKF) improves the performance of the standard Kalman Filter by optimizing both its architecture and parameters, making it competitive with non-linear models like neural networks.


Fine-grained Late-interaction Multi-modal Retrieval for Retrieval Augmented Visual Question Answering

Weizhe Lin, Jinghong Chen, Jingbiao Mei, Alexandru Coca, Bill Byrne

https://openreview.net/forum?id=IWWWulAX7g

Keywords: knowledge-based visual question answering, knowledge retrieval, multi-modality, vision-and-language

Compressor summary: FLMR improves knowledge retrieval in RA-VQA by using finer-grained image and question embeddings and complementary image representations from a vision model.


Are Emergent Abilities of Large Language Models a Mirage?

Rylan Schaeffer, Brando Miranda, Sanmi Koyejo

https://openreview.net/forum?id=ITw9edRDlD

Keywords: large language models, foundation models, natural language processing, language modeling, emergent abilities

Compressor summary: The paper suggests that emergent abilities in large language models are due to the choice of metric rather than scale, and provides evidence for this claim through various analyses and experiments.


Implicit Bias of Gradient Descent for Logistic Regression at the Edge of Stability

Jingfeng Wu, Vladimir Braverman, Jason D. Lee

https://openreview.net/forum?id=IT9mWLYNpQ

Keywords: gd; implicit bias; edge of stability

Compressor summary: This paper studies how gradient descent with constant stepsize behaves on linearly separable data with non-monotonic losses, proving its ability to minimize logistic loss and revealing its divergence under exponential loss in the edge of stability regime.


CD-GraB: Coordinating Distributed Example Orders for Provably Accelerated Training

A. Feder Cooper, Wentao Guo, Khiem Pham, Tiancheng Yuan, Charlie F. Ruan, Yucheng Lu, Christopher De Sa

https://openreview.net/forum?id=ISRyILhAyS

Keywords: permuted example ordering, distributed training, scalable training, herding

Compressor summary: CD-GraB is a new method that improves the speed of distributed machine learning by using example ordering based on stale gradients, leading to faster convergence than existing methods.


Mutual-Information Regularized Multi-Agent Policy Iteration

Jiangxing Wang, Deheng Ye, Zongqing Lu

https://openreview.net/forum?id=IQRc3FrYOG

Keywords: Multi-Agent Reinforcement Learning

Compressor summary: The text proposes using mutual information as an augmented reward to help multi-agent reinforcement learning algorithms adapt to changing team compositions and perform well in complex cooperative tasks.


Towards Characterizing the First-order Query Complexity of Learning (Approximate) Nash Equilibria in Zero-sum Matrix Games

Hedi Hadiji, Sarah Sachs, Tim van Erven, Wouter M Koolen

https://openreview.net/forum?id=IPNg84RF1k

Keywords: game theory, minimax optimization, lower bounds

Compressor summary: The paragraph discusses the query complexity of learning equilibria in zero-sum matrix games, presenting new lower bounds and upper bounds for different values of K and epsilon, and proposing future research directions.


HQA-Attack: Toward High Quality Black-Box Hard-Label Adversarial Attack on Text

Han Liu, Zhi Xu, Xiaotong Zhang, Feng Zhang, Fenglong Ma, Hongyang Chen, Hong Yu, Xianchao Zhang

https://openreview.net/forum?id=IOuuLBrGJR

Keywords: High-quality adversarial example, Black-box hard-label textual adversarial attack

Compressor summary: HQA-Attack is a simple framework for generating high quality adversarial examples on text data under black-box hard-label attack scenarios, by substituting original words and using synonym sets to optimize the example while minimizing perturbation rate and query number.


DSR: Dynamical Surface Representation as Implicit Neural Networks for Protein

Daiwen Sun, He Huang, Yao Li, Xinqi Gong, Qiwei Ye

https://openreview.net/forum?id=IOSaJ7ukgf

Keywords: Protein molecular dynamics, Protein surface representation, Implicit neural representation, Signed distance function, Continuous time modeling

Compressor summary: The authors propose a new neural network method for modeling protein dynamics using an implicit representation of protein surfaces based on signed distance functions, which can accurately capture and interpolate large-scale protein motions, and offer advantages over existing methods.


TopoSRL: Topology preserving self-supervised Simplicial Representation Learning

Hiren Madhu, Sundeep Prabhakar Chepuri

https://openreview.net/forum?id=INS3ltgjg7

Keywords: Simplicial representation learning, Self-supervised learning, Message passing simplicial networks

Compressor summary: $\texttt{TopoSRL}$ is a new self-supervised learning method for simplicial complexes that captures higher-order interactions and preserves topology, outperforming existing graph-based methods and supervised models.


Task-aware world model learning with meta weighting via bi-level optimization

Huining Yuan, Hongkun Dou, Xingyu Jiang, Yue Deng

https://openreview.net/forum?id=IN3hQx1BrC

Keywords: Model-based reinforcement learning, world model, generative model, meta-learning, bi-level optimization

Compressor summary: TEMPO is a bi-level model learning framework for reinforcement learning that combines maximum-likelihood and value-equivalent models using a meta weighter network to improve task awareness and semantic information.


“Why Not Looking backward?” A Robust Two-Step Method to Automatically Terminate Bayesian Optimization

Shuang Li, Ke Li, Wei Li

https://openreview.net/forum?id=IMiGRqltQQ

Keywords: Bayesian Optimization, Termination Criterion, Looking Backward

Compressor summary: The paper presents a novel, theoretically grounded method to terminate Bayesian Optimization early when it is in a convex region, saving computation while maintaining good performance.


EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought

Yao Mu, Qinglong Zhang, Mengkang Hu, Wenhai Wang, Mingyu Ding, Jun Jin, Bin Wang, Jifeng Dai, Yu Qiao, Ping Luo

https://openreview.net/forum?id=IL5zJqfxAa

Keywords: Embodied AI, Multi-modal Foundation Model, Embodied Control

Compressor summary: EmbodiedGPT is an end-to-end multi-modal model for embodied AI that uses a large language model to generate plans and extract features for effective embodied tasks.


Initialization Matters: Privacy-Utility Analysis of Overparameterized Neural Networks

Jiayuan Ye, Zhenyu Zhu, Fanghui Liu, Reza Shokri, Volkan Cevher

https://openreview.net/forum?id=IKvxmnHjkL

Keywords: overparameterized neural network, privacy

Compressor summary: This paper analyzes how over-parameterized models in machine learning algorithms can leak information about their training data and explores how initialization distribution affects this privacy loss depending on the model's depth.


Towards Label Position Bias in Graph Neural Networks

Haoyu Han, Xiaorui Liu, Feng Shi, MohamadAli Torkamani, Charu C. Aggarwal, Jiliang Tang

https://openreview.net/forum?id=IKjOMA8olL

Keywords: Graph Neural Networks, Label Position Bias, Graph Structure Learning

Compressor summary: The paper introduces label position bias as a new bias in graph neural networks and proposes an optimization framework to reduce it.


QuACK: Accelerating Gradient-Based Quantum Optimization with Koopman Operator Learning

Di Luo, Jiayu Shen, Rumen Dangovski, Marin Soljacic

https://openreview.net/forum?id=IKQOS8rqwr

Keywords: Koopman operator, quantum optimization, machine learning

Compressor summary: QuACK is a novel framework that uses Koopman operator theory to significantly accelerate quantum optimization and machine learning by efficiently predicting gradient dynamics on quantum computers.


Goal-Conditioned Predictive Coding for Offline Reinforcement Learning

Zilai Zeng, Ce Zhang, Shijie Wang, Chen Sun

https://openreview.net/forum?id=IJblKO45YU

Keywords: reinforcement learning, offline RL, self-supervised learning

Compressor summary: This paper explores how sequence modeling can improve policy learning by condensing trajectories into useful representations using Goal-Conditioned Predictive Coding (GCPC).


Leveraging sparse and shared feature activations for disentangled representation learning

Marco Fumero, Florian Wenzel, Luca Zancato, Alessandro Achille, Emanuele Rodolà, Stefano Soatto, Bernhard Schölkopf, Francesco Locatello

https://openreview.net/forum?id=IHR83ufYPy

Keywords: disentanglement, OOD generalization, multitask learning

Compressor summary: The authors propose a method to learn disentangled features from multiple supervised tasks without directly observing the factors of variation, and show its effectiveness on various real-world data.


Connecting Multi-modal Contrastive Representations

Zehan Wang, Yang Zhao, Xize Cheng, Haifeng Huang, Jiageng Liu, Aoxiong Yin, Li Tang, Linjun Li, Yongqi Wang, Ziang Zhang, Zhou Zhao

https://openreview.net/forum?id=IGTbT9P1ti

Keywords: multi-modal, representation learning, contrastive learning

Compressor summary: The paper proposes a method called C-MCR to learn multi-modal representations without paired data by connecting existing MCRs trained on different modality pairs using overlapping modalities and enhancing semantic consistency.


SHAP-IQ: Unified Approximation of any-order Shapley Interactions

Fabian Fumagalli, Maximilian Muschalik, Patrick Kolpaczki, Eyke Hüllermeier, Barbara Eva Hammer

https://openreview.net/forum?id=IEMLNF4gK4

Keywords: Explainable Artificial Intelligence, Feature Interaction, Shapley Interaction, Shapley Value

Compressor summary: The paper introduces SHAP-IQ, a method for computing Shapley interactions in XAI using a novel representation and sampling, with theoretical guarantees and improved calculations.


Estimating Causal Effects Identifiable from a Combination of Observations and Experiments

Yonghan Jung, Ivan Diaz, Jin Tian, Elias Bareinboim

https://openreview.net/forum?id=IEJzoOBM0z

Keywords: Causal Effect Estimation, Causal Effect Identification, Data Fusion, Double Machine Learning, Doubly Robust Estimator

Compressor summary: The paper proposes a new estimator for learning causal relations from observational and interventional data that is robust to bias and can handle multiple outcomes.


Evaluating and Inducing Personality in Pre-trained Language Models

Guangyuan Jiang, Manjie Xu, Song-Chun Zhu, Wenjuan Han, Chi Zhang, Yixin Zhu

https://openreview.net/forum?id=I9xE1Jsjfx

Keywords: machine personality, machine behavior, personality trait theory, psychometric, large language models, prompt

Compressor summary: The authors propose using a Machine Personality Inventory (MPI) tool to study and induce specific personalities in large language models (LLMs), inspired by human psychometric tests and the Big Five Personality Factors theory.


Formulating Discrete Probability Flow Through Optimal Transport

Pengze Zhang, Hubery Yin, Chen Li, Xiaohua Xie

https://openreview.net/forum?id=I9GNrInbdf

Keywords: Discrete Probability Flow, Optimal Transport

Compressor summary: The paper introduces a new theory for discrete diffusion models, based on optimal transport principles, and shows how it leads to a better sampling method with more certain outcomes.


State Regularized Policy Optimization on Data with Dynamics Shift

Zhenghai Xue, Qingpeng Cai, Shuchang Liu, Dong Zheng, Peng Jiang, Kun Gai, Bo An

https://openreview.net/forum?id=I8t9RKDnz2

Keywords: Reinforcement Learning, Dynamics Shift, Stationary State Distribution, Offline RL, Off-Policy RL

Compressor summary: The paper proposes a new Reinforcement Learning algorithm (SRPO) that uses stationary state distributions to regularize policies, enabling efficient data reuse and improving the performance of context-based algorithms in different environments with similar structures but different dynamics.


Robust Concept Erasure via Kernelized Rate-Distortion Maximization

Somnath Basu Roy Chowdhury, Nicholas Monath, Kumar Avinava Dubey, Amr Ahmed, Snigdha Chaturvedi

https://openreview.net/forum?id=I6aOjhpcNQ

Keywords: Concept Erasure, Representation Learning, Rate distortion, Fairness, Debiasing

Compressor summary: The paper introduces KRaM, a new objective for concept erasure in distributed representations, which uses a rate-distortion function to transform representations while preserving other information. KRaM can handle various concepts and domains and shows good results in experiments.


Deductive Verification of Chain-of-Thought Reasoning

Zhan Ling, Yunhao Fang, Xuanlin Li, Zhiao Huang, Mingu Lee, Roland Memisevic, Hao Su

https://openreview.net/forum?id=I5rsM4CY2z

Keywords: Chain-of-thought, Large language model, Reasoning

Compressor summary: The text discusses a proposed method to improve large language models' deductive reasoning through natural language-based reasoning verification in step-by-step subprocesses.


On Robust Streaming for Learning with Experts: Algorithms and Lower Bounds

David Woodruff, Fred Zhang, Samson Zhou

https://openreview.net/forum?id=I5SM5y57k2

Keywords: online learning, memory efficiency, sub-linear algorithm, communication lower bound

Compressor summary: The paper studies robust algorithms for online learning with experts under memory constraints and shows a trade-off between space and regret, while also providing a lower bound on space usage for deterministic algorithms.


Provably Bounding Neural Network Preimages

Suhas Kotha, Christopher Brix, J Zico Kolter, Krishnamurthy Dj Dvijotham, Huan Zhang

https://openreview.net/forum?id=I50HbChk3U

Keywords: Trustworthy ML, Formal Verification, Safe Control, OOD Detection

Compressor summary: The INVPROP algorithm efficiently over-approximates inputs that lead to specific outputs for neural networks and other systems, outperforming prior work in speed and precision.


Randomized and Deterministic Maximin-share Approximations for Fractionally Subadditive Valuations

Hannaneh Akrami, Kurt Mehlhorn, Masoud Seddighin, Golnoosh Shahkarami

https://openreview.net/forum?id=I3k2NHt1zu

Keywords: Algorithmic game theory, Fairness, Randomized, Allocation, Maximin-share, Fractionally Subadditive

Compressor summary: The paper studies how to fairly allocate indivisible items to agents with subadditive valuations, improving both deterministic and randomized allocation methods.


GeoCLIP: Clip-Inspired Alignment between Locations and Images for Effective Worldwide Geo-localization

Vicente Vivanco Cepeda, Gaurav Kumar Nayak, Mubarak Shah

https://openreview.net/forum?id=I18BXotQ7j

Keywords: Geo-localization, Image-to-GPS retrieval, CLIP, Random Fourier Features

Compressor summary: GeoCLIP is a new method for pinpointing images' locations on Earth using GPS encoding and a hierarchical representation, overcoming limitations of previous approaches.


Future-Dependent Value-Based Off-Policy Evaluation in POMDPs

Masatoshi Uehara, Haruka Kiyohara, Andrew Bennett, Victor Chernozhukov, Nan Jiang, Nathan Kallus, Chengchun Shi, Wen Sun

https://openreview.net/forum?id=HwhRehMr4a

Keywords: Reinforcement learning theory, POMDP, PAC RL, Off-policy evaluation, Offilne reinforcement learning

Compressor summary: The paper proposes a new model-free off-policy evaluation method for partially observable MDPs using future-dependent value functions and conditional moment equations, with PAC result and Bellman completeness guarantees.


Synthetic-to-Real Pose Estimation with Geometric Reconstruction

Qiuxia Lin, Kerui Gu, Linlin Yang, Angela Yao

https://openreview.net/forum?id=HvhagNdf5z

Keywords: pose estimation, domain adaptation

Compressor summary: The paper proposes a reconstruction-based strategy to adapt models trained on synthetic data to real-world domains using unlabelled data, improving keypoint localization accuracy for hand and human pose estimation tasks.


Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models

Pan Lu, Baolin Peng, Hao Cheng, Michel Galley, Kai-Wei Chang, Ying Nian Wu, Song-Chun Zhu, Jianfeng Gao

https://openreview.net/forum?id=HtqnVSCj3q

Keywords: large language models, compositional reasoning, tool use, multi-modal reasoning, mathematical reasoning

Compressor summary: Chameleon is an AI system that augments large language models with plug-and-play modules for compositional reasoning, enabling them to access up-to-date information and perform complex tasks.


Understanding and Mitigating Copying in Diffusion Models

Gowthami Somepalli, Vasu Singla, Micah Goldblum, Jonas Geiping, Tom Goldstein

https://openreview.net/forum?id=HtMXRGbUMt

Keywords: diffusion models, memorization, data replication, model safety

Compressor summary: The paper analyzes how text-to-image diffusion models can unintentionally copy images from their training data and proposes techniques to reduce this problem by modifying image captions.


Addressing the speed-accuracy simulation trade-off for adaptive spiking neurons

Luke Taylor, Andrew J King, Nicol Spencer Harper

https://openreview.net/forum?id=Ht79ZTVMsn

Keywords: spiking neural network, surrogate gradient descent, adaptive leaky integrate and fire neuron, speed-accuracy trade-off, electrophysiological recordings

Compressor summary: The ALIF model is a brain simulation tool that can be made faster and more accurate by using parallel processing on GPUs and smaller time-steps.


Causal Component Analysis

Wendong Liang, Armin Kekić, Julius von Kügelgen, Simon Buchholz, Michel Besserve, Luigi Gresele, Bernhard Schölkopf

https://openreview.net/forum?id=HszLRiHyfO

Keywords: Causality, independent component analysis, causal inference, interventions, latent variable models, identifiability

Compressor summary: Causal Component Analysis (CauCA) is an extension of Independent Component Analysis (ICA) that incorporates causal relationships among latent variables, and can be seen as a special case of Causal Representation Learning (CRL).


Assessor360: Multi-sequence Network for Blind Omnidirectional Image Quality Assessment

Tianhe Wu, Shuwei Shi, Haoming Cai, Mingdeng Cao, Jing Xiao, Yinqiang Zheng, Yujiu Yang

https://openreview.net/forum?id=HrL1oblm1a

Keywords: Blind omnidirectional image quality assessment, Multi-sequence network, Viewport sequence

Compressor summary: The paper introduces Assessor360, a novel network for objectively assessing the quality of omnidirectional images (ODIs) used in virtual reality, by modeling the observer's browsing process and using a multi-sequence approach with recursive probability sampling, multi-scale feature aggregation, and temporal modeling.


Construction of Hierarchical Neural Architecture Search Spaces based on Context-free Grammars

Simon Schrodi, Danny Stoll, Binxin Ru, Rhea Sanjay Sukthanker, Thomas Brox, Frank Hutter

https://openreview.net/forum?id=Hpt1i5j6wh

Keywords: Neural Architecture Search, Hierarchical Search Space, Context-free Grammars, Bayesian Optimization

Compressor summary: The paper introduces a new framework for generating large hierarchical search spaces for Neural Architecture Search using context-free grammars and an efficient kernel design for Bayesian Optimization.


Ensemble-based Deep Reinforcement Learning for Vehicle Routing Problems under Distribution Shift

Yuan Jiang, Zhiguang Cao, Yaoxin Wu, Wen Song, Jie Zhang

https://openreview.net/forum?id=HoBbZ1vPAh

Keywords: Vehicle Routing Problem, Distribution shift, Deep Reinforcement Learning, Ensemble Learning

Compressor summary: The authors propose a diverse ensemble of sub-policies using deep reinforcement learning to improve vehicle routing performance across different instance distributions.


Learning in the Presence of Low-dimensional Structure: A Spiked Random Matrix Perspective

Jimmy Ba, Murat A Erdogdu, Taiji Suzuki, Zhichao Wang, Denny Wu

https://openreview.net/forum?id=HlIAoCHDWW

Keywords: random matrix theory, high-dimensional statistics, neural network, kernel method, feature learning

Compressor summary: The paper investigates how the strength of a low-dimensional component in high-dimensional data affects the learning performance of kernel methods and neural networks with gradient descent.


Diffusion-Based Probabilistic Uncertainty Estimation for Active Domain Adaptation

Zhekai Du, Jingjing Li

https://openreview.net/forum?id=HffQOS3mk8

Keywords: diffusion-based models, active learning, domain adaptation, source-free domain adaptation, uncertainty estimation

Compressor summary: The paragraph describes a probabilistic framework for active domain adaptation that uses variational inference to capture data-level and prediction-level uncertainties, enabling efficient sampling of all possible predictions and selective annotation of informative target samples based on p-values.


ID and OOD Performance Are Sometimes Inversely Correlated on Real-world Datasets

Damien Teney, LIN Yong, Seong Joon Oh, Ehsan Abbasnejad

https://openreview.net/forum?id=HZQZli6amV

Keywords: Generalisation; machine learning

Compressor summary: The paper reveals that there are cases where in-distribution and out-of-distribution performance trade-offs exist, contrary to previous findings, and suggests that relying solely on in-distribution performance for model selection can lead to suboptimal out-of-distribution generalization.


SLIBO-Net: Floorplan Reconstruction via Slicing Box Representation with Local Geometry Regularization

Jheng-Wei Su, Kuei-Yu Tung, Chi-Han Peng, Peter Wonka, Hung-Kuo Chu

https://openreview.net/forum?id=HYo2Ao3hP8

Keywords: deep learning, layout reconstruction

Compressor summary: The paper introduces SLIBO-Net, a novel transformer-based approach to reconstruct 2D floorplans from unstructured 3D point clouds with improved semantic quality, efficient representation, and local geometric details, achieving state-of-the-art results on the Structure3D dataset.


Alignment with human representations supports robust few-shot learning

Ilia Sucholutsky, Thomas L. Griffiths

https://openreview.net/forum?id=HYGnmSLBCf

Keywords: representation learning, supervised learning, human alignment, few-shot learning

Compressor summary: AI systems with representations similar to humans perform better on few-shot learning tasks, are more robust to attacks and domain shifts, but having human alignment is not always necessary.


Video Prediction Models as Rewards for Reinforcement Learning

Alejandro Escontrela, Ademi Adeniji, Wilson Yan, Ajay Jain, Xue Bin Peng, Ken Goldberg, Youngwoon Lee, Danijar Hafner, Pieter Abbeel

https://openreview.net/forum?id=HWNl9PAYIP

Keywords: Reinforcement learning, generative modeling, learning from demonstrations, video prediction, unsupervised reinforcement learning

Compressor summary: VIPER is an algorithm that learns expert-level behaviors for reinforcement learning agents by extracting preferences from unlabeled videos using pretrained video prediction models.


On the Convergence and Sample Complexity Analysis of Deep Q-Networks with $\epsilon$-Greedy Exploration

Shuai Zhang, Hongkang Li, Meng Wang, Miao Liu, Pin-Yu Chen, Songtao Lu, Sijia Liu, Keerthiram Murugesan, Subhajit Chaudhury

https://openreview.net/forum?id=HWGWeaN76q

Keywords: Reinforcement learning, Deep Q Network, Convergence analysis, Sample complexity, Generalization analysis

Compressor summary: The paper analyzes the theoretical aspects of deep Q-Network with epsilon-greedy exploration in reinforcement learning and shows how its convergence and sample complexity depend on the epsilon value.


Online RL in Linearly $q^\pi$-Realizable MDPs Is as Easy as in Linear MDPs If You Learn What to Ignore

Gellért Weisz, András György, Csaba Szepesvari

https://openreview.net/forum?id=HV85SiyrsV

Keywords: Reinforcement learning, linear function approximation, online learning

Compressor summary: The paper proposes a novel online reinforcement learning algorithm for linearly $q^\pi$-realizable MDPs that can efficiently learn policies even in the presence of states with nearly equal action-values and misspecified features.


Detecting hidden confounding in observational data using multiple environments

Rickard Karlsson, JH Krijthe

https://openreview.net/forum?id=HUuEMMM8Ik

Keywords: causal inference, hidden confounding, multiple environments, independent causal mechansisms, independence testing

Compressor summary: The paper proposes a method to detect hidden confounding factors in multiple observational datasets by testing conditional independencies and evaluates its performance using simulations and real-world data.


Guiding The Last Layer in Federated Learning with Pre-Trained Models

Gwen Legate, Nicolas Bernier, Lucas Caccia, Edouard Oyallon, Eugene Belilovsky

https://openreview.net/forum?id=HRGd5dcVfw

Keywords: federated learning; transfer learning; nearest mean classifier; continual learning;

Compressor summary: The text describes how using pre-trained models and efficient transfer learning methods can improve federated learning by reducing costs and improving performance.


Direct Preference Optimization: Your Language Model is Secretly a Reward Model

Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher D Manning, Stefano Ermon, Chelsea Finn

https://openreview.net/forum?id=HPuSIXJaa9

Keywords: reinforcement learning from human feedback, language models, RLHF, preferences

Compressor summary: DPO is a simple and effective method for aligning large unsupervised language models with human preferences, outperforming existing methods like RLHF.


Learning Dense Flow Field for Highly-accurate Cross-view Camera Localization

Zhenbo Song, XiangHui Ze, Jianfeng Lu, Yujiao Shi

https://openreview.net/forum?id=HPrd17Qvbp

Keywords: Optical flow, correspondence learning, cross-view, camera localization

Compressor summary: The paper proposes a novel method to estimate camera pose using pixel-wise flow fields learned from ground and satellite images, improving localization accuracy significantly.


A Heat Diffusion Perspective on Geodesic Preserving Dimensionality Reduction

Guillaume Huguet, Alexander Tong, Edward De Brouwer, Yanlei Zhang, Guy Wolf, Ian Adelstein, Smita Krishnaswamy

https://openreview.net/forum?id=HNd4qTJxkW

Keywords: manifold learning, heat diffusion, geodesic, metric preserving, dimensionality reduction, embedding

Compressor summary: The authors propose a new heat geodesic embedding method for manifold learning and denoising, which connects heat diffusion to manifold distances using Riemannian geometry and outperforms existing methods in preserving ground truth distances and cluster structure.


A Simple Yet Effective Strategy to Robustify the Meta Learning Paradigm

Cheems Wang, Yiqin Lv, Yanghe Feng, Zheng Xie, Jincai Huang

https://openreview.net/forum?id=HMqGYxnlpv

Keywords: meta learning, robust fast adaptation, model agnostic meta learning

Compressor summary: This paper proposes a distributionally robust optimization approach for meta learning, which improves its robustness to task distributions and reduces the worst fast adaptation risk.


Unifying GANs and Score-Based Diffusion as Generative Particle Models

Jean-Yves Franceschi, Mike Gartrell, Ludovic Dos Santos, Thibaut Issenhuth, Emmanuel de Bezenac, Mickael Chen, Alain Rakotomamonjy

https://openreview.net/forum?id=HMhEFKDQ6J

Keywords: deep learning, generative models, GANs, generative adversarial networks, diffusion, score-based, gradient flows

Compressor summary: The paper proposes a unified framework for particle and adversarial generative models, suggesting that a generator is an optional addition to any generative model, and empirically tests its applications.


Embedding Space Interpolation Beyond Mini-Batch, Beyond Pairs and Beyond Examples

Shashanka Venkataramanan, Ewa Kijak, laurent amsaleg, Yannis Avrithis

https://openreview.net/forum?id=HKueO74ZTB

Keywords: Interpolation based data augmentation, mixup, dense interpolation, robustness, representation learning

Compressor summary: The authors propose MultiMix, a data augmentation method that generates more interpolated examples in the embedding space than previous methods, leading to better performance on sequence data tasks.


Aligning Language Models with Human Preferences via a Bayesian Approach

Jiashuo WANG, Haozhao Wang, Shichao Sun, Wenjie Li

https://openreview.net/forum?id=HGFcM3UU50

Keywords: Aligned Models; Human-centric NLG

Compressor summary: The paper proposes a novel Bayesian framework called d-PM to account for the distribution of disagreements among human preferences in NLG systems and uses contrastive learning to train the model efficiently, achieving better results than previous methods.


ResMem: Learn what you can and memorize the rest

Zitong Yang, Michal Lukasik, Vaishnavh Nagarajan, Zonglin Li, Ankit Singh Rawat, Manzil Zaheer, Aditya Krishna Menon, Sanjiv Kumar

https://openreview.net/forum?id=HFQFAyNucq

Keywords: deep learning, generalization, memorization, deep learning theory, boosting, nearest neighbor

Compressor summary: ResMem is a method to improve model generalization by explicitly memorizing training labels through fitting residuals with a nearest-neighbor based regressor.


On quantum backpropagation, information reuse, and cheating measurement collapse

Amira Abbas, Robbie King, Hsin-Yuan Huang, William J. Huggins, Ramis Movassagh, Dar Gilboa, Jarrod Ryan McClean

https://openreview.net/forum?id=HF6bnhfSqH

Keywords: Backpropagation, quantum computing, shadow tomography, gentle measurement

Compressor summary: The paper explores how to train quantum neural networks efficiently by using multiple copies of a state and shadow tomography, similar to backpropagation in classical deep learning.


Training Transformers with 4-bit Integers

Haocheng Xi, ChangHao Li, Jianfei Chen, Jun Zhu

https://openreview.net/forum?id=H9hWlfMT6O

Keywords: neural network quantization, transformer, matrix multiplication, randomized numerical linear algebra

Compressor summary: The paper proposes a new method to train transformer models with 4-bit precision, which is faster and can work on current GPUs, by using dedicated quantizers for activations, weights, and gradients.


Failure-Aware Gaussian Process Optimization with Regret Bounds

Shogo Iwazaki, Shion Takeno, Tomohiko Tanabe, Mitsuru Irie

https://openreview.net/forum?id=H5pwAeYAun

Keywords: Gaussian process optimization, regret analysis, black-box optimization, Bayesian optimization

Compressor summary: F-GP-UCB is a black-box optimization method that handles observation failure with complex latent constraints and provides regret upper bounds and convergence guarantees.


Facilitating Graph Neural Networks with Random Walk on Simplicial Complexes

Cai Zhou, Xiyuan Wang, Muhan Zhang

https://openreview.net/forum?id=H57w5EOj6O

Keywords: random walk on simplicials, Hodge Laplacian, graph neural networks, edge-level positional encoding

Compressor summary: The paper studies how random walk on different orders of simplicial complexes improves Graph Neural Networks' theoretical expressivity and proposes new positional encoding methods based on Hodge Laplacians.


On the Overlooked Structure of Stochastic Gradients

Zeke Xie, Qian-Yuan Tang, Mingming Sun, Ping Li

https://openreview.net/forum?id=H4GsteoL0M

Keywords: Gradient Noise, SGD, Deep Learning

Compressor summary: This paper investigates the heavy-tail properties of stochastic gradients in deep neural networks and reveals new insights into their covariance structures.


Grassmann Manifold Flows for Stable Shape Generation

Ryoma Yataka, Kazuki Hirashima, Masashi Shiraishi

https://openreview.net/forum?id=H2udtfMbl4

Keywords: Generative Models, Geometric Deep Learning, Normalizing Flows, Shape Analysis, Grassmann Manifold

Compressor summary: This paper proposes a novel method to generate stable shapes using Grassmann manifolds and continuous normalization flows in machine learning, outperforming existing methods.


Tree-Based Diffusion Schrödinger Bridge with Applications to Wasserstein Barycenters

Maxence Noble, Valentin De Bortoli, Arnaud Doucet, Alain Durmus

https://openreview.net/forum?id=H2SuXHbFIn

Keywords: Schrödinger bridge, optimal transport, diffusion model, Wasserstein barycenter

Compressor summary: The paper introduces TreeDSB, an extension of DSB for solving entropic multi-marginal optimal transport problems with tree-structured quadratic costs, which can be used for computing Wasserstein barycenters and applying to high-dimensional tasks like image interpolation and Bayesian fusion.


Accelerated Training via Incrementally Growing Neural Networks using Variance Transfer and Learning Rate Adaptation

Xin Yuan, Pedro Henrique Pamplona Savarese, Michael Maire

https://openreview.net/forum?id=H1a7bVVnPK

Keywords: network growing, efficient network training

Compressor summary: The approach grows neural networks efficiently by designing parameterization and optimization strategies for the training dynamics, achieving comparable or better accuracy than fixed-size models with less computation budget and faster training speed.


CBD: A Certified Backdoor Detector Based on Local Dominant Probability

Zhen Xiang, Zidi Xiong, Bo Li

https://openreview.net/forum?id=H1CQZqpgdQ

Keywords: backdoor, Trojan, certification, adversarial learning, deep neural network, conformal prediction

Compressor summary: The paper introduces a novel certified backdoor detector (CBD) that uses conformal prediction and local dominant probability statistics to detect and guarantee the detection of various backdoor attacks on deep neural networks.


Training Fully Connected Neural Networks is $\exists\mathbb{R}$-Complete

Daniel Bertschinger, Christoph Hertrich, Paul Jungeblut, Tillmann Miltzow, Simon Weber

https://openreview.net/forum?id=H15KtcyHvn

Keywords: Neural Network Training, Computational Complexity, Existential Theory of the Reals, Algebraic Universality, Empirical Risk Minimization

Compressor summary: The paper shows that finding the optimal weights and biases for a two-layer neural network is as hard as solving real roots of multivariate polynomials with integer coefficients.


Functional Renyi Differential Privacy for Generative Modeling

Dihong Jiang, Sun Sun, Yaoliang Yu

https://openreview.net/forum?id=GzlDKZlwie

Keywords: Renyi differential privacy, RKHS, MMD, Gaussian process, generative model

Compressor summary: The paper extends differential privacy to functional outputs and applies it to a generative model, improving the privacy-utility trade-off.


Distributional Learning of Variational AutoEncoder: Application to Synthetic Data Generation

Seunghwan An, Jong-June Jeon

https://openreview.net/forum?id=GxL6PrmEUw

Keywords: Variational AutoEncoder, distributional learning, synthetic data generation, CRPS, asymmetric Laplace distribution

Compressor summary: The paper introduces a new VAE model with an infinite mixture of asymmetric Laplace distribution that can fit any continuous distribution and has better quantile estimation and privacy control properties.


Unified Segment-to-Segment Framework for Simultaneous Sequence Generation

Shaolei Zhang, Yang Feng

https://openreview.net/forum?id=GuErIOGLie

Keywords: Machine Translation, Speech Translation, Speech Recognition, Simultaneous Generation, Simultaneous Translation

Compressor summary: The paper introduces a framework called Seg2Seg that learns to generate target sequences adaptively and simultaneously from source sequences in different tasks, using latent segments as the bridge between them.


Robustifying Generalizable Implicit Shape Networks with a Tunable Non-Parametric Model

Amine Ouasfi, Adnane Boukhayma

https://openreview.net/forum?id=Gtse4R6iS4

Keywords: implicit neural representations, 3D reconstruction from unoriented point could, kernel ridge regression

Compressor summary: The authors propose a method to improve implicit shape reconstruction from point clouds by combining inter-shape and intra-shape priors, achieving better stability and efficiency.


Joint Data-Task Generation for Auxiliary Learning

Hong Chen, Xin Wang, Yuwei Zhou, Yijian Qin, Chaoyu Guan, Wenwu Zhu

https://openreview.net/forum?id=GtgFo5lmOB

Keywords: auxiliary learning, data-task joint generation

Compressor summary: The paper proposes a new method for auxiliary learning that generates data and tasks jointly to improve the primary task performance and avoid relying on domain knowledge for data collection.


Robust covariance estimation with missing values and cell-wise contamination

gregoire pacreau, Karim Lounici

https://openreview.net/forum?id=GtYlxtwO74

Keywords: robust statistics, missing values, cell-wise contamination

Compressor summary: The paper proposes a new estimator for covariance that works well with missing data and cell-wise outliers, without needing to remove or impute any data points.


REFINE: A Fine-Grained Medication Recommendation System Using Deep Learning and Personalized Drug Interaction Modeling

Suman Bhoi, Mong-Li Lee, Wynne Hsu, Ngiap Chuan Tan

https://openreview.net/forum?id=GsCTjmYe5v

Keywords: Fine-grained Medication recommendation, Drug Interaction Severity

Compressor summary: REFINE is a deep learning system that recommends fine-grained medications to patients with co-morbidities, considering their health conditions and drug interactions, and improves treatment outcomes and safety.


STXD: Structural and Temporal Cross-Modal Distillation for Multi-View 3D Object Detection

Sujin Jang, Dae Ung Jo, Sung Ju Hwang, Dongwook Lee, Daehyun Ji

https://openreview.net/forum?id=Grz2ijKrWI

Keywords: knowledge distillation, cross-modal learning, 3d object detection

Compressor summary: The paper proposes a novel framework called STXD to improve multi-view 3D object detection by transferring structural, temporal, and output knowledge from LiDAR-based teachers using cross-modal distillation.


Demographic Parity Constrained Minimax Optimal Regression under Linear Model

Kazuto Fukuchi, Jun Sakuma

https://openreview.net/forum?id=GrFsx4mBWF

Keywords: demographic parity, regression, minimax optimal

Compressor summary: We explore the minimax optimal error associated with a demographic parity-constrained regression problem within the context of a linear model. Our proposed model encompasses a broader range of discriminatory bias sources compared to the model presented by Chzhen and Schreuder. Our analysis reveals that the minimax optimal error for the demographic parity-constrained regression problem under our model is characterized by $\Theta(\frac{dM}{n})$, where $n$ denotes the sample size, $d$ represents the dimensionality, and $M$ signifies the number of demographic groups arising from sensitive attributes. Moreover, we demonstrate that the minimax error increases in conjunction with a larger bias present in the model.


Score-based Generative Modeling through Stochastic Evolution Equations in Hilbert Spaces

Sungbin Lim, Eunbi Yoon, Taehyun Byun, Taewon Kang, Seungwoo Kim, Kyungjae Lee, Sungjoon Choi

https://openreview.net/forum?id=GrElRvXnEj

Keywords: Generative Model, Score-based Method, Diffusion Model

Compressor summary: The paper introduces Hilbert Diffusion Model (HDM), which extends score-based generative models to function spaces by using stochastic evolution equations in Hilbert spaces and achieves better performance in sampling functions and synthesizing motion.


$\varepsilon$-fractional core stability in Hedonic Games.

Simone Fioravanti, Michele Flammini, Bojana Kodric, Giovanna Varricchio

https://openreview.net/forum?id=GqtpYUCwnu

Keywords: Game Theory, Hedonic Games, Core stability, Coalition Formation, Social Choice, PAC learning

Compressor summary: The text introduces a new concept called $\varepsilon$-fractional core-stability to address the challenges of finding stable coalition structures in hedonic games, and presents efficient algorithms for two classes of games.


Defending Pre-trained Language Models as Few-shot Learners against Backdoor Attacks

Zhaohan Xi, Tianyu Du, Changjiang Li, Ren Pang, Shouling Ji, Jinghui Chen, Fenglong Ma, Ting Wang

https://openreview.net/forum?id=GqXbfVmEPW

Keywords: few-shot learning, prompt learning, language model, backdoor defense

Compressor summary: The authors propose a defense method (MDP) for language models that are vulnerable to backdoor attacks in few-shot learning scenarios, which identifies poisoned samples by comparing their masking-sensitivity with clean samples.


Exploiting Contextual Objects and Relations for 3D Visual Grounding

Li Yang, Chunfeng Yuan, Ziqi Zhang, Zhongang Qi, Yan Xu, Wei Liu, Ying Shan, Bing Li, Weiping Yang, Peng Li, Yan Wang, Weiming Hu

https://openreview.net/forum?id=GlWzQhf2lV

Keywords: 3D Visual Grounding, Contextual Object, Contextual Relation

Compressor summary: The paper presents a new model, CORE-3DVG, that learns to identify objects in 3D scenes from natural language inputs by explicitly learning contextual information and achieves state-of-the-art performance on three datasets.


Multitask Learning with No Regret: from Improved Confidence Bounds to Active Learning

Pier Giuseppe Sessa, Pierre Laforgue, Nicolò Cesa-Bianchi, Andreas Krause

https://openreview.net/forum?id=GjJRbEZ1dc

Keywords: multitask learning, confidence intervals, online learning theory, active learning, regret

Compressor summary: The text discusses novel confidence intervals for multitask regression in the agnostic setting, regret guarantees, an online learning algorithm that adapts to task similarity, and a multitask active learning setup with no-regret algorithms.


Neural Latent Geometry Search: Product Manifold Inference via Gromov-Hausdorff-Informed Bayesian Optimization

Haitz Sáez de Ocáriz Borde, Alvaro Arroyo, Ismael Morales López, Ingmar Posner, Xiaowen Dong

https://openreview.net/forum?id=Gij638d76O

Keywords: Representation Learning, Product Manifolds, Bayesian Optimization, Gromov-Hausdorff Distance

Compressor summary: The paper proposes NLGS, a method to automatically find the best latent geometry for machine learning models using Bayesian optimization and metric geometry techniques.


Multi-Player Zero-Sum Markov Games with Networked Separable Interactions

Chanwoo Park, Kaiqing Zhang, Asuman E. Ozdaglar

https://openreview.net/forum?id=GiiOpKinGm

Keywords: Markov Games, Local Interaction, PPAD-Hardness, Fictitious Play

Compressor summary: The paragraph introduces a new class of Markov games called zero-sum NMGs, which model non-cooperative multi-agent sequential decision-making with networked separable interactions, and studies their equilibrium properties and learning dynamics.


Bi-Level Offline Policy Optimization with Limited Exploration

Wenzhuo Zhou

https://openreview.net/forum?id=GiUe0ZFiVe

Keywords: Offline Reinforcement Learning, Sample Efficiency, Regret Bound, Data Coverage

Compressor summary: The paper proposes a bi-level structured policy optimization algorithm for offline reinforcement learning that addresses the distributional shift problem by constructing a confidence set of value estimates and maximizing a conservative value estimate at the upper level.


Combining Behaviors with the Successor Features Keyboard

Wilka Carvalho, Andre Saraiva, Angelos Filos, Andrew Kyle Lampinen, Loic Matthey, Richard Lewis, Honglak Lee, Satinder Singh, Danilo Jimenez Rezende, Daniel Zoran

https://openreview.net/forum?id=GhNCFtLSsy

Keywords: deep reinforcement learning, successor features, transfer, generalization, feature-discovery

Compressor summary: The paper proposes a new method, Successor Features Keyboard (SFK), for transferring behavioral knowledge across tasks using discovered state-features and task encodings, and demonstrates its effectiveness in a 3D environment.


PreDiff: Precipitation Nowcasting with Latent Diffusion Models

Zhihan Gao, Xingjian Shi, Boran Han, Hao Wang, Xiaoyong Jin, Danielle C. Maddix, Yi Zhu, Mu Li, Bernie Wang

https://openreview.net/forum?id=Gh67ZZ6zkS

Keywords: Machine Learning for Earth Science, Spatiotemporal Forecasting, Generative Models, Diffusion Models

Compressor summary: The authors propose a two-stage pipeline for probabilistic spatiotemporal forecasting using a conditional latent diffusion model (PreDiff) and a knowledge alignment mechanism to handle uncertainty and incorporate physical constraints.


CoDrug: Conformal Drug Property Prediction with Density Estimation under Covariate Shift

Siddhartha Laghuvarapu, Zhen Lin, Jimeng Sun

https://openreview.net/forum?id=GgdFLb94Ld

Keywords: drug discovery, molecule property prediction, conformal prediction

Compressor summary: CoDrug is a method that uses an energy-based model and Kernel Density Estimation to create reliable uncertainty estimates for drug molecules, accounting for distribution shifts in drug discovery tasks.


Consistent Diffusion Models: Mitigating Sampling Drift by Learning to be Consistent

Giannis Daras, Yuval Dagan, Alex Dimakis, Constantinos Costis Daskalakis

https://openreview.net/forum?id=GfZGdJHj27

Keywords: diffusion models, sampling drift, Fokker-Planck, invariances, Stochastic Differential Equations, Martingales

Compressor summary: The paper proposes a method to improve the quality of generated images by enforcing a Consistency property that ensures predictions on generated data are consistent across time, and shows empirical results on various image datasets.


Human-Aligned Calibration for AI-Assisted Decision Making

Nina L. Corvelo Benz, Manuel Gomez Rodriguez

https://openreview.net/forum?id=GfITbjrIOd

Keywords: Calibration, Trustworthy Machine Learning, Human-Centric ML, Probabilistic Models and Methods

Compressor summary: The paper explores why decision makers struggle with confidence values provided by binary classifiers and proposes constructing more useful confidence values based on alignment with the decision maker's own confidence.


May the Force be with You: Unified Force-Centric Pre-Training for 3D Molecular Conformations

Rui Feng, Qi Zhu, Huan Tran, Binghong Chen, Aubrey Toland, Rampi Ramprasad, Chao Zhang

https://openreview.net/forum?id=Ge8Mhggq0z

Keywords: molecular pretraining, molecular representation learning

Compressor summary: The paper proposes a new force-centric pretraining model for 3D molecular conformations that covers both equilibrium and off-equilibrium data, improving forces accuracy, simulation performance, and inference speed.


Cal-QL: Calibrated Offline RL Pre-Training for Efficient Online Fine-Tuning

Mitsuhiko Nakamoto, Yuexiang Zhai, Anikait Singh, Max Sobol Mark, Yi Ma, Chelsea Finn, Aviral Kumar, Sergey Levine

https://openreview.net/forum?id=GcEIvidYSw

Keywords: offline reinforcement learning, online fine-tuning

Compressor summary: The paper proposes calibrated Q-learning, a method that initializes offline RL policies with conservative value functions, enabling fast online fine-tuning and improving performance on benchmark tasks.


Conditional Matrix Flows for Gaussian Graphical Models

Marcello Massimo Negri, Fabricio Arend Torres, Volker Roth

https://openreview.net/forum?id=GYnbubCXhE

Keywords: normalizing flow, variational inference, graphical lasso, gaussian graphical model, bayesian inference

Compressor summary: The paper presents a new method to study conditional independence among many variables using Gaussian Graphical Models that unifies frequentist and Bayesian approaches and handles different sparsity-inducing norms.


FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and Editing

Mingyuan Zhang, Huirong Li, Zhongang Cai, Jiawei Ren, Lei Yang, Ziwei Liu

https://openreview.net/forum?id=GYjV1M5s0D

Keywords: Motion Generation, Diffusion Model

Compressor summary: FineMoGen is a diffusion-based framework that generates and edits fine-grained motions using a novel spatio-temporal attention mechanism and a large dataset, enabling zero-shot motion editing with language models.


Structured Neural-PI Control with End-to-End Stability and Output Tracking Guarantees

Wenqi Cui, Yan Jiang, Baosen Zhang, Yuanyuan Shi

https://openreview.net/forum?id=GWIRpKF6yU

Keywords: Control, Stability, Tracking, Passivity, Neural network-based controllers, Power systems

Compressor summary: The paper proposes neural PI controllers with provable stability and output tracking guarantees for multiple-input and multiple-output systems using equilibrium-independent passivity and strictly convex neural networks.


ChatGPT-Powered Hierarchical Comparisons for Image Classification

Zhiyuan Ren, Yiyang Su, Xiaoming Liu

https://openreview.net/forum?id=GTYaYNsFyv

Keywords: ChatGPT, Hierarchical Comparisons, Image Classification, Zero shot

Compressor summary: The authors propose a novel image classification framework using hierarchical comparisons of text embeddings to overcome CLIP's limitations and achieve better accuracy.


4D Panoptic Scene Graph Generation

Jingkang Yang, Jun CEN, Wenxuan Peng, Shuai Liu, Fangzhou Hong, Xiangtai Li, Kaiyang Zhou, Qifeng Chen, Ziwei Liu

https://openreview.net/forum?id=GRHZiTbDDI

Keywords: Scene Graph Generation, 4D Understanding, 4D Perception.

Compressor summary: The paragraph introduces PSG-4D, a new 4D visual representation for AI, and presents a dataset, model, and application example for it.


Adversarial Training for Graph Neural Networks: Pitfalls, Solutions, and New Directions

Lukas Gosch, Simon Geisler, Daniel Sturm, Bertrand Charpentier, Daniel Zügner, Stephan Günnemann

https://openreview.net/forum?id=GPtroppvUM

Keywords: adversarial training, adversarial examples, robust graph learning, graph machine learning, graph neural networks, graphs

Compressor summary: The authors propose and evaluate a flexible graph neural network model with learnable graph diffusion that can defend against adversarial perturbations in the graph structure, while also introducing a new attack method for such perturbations.


D4Explainer: In-distribution Explanations of Graph Neural Network via Discrete Denoising Diffusion

Jialin Chen, Shirley Wu, Abhijit Gupta, Zhitao Ying

https://openreview.net/forum?id=GJtP1ZEzua

Keywords: Explainability, Graph Neural Network, Diffusion Model

Compressor summary: D4Explainer is a novel approach for generating reliable and diverse explanations for Graph Neural Networks by learning graph distributions that conform to the in-distribution property.


Estimating Koopman operators with sketching to provably learn large scale dynamical systems

Giacomo Meanti, Antoine Chatalic, Vladimir R Kostic, Pietro Novelli, massimiliano pontil, Lorenzo Rosasco

https://openreview.net/forum?id=GItLpB1vhK

Keywords: dynamical systems, kernel methods, koopman operator, sketching, molecular dynamics, efficient machine learning

Compressor summary: The paper presents new methods to learn complex dynamics from data using Koopman operators and kernel methods, making them more efficient with random projections.


Two-Stage Learning to Defer with Multiple Experts

Anqi Mao, Christopher Mohri, Mehryar Mohri, Yutao Zhong

https://openreview.net/forum?id=GIlsH0T4b2

Keywords: learning to defer, learning theory

Compressor summary: The authors propose a new method to learn from multiple experts in two stages, using surrogate loss functions that ensure consistency and perform well in experiments.


Machine learning detects terminal singularities

Tom Coates, Alexander M. Kasprzyk, Sara Veneziale

https://openreview.net/forum?id=GI4Pp01prW

Keywords: mathematics, geometry, Fano varieties, terminal singularities, theorem discovery, neural network classifier, supervised learning

Compressor summary: The paper uses machine learning to classify eight-dimensional positively curved algebraic varieties with toric symmetry and Picard rank two, revealing the landscape of Q-Fano varieties and providing new evidence for machine learning's role in mathematical discovery.


High dimensional, tabular deep learning with an auxiliary knowledge graph

Camilo Ruiz, Hongyu Ren, Kexin Huang, Jure Leskovec

https://openreview.net/forum?id=GGylthmehy

Keywords: Tabular Data, Deep Learning, Knowledge Graph, Regularization

Compressor summary: PLATO is a method that improves machine learning performance on tabular data with high dimensions and low samples by using a heterogeneous knowledge graph as an auxiliary source of information to regularize a multilayer perceptron.


Domain Adaptive Imitation Learning with Visual Observation

Sungho Choi, Seungyul Han, Woojun Kim, Jongseong Chae, Whiyoung Jung, Youngchul Sung

https://openreview.net/forum?id=GGbBXSkX3r

Keywords: Reinforcement Learning, Deep Reinforcement Learning, Imitation Learning

Compressor summary: The paper presents a new method for teaching a robot to perform a task by observing other robots, even if they look different or are viewed from different angles.


CORNN: Convex optimization of recurrent neural networks for rapid inference of neural dynamics

Fatih Dinc, Adam Shai, Mark Schnitzer, Hidenori Tanaka

https://openreview.net/forum?id=GGIA1p9fDT

Keywords: brain-machine interfaces, recurrent neural networks, convex optimization, computational neuroscience

Compressor summary: CORNN is a new training method for recurrent neural networks that enables faster and more accurate modeling of large neural datasets, potentially enabling real-time analysis and control of animal behavior.


iSCAN: Identifying Causal Mechanism Shifts among Nonlinear Additive Noise Models

Tianyu Chen, Kevin Bello, Bryon Aragam, Pradeep Kumar Ravikumar

https://openreview.net/forum?id=GEtXhqKW6X

Keywords: distribution shifts, heterogeneous data, feature-shift, structural causal models, additive noise models, causality

Compressor summary: This paper proposes a method for identifying causal mechanism shifts in related datasets without estimating their full causal structure, assuming nonlinear additive noise models.


Transition-constant Normalization for Image Enhancement

Jie Huang, Man Zhou, JingHao Zhang, Gang Yang, Mingde Yao, Chongyi Li, Zhiwei Xiong, Feng Zhao

https://openreview.net/forum?id=GEWzHeHpLr

Keywords: Image Enhancement, Normalization, Image Restoration

Compressor summary: The authors propose a novel normalization technique, Transition-Constant Normalization (TCN), for various image enhancement tasks that has several advantages and shows performance improvements in multiple experiments.


$L_2$-Uniform Stability of Randomized Learning Algorithms: Sharper Generalization Bounds and Confidence Boosting

Xiaotong Yuan, Ping Li

https://openreview.net/forum?id=GEQZ52oqxa

Keywords: Uniform stability, Randomized learning algorithms, Confidence boosting, Generalization bounds, Stochastic gradient methods

Compressor summary: The paper introduces a new concept of uniform stability and proves strong exponential bounds on generalization error, leading to near-optimal results for randomized learning algorithms like SGD.


Smooth Flipping Probability for Differential Private Sign Random Projection Methods

Ping Li, Xiaoyun Li

https://openreview.net/forum?id=GEMHw2sd9S

Keywords: Differential Privacy, Random Projection

Compressor summary: The text introduces new differential privacy algorithms using random projection and sign random projection methods, improving performance and utility for data protection in machine learning and search applications.


Facing Off World Model Backbones: RNNs, Transformers, and S4

Fei Deng, Junyeong Park, Sungjin Ahn

https://openreview.net/forum?id=GDYuzX0rwj

Keywords: world models, structured state space sequence models, S4, long-term memory, model-based reinforcement learning

Compressor summary: This paper explores alternative world model backbones for improving long-term memory in model-based reinforcement learning, comparing Transformers, Structured State Space Sequence models, and recurrent neural networks across various tasks.


TaskMet: Task-driven Metric Learning for Model Learning

Dishank Bansal, Ricky T. Q. Chen, Mustafa Mukadam, Brandon Amos

https://openreview.net/forum?id=GCY9C43A4L

Keywords: task-based learning, decision-focused learning

Compressor summary: The proposed method improves deep learning models' performance on downstream tasks by using a task loss to learn a metric that emphasizes important information for the desired task without altering the original prediction model.


KAKURENBO: Adaptively Hiding Samples in Deep Neural Network Training

Thao Nguyen Truong, Balazs Gerofi, Edgar Josafat Martinez-Noriega, François Trahay, Mohamed Wahib

https://openreview.net/forum?id=GAsRl2ElHk

Keywords: Deep learning, Sampling

Compressor summary: The paper presents a method to hide less important samples in deep neural network training to save time without much accuracy loss.


Guide Your Agent with Adaptive Multimodal Rewards

Changyeon Kim, Younggyo Seo, Hao Liu, Lisa Lee, Jinwoo Shin, Honglak Lee, Kimin Lee

https://openreview.net/forum?id=G8nal7MpIQ

Keywords: Reinforcement Learning, Multimodal Representation, Imitation Learning

Compressor summary: The paper proposes ARP, an imitation learning framework that uses multimodal rewards based on natural language descriptions and visual observations to enhance generalization and adapt to unseen environments.


On the Pareto Front of Multilingual Neural Machine Translation

Liang Chen, Shuming Ma, Dongdong Zhang, Furu Wei, Baobao Chang

https://openreview.net/forum?id=G7sQlfTzmY

Keywords: Multilinugal Neural Machine Translation, Multitask Learning, Pareto Optimization

Compressor summary: The paper investigates how different translation directions perform in multilingual neural machine translation, proposes a method to predict and optimize their performance trade-offs, and shows that it improves over existing methods.


CSLP-AE: A Contrastive Split-Latent Permutation Autoencoder Framework for Zero-Shot Electroencephalography Signal Conversion

Anders Vestergaard Nørskov, Alexander Neergaard Zahid, Morten Mørup

https://openreview.net/forum?id=G7Y145tm2F

Keywords: zero-shot conversion, representations learning, contrastive learning, electroencephalography, autoencoder, subject variability, permutation invariant training

Compressor summary: The authors propose a new method (CSLP-AE) to convert EEG signals between tasks and subjects by extracting latent representations that account for both content and style, inspired by voice conversion technologies.


Pick-a-Pic: An Open Dataset of User Preferences for Text-to-Image Generation

Yuval Kirstain, Adam Polyak, Uriel Singer, Shahbuland Matiana, Joe Penna, Omer Levy

https://openreview.net/forum?id=G5RwHpBUv0

Keywords: text-to-image, human-preferences, dataset

Compressor summary: The authors create a web app to collect user preferences for text-to-image generation and use it to build a large dataset (Pick-a-Pic) and a scoring function (PickScore) that predicts human preferences better than existing metrics and can improve existing models.


Implicit Bias of Gradient Descent for Two-layer ReLU and Leaky ReLU Networks on Nearly-orthogonal Data

Yiwen Kou, Zixiang Chen, Quanquan Gu

https://openreview.net/forum?id=G560qr59Gi

Keywords: ReLU Neural Networks, Implicit Bias, Deep Learning Theory

Compressor summary: The paper investigates how gradient descent impacts the implicit bias of non-smooth neural networks, especially for leaky ReLU activation functions, and shows that it leads to stable rank and uniform normalized margin in some cases.


On the Sublinear Regret of GP-UCB

Justin Whitehouse, Aaditya Ramdas, Steven Wu

https://openreview.net/forum?id=G3aubF5Wnw

Keywords: Kernel Bandits, Online Learning, Self-Normalized Concentration, Online Regression

Compressor summary: The paper presents a new analysis of the popular Gaussian Process Upper Confidence Bound (GP-UCB) algorithm that shows it achieves nearly optimal regret and improves over previous results for the Matern kernel.


Evolutionary Neural Architecture Search for Transformer in Knowledge Tracing

Shangshang Yang, Xiaoshan Yu, Ye Tian, Xueming Yan, Haiping Ma, Xingyi Zhang

https://openreview.net/forum?id=G14N38AjpU

Keywords: Knowledge tracing, intelligent education, neural architecture search, Transformer

Compressor summary: The paper proposes a new approach for knowledge tracing that uses convolution operations in Transformers, automates input feature selection, and applies evolutionary neural architecture search to balance local and global context modelling.


A Dual-Stream Neural Network Explains the Functional Segregation of Dorsal and Ventral Visual Pathways in Human Brains

Minkyu Choi, Kuan Han, Xiaokai Wang, Yizhen Zhang, Zhongming Liu

https://openreview.net/forum?id=Fy1S3v4UAk

Keywords: brain-inspired AI, retina transformation, eye movements, deep neural networks

Compressor summary: The authors developed a dual-stream vision model inspired by human vision, which uses two CNN branches for spatial processing and object recognition, and found that it matches the dorsal and ventral pathways of the visual cortex in terms of functional alignment.


Neural Image Compression: Generalization, Robustness, and Spectral Biases

Kelsey Lieberman, James Diffenderfer, Charles Godfrey, Bhavya Kailkhura

https://openreview.net/forum?id=FxRfAIj4s2

Keywords: image compression, robustness, generalization

Compressor summary: The paper introduces a benchmark suite, inspection tools, and theoretical analysis to evaluate the out-of-distribution performance of neural image compression methods for real-world applications.


Data Quality in Imitation Learning

Suneel Belkhale, Yuchen Cui, Dorsa Sadigh

https://openreview.net/forum?id=FwmvbuDiMk

Keywords: Imitation Learning, Robotics, Data Quality

Compressor summary: The paper proposes two properties for defining high quality datasets in imitation learning: action divergence and transition diversity, which help the learned policy to stay in distribution at test time.


On Differentially Private Sampling from Gaussian and Product Distributions

Badih Ghazi, Xiao Hu, Ravi Kumar, Pasin Manurangsi

https://openreview.net/forum?id=FviF8vuz5B

Keywords: privacy, sampling, Gaussian distribution, product distributions

Compressor summary: The paper proposes new differentially private algorithms for generating samples from multi-dimensional Gaussian distributions with different covariance assumptions and improves the performance in some settings.


Balancing memorization and generalization in RNNs for high performance brain-machine Interfaces

Joseph T Costello, Hisham Temmar, Luis H Cubillos, Matthew J Mender, Dylan M Wallace, Matthew S Willsey, Parag G Patil, Cynthia Chestek

https://openreview.net/forum?id=FujJO3dsNj

Keywords: brain computer interface, brain machine interface, neural decoding, prosthetic control, recurrent neural network, RNN, transformer, real time, closed-loop, user interface

Compressor summary: The study evaluated recurrent neural networks (RNNs) for real-time brain-machine interface control, showing improved performance over other architectures in decoding finger movements from primate neural signals.


Dynamic Pricing and Learning with Bayesian Persuasion

Shipra Agrawal, Yiding Feng, Wei Tang

https://openreview.net/forum?id=FtZ7lUwH99

Keywords: dynamic pricing, information design, regret minimization

Compressor summary: The paper proposes an algorithm that learns optimal pricing and advertising strategies for a seller in a dynamic setting with linear buyer valuations.


Exact Bayesian Inference on Discrete Models via Probability Generating Functions: A Probabilistic Programming Approach

Fabian Zaiser, Andrzej S Murawski, Luke Ong

https://openreview.net/forum?id=FtNruwFEs3

Keywords: Bayesian statistics, probabliistic programming, exact inference, discrete models, probability generating functions

Compressor summary: The authors present an exact Bayesian inference method for discrete models using a probabilistic programming language and probability generating functions, which they implement in a tool called Genfer that outperforms existing tools on many problems.


RiskQ: Risk-sensitive Multi-Agent Reinforcement Learning Value Factorization

Siqi Shen, Chennan Ma, Chao Li, Weiquan Liu, Yongquan Fu, Songzhu Mei, Xinwang Liu, Cheng Wang

https://openreview.net/forum?id=FskZtRvMJI

Keywords: multi-agent reinforcement learning, value factorization, individual global max, risk-sensitive

Compressor summary: RIGM principle introduces a new approach to coordinate risk-sensitive MARL policies, and RiskQ is a value factorization method that satisfies the RIGM principle for various risk metrics.


Differentiable Blocks World: Qualitative 3D Decomposition by Rendering Primitives

Tom Monnier, Jake Austin, Angjoo Kanazawa, Alexei A Efros, Mathieu Aubry

https://openreview.net/forum?id=FsQWxU5TOL

Keywords: 3D decomposition, 3D reconstruction, MVS, primitives, qualitative 3D

Compressor summary: The authors present a method to generate simple 3D world representations from images using textured superquadric meshes, which model transparency and handle varying numbers of primitives. The method operates on images through differentiable rendering and surpasses existing approaches on various scenes.


Faster approximate subgraph counts with privacy

Dung Nguyen, Mahantesh M Halappanavar, Venkatesh Srinivasan, Anil Vullikanti

https://openreview.net/forum?id=Fqg9vGWy4k

Keywords: differential privacy, subgraph counting, smooth sensitivity, local sensitivity

Compressor summary: The paper proposes new algorithms for privately counting subgraphs in graph data, using approximated sensitivity metrics and reducing noise.


3D-IntPhys: Towards More Generalized 3D-grounded Visual Intuitive Physics under Challenging Scenes

Haotian Xue, Antonio Torralba, Joshua B. Tenenbaum, Daniel LK Yamins, Yunzhu Li, Hsiao-Yu Tung

https://openreview.net/forum?id=Fp5uC6YHwe

Keywords: Intuitive Physics, Computer Vision

Compressor summary: The paper presents a framework to learn 3D visual intuitive physics models from videos of complex scenes with fluids using a conditional NeRF-style visual frontend and a point-based dynamics prediction backend, which can handle challenging scenarios without relying on dense point trajectory supervision.


NU-MCC: Multiview Compressive Coding with Neighborhood Decoder and Repulsive UDF

Stefan Lionar, Xiangyu Xu, Min Lin, Gim Hee Lee

https://openreview.net/forum?id=FmpH0CYWiX

Keywords: single-view 3d reconstruction, neural fields, 3d reconstruction

Compressor summary: NU-MCC is a new approach for single-view 3D reconstruction that improves efficiency and detail recovery by using a neighborhood decoder and a repulsive unsigned distance function, achieving state-of-the-art results on the CO3D-v2 dataset.


Robust Multi-Agent Reinforcement Learning via Adversarial Regularization: Theoretical Foundation and Stable Algorithms

Alexander Bukharin, Yan Li, Yue Yu, Qingru Zhang, Zhehui Chen, Simiao Zuo, Chao Zhang, Songan Zhang, Tuo Zhao

https://openreview.net/forum?id=FmZVRe0gn8

Keywords: Multi-Agent Reinforcement Learning, Theory of Robust Reinforcement Learning, Adversarial Regularization

Compressor summary: The paper introduces ERNIE, a robust multi-agent reinforcement learning framework that uses adversarial regularization to control policy's Lipschitz constant and improve stability against environmental changes.


Entropy-dissipation Informed Neural Network for McKean-Vlasov Type PDEs

Zebang Shen, Zhenfu Wang

https://openreview.net/forum?id=FkpMm9avyP

Keywords: Entropy-dissipation, McKean-Vlasov, Navier-Stokes, PDE, Coulomb, singular interaction

Compressor summary: The paper proposes a new neural network-based method to solve McKean-Vlasov equations involving singular interactions, using entropy dissipation as a control criterion.


Faith and Fate: Limits of Transformers on Compositionality

Nouha Dziri, Ximing Lu, Melanie Sclar, Xiang Lorraine Li, Liwei Jiang, Bill Yuchen Lin, Sean Welleck, Peter West, Chandra Bhagavatula, Ronan Le Bras, Jena D. Hwang, Soumya Sanyal, Xiang Ren, Allyson Ettinger, Zaid Harchaoui, Yejin Choi

https://openreview.net/forum?id=Fkckkr3ya8

Keywords: Natural language processing, large language models, multi-step reasoning

Compressor summary: The authors examine transformer LLMs' ability to perform complex tasks requiring multi-step reasoning and find that these models rely on subgraph matching instead of developing problem-solving skills, which may limit their performance in more challenging tasks.


Direct Preference-based Policy Optimization without Reward Modeling

Gaon An, Junhyeok Lee, Xingdong Zuo, Norio Kosaka, Kyung-Min Kim, Hyun Oh Song

https://openreview.net/forum?id=FkAwlqBuyO

Keywords: Preference-based reinforcement learning, Contrastive learning, Offline reinforcement learning, RLHF

Compressor summary: The paper proposes a preference-based reinforcement learning (PbRL) method that uses contrastive learning to directly learn from human preferences without requiring a reward model.


A Unified Approach to Domain Incremental Learning with Memory: Theory and Algorithm

Haizhou Shi, Hao Wang

https://openreview.net/forum?id=FiClXlUqA7

Keywords: Domain Incremental Learning, Continual Learning, Theory

Compressor summary: The paper proposes a unified framework for domain incremental learning that adapts to different domains with memory, improves generalization error bounds, and outperforms existing methods on various datasets.


A Metadata-Driven Approach to Understand Graph Neural Networks

Ting Wei Li, Qiaozhu Mei, Jiaqi Ma

https://openreview.net/forum?id=FgakGFpll1

Keywords: Graph Neural Networks, Metadata-Driven Analysis, Gini Coefficient of Degree Distribution

Compressor summary: This paper proposes a metadata-driven approach to analyze how graph data properties affect the performance of Graph Neural Networks (GNNs) using regression analysis, theoretical analysis, and experiments.


Gradient-Based Feature Learning under Structured Data

Alireza Mousavi-Hosseini, Denny Wu, Taiji Suzuki, Murat A Erdogdu

https://openreview.net/forum?id=Fe8PxP2F2p

Keywords: feature learning, neural networks, single-index model, gradient descent

Compressor summary: The paper studies how adding structure to the input data affects gradient-based learning of single index models and shows that appropriate normalization and exploiting the alignment between the input covariance and the target can improve performance.


On the Trade-off of Intra-/Inter-class Diversity for Supervised Pre-training

Jieyu Zhang, Bohan Wang, Zhengyu Hu, Pang Wei Koh, Alexander Ratner

https://openreview.net/forum?id=Fe6fDq65aZ

Keywords: data-centric study, supervised pretraining, transfer learning

Compressor summary: The paper studies how the balance between intra-class and inter-class diversity in pre-training datasets affects downstream performance, and shows that the optimal class-to-sample ratio is independent of the pre-training dataset size.


Constraint-Conditioned Policy Optimization for Versatile Safe Reinforcement Learning

Yihang Yao, Zuxin Liu, Zhepeng Cen, Jiacheng Zhu, Wenhao Yu, Tingnan Zhang, Ding Zhao

https://openreview.net/forum?id=FdtdjQpAwJ

Keywords: Safe Reinforcement Learning, Conditioned Reinforcement Learning, Multi-task Reinforcement Learning

Compressor summary: The paper proposes a new framework, CCPO, that learns safe policies in reinforcement learning that can adapt to changing safety constraints without retraining and achieve good performance.


PaintSeg: Painting Pixels for Training-free Segmentation

Xiang Li, Chung-Ching Lin, Yinpeng Chen, Zicheng Liu, Jinglu Wang, Rita Singh, Bhiksha Raj

https://openreview.net/forum?id=FdsS51iif3

Keywords: Prompt-guided Segmentation, Generative models, Training-free

Compressor summary: PaintSeg is an unsupervised method that uses adversarial masked contrastive painting to create contrast between original and painted images, advancing the target segmentation mask towards ground truth without supervision or training.


Mnemosyne: Learning to Train Transformers with Transformers

Deepali Jain, Krzysztof Marcin Choromanski, Kumar Avinava Dubey, Sumeet Singh, Vikas Sindhwani, Tingnan Zhang, Jie Tan

https://openreview.net/forum?id=Fdfyga5i0A

Keywords: learnable optimizers, Transformers, efficient attention, spatio-temporal attention

Compressor summary: Mnemosyne is a new learnable optimizer that can train neural networks without task-specific tuning, performs well compared to popular LSTM optimizers, and has space complexity comparable to hand-designed optimizers.


Towards In-context Scene Understanding

Ivana Balazevic, David Steiner, Nikhil Parthasarathy, Relja Arandjelovic, Olivier J Henaff

https://openreview.net/forum?id=FasIQqsJhe

Keywords: transfer learning, adaptation, self-supervised learning, contrastive learning, scene understanding, representation learning, in-context learning, vision transformers

Compressor summary: The paragraph discusses a new approach called Hummingbird, which enables in-context learning for computer vision tasks such as semantic segmentation and depth estimation using nearest neighbor retrieval and attention-based pretraining.


StyleGAN knows Normal, Depth, Albedo, and More

Anand Bhattad, Daniel McKee, Derek Hoiem, David Forsyth

https://openreview.net/forum?id=FYqqvQdXhZ

Keywords: Generative models, StyleGAN, Depth, Normals, Segmentation, Intrinsic Images, Albedo, Shading

Compressor summary: This paper shows how StyleGAN can generate different types of intrinsic images easily and demonstrates its advantages over existing methods in terms of quality, robustness, and efficiency.


Episodic Multi-Task Learning with Heterogeneous Neural Processes

Jiayi Shen, Xiantong Zhen, Cheems Wang, Marcel Worring

https://openreview.net/forum?id=FXU4aR2uif

Keywords: data-insufficiency problem, episodic training, multi-task learning and neural processes

Compressor summary: The paper introduces HNPs, a method that uses meta-knowledge, heterogeneous information, and task-relatedness to adapt to novel tasks in episodic multi-task learning with limited data, and shows its superior performance over baselines.


Non-stationary Experimental Design under Linear Trends

David Simchi-Levi, Chonghuan Wang, Zeyu Zheng

https://openreview.net/forum?id=FV4ngfUlY0

Keywords: Adaptive Experimental Design, Non-stationary, Online Learning, Treatment Effect

Compressor summary: The paper proposes an efficient design for non-stationary experiments with linear trends, aiming to estimate the dynamic treatment effect and minimize welfare loss, while highlighting the challenge and trade-off between these objectives.


Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Minsoo Kim, Sihwa Lee, Janghwan Lee, Sukjin Hong, Du-Seong Chang, Wonyong Sung, Jungwook Choi

https://openreview.net/forum?id=FUnEkOkodU

Keywords: Generative Language Model, Quantization, QAT, Knowledge Distillation, Causal Attention, Language Modeling

Compressor summary: The paper proposes a new method for training large generative language models with less memory and maintaining accuracy in various tasks.


Markovian Sliced Wasserstein Distances: Beyond Independent Projections

Khai Nguyen, Tongzheng Ren, Nhat Ho

https://openreview.net/forum?id=FT2q2B4cKZ

Keywords: Sliced Wasserstein, Generative Models, Optimal Transport

Compressor summary: The Markovian sliced Wasserstein (MSW) distance is a new family of sliced Wasserstein distances that imposes a first-order Markov structure on projecting directions, improving metricity, computational efficiency, and effectiveness over previous methods.


Streaming PCA for Markovian Data

Syamantak Kumar, Purnamrita Sarkar

https://openreview.net/forum?id=FQGRkwmRzm

Keywords: Streaming PCA, Markov Chain, Mixing, Oja's algorithm

Compressor summary: The paper proposes a new method to estimate the top eigenvector of a covariance matrix using Oja's algorithm on data streams sampled from a Markov chain without downsampling.


BCDiff: Bidirectional Consistent Diffusion for Instantaneous Trajectory Prediction

Rongqing Li, Changsheng Li, Dongchun Ren, Guangyi Chen, Ye Yuan, Guoren Wang

https://openreview.net/forum?id=FOFJmR1oxt

Keywords: Trajectory prediction, instantaneous observation

Compressor summary: BCDiff is a novel framework for predicting pedestrian trajectories from instantaneous observations, using two coupled diffusion models that mutually guide each other to generate more accurate predictions.


Hyperbolic VAE via Latent Gaussian Distributions

Seunghyuk Cho, Juyong Lee, Dongwoo Kim

https://openreview.net/forum?id=FNn4zibGvw

Keywords: Hyperbolic space, VAE, Distribution on hyperbolic space, Hierarchical representation learning, Reinforcement Learning

Compressor summary: The GM-VAE is a new model that uses Gaussian distributions to represent data in a latent space shaped like a hyperbolic space, which improves density estimation and reinforcement learning tasks.


Core-sets for Fair and Diverse Data Summarization

Sepideh Mahabadi, Stojan Trajanovski

https://openreview.net/forum?id=FM8thAWqiO

Keywords: Constrained Diversity Maximization, Fairness, Data Summarization, Core-sets, Approximation Algorithms

Compressor summary: The paragraph discusses algorithms that maximize diversity in partitioned sets under fairness constraints and shows their effectiveness in summarizing timed messages for a communication platform.


FedL2P: Federated Learning to Personalize

Royson Lee, Minyoung Kim, Da Li, Xinchi Qiu, Timothy Hospedales, Ferenc Huszár, Nicholas Donald Lane

https://openreview.net/forum?id=FM81CI68Iz

Keywords: federated learning; meta-learning; hyperparameter optimization

Compressor summary: The paper proposes a federated meta-learning approach to learn personalization strategies for different clients in federated learning problems, using meta-nets to adjust batch-norm and learning rate parameters based on local data statistics.


Scaling Riemannian Diffusion Models

Aaron Lou, Minkai Xu, Adam Farris, Stefano Ermon

https://openreview.net/forum?id=FLTg8uA5xI

Keywords: Diffusion Models, Geometric Deep Learning, Manifolds, Numerical Algorithms

Compressor summary: The authors improve Riemannian diffusion models by leveraging symmetry and computationally efficient approximations, enabling their application to high dimensional tasks and reducing representation collapse.


Meta-learning families of plasticity rules in recurrent spiking networks using simulation-based inference

Basile Confavreux, Poornima Ramesh, Pedro J. Goncalves, Jakob H. Macke, Tim P. Vogels

https://openreview.net/forum?id=FLFasCFJNo

Keywords: synaptic plasticity, spiking network, meta-learning, computational neuroscience

Compressor summary: The authors develop a simulation-based inference method (filter SBI) that allows them to infer complex and co-active plasticity rules in spiking networks, which can be used for deeper insights into brain function.


Pseudo-Likelihood Inference

Theo Gruner, Boris Belousov, Fabio Muratore, Daniel Palenicek, Jan Peters

https://openreview.net/forum?id=FIv84qGPFT

Keywords: simulation-based inference, approximate Bayesian computation

Compressor summary: Pseudo-Likelihood Inference (PLI) is a new method that combines neural approximation with Approximate Bayesian Computation (ABC) to perform better on high-dimensional Bayesian system identification tasks, especially when more data is available.


On the spectral bias of two-layer linear networks

Aditya Vardhan Varre, Maria-Luiza Vladarean, Loucas Pillaud-Vivien, Nicolas Flammarion

https://openreview.net/forum?id=FFdrXkm3Cz

Keywords: linear networks, spectral bias, low rank, singular values, mirror flow

Compressor summary: The paper analyzes how two-layer neural networks with linear activations and gradient flow training behave differently depending on initialization scale, revealing that low-rank structures arise in small scale regimes.


C-Disentanglement: Discovering Causally-Independent Generative Factors under an Inductive Bias of Confounder

Xiaoyu Liu, Jiaxin Yuan, Bang An, Yuancheng Xu, Yifan Yang, Furong Huang

https://openreview.net/forum?id=FFOYWUpBca

Keywords: causal disentanglement, causal generative process, generative factors, confounder, inductive bias, disentanglement, causal inference

Compressor summary: The paper proposes C-Disentanglement, a framework that incorporates expert knowledge to identify causally disentangled factors in latent space, improving data generation and generalization.


Squared Neural Families: A New Class of Tractable Density Models

Russell Tsuchida, Cheng Soon Ong, Dino Sejdinovic

https://openreview.net/forum?id=FDzQQTPqEJ

Keywords: probabilistic modelling; density estimation; exponential family;

Compressor summary: SNEFYs are a new type of probability distributions that can model complex data with flexible and tractable density models, and are applicable to many machine learning problems.


Anytime-Competitive Reinforcement Learning with Policy Prior

Jianyi Yang, Pengfei Li, Tongxin Li, Adam Wierman, Shaolei Ren

https://openreview.net/forum?id=FCwfZj1bQl

Keywords: Markov Decision Process, Constrained Reinforcement Learning, Anytime Competitive Constraints

Compressor summary: The paper introduces ACRL, a new algorithm for A-CMDP that guarantees cost bounds and optimizes expected reward, with experiments showing its effectiveness in carbon-intelligent computing.


Directed Cyclic Graph for Causal Discovery from Multivariate Functional Data

Saptarshi Roy, Raymond K. W. Wong, Yang Ni

https://openreview.net/forum?id=FCwF5431IY

Keywords: Causal Embedding, Causal Discovery, Multivariate Functional Data, Directed Cyclic Graph, Causal Structure Learning, Bayesian Inference

Compressor summary: The article presents a functional linear structural equation model for learning causal structures from multivariate functional data, which preserves causal information in a low-dimensional space and can be inferred using a Bayesian framework.


Cognitive Steering in Deep Neural Networks via Long-Range Modulatory Feedback Connections

Talia Konkle, George A. Alvarez

https://openreview.net/forum?id=FCIj5KMn2m

Keywords: convolutional neural networks, steerability, computer vision

Compressor summary: The paragraph discusses how adding long-range modulatory pathways to vision models can improve image recognition, robustness, and alignment, as well as allow for goal-directed visual encoding using output space vectors.


Variational Monte Carlo on a Budget — Fine-tuning pre-trained Neural Wavefunctions

Michael Scherbela, Leon Gerard, Philipp Grohs

https://openreview.net/forum?id=FBNyccPfAu

Keywords: Computational Physics, Machine Learning for Science, Quantum Monte Carlo, Fermionic Neural Networks

Compressor summary: The authors propose a deep learning-based method for quantum chemistry that uses pre-training and fine-tuning to achieve high accuracy with less computation than conventional methods.


A Privacy-Friendly Approach to Data Valuation

Jiachen T. Wang, Yuqing Zhu, Yu-Xiang Wang, Ruoxi Jia, Prateek Mittal

https://openreview.net/forum?id=FAZ3i0hvm0

Keywords: Data Valuation, Differential Privacy

Compressor summary: The paper introduces TKNN-Shapley, a privacy-friendly variant of KNN-Shapley, which is a method for data valuation in machine learning models, and shows that it offers better trade-offs between privacy and utility than the original method.


Debiased and Denoised Entity Recognition from Distant Supervision

Haobo Wang, Yiwen Dong, Ruixuan Xiao, Fei Huang, Gang Chen, Junbo Zhao

https://openreview.net/forum?id=FAGY52HbyV

Keywords: Distant Supervision; Named Entity-Recognition; Biased Learning

Compressor summary: DesERT is a novel self-training framework that addresses two types of biases in distant supervision for named entity recognition and achieves state-of-the-art performance on five benchmark datasets.


Toward Better PAC-Bayes Bounds for Uniformly Stable Algorithms

Sijia Zhou, Yunwen Lei, Ata Kaban

https://openreview.net/forum?id=F6j16Qr6Vk

Keywords: PAC-Bayesian Bounds, Uniform Stability, Generalization Analysis

Compressor summary: The paper improves bounds for stable randomized algorithms using concentration of weakly dependent variables and introduces a sub-exponential stability parameter assumption, applicable to stochastic gradient descent and randomized coordinate descent.


Decision Tree for Locally Private Estimation with Public Data

Yuheng Ma, Han Zhang, Yuchao Cai, Hanfang Yang

https://openreview.net/forum?id=F5FVsfCxt8

Keywords: Local differential privacy, non-parametric regression, decision tree, public data

Compressor summary: The paper introduces a new algorithm called Locally differentially Private Decision Tree (LPDT) for enhancing private estimation using public data, which improves convergence rates and performance compared to existing methods.


GRAND-SLAMIN’ Interpretable Additive Modeling with Structural Constraints

Shibal Ibrahim, Gabriel Isaac Afriat, Kayhan Behdin, Rahul Mazumder

https://openreview.net/forum?id=F5DYsAc7Rt

Keywords: Generalized additive models, component selection, hierarchy, interpretability

Compressor summary: The paper introduces GRAND-SLAMIN, a framework for learning generalized additive models (GAMs) with sparse interactions and structural constraints using differentiable optimization and novel prediction bounds.


Invariant Anomaly Detection under Distribution Shifts: A Causal Perspective

João B. S. Carvalho, Mengtao Zhang, Robin Geyer, Carlos Cotrini, Joachim M. Buhmann

https://openreview.net/forum?id=F1mv2L7Rkb

Keywords: anomaly detection, causal inference, distribution shifts

Compressor summary: The paper proposes a new regularization term for anomaly detection that improves its performance under distribution shifts, using tools from causal inference and ensuring invariant representations.


GlucoSynth: Generating Differentially-Private Synthetic Glucose Traces

Josephine Lamp, Mark Derdzinski, Christopher Hannemann, Joost Van der Linden, Lu Feng, Tianhao Wang, David Evans

https://openreview.net/forum?id=Eysb8t3MJ5

Keywords: Synthetic Data, Time Series, Generative Adversarial Networks, Differential Privacy, Glucose, Diabetes

Compressor summary: The paper introduces GlucoSynth, a novel GAN framework that generates private and high-quality synthetic glucose traces by preserving the relationships among glucose events and temporal dynamics.


Kronecker-Factored Approximate Curvature for Modern Neural Network Architectures

Runa Eschenhagen, Alexander Immer, Richard E Turner, Frank Schneider, Philipp Hennig

https://openreview.net/forum?id=Ex3oJEKS53

Keywords: deep learning, second-order, optimization, natural gradient, fisher, gauss-newton, k-fac, weight-sharing

Compressor summary: The paper proposes two flavors of K-FAC for linear weight-sharing layers in neural networks and shows that they can speed up training and reduce computational costs, especially for deep vision tasks.


Spiking PointNet: Spiking Neural Networks for Point Clouds

Dayong Ren, Zhe Ma, Yuanpei Chen, Weihang Peng, Xiaode Liu, Yuhan Zhang, Yufei Guo

https://openreview.net/forum?id=Ev2XuqvJCy

Keywords: Spiking Neural Networks, Point Clouds

Compressor summary: The paper introduces Spiking PointNet, a spiking neural network model for efficient 3D point cloud recognition, overcoming the challenges of SNNs' optimization and PointNet's memory and computation cost.


OpenShape: Scaling Up 3D Shape Representation Towards Open-World Understanding

Minghua Liu, Ruoxi Shi, Kaiming Kuang, Yinhao Zhu, Xuanlin Li, Shizhong Han, Hong Cai, Fatih Porikli, Hao Su

https://openreview.net/forum?id=Eu4Kkefq7p

Keywords: 3d, shape understanding, open-world understanding, zero-shot 3D classification, vision-language model

Compressor summary: OpenShape is a method that learns joint representations of text, image, and point clouds to enable open-world 3D shape understanding and achieve superior zero-shot classification performance on several benchmarks.


Logarithmic Bayes Regret Bounds

Alexia Atsidakou, Branislav Kveton, Sumeet Katariya, Constantine Caramanis, sujay sanghavi

https://openreview.net/forum?id=ErAP8kF4tG

Keywords: Bayesian bandits, logarithmic regret bounds, multi-armed bandits, linear bandits

Compressor summary: The paper presents new upper bounds for finite-time logarithmic Bayes regret in Bayesian bandits and applies simple and general techniques to linear bandits.


Does Invariant Graph Learning via Environment Augmentation Learn Invariance?

Yongqiang Chen, Yatao Bian, Kaiwen Zhou, Binghui Xie, Bo Han, James Cheng

https://openreview.net/forum?id=EqpR9Vtt13

Keywords: Graph Neural Networks, Out-of-Distribution Generalization, Invariant Learning

Compressor summary: The paper proposes a new framework, GALA, to learn invariant graph representations by incorporating an assistant model that detects environment changes and identifies maximally invariant subgraphs for out-of-distribution generalization.


Sample Complexity of Goal-Conditioned Hierarchical Reinforcement Learning

Arnaud Robert, Ciara Pike-Burke, Aldo A. Faisal

https://openreview.net/forum?id=EqnZqrbFrc

Keywords: Hierarchical Reinforcement Learning, Sample Complexity

Compressor summary: The paper derives a lower bound on the sample complexity for goal-conditioned Hierarchical Reinforcement Learning (HRL) algorithms and proposes a new algorithm that leverages hierarchical decompositions, which is empirically validated on various tasks.


Unbounded Differentially Private Quantile and Maximum Estimation

David Durfee

https://openreview.net/forum?id=Eq9AFZlAjt

Keywords: Differential privacy, Theory, Spars Vector Technique, Quantile

Compressor summary: The paper presents an efficient and accurate method for computing differential privacy using $\texttt{AboveThreshold}$ subroutine on unbounded data with improved privacy guarantees.


Reconciling Competing Sampling Strategies of Network Embedding

Yuchen Yan, Baoyu Jing, Lihui Liu, Ruijie Wang, Jinning Li, Tarek Abdelzaher, Hanghang Tong

https://openreview.net/forum?id=EoDpq18R30

Keywords: Network embedding

Compressor summary: The paper studies the trade-off between discrimination and monotonicity properties for node embeddings under different sampling strategies and proposes a new model (SENSEI) that improves network embedding performance.


Transfer learning for atomistic simulations using GNNs and kernel mean embeddings

John Isak Texas Falk, Luigi Bonati, Pietro Novelli, Michele Parrinello, massimiliano pontil

https://openreview.net/forum?id=Enzew8XujO

Keywords: GNN, Mean Embedding, Kernels, Atomistic Simulations, OCP, Transfer Learning, Molecular Dynamics, Kernel Ridge Regression, Neural Networks

Compressor summary: The authors propose a method to learn interatomic potentials using graph neural networks and kernel mean embeddings, which improves accuracy and interpretability for atomistic simulations of catalytic processes.


Honesty Is the Best Policy: Defining and Mitigating AI Deception

Francis Rhys Ward, Francesca Toni, Francesco Belardinelli, Tom Everitt

https://openreview.net/forum?id=EmxpDiPgRu

Keywords: Deception, Causality, Game Theory

Compressor summary: The paper introduces a formal definition of deception in structural causal games, based on philosophy literature, and provides graphical criteria for detecting and mitigating deception in AI systems like reinforcement learners and language models.


Conditional Mutual Information for Disentangled Representations in Reinforcement Learning

Mhairi Dunion, Trevor McInroe, Kevin Sebastian Luck, Josiah P. Hanna, Stefano V Albrecht

https://openreview.net/forum?id=EmYWJsyad4

Keywords: Reinforcement Learning, Representation Learning, Disentanglement

Compressor summary: The paper proposes a method to teach reinforcement learning agents to separate and represent correlated features in their latent representation, which improves their ability to generalize when the environment changes or when deployed in the real world.


ImageBrush: Learning Visual In-Context Instructions for Exemplar-Based Image Manipulation

Yasheng SUN, Yifan Yang, Houwen Peng, Yifei Shen, Yuqing Yang, Han Hu, Lili Qiu, Hideki Koike

https://openreview.net/forum?id=EmOIP3t9nk

Keywords: Image Manipulation, Visual Instruction

Compressor summary: The paper proposes ImageBrush, a novel image manipulation method that learns visual instructions from transformation images to accurately edit images without using language descriptions.


Does Localization Inform Editing? Surprising Differences in Causality-Based Localization vs. Knowledge Editing in Language Models

Peter Hase, Mohit Bansal, Been Kim, Asma Ghandeharioun

https://openreview.net/forum?id=EldbUlZtbd

Keywords: localization, model editing, mechanistic interpretability, language models

Compressor summary: The paper finds that editing facts in language models does not always follow the localization conclusions from representation denoising and suggests that a better understanding of model behavior may not be helpful for editing tasks.


Recaptured Raw Screen Image and Video Demoiréing via Channel and Spatial Modulations

Huanjing Yue, Yijia Cheng, Xin Liu, Jingyu Yang

https://openreview.net/forum?id=EkcO9tHm6S

Keywords: Raw image demoiréing, raw video demoiréing, video demoiréing dataset

Compressor summary: The authors propose a new network for removing moiré patterns from raw images and videos, which outperforms existing methods and introduce a new dataset with an alignment method.


Equivariant Neural Operator Learning with Graphon Convolution

Chaoran Cheng, Jian Peng

https://openreview.net/forum?id=EjiA3uWpnc

Keywords: Neural Operator Learning, Spectral Graph Theory, Graphon

Compressor summary: The paragraph describes a new model called InfGCN that uses a combination of continuous and discrete graph structures to learn mappings between 3D functions while preserving equivariance, and shows it performs better than existing architectures on electron density datasets.


Timewarp: Transferable Acceleration of Molecular Dynamics by Learning Time-Coarsened Dynamics

Leon Klein, Andrew Y. K. Foong, Tor Erlend Fjelde, Bruno Kacper Mlodozeniec, Marc Brockschmidt, Sebastian Nowozin, Frank Noe, Ryota Tomioka

https://openreview.net/forum?id=EjMLpTgvKH

Keywords: Molecular Dynamics, Normalizing Flows, MCMC

Compressor summary: Timewarp is a method that uses a flow trained on MD trajectories to make large steps in time, enabling efficient simulation of processes like binding and folding over longer timescales for different molecular systems.


Information-guided Planning: An Online Approach for Partially Observable Problems

Matheus Aparecido Do Carmo Alves, Amokh Varma, Yehia Elkhatib, Leandro Soriano Marcolino

https://openreview.net/forum?id=EjG2G1PT2v

Keywords: Information-guided planning, Planning under uncertainty, Sequential decision making

Compressor summary: IB-POMCP is an online planning algorithm that uses entropy estimates of the world belief to guide a tree search process and improve decision-making in partially observable environments, outperforming existing methods in reward, time, and convergence.


Generative Modelling of Stochastic Actions with Arbitrary Constraints in Reinforcement Learning

Changyu Chen, Ramesha Karunasena, Thanh Hong Nguyen, Arunesh Sinha, Pradeep Varakantham

https://openreview.net/forum?id=Ehzj9F2Kmj

Keywords: Action constrained reinforcement learning, Normalizing flow, Generative modelling

Compressor summary: The authors propose a method for solving large discrete action space problems in Reinforcement Learning with validity constraints by using a conditional normalizing flow to represent a stochastic policy and an invalid action rejection technique to update the base policy.


f-Policy Gradients: A General Framework for Goal-Conditioned RL using f-Divergences

Siddhant Agarwal, Ishan Durugkar, Peter Stone, Amy Zhang

https://openreview.net/forum?id=EhhPtGsVAv

Keywords: Goal Conditioned Reinforcement Learning, Shaping Rewards, Reward Design

Compressor summary: The paper proposes $f$-Policy Gradients, a method to encourage exploration in sparse reward RL problems by minimizing the f-divergence between the agent's state visitation distribution and the goal, which can lead to an optimal policy.


SwapPrompt: Test-Time Prompt Adaptation for Vision-Language Models

Xiaosong Ma, Jie ZHANG, Song Guo, Wenchao Xu

https://openreview.net/forum?id=EhdNQiOWgQ

Keywords: Test-Time Adaptation, Prompt Learning, Unsupervised Representation Learning

Compressor summary: SwapPrompt is a novel framework for test-time adaptation that leverages self-supervised contrastive learning to improve the performance of pre-trained vision-language models on unseen test domains, achieving state-of-the-art results and comparable performance with supervised methods.


Training Private Models That Know What They Don’t Know

Stephan Rabanser, Anvith Thudi, Abhradeep Guha Thakurta, Krishnamurthy Dj Dvijotham, Nicolas Papernot

https://openreview.net/forum?id=EgCjf1vjMB

Keywords: differential privacy, selective classification, selective prediction, abstain option, reject option, uncertainty quantification, misclassification detection

Compressor summary: The authors investigate selective classifiers under differential privacy and find that some methods are ineffective due to increased privacy leakage, while a recent approach works well; however, performance degrades with lower privacy levels.


STREAMER: Streaming Representation Learning and Event Segmentation in a Hierarchical Manner

Ramy Mounir, Sujal Vijayaraghavan, Sudeep Sarkar

https://openreview.net/forum?id=EfTMRQn00d

Keywords: predictive learning, hierarchical event segmentation, self-supervised learning, streaming processing, perceptual inputs, biologically-plausible.

Compressor summary: The paragraph describes STREAMER, a self-supervised architecture for hierarchical representation learning and segmentation of streaming inputs, which is evaluated on the egocentric EPIC-KITCHENS dataset for temporal event segmentation.


Speculative Decoding with Big Little Decoder

Sehoon Kim, Karttikeya Mangalam, Suhong Moon, Jitendra Malik, Michael W. Mahoney, Amir Gholami, Kurt Keutzer

https://openreview.net/forum?id=EfMyf9MC3t

Keywords: Transformer, efficient inference, efficient model, decoding

Compressor summary: BiLD improves text generation efficiency by using a small model for autoregressive tasks and a large model for refining predictions non-autoregressively, with two policies to coordinate them.


Isometric Quotient Variational Auto-Encoders for Structure-Preserving Representation Learning

In Huh, changwook jeong, Jae Myung Choe, Young-Gu Kim, Dae Sin Kim

https://openreview.net/forum?id=EdgPb3ngR4

Keywords: representation learning, auto-encoders, geometry, symmetry

Compressor summary: The paper introduces IQVAEs, a novel auto-encoding framework that learns symmetry-preserving representations of data manifolds embedded in high-dimensional spaces using variational auto-encoders.


RRHF: Rank Responses to Align Language Models with Human Feedback

Hongyi Yuan, Zheng Yuan, Chuanqi Tan, Wei Wang, Songfang Huang, Fei Huang

https://openreview.net/forum?id=EdIGMCHk4l

Keywords: Large Language Model, Human Alignment

Compressor summary: RRHF is a novel learning paradigm for aligning large language models with human preferences using ranking loss and sampling responses from various sources, outperforming PPO in simplicity, efficiency, and performance.


Better Correlation and Robustness: A Distribution-Balanced Self-Supervised Learning Framework for Automatic Dialogue Evaluation

Peiwen Yuan, Xinglin Wang, Jiayi Shi, Bin Sun, Yiwei Li, Kan Li

https://openreview.net/forum?id=Ecv1GMiXSk

Keywords: Natural language process, Automatic dialog evaluation

Compressor summary: BCR is a distribution-balanced self-supervised learning framework that improves the correlation and robustness of turn-level dialogue evaluation models using coherence-balanced training signals and a novel adaptive loss function.


Mobilizing Personalized Federated Learning in Infrastructure-Less and Heterogeneous Environments via Random Walk Stochastic ADMM

Ziba Parsons, Fei Dou, Houyi Du, Zheng Song, Jin Lu

https://openreview.net/forum?id=EcmqyXekuP

Keywords: Mobilized Federated Networks, Personalized Federated Learning, Random Walk, Stochastic ADMM

Compressor summary: The paper proposes a novel FL approach for isolated nodes with data heterogeneity and wireless links, using Random Walk SADMM optimization algorithm that improves convergence, accuracy, and communication efficiency.


Optimal Treatment Allocation for Efficient Policy Evaluation in Sequential Decision Making

Ting Li, Chengchun Shi, Jianing Wang, Fan Zhou, Hongtu Zhu

https://openreview.net/forum?id=EcReRm7q9p

Keywords: Average treatment effect, Experimental design, Off-policy evaluation, Optimal treatment allocation

Compressor summary: The paper proposes three optimal allocation strategies for A/B testing in dynamic settings to minimize variance and improve accuracy of treatment effect estimators using off-policy evaluation methods.


Learning Time-Invariant Representations for Individual Neurons from Population Dynamics

Lu Mi, Trung Le, Tianxing He, Eli Shlizerman, Uygar Sümbül

https://openreview.net/forum?id=EcN3l6Xmnx

Keywords: population dynamics, neuronal representation, calcium imaging, cell types

Compressor summary: The authors propose a self-supervised learning method to assign stable representations to neurons based on population activity and improve predictions of gene expression and neuronal identity.


Disentangled Wasserstein Autoencoder for T-Cell Receptor Engineering

Tianxiao Li, Hongyu Guo, Filippo Grazioli, Mark Gerstein, Martin Renqiang Min

https://openreview.net/forum?id=Eb74zfBkWa

Keywords: protein engineering, disentangled representation, T cell receptor

Compressor summary: The proposed method uses a Wasserstein autoencoder to automatically identify functional residues in proteins and edit them without affecting the overall structure, improving efficiency and quality for protein engineering applications like T-cell receptor modification.


Context-lumpable stochastic bandits

Chung-Wei Lee, Qinghua Liu, Yasin Abbasi-Yadkori, Chi Jin, Tor Lattimore, Csaba Szepesvari

https://openreview.net/forum?id=EY7Hpj8Ok6

Keywords: Contextual bandits, low-rank bandits, latent bandits, clustering bandits, stochastic bandit problems, context-lumpable bandits

Compressor summary: The paper proposes an algorithm for a contextual bandit problem where the learner groups the contexts by similarity and achieves near-optimal sample complexity and minimax regret.


Reward Scale Robustness for Proximal Policy Optimization via DreamerV3 Tricks

Ryan Sullivan, Akarsh Kumar, Shengyi Huang, John P Dickerson, Joseph Suarez

https://openreview.net/forum?id=EY4OHikuBm

Keywords: Reinforcement Learning, Proximal Policy Optimization, Reward Normalization

Compressor summary: The paper applies DreamerV3's tricks to PPO and shows that they don't generally improve PPO, but have some specific cases where they work well.


Generative Neural Fields by Mixtures of Neural Implicit Functions

Tackgeun You, Mijeong Kim, Jungtaek Kim, Bohyung Han

https://openreview.net/forum?id=EWNtYvepJh

Keywords: generative neural fields; implicit neural representation; model averaging

Compressor summary: The proposed method learns implicit neural representations and their coefficients to enlarge the capacity of generative neural fields, improving efficiency and effectiveness in generating diverse data.


Better Private Linear Regression Through Better Private Feature Selection

Travis Dick, Jennifer Gillenwater, Matthew Joseph

https://openreview.net/forum?id=EUiIbwV379

Keywords: differential privacy, linear regression, sparse, feature selection, kendall

Compressor summary: The authors propose a differentially private feature selection method based on Kendall rank correlation to broaden the applicability of private linear regression algorithms in high-dimensional problems.


SlotDiffusion: Object-Centric Generative Modeling with Diffusion Models

Ziyi Wu, Jingyu Hu, Wuyue Lu, Igor Gilitschenski, Animesh Garg

https://openreview.net/forum?id=ETk6cfS3vk

Keywords: Unsupervised object-centric learning, diffusion model, generative modeling

Compressor summary: SlotDiffusion is a new object-centric Latent Diffusion Model that improves visual generation and object manipulation in images and videos by enhancing slot-to-image decoding.


Open-Vocabulary Semantic Segmentation via Attribute Decomposition-Aggregation

Chaofan Ma, Yuhuan Yang, Chen Ju, Fei Zhang, Ya Zhang, Yanfeng Wang

https://openreview.net/forum?id=ESEM1lNoeS

Keywords: Open-Vocabulary Semantic Segmentation, Attributes, Decomposition and Aggregation

Compressor summary: The paragraph describes a novel method for semantic segmentation that uses attribute decomposition and aggregation to handle ambiguous or incomplete textual category names, inspired by human cognition.


PrimDiffusion: Volumetric Primitives Diffusion for 3D Human Generation

Zhaoxi Chen, Fangzhou Hong, Haiyi Mei, Guangcong Wang, Lei Yang, Ziwei Liu

https://openreview.net/forum?id=ESCafo3oD5

Keywords: neural rendering, 3D generative model, diffusion model, volumetric primitives, 3D human generation

Compressor summary: PrimDiffusion is a diffusion-based framework for generating 3D humans using volumetric primitives, which enables efficient, high-quality, and flexible rendering.


H3T: Efficient Integration of Memory Optimization and Parallelism for Large-scale Transformer Training

Yuzhong Wang, Xu Han, Weilin Zhao, Guoyang Zeng, Zhiyuan Liu, Maosong Sun

https://openreview.net/forum?id=ES32O8mBK3

Keywords: ML System, Parallelism Learning, Memory Optimization, Data Parallelism, Model Parallelism, Parameter Parallelism, ZeRO, Rematerialization, Checkpointing, Tensor Offloading, Dynamic Programming

Compressor summary: The paper proposes a framework called H3T that automatically integrates memory optimization and parallelism to improve the efficiency of training large Transformer-based models, achieving significant speedups and reduced memory overhead.


Byzantine-Tolerant Methods for Distributed Variational Inequalities

Nazarii Tupitsa, Abdulla Jasem Almansoori, Yanlin Wu, Martin Takáč, Karthik Nandakumar, Samuel Horváth, Eduard Gorbunov

https://openreview.net/forum?id=ER0bcYXvvo

Keywords: byzantine robustness, variational inequalities, min-max problems

Compressor summary: The authors present new methods for solving Byzantine-robust distributed variational inequalities, improving upon existing work and providing empirical evidence.


AUDIT: Audio Editing by Following Instructions with Latent Diffusion Models

Yuancheng Wang, Zeqian Ju, Xu Tan, Lei He, Zhizheng Wu, Jiang Bian, sheng zhao

https://openreview.net/forum?id=EO1KuHoR0V

Keywords: audio editing, text-to-audio generation, diffusion models

Compressor summary: AUDIT is an instruction-guided audio editing model that uses diffusion and denoising processes, triplet training data, and only needs edit instructions as input to achieve state-of-the-art results in various audio editing tasks.


Neural Graph Generation from Graph Statistics

Kiarash Zahirnia, Yaochen Hu, Mark Coates, Oliver Schulte

https://openreview.net/forum?id=EI6BHFKA5p

Keywords: Graph Generation, Local Differential Privacy, Graph Statistics, Latent Adjacency Matrix

Compressor summary: The text describes a new method for learning a deep graph generative model from aggregate statistics without using the graph adjacency matrix, which can protect privacy and generate realistic graphs.


A Long $N$-step Surrogate Stage Reward for Deep Reinforcement Learning

Junmin Zhong, Ruofan Wu, Jennie Si

https://openreview.net/forum?id=EGfYnTyEGv

Keywords: Deep reinforcement learning, Reward Estimation

Compressor summary: The paper introduces a new stage reward estimator, LNSS, for deep RL that mitigates high variance problem and improves performance in continuous control tasks using various baseline algorithms.


LayoutPrompter: Awaken the Design Ability of Large Language Models

Jiawei Lin, Jiaqi Guo, Shizhao Sun, Zijiang James Yang, Jian-Guang Lou, Dongmei Zhang

https://openreview.net/forum?id=EF56cv8B3b

Keywords: graphic design, graphic layout, large language models, in-context learning

Compressor summary: The paper introduces LayoutPrompter, which uses large language models to generate layouts efficiently and versatilely without training or fine-tuning.


Harnessing the power of choices in decision tree learning

Guy Blanc, Jane Lange, Chirag Pabbaraju, Colin Sullivan, Li-Yang Tan, Mo Tiwari

https://openreview.net/forum?id=EEtJTfvNZx

Keywords: Decision Trees, Decision Tree Learning, Top-$k$, ID3, Greedy Algorithms

Compressor summary: Top-$k$ is a simple generalization of decision tree learning algorithms that considers multiple best attributes for splitting and shows improved accuracy, scalability, and applicability compared to existing methods.


Auditing Fairness by Betting

Ben Chugg, Santiago Cortes-Gomez, Bryan Wilder, Aaditya Ramdas

https://openreview.net/forum?id=EEVpt3dJQj

Keywords: fairness, auditing, sequential analysis, martingales, testing by betting

Compressor summary: The authors propose efficient, sequential, and probabilistic methods for auditing the fairness of classification and regression models in real-world systems using inferential and game-theoretic frameworks.


PICProp: Physics-Informed Confidence Propagation for Uncertainty Quantification

Qianli Shen, Wai Hoh Tang, Zhun Deng, Apostolos Psaros, Kenji Kawaguchi

https://openreview.net/forum?id=EETqXXdqkI

Keywords: physics-informed learning, uncertainty quantification, deep learning

Compressor summary: The paper presents PICProp, a bi-level optimization method for estimating confidence intervals in physics-informed learning without strong assumptions.


Bandit Task Assignment with Unknown Processing Time

Shinji Ito, Daisuke Hatano, Hanna Sumita, Kei Takemura, Takuro Fukunaga, Naonori Kakimura, Ken-Ichi Kawarabayashi

https://openreview.net/forum?id=EE1Uiu3Ryb

Keywords: bandit, combinatorial semi-bandits, bandits with badget

Compressor summary: The paper introduces a new bandit task assignment problem with processing time and combinatorial constraints, and proposes an UCB-based algorithm with phased updates that achieves near-optimal regret bounds.


Visual Explanations of Image-Text Representations via Multi-Modal Information Bottleneck Attribution

Ying Wang, Tim G. J. Rudner, Andrew Gordon Wilson

https://openreview.net/forum?id=ECvtxmVP0x

Keywords: Interpretability, Attribution Maps, Information Bottleneck, Multi-Modal Learning, Vision-Language Pretrained Models

Compressor summary: M2IB is a method to improve interpretability of vision-language models like CLIP by compressing irrelevant information and preserving relevant features, without requiring ground truth labels.


Prototype-based Aleatoric Uncertainty Quantification for Cross-modal Retrieval

Hao Li, Jingkuan Song, Lianli Gao, Xiaosu Zhu, Heng Tao Shen

https://openreview.net/forum?id=ECRgBK6sk1

Keywords: multimodal learning, cross-modal retrieval, robust learning, uncertainty

Compressor summary: The paper introduces a new framework to quantify and reduce uncertainty in cross-modal retrieval methods by using learnable prototypes and Subjective Logic Theory.


Zero-Regret Performative Prediction Under Inequality Constraints

Wenjing YAN, Xuanyu Cao

https://openreview.net/forum?id=ECBK3TVmZl

Keywords: Performative prediction, decision-dependent distribution, inequality constraints, primal-dual algorithm.

Compressor summary: The paper studies how to optimize performative prediction problems with inequality constraints using a robust primal-dual framework and proposes an adaptive algorithm that achieves low regret and constraint violations.


Learning from Visual Observation via Offline Pretrained State-to-Go Transformer

Bohan Zhou, Ke Li, Jiechuan Jiang, Zongqing Lu

https://openreview.net/forum?id=E58gaxJN1d

Keywords: Learning from Observations, Offline Learning from Visual Observations, State-to-Go Transformer

Compressor summary: The paper proposes a two-stage framework that uses a pretrained State-to-Go Transformer to predict latent transitions from visual observation and provide intrinsic rewards for reinforcement learning tasks, achieving performance comparable to policy learned from environmental rewards in some cases.


On the Asymptotic Learning Curves of Kernel Ridge Regression under Power-law Decay

Yicheng Li, Haobo Zhang, Qian Lin

https://openreview.net/forum?id=E4P5kVSKlT

Keywords: generalization, reproducing kernel Hilbert space, bias-variance trade-off

Compressor summary: The paper studies the learning curve of kernel ridge regression under realistic assumptions and shows how it relates to the choice of parameters, source condition, and noise, with implications for the benign overfitting phenomenon in over-parametrized neural networks.


Strategic Distribution Shift of Interacting Agents via Coupled Gradient Flows

Lauren E Conger, Franca Hoffman, Eric Mazumdar, Lillian J Ratliff

https://openreview.net/forum?id=E3ZUEaeFYS

Keywords: distribution shift, partial differential equations

Compressor summary: The paper proposes a new framework to analyze how learning algorithms and their environments influence each other over time, accounting for complex dynamics and proving its convergence in different settings.


Reward Imputation with Sketching for Contextual Batched Bandits

Xiao Zhang, Ninglu Shao, Zihua Si, Jun Xu, Wenhan Wang, Hanjing Su, Ji-Rong Wen

https://openreview.net/forum?id=E2zoGTkTbW

Keywords: batched bandit, sketching, reward imputation, regret bound, ridge regression

Compressor summary: SPUIR is an efficient batched bandit algorithm that imputes unobserved rewards using sketching and achieves low regret bounds in both theoretical and practical settings.


VeriX: Towards Verified Explainability of Deep Neural Networks

Min Wu, Haoze Wu, Clark Barrett

https://openreview.net/forum?id=E2TJI6CKm0

Keywords: trustworthy machine learning, deep neural networks, explainability, interpretability, formal methods, automated verification

Compressor summary: VeriX is a system that creates optimal robust explanations and counterfactuals for machine learning models using constraint solving and feature-level sensitivity ranking.


Federated Conditional Stochastic Optimization

Xidong Wu, Jianhui Sun, Zhengmian Hu, Junyi Li, Aidong Zhang, Heng Huang

https://openreview.net/forum?id=E0Gw1uz7lU

Keywords: Federated Learning, Conditional Stochastic Optimization, Nonconvex Optimization

Compressor summary: The paper introduces FCSG and FCSG-M, two new federated learning algorithms for nonconvex conditional stochastic optimization, and Acc-FCSG-M, an accelerated version with lower complexity, which are tested and validated on various tasks.


Modelling Cellular Perturbations with the Sparse Additive Mechanism Shift Variational Autoencoder

Michael Bereket, Theofanis Karaletsos

https://openreview.net/forum?id=DzaCE00jGV

Keywords: Disentagled representation learning, VAE, generative models, sparse mechanism shift, perturbation modeling, cellular modeling

Compressor summary: SAMS-VAE is a novel generative model that combines compositionality, disentanglement, and interpretability for perturbation models in drug discovery, and shows strong performance and correlations to known biological mechanisms.


Convergence Analysis of Sequential Federated Learning on Heterogeneous Data

Yipeng Li, Xinchen Lyu

https://openreview.net/forum?id=Dxhv8Oja2V

Keywords: Federated Learning, Convergence analysis

Compressor summary: The paper analyzes the convergence theory of sequential federated learning (SFL) for non-convex objectives on heterogeneous data and shows that SFL performs better than parallel federated learning (PFL) in cross-device settings.


Distribution Learnability and Robustness

Shai Ben-David, Alex Bie, Gautam Kamath, Tosca Lechner

https://openreview.net/forum?id=Dx99y3okbL

Keywords: robustness, distribution learning

Compressor summary: The paper investigates how different types of contamination affect the relationship between learnability and robust learnability in distribution learning, and discusses implications for other settings like compression and privacy.


Curvature Filtrations for Graph Generative Model Evaluation

Joshua Southern, Jeremy Wayland, Michael M. Bronstein, Bastian Rieck

https://openreview.net/forum?id=Dt71xKyabn

Keywords: Curvature, topology, persistent homology, graph learning, generative model, machine learning, geometric deep learning

Compressor summary: The authors propose using graph curvature and topological data analysis to create effective ways of assessing graph generative models.


Ecosystem-level Analysis of Deployed Machine Learning Reveals Homogeneous Outcomes

Connor Toups, Rishi Bommasani, Kathleen Creel, Sarah H Bana, Dan Jurafsky, Percy Liang

https://openreview.net/forum?id=Ds8iLujo3p

Keywords: homogenous outcomes, societal impact of ML, deployed ML, systemic failure

Compressor summary: The authors propose a new approach called ecosystem-level analysis to study the societal impact of machine learning models by considering their collective effects on users, and they apply it to three modalities (text, images, speech) and find that systemic failure is common and not reduced by individual model improvements.


Faster Query Times for Fully Dynamic $k$-Center Clustering with Outliers

Leyla Biabani, Annika Hennes, Morteza Monemizadeh, Melanie Schmidt

https://openreview.net/forum?id=Ds7Vd83HlC

Keywords: $k$-center clustering, outliers, dynamic algorithms

Compressor summary: The paragraph describes a problem of finding the optimal set of centers for a point cloud with some outliers, and presents an efficient data structure that can handle dynamic updates and provide approximate solutions in sublinear time.


The Impact of Positional Encoding on Length Generalization in Transformers

Amirhossein Kazemnejad, Inkit Padhi, Karthikeyan Natesan, Payel Das, Siva Reddy

https://openreview.net/forum?id=Drrl2gcjzl

Keywords: Transformers, Positional Encoding, Length Generalization

Compressor summary: The paper compares five positional encoding methods for decoder-only Transformers and finds that no explicit encoding is better than others for length generalization in downstream tasks.


Gaussian Differential Privacy on Riemannian Manifolds

Yangdi Jiang, Xiaotian Chang, Yi Liu, Lei Ding, Linglong Kong, Bei Jiang

https://openreview.net/forum?id=DrIZZwEZtM

Keywords: Differential Privacy, Gaussian Differential Privacy, Differential Geometry, Riemannian Manifold, Homogeneous Riemannian Manifold, Frechet Mean

Compressor summary: The authors develop a new method for preserving privacy in data analysis on curved spaces using a special type of probability distribution that accounts for the geometry of the space.


Navigating the Pitfalls of Active Learning Evaluation: A Systematic Framework for Meaningful Performance Assessment

Carsten Tim Lüth, Till J. Bungert, Lukas Klein, Paul F Jaeger

https://openreview.net/forum?id=Dqn715Txgl

Keywords: Active Learning, Evaluation, Study

Compressor summary: The paper discusses the challenges of evaluating active learning methods, proposes a framework to address these issues, and presents an empirical study to demonstrate its usefulness.


Decentralized Randomly Distributed Multi-agent Multi-armed Bandit with Heterogeneous Rewards

Mengfan Xu, Diego Klabjan

https://openreview.net/forum?id=DqfdhM64LI

Keywords: decentralized multi-agent MAB, heterogeneous light-tailed and heavy-tailed rewards, time dependent random graphs

Compressor summary: The paper proposes a novel algorithm for decentralized multi-agent bandit problems with time-varying graphs and varying rewards, achieving near-optimal regret bounds with high probability.


Selectively Sharing Experiences Improves Multi-Agent Reinforcement Learning

Matthias Gerstgrasser, Tom Danino, Sarah Keren

https://openreview.net/forum?id=DpuphOgJqh

Keywords: multi-agent reinforcement learning, reinforcement learning, deep q learning, cooperative ai

Compressor summary: Selective Multi-Agent Prioritized Experience Relay is a new method for multi-agent reinforcement learning where agents share a small number of relevant experiences to improve learning efficiency and performance.


Statistically Valid Variable Importance Assessment through Conditional Permutations

Ahmad Chamma, Denis Engemann, Bertrand Thirion

https://openreview.net/forum?id=DoE3esTIEM

Keywords: Interpretability, Variable Importance, Machine Learning, Deep Learning, Statistical Inference

Compressor summary: The authors propose CPI, a new method to assess variable importance in machine learning models, which overcomes the limitations of standard permutation importance by providing accurate type-I error control and performs well on benchmarks and real-world data analysis.


On the Implicit Bias of Linear Equivariant Steerable Networks

Ziyu Chen, Wei Zhu

https://openreview.net/forum?id=DnVjDRLwVu

Keywords: implicit bias, equivariant steerable networks, data augmentation, margin, generalization bound

Compressor summary: The text discusses how gradient flow helps linear equivariant steerable networks find group-invariant classifiers with a maximum margin, and shows that they are equivalent to data augmentation under certain conditions.


Approximation-Generalization Trade-offs under (Approximate) Group Equivariance

Mircea Petrache, Shubhendu Trivedi

https://openreview.net/forum?id=DnO6LTQ77U

Keywords: Equivariance, Invariance, Generalization, Equivariant Neural Networks, Approximation Error

Compressor summary: The paper investigates how incorporating task-specific symmetries in machine learning models improves generalization and how to handle approximate or partial symmetries between models and data distributions.


Data-Centric Learning from Unlabeled Graphs with Diffusion Model

Gang Liu, Eric Inae, Tong Zhao, Jiaxin Xu, Tengfei Luo, Meng Jiang

https://openreview.net/forum?id=DmakwvCJ7l

Keywords: Graph property prediction, Molecular property prediction, Diffusion model, Unlabeled data, Data augmentation, Transfer learning

Compressor summary: The paper proposes a method to use unlabeled graphs for property prediction tasks by generating task-specific graph examples and labels with a diffusion model.


Sharpness Minimization Algorithms Do Not Only Minimize Sharpness To Achieve Better Generalization

Kaiyue Wen, Zhiyuan Li, Tengyu Ma

https://openreview.net/forum?id=Dkmpa6wCIx

Keywords: Sharpness, Flatness, Generalization, Generalization Bound, SAM

Compressor summary: The paper investigates how flatness (sharpness) of two-layer ReLU neural networks relates to generalization and finds that it depends on data distribution and architecture, suggesting that sharpness minimization alone does not guarantee better generalization.


How to Scale Your EMA

Dan Busbridge, Jason Ramapuram, Pierre Ablin, Tatiana Likhomanenko, Eeshan Gunesh Dhekane, Xavier Suau, Russell Webb

https://openreview.net/forum?id=DkeeXVdQyu

Keywords: Optimization, scaling rules, EMA, exponential moving average, self-supervised learning, pseudo-labelling, semi-supervised learning, BYOL, distillation, speech, vision

Compressor summary: The paper proposes a scaling rule for optimization with model EMA that preserves training dynamics across batch sizes and improves the performance of various machine learning methods, including pseudo-labeling and SSL.


Optimal Rates for Bandit Nonstochastic Control

Y. Jennifer Sun, Stephen Newman, Elad Hazan

https://openreview.net/forum?id=DkKHSsmVuA

Keywords: Bandit control, online learning

Compressor summary: The paper proposes an algorithm for optimal control under adversarial conditions that achieves sublinear regret and introduces a novel memory-based technique for bandit convex optimization.


NAR-Former V2: Rethinking Transformer for Universal Neural Network Representation Learning

Yun Yi, Haokui Zhang, Rong Xiao, Nannan Wang, Xiaoyu Wang

https://openreview.net/forum?id=DjX2Nr15kY

Keywords: Transformer, Graph Neural Network, neural network encoding, representation learning, neural architecture search, neural network deployment

Compressor summary: The paper proposes a modified Transformer-based model that can learn efficient representations from both cell-structured and entire networks by incorporating inductive representation learning capability of GNNs and enhancing Transformer's ability to work with graph structures, achieving better performance in predicting latency and accuracy.


Automatic Integration for Spatiotemporal Neural Point Processes

Zihao Zhou, Rose Yu

https://openreview.net/forum?id=Deb1yP1zMN

Keywords: spatiotemporal modeling, neural point processes, integration method

Compressor summary: The paper introduces `Auto-STPP`, a novel method for efficiently integrating flexible spatiotemporal neural point processes using a decomposable parametrization and ProdNet, which outperforms existing methods in recovering complex intensity functions from irregular events.


Sketchy: Memory-efficient Adaptive Regularization with Frequent Directions

Vladimir Feinberg, Xinyi Chen, Y. Jennifer Sun, Rohan Anil, Elad Hazan

https://openreview.net/forum?id=DeZst6dKyi

Keywords: online convex optimization, deep learning, matrix sketching, frequent directions

Compressor summary: The paper proposes a low-rank sketching approach using Frequent Directions (FD) to reduce memory and compute requirements of matrix preconditioners in deep learning training tasks, achieving similar performance to Shampoo and Adam with less memory.


Uncovering Prototypical Knowledge for Weakly Open-Vocabulary Semantic Segmentation

Fei Zhang, Tianfei Zhou, Boyang Li, Hao He, Chaofan Ma, Tianjiao Zhang, Jiangchao Yao, Ya Zhang, Yanfeng Wang

https://openreview.net/forum?id=DdViWdxCTs

Keywords: Weakly (Text-based) Open-Vocabulary Semantic Segmentation, Vision-Language Pretraining, Prototypical Knowledge

Compressor summary: This paper introduces a new method for semantic segmentation of arbitrary classes using image-text pairs, which improves the performance by incorporating explicit supervision and multi-modal regularization for group tokens.


How2comm: Communication-Efficient and Collaboration-Pragmatic Multi-Agent Perception

Dingkang Yang, Kun Yang, Yuzheng Wang, Jing Liu, Zhi Xu, Rongbin Yin, Peng Zhai, Lihua Zhang

https://openreview.net/forum?id=Dbaxm9ujq6

Keywords: Collaborative perception, multi-agent communication

Compressor summary: How2comm is a collaborative perception framework for driving scenarios that balances performance and communication bandwidth by using mutual information-aware communication, spatial-channel filtering, flow-guided delay compensation, and a collaboration transformer to handle various noises.


AiluRus: A Scalable ViT Framework for Dense Prediction

Jin Li, Yaoming Wang, XIAOPENG ZHANG, Bowen Shi, Dongsheng Jiang, Chenglin Li, Wenrui Dai, Hongkai Xiong, Qi Tian

https://openreview.net/forum?id=DVm0xxaEq1

Keywords: vision transformer, dense prediction

Compressor summary: The paper proposes an adaptive resolution method for vision transformers to speed up dense prediction tasks like semantic segmentation or object detection by clustering tokens based on importance and merging adjacent ones.


RoboCLIP: One Demonstration is Enough to Learn Robot Policies

Sumedh Anand Sontakke, Jesse Zhang, Séb Arnold, Karl Pertsch, Erdem Biyik, Dorsa Sadigh, Chelsea Finn, Laurent Itti

https://openreview.net/forum?id=DVlawv2rSI

Keywords: Reinforcement Learning, Vision and Language Models

Compressor summary: RoboCLIP is an online imitation learning method that uses VLMs to generate rewards from a single video or text demonstration, enabling high zero-shot performance in robot manipulation tasks without manual reward design or domain matching.


Chasing Fairness Under Distribution Shift: A Model Weight Perturbation Approach

Zhimeng Jiang, Xiaotian Han, Hongye Jin, Guanchu Wang, Rui Chen, Na Zou, Xia Hu

https://openreview.net/forum?id=DVjyq5eCAD

Keywords: Model Weight Perturbation, fairness, distribution shift

Compressor summary: The paper proposes a method called robust fairness regularization (RFR) that improves algorithmic fairness under distribution shifts by considering the worst case weight perturbation for each sensitive attribute group.


R-divergence for Estimating Model-oriented Distribution Discrepancy

Zhilin Zhao, Longbing Cao

https://openreview.net/forum?id=DVWIA9v9Jm

Keywords: non-IID, Distribution Discrepancy, Data Divergence, Two-sample Test

Compressor summary: R-divergence is a method to measure how similar or different two datasets' probability distributions are by learning a minimum hypothesis on mixed data and comparing their expected risks.


Causal Interpretation of Self-Attention in Pre-Trained Transformers

Raanan Yehezkel Rohekar, Yaniv Gurwicz, Shami Nisimov

https://openreview.net/forum?id=DS4rKySlYC

Keywords: Self-Attention, Causal Discovery, Reasoning, Explainability, Zero-shot, Transformer

Compressor summary: The authors propose a way to understand self-attention in Transformers as a causal structure over input symbols, allowing them to use pre-trained models for zero-shot causal discovery and provide explanations for the outcomes of two tasks.


ZoomTrack: Target-aware Non-uniform Resizing for Efficient Visual Tracking

Yutong Kou, Jin Gao, Bing Li, Gang Wang, Weiming Hu, Yizheng Wang, Liang Li

https://openreview.net/forum?id=DQgTewaKzt

Keywords: Visual tracking, non-uniform resizing, HVS-inspired processing

Compressor summary: The paper proposes a method to improve the speed and performance of trackers by non-uniformly resizing the input image, allowing for better target detection despite smaller input size.


Transfer Learning with Affine Model Transformation

Shunya Minami, Kenji Fukumizu, Yoshihiro Hayashi, Ryo Yoshida

https://openreview.net/forum?id=DPeBX79eNz

Keywords: Machine learning, Transfer learning

Compressor summary: The paper introduces a new transfer learning regression method called affine model transfer that has theoretical properties and outperforms existing methods in scenarios with scarce data.


A Theory of Unsupervised Translation Motivated by Understanding Animal Communication

Shafi Goldwasser, David Gruber, Adam Tauman Kalai, Orr Paradise

https://openreview.net/forum?id=DP2lioYIYl

Keywords: Theory, Unsupervised Machine Translation

Compressor summary: The paragraph discusses a theoretical framework for analyzing unsupervised machine translation (UMT) between two languages with different structures, and how it could potentially help understand animal communication systems.


Off-Policy Evaluation for Human Feedback

Qitong Gao, Ge Gao, Juncheng Dong, Vahid Tarokh, Min Chi, Miroslav Pajic

https://openreview.net/forum?id=DOdaV0Hqdy

Keywords: Off-policy evaluation (OPE), Variational latent model for trajectory representation learning, Reinforcement learning and OPE for adaptive neurostimulation

Compressor summary: The paper proposes an OPEHF framework for estimating human feedback signals in RL using offline data, which outperforms existing OPE methods in three real-world and simulation experiments.


Weakly Coupled Deep Q-Networks

Ibrahim El Shar, Daniel R. Jiang

https://openreview.net/forum?id=DNubFPV5Dy

Keywords: Reinforcement learning, Deep Reinforcement Learning, Weakly Coupled MDPs

Compressor summary: The paper introduces WCDQN, a reinforcement learning algorithm that improves performance in WCMDPs by training multiple subagents and combining their solutions to guide the main agent towards optimality.


Crystal Structure Prediction by Joint Equivariant Diffusion

Rui Jiao, Wenbing Huang, Peijia Lin, Jiaqi Han, Pin Chen, Yutong Lu, Yang Liu

https://openreview.net/forum?id=DNdN26m2Jk

Keywords: crystal structure prediction, equivariant graph neural networks, diffusion generative models

Compressor summary: DiffCSP is a novel diffusion model that leverages fractional coordinates and equivariant denoising to efficiently predict crystal structures with high accuracy.


On the Stability-Plasticity Dilemma in Continual Meta-Learning: Theory and Algorithm

Qi CHEN, Changjian Shui, Ligong Han, Mario Marchand

https://openreview.net/forum?id=DNHGKeOhLl

Keywords: continual meta-learning; transfer learning; stability-plasticity dilemma;

Compressor summary: The paragraph describes Continual Meta-Learning (CML), which aims to balance stability and plasticity in learning from non-i.i.d. tasks, and proposes a novel algorithm that adapts meta-parameter and learning rate to environment change.


Optimal Convergence Rate for Exact Policy Mirror Descent in Discounted Markov Decision Processes

Emmeran Johnson, Ciara Pike-Burke, Patrick Rebeschini

https://openreview.net/forum?id=DKHEkP7Idx

Keywords: Reinforcement Learning Theory, Policy Mirror Descent, Policy Gradient

Compressor summary: The paper introduces and analyzes Policy Mirror Descent (PMD), a general family of algorithms in reinforcement learning that achieves the optimal convergence rate and step-size when using exact policy evaluation.


Don’t just prune by magnitude! Your mask topology is a secret weapon

Duc N.M Hoang, Souvik Kundu, Shiwei Liu, Zhangyang Wang

https://openreview.net/forum?id=DIBcdjWV7k

Keywords: Pruning at Initialization, Pruning at Training, LTH, DST, Ramanujan, graph

Compressor summary: This paper explores how the connectivity of a deep network's architecture as a graph affects its performance, and proposes a new pruning method that maximizes the distance between sparse subnetworks and their original initialization.


First Order Stochastic Optimization with Oblivious Noise

Ilias Diakonikolas, Sushrut Karmalkar, Jongho Park, Christos Tzamos

https://openreview.net/forum?id=DI6KQhgqUr

Keywords: Oblivious noise, Robust Statistics, Heavy-tailed Stochastic Optimization, Approximate Gradients, Inexact Gradients

Compressor summary: The study explores stochastic optimization with oblivious noise and presents an efficient list-decodable learner that finds a close candidate or a single solution depending on the noise level.


Zero-One Laws of Graph Neural Networks

Sam Adam-Day, Theodor-Mihai Iliant, Ismail Ilkan Ceylan

https://openreview.net/forum?id=DGmxTUCHYs

Keywords: graph neural networks, graph convolutional networks, zero-one law, expressivity, asymptotic behavior

Compressor summary: The paper analyzes how graph neural networks (GNNs) represent and extrapolate information on large graphs, finding theoretical and empirical limitations to their performance.


Propagating Knowledge Updates to LMs Through Distillation

Shankar Padmanabhan, Yasumasa Onoe, Michael JQ Zhang, Greg Durrett, Eunsol Choi

https://openreview.net/forum?id=DFaGf3O7jf

Keywords: Knowledge editing, NLP, Distillation, deep learning, fine-tuning

Compressor summary: The authors propose a method for updating language models with new entity definitions and show that their approach enables better inferences based on these definitions without sacrificing performance in other contexts.


DreamSim: Learning New Dimensions of Human Visual Similarity using Synthetic Data

Stephanie Fu, Netanel Yakir Tamir, Shobhita Sundaram, Lucy Chai, Richard Zhang, Tali Dekel, Phillip Isola

https://openreview.net/forum?id=DEiNSfh1k7

Keywords: perceptual similarity, foundation model, perception, computer vision, image metric

Compressor summary: The paper introduces DreamSim, a new perceptual similarity metric that assesses images holistically by capturing mid-level similarities in layout, object pose, and semantic content, and shows its effectiveness on retrieval and reconstruction tasks using a new dataset of human similarity judgments.


Coupled Reconstruction of Cortical Surfaces by Diffeomorphic Mesh Deformation

Hao Zheng, Hongming Li, Yong Fan

https://openreview.net/forum?id=DEC7NxDJLh

Keywords: Brain MRIs, cortical surface reconstruction, deep learning

Compressor summary: The authors propose a new deep learning framework that reconstructs cortical surfaces from brain MRIs by jointly estimating the midthickness surface and diffeomorphic flows to optimize and deform it into inner, outer, and midthickness surfaces while accounting for topological correctness and estimating cortical thickness.


Discrete-Smoothness in Online Algorithms with Predictions

Yossi Azar, Debmalya Panigrahi, Noam Touitou

https://openreview.net/forum?id=DDmH3H78iJ

Keywords: Online, Predictions, Learning-augmented, Facility Location, Set Cover

Compressor summary: The paper proposes discrete-smooth algorithms for online covering problems, such as facility location and set cover, which balance consistency, robustness, and smoothness in machine-learning predictions.


Brant: Foundation Model for Intracranial Neural Signal

Daoze Zhang, Zhizhang Yuan, Yang Yang, Junru Chen, Jingjing Wang, Yafeng Li

https://openreview.net/forum?id=DDkl9vaJyE

Keywords: Foundation model, Brain signal, Pretraining, Medicine

Compressor summary: Brant is a pre-trained foundation model that learns powerful representations of intracranial neural signals, achieving state-of-the-art performance on different brain signal tasks.


ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP

Lu Yan, ZHUO ZHANG, Guanhong Tao, Kaiyuan Zhang, Xuan Chen, Guangyu Shen, Xiangyu Zhang

https://openreview.net/forum?id=DD0QJvPbTD

Keywords: NLP, backdoor attack, fuzzing

Compressor summary: The paper proposes a method to detect stealthy backdoor attacks in NLP models by paraphrasing inputs to remove triggers and comparing model predictions before and after paraphrasing.


Payoff-based Learning with Matrix Multiplicative Weights in Quantum Games

Kyriakos Lotidis, Panayotis Mertikopoulos, Nicholas Bambos, Jose Blanchet

https://openreview.net/forum?id=DCIsNIUCV7

Keywords: quantum games, Matrix Multiplicative Weights, zero-sum games, Nash equilibrium

Compressor summary: The paper presents new methods for learning in quantum games and other semidefinite games with minimal information feedback, using zeroth-order gradient samplers and achieving different convergence rates depending on the information framework.


Provably Fast Finite Particle Variants of SVGD via Virtual Particle Stochastic Approximation

Aniket Das, Dheeraj Mysore Nagaraj

https://openreview.net/forum?id=DBz9E5aZey

Keywords: Stein Variational Gradient Descent, Variational Inference, Sampling

Compressor summary: The paper introduces virtual particles to improve the efficiency and convergence rate of Stein Variational Gradient Descent (SVGD), a particle-based variational inference algorithm, in the finite-particle regime.


Brain-like Flexible Visual Inference by Harnessing Feedback Feedforward Alignment

Tahereh Toosi, Elias Issa

https://openreview.net/forum?id=DBlkX8Nczr

Keywords: Visual inference, Bio-plausible learning algorithm, Feedback connections, Visual imagery, Occlusions, Noise

Compressor summary: The paper proposes a learning algorithm, Feedback-Feedforward Alignment (FFA), that aligns feedforward and feedback pathways in neural networks to co-optimize tasks and enable emergent visual inference functions similar to those in natural vision.


Optimal Transport for Treatment Effect Estimation

Hao Wang, Jiajun Fan, Zhichao Chen, Haoxuan Li, Weiming Liu, Tianqiao Liu, Quanyu Dai, Yichao Wang, Zhenhua Dong, Ruiming Tang

https://openreview.net/forum?id=DAdfU1ASLb

Keywords: treatment effect estimation, optimal transport, wasserstein, causal inference, counterfactual

Compressor summary: The paper proposes a new method called Entire Space CounterFactual Regression (ESCFR) to better estimate individual treatment effects from observational data by addressing two common issues in existing methods.


A Smooth Binary Mechanism for Efficient Private Continual Observation

Joel Daniel Andersson, Rasmus Pagh

https://openreview.net/forum?id=DAKAkMhjSR

Keywords: differential privacy, continual observation, binary mechanism

Compressor summary: The paper proposes an efficient alternative method for releasing differentially private estimates based on a dataset that evolves over time, improving on previous methods in terms of noise distribution and computation time.


MGDD: A Meta Generator for Fast Dataset Distillation

Songhua Liu, Xinchao Wang

https://openreview.net/forum?id=D9CMRR5Lof

Keywords: Dataset Distillation, Dataset Condensation, Efficient Learning, Conditional Generation, Meta Learning

Compressor summary: The paper introduces a faster and more flexible dataset distillation method using a generative approach and meta-learning, achieving significant speedup and improvement over existing techniques.


A One-Size-Fits-All Approach to Improving Randomness in Paper Assignment

Yixuan Even Xu, Steven Jecmen, Zimeng Song, Fei Fang

https://openreview.net/forum?id=D94QKZA7UP

Keywords: peer review, randomized paper assignment, mitigating malicious behavior, convex optimization

Compressor summary: The paper proposes a practical method for randomized paper assignment in peer review processes that balances various considerations like expertise, robustness, diversity, and anonymity.


Fairness-guided Few-shot Prompting for Large Language Models

Huan Ma, Changqing Zhang, Yatao Bian, Lemao Liu, Zhirui Zhang, Peilin Zhao, Shu Zhang, Huazhu Fu, Qinghua Hu, Bingzhe Wu

https://openreview.net/forum?id=D8oHQ2qSTj

Keywords: large language models, prompts, classification

Compressor summary: The paper introduces a metric to measure predictive bias of prompts for in-context learning and proposes a search strategy to find near-optimal prompts, improving the performance on various downstream tasks.


On Transfer of Adversarial Robustness from Pretraining to Downstream Tasks

Laura Fee Nern, Harsh Raj, Maurice Georgi, Yash Sharma

https://openreview.net/forum?id=D8nAMRRCLS

Keywords: Machine Learning Theory, Transfer Learning, Adversarial Robustness

Compressor summary: The study shows how to measure and improve the robustness of linear predictors on downstream tasks by using representations with desired robustness properties.


Optimal Transport Model Distributional Robustness

Van-Anh Nguyen, Trung Le, Anh Tuan Bui, Thanh-Toan Do, Dinh Phung

https://openreview.net/forum?id=D7LdL2SCCi

Keywords: Distributional Robustness, Sharpness-aware, SAM

Compressor summary: The paper proposes an optimal transport-based distributional robustness framework for training deep learning models that can be applied to single, ensemble, or Bayesian models and shows improved performance in various experiments.


Flow Matching for Scalable Simulation-Based Inference

Jonas Bernhard Wildberger, Maximilian Dax, Simon Buchholz, Stephen R Green, Jakob H. Macke, Bernhard Schölkopf

https://openreview.net/forum?id=D2cS6SoYlP

Keywords: simulation-based inference, likelihood-free inference, machine learning for physical sciences

Compressor summary: FMPE is a continuous normalizing flows technique for simulation-based inference that offers flexibility, exact density evaluation, fast training, and scalability, and outperforms discrete flows methods in gravitational-wave inference.


Temporal Dynamic Quantization for Diffusion Models

Junhyuk So, Jungwon Lee, Daehyun Ahn, Hyungjun Kim, Eunhyeok Park

https://openreview.net/forum?id=D1sECc9fiG

Keywords: deep learning optimization, quantization, diffusion model, generative model

Compressor summary: The paper introduces a novel quantization method for diffusion models that adapts to time steps, improving performance without increasing computation cost.


Selectivity Drives Productivity: Efficient Dataset Pruning for Enhanced Transfer Learning

Yihua Zhang, Yimeng Zhang, Aochuan Chen, Jinghan Jia, Jiancheng Liu, Gaowen Liu, Mingyi Hong, Shiyu Chang, Sijia Liu

https://openreview.net/forum?id=D0MII7rP3R

Keywords: Dataset pruning, transfer learning

Compressor summary: The authors propose new dataset pruning methods for transfer learning that reduce data usage by up to 80% without compromising performance on downstream tasks, leading to significant speed-ups during pretraining.


Optimal cross-learning for contextual bandits with unknown context distributions

Jon Schneider, Julian Zimmert

https://openreview.net/forum?id=CzkOzKWpMa

Keywords: bandits, first-price auction, sleeping bandits, contextual bandits

Compressor summary: The paper proposes an efficient algorithm for designing contextual bandit algorithms with a nearly tight regret bound, which removes correlations between estimation of the unknown context distribution and the actions played by the algorithm.


Coneheads: Hierarchy Aware Attention

Albert Tseng, Tao Yu, Toni J.B. Liu, Christopher De Sa

https://openreview.net/forum?id=CzAFnfwbGd

Keywords: Hyperbolic Entailment Cones, Hyperbolic Space, Entailment Cones, Attention, Dot Product, Hierarchy, Transformers

Compressor summary: Cone attention is a new attention mechanism for neural networks that uses hyperbolic cones to model the hierarchy between data points, improving performance over dot product attention and requiring fewer parameters.


Exploring and Interacting with the Set of Good Sparse Generalized Additive Models

Chudi Zhong, Zhi Chen, Jiachang Liu, Margo Seltzer, Cynthia Rudin

https://openreview.net/forum?id=CzAAbKOHQW

Keywords: Interpretability, human-model interaction, generalized additive model, Rashomon set

Compressor summary: The text discusses how approximating the Rashomon set of models allows for better interaction between machine learning models and domain experts, and presents algorithms to efficiently approximate this set for sparse, generalized additive models with fixed support sets.


Neural Priming for Sample-Efficient Adaptation

Matthew Wallingford, Vivek Ramanujan, Alex Fang, Aditya Kusupati, Roozbeh Mottaghi, Aniruddha Kembhavi, Ludwig Schmidt, Ali Farhadi

https://openreview.net/forum?id=Cyn1PvuZsB

Keywords: Transfer Learning, Distribution Shift, Test-Time Training

Compressor summary: Neural Priming is a technique that adapts pretrained models to new tasks and distributions using recalled data from previous training, improving accuracy in various benchmarks.


SALSA VERDE: a machine learning attack on LWE with sparse small secrets

Cathy Yuanchen Li, Emily Wenger, Zeyuan Allen-Zhu, Francois Charton, Kristin E. Lauter

https://openreview.net/forum?id=CxzCoFDeQf

Keywords: machine learning, cryptography, cryptanalysis

Compressor summary: VERDE is an improved machine learning attack on the Learning with Errors problem, which can recover sparse secrets and work with larger dimensions and smaller moduli, while using less time and power.


Neural Circuits for Fast Poisson Compressed Sensing in the Olfactory Bulb

Jacob A Zavatone-Veth, Paul Masset, William Lingxiao Tong, Joseph Zak, Venkatesh N Murthy, Cengiz Pehlevan

https://openreview.net/forum?id=Cxn1FpnNvG

Keywords: Olfaction, Bayesian inference, neural circuits, normative models, population geometry

Compressor summary: The proposed rate-based Poisson compressed sensing circuit model for the olfactory bulb accurately detects tens of odors within a single sniff and performs uncertainty estimation by matching receptor properties with the geometry of the neural code.


Non-Asymptotic Analysis of a UCB-based Top Two Algorithm

Marc Jourdan, Rémy Degenne

https://openreview.net/forum?id=CxjmYRP9Ji

Keywords: multi-armed bandits, best-arm identification, Gaussian bandits, Top Two algorithm, fixed confidence, finite confidence

Compressor summary: The paper derives the first non-asymptotic upper bound on the expected sample complexity of a Top Two bandit algorithm with any error level, and proposes a UCB-based algorithm that has both theoretical guarantees and good empirical performance.


Diffusion Probabilistic Models for Structured Node Classification

Hyosoon Jang, Seonghyun Park, Sangwoo Mo, Sungsoo Ahn

https://openreview.net/forum?id=CxUuCydMDU

Keywords: diffusion model, graph neural network, structured prediction, node classification

Compressor summary: The paper proposes a new framework for structured node classification on graphs that considers label dependencies and uses a diffusion probabilistic model to learn from partially labeled data.


Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization

Jameel Hassan Abdul Samadh, Hanan Gani, Noor Hazim Hussein, Muhammad Uzair Khattak, Muzammal Naseer, Fahad Khan, Salman Khan

https://openreview.net/forum?id=CusNOTRkQw

Keywords: Vision-Language models, Prompt Learning, Test-Time Adaptation

Compressor summary: The paper proposes a method to improve zero-shot generalization of vision-language models using prompt tuning to align test sample statistics with source data, achieving better results than existing techniques on domain generalization benchmarks.


Online POMDP Planning with Anytime Deterministic Guarantees

Moran Barenboim, Vadim Indelman

https://openreview.net/forum?id=Cupr2yTFSx

Keywords: POMDPs, Planning under uncertainty, Robotics

Compressor summary: The authors propose a method to find a simplified solution for planning under uncertainty in POMDPs that has deterministic guarantees and can be integrated with existing solvers.


Convolution Monge Mapping Normalization for learning on sleep data

Theo Gnassounou, Rémi Flamary, Alexandre Gramfort

https://openreview.net/forum?id=CuHymkHRus

Keywords: Neuroscience, Domain adaptation, Optimal Transport

Compressor summary: In many machine learning applications on signals and biomedical data, especially electroencephalogram (EEG), one major challenge is the variability of the data across subjects, sessions, and hardware devices. In this work, we propose a new method called Convolutional Monge Mapping Normalization ($\texttt{CMMN}$), which consists in filtering the signals in order to adapt their power spectrum density (PSD) to a Wasserstein barycenter estimated on training data. $\texttt{CMMN}$ relies on novel closed-form solutions for optimal transport mappings and barycenters and provides individual test time adaptation to new data without needing to retrain a prediction model. Numerical experiments on sleep EEG data show that $\texttt{CMMN}$ leads to significant and consistent performance gains independent from the neural network architecture when adapting between subjects, sessions, and even datasets collected with different hardware. Notably our performance gain is on par with much more numerically intensive Domain Adaptation (DA) methods and can be used in conjunction with those for even better performances.


V-InFoR: A Robust Graph Neural Networks Explainer for Structurally Corrupted Graphs

Senzhang Wang, Jun Yin, Chaozhuo Li, Xing Xie, Jianxin Wang

https://openreview.net/forum?id=CtXXOaxDw7

Keywords: Explainable AI, Graph Neural Networks, Machine Learning

Compressor summary: V-InfoR is a robust graph neural network (GNN) explainer that uses variational inference to extract graph representations and graph information bottleneck optimization to generate explanations for corrupted graphs.


Saving 100x Storage: Prototype Replay for Reconstructing Training Sample Distribution in Class-Incremental Semantic Segmentation

Jinpeng Chen, Runmin Cong, Yuxuan LUO, Horace Ip, Sam Kwong

https://openreview.net/forum?id=Ct0zPIe3xs

Keywords: Continual learning, Class-incremental semantic segmentation, Prototype replay

Compressor summary: STAR is a method for semantic segmentation that addresses class imbalance by storing compact prototypes and repeating background pixels, achieving better performance with less storage than previous methods.


Imitation Learning from Vague Feedback

Xin-Qiang Cai, Yu-Jie Zhang, Chao-Kai Chiang, Masashi Sugiyama

https://openreview.net/forum?id=CswEebv5Hn

Keywords: imitation learning, vague feedback, risk rewriting, mixture propotion estimation

Compressor summary: The paragraph discusses imitation learning from human feedback, which aims to train agents using relative comparisons of demonstrations, and proposes a method for handling vague feedback in this setting.


List and Certificate Complexities in Replicable Learning

Peter Dixon, A. Pavan, Jason Vander Woude, N V Vinodchandran

https://openreview.net/forum?id=Cs9ea2Gbgx

Keywords: Replicability, learning algorithms, sample complexity, PAC Learning

Compressor summary: The authors study replicable learning algorithms and design optimal list and certificate complexities for estimating coin biases and learning hypothesis classes with nonadaptive statistical queries, while minimizing sample complexity.


Finding Local Minima Efficiently in Decentralized Optimization

Wenhan Xian, Heng Huang

https://openreview.net/forum?id=CruxS0C0LS

Keywords: second-order optimality, decentralized optimization

Compressor summary: The paper introduces PEDESTAL, a new decentralized stochastic algorithm for nonconvex optimization that achieves second-order optimality with theoretical guarantees and matches state-of-the-art results in centralized methods.


On the Relationship Between Relevance and Conflict in Online Social Link Recommendations

Yanbang Wang, Jon Kleinberg

https://openreview.net/forum?id=CrpL8mGa0Q

Keywords: social networks, spectral analysis, link recommendation, polarization and conflict

Compressor summary: The paper analyzes how link recommendations in online social networks affect relevance and conflict, finding that some algorithms reduce conflict more effectively than others.


HiNeRV: Video Compression with Hierarchical Encoding-based Neural Representation

Ho Man Kwan, Ge Gao, Fan Zhang, Andy Gower, David Bull

https://openreview.net/forum?id=CpoS56pYnU

Keywords: Video compression, Implicit neural representations

Compressor summary: The paper introduces HiNeRV, a new Implicit Neural Representation for video compression that combines light weight layers with hierarchical positional encodings and achieves significant improvement over existing methods.


Newton–Cotes Graph Neural Networks: On the Time Evolution of Dynamic Systems

Lingbing Guo, Weiqing Wang, Zhuo Chen, Ningyu Zhang, Zequn Sun, Yixuan Lai, Qiang Zhang, Huajun Chen

https://openreview.net/forum?id=CnvZ7FIyAD

Keywords: Equivariant Graph Neural Networks, Molecular Dynamics, N-body System, Human Motion

Compressor summary: The paper introduces a new method for predicting the future state of system dynamics using velocity estimations and Newton–Cotes formulas, which improves upon existing GNNs-based approaches.


Reusable Slotwise Mechanisms

Trang Nguyen, Amin Mansouri, Kanika Madan, Nguyen Duy Khuong, Kartik Ahuja, Dianbo Liu, Yoshua Bengio

https://openreview.net/forum?id=CniUitfEY3

Keywords: Out-of-Distribution Generalization, Slotwise, Visual Reasoning, Video Prediction, Reusable Mechanism, Dynamics modeling

Compressor summary: The paper proposes a framework called Reusable Slotwise Mechanisms (RSM) that models object dynamics using communication among slots and a modular architecture, achieving improved robustness, generalization, and performance in various tasks.


When Visual Prompt Tuning Meets Source-Free Domain Adaptive Semantic Segmentation

Xinhong Ma, Yiming Wang, Hao Liu, Tianyu Guo, Yunhe Wang

https://openreview.net/forum?id=ChGGbmTNgE

Keywords: unsupervised domain adaptation, semantic segmentation, visual prompt tuning

Compressor summary: The text introduces a new method (Uni-UVPT) for adapting a semantic segmentation model to an unlabeled target domain without using the source data, by using a lightweight prompt adapter and a pseudo-label correction strategy.


PAPR: Proximity Attention Point Rendering

Yanshu Zhang, Shichong Peng, Seyed Alireza Moazenipourasil, Ke Li

https://openreview.net/forum?id=CgJJvuLjec

Keywords: point cloud learning, point cloud rendering

Compressor summary: PAPR is a novel point-based scene representation and differentiable renderer that learns accurate scene geometry and texture using a parsimonious set of points and demonstrates practical applications.


Smooth, exact rotational symmetrization for deep learning on point clouds

Sergey Pozdnyakov, Michele Ceriotti

https://openreview.net/forum?id=CdSRFn1fVe

Keywords: geometric deep learning, point clouds, equivariance, machine learning potentials, GNN, transformer, atomic-scale modeling

Compressor summary: The authors propose a method to add rotational symmetry to point-cloud models for chemical and materials modeling, and introduce the Point Edge Transformer (PET) architecture that achieves state-of-the-art performance on molecular and solid datasets.


Spuriosity Didn’t Kill the Classifier: Using Invariant Predictions to Harness Spurious Features

Cian Eastwood, Shashank Singh, Andrei Liviu Nicolicioiu, Marin Vlastelica, Julius von Kügelgen, Bernhard Schölkopf

https://openreview.net/forum?id=Cc2fjBBlBD

Keywords: invariant prediction, spurious correlations, out-of-distribution generalization, domain generalization, domain adaptation, test-time domain adaptation

Compressor summary: The paper proposes a method to improve performance in test domain by using pseudo-labels based on stable features, which are conditionally independent from unstable ones given the label.


In-Context Impersonation Reveals Large Language Models' Strengths and Biases

Leonard Salewski, Stephan Alaniz, Isabel Rio-Torto, Eric Schulz, Zeynep Akata

https://openreview.net/forum?id=CbsJ53LdKc

Keywords: large language models, impersonation, vision language models, reasoning

Compressor summary: The text explores how LLMs can take on different roles by assuming personas based on social identity or domain expertise, and how this affects their performance in various tasks, revealing both strengths and biases.


MMGP: a Mesh Morphing Gaussian Process-based machine learning method for regression of physical problems under nonparametrized geometrical variability

Fabien Casenave, Brian Staber, Xavier Roynard

https://openreview.net/forum?id=Ca78M3awPw

Keywords: Gaussian process, mesh morphing, mesh parametrization, finite element interpolation, simulation, physics, predictive uncertainties, nonparametrized geometries

Compressor summary: The authors propose a machine learning method that uses mesh morphing, dimensionality reduction, and Gaussian processes to model physical phenomena in industrial designs without relying on graph neural networks, providing predictive uncertainties and handling large meshes efficiently.


Survival Permanental Processes for Survival Analysis with Time-Varying Covariates

Hideaki Kim

https://openreview.net/forum?id=CYCzfXn6cZ

Keywords: survival analysis, temporal point process, Bayesian estimation, permanental process, representer theorem, kernel method

Compressor summary: The paper proposes a non-parametric Bayesian survival model for analyzing nonlinear dependence on time-varying covariates using a permanental process, which is fast and accurate.


DiffSketcher: Text Guided Vector Sketch Synthesis through Latent Diffusion Models

XiMing Xing, Chuang Wang, Haitao Zhou, Jing Zhang, Qian Yu, Dong Xu

https://openreview.net/forum?id=CY1xatvEQj

Keywords: Sketch; Vector Sketch; Sketch Generation; Diffusion Models

Compressor summary: The paper introduces DiffSketcher, an algorithm that creates vectorized free-hand sketches from natural language input using a pre-trained text-to-image diffusion model and optimizing Bézier curves with a modified loss function.


PETAL: Physics Emulation Through Averaged Linearizations for Solving Inverse Problems

Jihui Jin, Etienne Ollivier, Richard Touret, Matthew McKinley, Karim Sabra, Justin Romberg

https://openreview.net/forum?id=CXrRMfs5eY

Keywords: Inverse Problems, Neural Adjoint, Hybrid Machine Learning, Physics

Compressor summary: The paragraph discusses a new method for recovering an underlying signal using a learned weighted average model that incorporates physics-based linearizations, which improves accuracy and provides better gradient information for the inverse problem of ocean acoustic tomography.


SPAE: Semantic Pyramid AutoEncoder for Multimodal Generation with Frozen LLMs

Lijun Yu, Yong Cheng, Zhiruo Wang, Vivek Kumar, Wolfgang Macherey, Yanping Huang, David A Ross, Irfan Essa, Yonatan Bisk, Ming-Hsuan Yang, Kevin Patrick Murphy, Alexander G Hauptmann, Lu Jiang

https://openreview.net/forum?id=CXPUg86A1D

Keywords: multimodal, generation, large language model

Compressor summary: The paper presents SPAE, a technique that allows frozen large language models (LLMs) to perform image understanding and generation tasks by converting raw pixels to lexical tokens.


Fine-Grained Human Feedback Gives Better Rewards for Language Model Training

Zeqiu Wu, Yushi Hu, Weijia Shi, Nouha Dziri, Alane Suhr, Prithviraj Ammanabrolu, Noah A. Smith, Mari Ostendorf, Hannaneh Hajishirzi

https://openreview.net/forum?id=CSbGXyCswu

Keywords: Language Model, Reinforcement Learning with Human Feedback, Long-Form Text Generation

Compressor summary: The paper proposes Fine-Grained RLHF, a framework that uses reward functions based on fine-grained human feedback to train language models for better text generation.


Online Clustering of Bandits with Misspecified User Models

Zhiyong Wang, Jize Xie, Xutong Liu, Shuai Li, John C.S. Lui

https://openreview.net/forum?id=CQuRzAgjg9

Keywords: online clustering of bandits

Compressor summary: This paper introduces two robust clustering of bandits algorithms that can handle misspecified user models, achieving near-optimal regret bounds and outperforming existing methods in simulations.


Unbiased learning of deep generative models with structured discrete representations

Harry Bendekgey, Gabriel Hope, Erik B. Sudderth

https://openreview.net/forum?id=CQqBt46FUD

Keywords: Generative Models, Graphical Models, Variational Inference, Amortized Inference

Compressor summary: The authors propose a new generative model that combines graphical models and deep learning, handle multimodal uncertainty with latent variables, and improve optimization techniques to learn interpretable discrete data representations.


Counting Distinct Elements in the Turnstile Model with Differential Privacy under Continual Observation

Palak Jain, Iden Kalemaj, Sofya Raskhodnikova, Satchit Sivakumar, Adam Smith

https://openreview.net/forum?id=CQ38aC92WY

Keywords: distinct elements, differential privacy, continual release, turnstile streams

Compressor summary: The paper discusses privacy challenges and error bounds for systems that learn from data streams with inserts and deletions, and proposes a new item-level differentially private mechanism with low additive error for streams with low maximum flippancy.


Norm-based Generalization Bounds for Sparse Neural Networks

Tomer Galanti, Mengjia Xu, Liane Galanti, Tomaso Poggio

https://openreview.net/forum?id=COPzNA10hZ

Keywords: generalization bounds, convolution, rademacher, generalization, sparsity

Compressor summary: The paper develops norm-based generalization bounds for sparse ReLU neural networks that account for their sparse structure and convolutional filters, leading to tighter bounds than previous methods and better estimation of generalization.


Divide, Evaluate, and Refine: Evaluating and Improving Text-to-Image Alignment with Iterative VQA Feedback

Jaskirat Singh, Liang Zheng

https://openreview.net/forum?id=CLjBBd8u2j

Keywords: Text-to-Image Generation

Compressor summary: The paper proposes a decompositional approach to evaluate and improve text-to-image alignment, using a novel score that measures how well each assertion in a complex caption aligns with the generated image.


A Theory of Transfer-Based Black-Box Attacks: Explanation and Implications

Yanbo Chen, Weiwei Liu

https://openreview.net/forum?id=CJY7NEXVwC

Keywords: Learning Theory

Compressor summary: This paper proposes a theoretical model that explains the transferability of adversarial examples in black-box attacks and shows that it depends on the curvature of the data manifold.


Differentiable Random Partition Models

Thomas M. Sutter, Alain Ryser, Joram Liebeskind, Julia E Vogt

https://openreview.net/forum?id=CJWQGDwa6u

Keywords: random partition model, continuous relaxation, reparameterization, generative models, vae, representation learning, weak supervision, variational clustering, deep learning

Compressor summary: The paper proposes a novel method for inferring partitions in machine learning problems that allows gradient-based optimization and shows its versatility on three different challenging experiments.


Automatic Grouping for Efficient Cooperative Multi-Agent Reinforcement Learning

Yifan Zang, Jinmin He, Kai Li, Haobo Fu, QIANG FU, Junliang Xing, Jian Cheng

https://openreview.net/forum?id=CGj72TyGJy

Keywords: MARL, Cooperative Multi-Agent Reinforcement Learning, Coordination and Cooperation, Automatic Grouping, Group-Wise Learning

Compressor summary: The paper introduces GoMARL, a novel multi-agent reinforcement learning approach that learns automatic grouping for efficient cooperation in natural systems.


Dual Mean-Teacher: An Unbiased Semi-Supervised Framework for Audio-Visual Source Localization

Yuxin Guo, Shijie Ma, Hu Su, Zhiqing Wang, Yuhao Zhao, Wei Zou, Siyang Sun, Yun Zheng

https://openreview.net/forum?id=CFhpBJ8eZ5

Keywords: audio-visual learning, audio-visual source localization, semi-supervised learning

Compressor summary: The paper proposes a semi-supervised learning framework called Dual Mean-Teacher for audio-visual source localization, which improves localization accuracy by using two teachers to generate high-quality pseudo-labels and outperforms state-of-the-art methods with limited labeled data.


Adversarially Robust Learning with Uncertain Perturbation Sets

Tosca Lechner, Vinayak Pathak, Ruth Urner

https://openreview.net/forum?id=CFQBcz7k8n

Keywords: adversarially robust learning

Compressor summary: The paper studies a new setting of learning with imperfect knowledge of adversarial perturbation sets and proposes methods to improve robustness against them.


Meet in the Middle: A New Pre-training Paradigm

Anh Tuan Nguyen, Nikos Karampatziakis, Weizhu Chen

https://openreview.net/forum?id=CEk6JK71Mb

Keywords: language modeling, pre-training, deep learning, NLP

Compressor summary: The paper introduces a new language model pre-training method called Meet in the Middle (MIM) that trains models in both left-to-right and right-to-left directions, improving data efficiency and performance on generation and infilling tasks.


Refined Mechanism Design for Approximately Structured Priors via Active Regression

Christos Boutsikas, Petros Drineas, Marios Mertzanidis, Alexandros Psomas, Paritosh Verma

https://openreview.net/forum?id=CDTifMbUNc

Keywords: mechanism design, revenue maximization, randomized linear algebra, active regression

Compressor summary: This paper proposes a method to design revenue-maximizing mechanisms for sellers with many items and strategic bidders, using topic models and active learning techniques from randomized linear algebra.


Video Dynamics Prior: An Internal Learning Approach for Robust Video Enhancements

Gaurav Shrivastava, Ser-Nam Lim, Abhinav Shrivastava

https://openreview.net/forum?id=CCq73CGMyV

Keywords: Computational Photography, Deep Internal Learning, low-level vision, video denoising, video super-resolution, video frame interpolation, video inpainting

Compressor summary: The paper proposes a robust framework for low-level vision tasks that learns from corrupted test sequences and uses a novel spatial pyramid loss to handle noise and degradation in videos, achieving state-of-the-art results on several tasks.


Equivariant Neural Simulators for Stochastic Spatiotemporal Dynamics

Koen Minartz, Yoeri Poels, Simon Martinus Koop, Vlado Menkovski

https://openreview.net/forum?id=CCVsGbhFdj

Keywords: stochastic simulation, equivariance, dynamical systems, probabilistic simulation, generative models

Compressor summary: Equivariant Probabilistic Neural Simulation (EPNS) is a framework for simulating stochastic systems with improved accuracy, efficiency, and stability by incorporating domain symmetries into probabilistic neural networks.


Neural Data Transformer 2: Multi-context Pretraining for Neural Spiking Activity

Joel Ye, Jennifer L Collinger, Leila Wehbe, Robert Gaunt

https://openreview.net/forum?id=CBBtMnlTGq

Keywords: Pretraining, Scaling Laws, Neuroscience, Brain-computer interfaces

Compressor summary: NDT2 is a spatiotemporal Transformer that uses unsupervised pretraining to learn rich representations from neural spiking data across sessions, subjects, and tasks for iBCI control.


Information Maximization Perspective of Orthogonal Matching Pursuit with Applications to Explainable AI

Aditya Chattopadhyay, Ryan Pilgrim, Rene Vidal

https://openreview.net/forum?id=CAF4CnUblx

Keywords: Information Maximization, Sparse Coding, Orthogonal Matching Pursuit, Explainable AI, Information Pursuit

Compressor summary: The paper explores using Orthogonal Matching Pursuit (OMP) as an alternative to Information Pursuit (IP) for greedily selecting queries and shows how IP with random projections can almost reduce to OMP, leading to a simple explainable AI algorithm that encodes images as sparse combinations of semantically meaningful dictionary atoms.


Pointwise uncertainty quantification for sparse variational Gaussian process regression with a Brownian motion prior

Luke Travis, Kolyan Ray

https://openreview.net/forum?id=CA8tMQiscx

Keywords: Gaussian process, sparse variational Bayes, uncertainty quantification, theoretical guarantees

Compressor summary: The paper investigates how well a specific type of Gaussian process prior performs in estimating and quantifying uncertainty, and under what conditions it is accurate or misleading.


Bicriteria Multidimensional Mechanism Design with Side Information

Siddharth Prasad, Nina Balcan, Tuomas Sandholm

https://openreview.net/forum?id=C9wlNF1Ooj

Keywords: mechanism design, revenue maximization, welfare maximization, side information, weakest competitors, algorithms with predictions, learning-augmented algorithms

Compressor summary: The paper presents a new method for multidimensional mechanism design using side information, which can generate high social welfare and revenue while accounting for errors and varying quality of the input data.


Causal discovery from observational and interventional data across multiple environments

Adam Li, Amin Jaber, Elias Bareinboim

https://openreview.net/forum?id=C9wTM5xyw2

Keywords: causal inference, causal discovery, transportability, multi-domain learning

Compressor summary: The paper proposes an approach to learn causal structures in non-Markovian systems using data from multiple domains, by connecting interventional distributions with graphical criteria and introducing a new causal discovery algorithm called S-FCI.


Fast Projected Newton-like Method for Precision Matrix Estimation under Total Positivity

Jian-Feng CAI, José Vinícius De Miranda Cardoso, Daniel P. Palomar, Jiaxi Ying

https://openreview.net/forum?id=C9cgwmJ8Pt

Keywords: MTP2, Total Positivity, Generalized graph Laplacian, Precision matrix estimation, Nonnegative partial correlations

Compressor summary: The paper proposes a new algorithm for estimating precision matrices in Gaussian distributions with specific properties, which is faster and theoretically proven to converge.


Conservative Offline Policy Adaptation in Multi-Agent Games

Chengjie Wu, Pingzhong Tang, Jun Yang, Yujing Hu, Tangjie Lv, Changjie Fan, Chongjie Zhang

https://openreview.net/forum?id=C8pvL8Qbfa

Keywords: reinforcement learning, opponent exploitation, multi-agent

Compressor summary: The paper introduces a novel approach for offline policy adaptation in multi-agent games using conservative reinforcement learning that can exploit weaknesses or enable cooperation without online interaction.


Towards Label-free Scene Understanding by Vision Foundation Models

Runnan Chen, Youquan Liu, Lingdong Kong, Nenglun Chen, Xinge ZHU, Yuexin Ma, Tongliang Liu, Wenping Wang

https://openreview.net/forum?id=C8JdyM7B8I

Keywords: label-free, scene understanding

Compressor summary: The paper explores using CLIP and SAM for label-free scene understanding and proposes a novel method to supervise 2D and 3D networks with noisy labels from these models, improving performance on several datasets.


Self-Consistent Velocity Matching of Probability Flows

Lingxiao Li, Samuel Hurault, Justin Solomon

https://openreview.net/forum?id=C6fvJ2RfsL

Keywords: JKO, mass-conservation, PDE, Fokker-Planck, scalable, discretization-free, neural ODE

Compressor summary: The authors propose a scalable framework for solving mass-conserving PDEs using an iterative formulation with a biased gradient estimator that overcomes computational challenges and outperforms existing methods.


Dynamic Sparsity Is Channel-Level Sparsity Learner

Lu Yin, Gen Li, Meng Fang, Li Shen, Tianjin Huang, Zhangyang Wang, Vlado Menkovski, Xiaolong Ma, Mykola Pechenizkiy, Shiwei Liu

https://openreview.net/forum?id=C6IIwFHWkF

Keywords: dynamic sparsity, dynamic sparse training, sparse training

Compressor summary: Chase is a new dynamic sparse training method that converts unstructured sparsity to GPU-friendly channel-level sparsity, improving inference speedup without sacrificing accuracy.


Clustering the Sketch: Dynamic Compression for Embedding Tables

Henry Tsang, Thomas Dybdahl Ahle

https://openreview.net/forum?id=C4rRqkXFyC

Keywords: Embedding table compression, Clustering and sketching, Memory-efficient training

Compressor summary: CCE combines clustering-based compression with dynamic methods to efficiently represent large embedding tables for recommendation systems, and has theoretical convergence guarantees.


On the explainable properties of 1-Lipschitz Neural Networks: An Optimal Transport Perspective

Mathieu Serrurier, Franck Mamalet, Thomas FEL, Louis Béthune, Thibaut Boissin

https://openreview.net/forum?id=ByDy2mlkig

Keywords: 1-Lipschitz neural network, explicability

Compressor summary: The paper shows that 1-Lipschitz neural networks, trained with optimal transportation problem's dual loss, produce saliency maps with low noise and high XAI properties, aligning well with human explanations and allowing counterfactual explanations that reveal the direction of transportation plans.


Lookup Table meets Local Laplacian Filter: Pyramid Reconstruction Network for Tone Mapping

Feng Zhang, Ming Tian, Zhiqiang Li, Bin Xu, Qingbo Lu, Changxin Gao, Nong Sang

https://openreview.net/forum?id=BxqPN7KuQS

Keywords: tone mapping; learnable local laplacian filter; laplacian pyramid; 3D lookup table

Compressor summary: The paper proposes a novel strategy for converting HDR to LDR images using image-adaptive 3D LUTs and local Laplacian filters learned from annotated data, achieving better tone mapping and edge preservation than existing methods.


Self-Evaluation Guided Beam Search for Reasoning

Yuxi Xie, Kenji Kawaguchi, Yiran Zhao, Xu Zhao, Min-Yen Kan, Junxian He, Qizhe Xie

https://openreview.net/forum?id=Bw82hwg5Q3

Keywords: Large Language Models, Multistep Reasoning, Stochastic Beam Search, LLM Self-Evaluation

Compressor summary: The paper proposes a decoding algorithm that integrates self-evaluation guidance to improve the reasoning quality of large language models and outperforms baselines on various benchmarks.


Improving the Knowledge Gradient Algorithm

Le Yang, Siyang Gao, Chin Pang Ho

https://openreview.net/forum?id=BvslVXlUvF

Keywords: best arm identification, knowledge gradient, asymptotic optimality, convergence rate

Compressor summary: The improved knowledge gradient (iKG) algorithm is a better version of the knowledge gradient (KG) algorithm for identifying the best arm in multiple-arm bandit problems, as it is asymptotically optimal and can handle more variants of the problem.


Incomplete Multimodality-Diffused Emotion Recognition

Yuanzhi Wang, Yong Li, Zhen Cui

https://openreview.net/forum?id=BuGFwUS9B3

Keywords: Multimodal emotion recognition, Incomplete multimodalities

Compressor summary: IMDer is a method to improve multimodal emotion recognition by recovering missing modalities using a score-based diffusion model and embedding available modalities as conditions for semantic alignment.


Pretraining task diversity and the emergence of non-Bayesian in-context learning for regression

Allan Raventos, Mansheej Paul, Feng Chen, Surya Ganguli

https://openreview.net/forum?id=BtAz4a5xDg

Keywords: in-context learning, Bayesian inference, transformers, task diversity, emergence

Compressor summary: The study investigates how task diversity during pretraining affects transformers' ability to learn new linear regression tasks without updating weights, finding a threshold beyond which they can optimally solve unseen tasks.


Optimizing Prompts for Text-to-Image Generation

Yaru Hao, Zewen Chi, Li Dong, Furu Wei

https://openreview.net/forum?id=BsZNWXD3a1

Keywords: prompt adaptation, automatic prompt engineering, text-to-image generation

Compressor summary: Prompt adaptation is a framework that automatically improves text-to-image models by adapting user input to model preferences using supervised fine-tuning and reinforcement learning, resulting in better images that match user intentions.


Augmenting Language Models with Long-Term Memory

Weizhi Wang, Li Dong, Hao Cheng, Xiaodong Liu, Xifeng Yan, Jianfeng Gao, Furu Wei

https://openreview.net/forum?id=BryMFPQ4L6

Keywords: large language models, long-term memory, long-text modeling and understanding, residual side-network, in-context learning

Compressor summary: The LongMem framework allows large language models to store and use long past contexts, improving performance on long-context modeling tasks and memory-augmented in-context learning.


SANFlow: Semantic-Aware Normalizing Flow for Anomaly Detection

Daehyun Kim, Sungyong Baik, Tae Hyun Kim

https://openreview.net/forum?id=BqZ70BEtuW

Keywords: Anomaly Detection, Visual Anomaly Detection, Computer Vision, Normalizing Flow, Anomaly Localization

Compressor summary: The authors propose a method to improve anomaly detection using normalizing flow to map different features of an image to different distributions, enhancing discriminability between normal and abnormal data.


Designing Robust Transformers using Robust Kernel Density Estimation

Xing Han, Tongzheng Ren, Tan Minh Nguyen, Khai Nguyen, Joydeep Ghosh, Nhat Ho

https://openreview.net/forum?id=BqTv1Mtuhu

Keywords: Transformers, Kernel Density Estimation, Robustness

Compressor summary: The paper proposes new transformer architectures that are more robust to adversarial attacks and data contamination by adapting classical kernel density estimation methods to the self-attention mechanism.


A Computationally Efficient Sparsified Online Newton Method

Fnu Devvrit, Sai Surya Duvvuri, Rohan Anil, Vineet Gupta, Cho-Jui Hsieh, Inderjit S Dhillon

https://openreview.net/forum?id=BopG5dhH7L

Keywords: Optimization, Second order methods; Deep Learning

Compressor summary: The paper introduces Sparsified Online Newton (SONew), a memory-efficient second-order algorithm that improves convergence and performance for deep neural network training using structured sparsity patterns.


Noise-Adaptive Thompson Sampling for Linear Contextual Bandits

Ruitu Xu, Yifei Min, Tianhao Wang

https://openreview.net/forum?id=BnV2M2WFaY

Keywords: Linear Contextual Bandit, Thompson Sampling, Noise-Adaptive

Compressor summary: The paper proposes a noise-adaptive Thompson sampling-style algorithm for linear contextual bandits with heteroscedastic noise that achieves better regret bounds in both constant-variance and deterministic scenarios.


Koopman Kernel Regression

Petar Bevanda, Max Beier, Armin Lederer, Stefan Georg Sosnowski, Eyke Hüllermeier, Sandra Hirche

https://openreview.net/forum?id=BmIW6U0rz8

Keywords: Kernel Methods, Regression, Statistical Learning Theory, Koopman Operator, Mode Decomposition, Dynamical Systems, Supervised Learning

Compressor summary: The paragraph discusses a new method called Koopman Kernel Regression (KKR) that uses linear time-invariant equations to simplify complex forecasts, providing better learning guarantees and performance than existing approaches.


Learning to Discover Skills through Guidance

Hyunseung Kim, Byungkun Lee, Hojoon Lee, Dongyoon Hwang, Sejik Park, Kyushik Min, Jaegul Choo

https://openreview.net/forum?id=Bkrmr9LjeI

Keywords: Unsupervised skill discovery, Reinforcement Learning

Compressor summary: The paper proposes a new unsupervised skill discovery algorithm that improves exploration by selecting a guide skill and guiding other skills to follow it while maximizing their discriminability in unexplored states, and shows its effectiveness in challenging environments.


How many samples are needed to leverage smoothness?

Vivien Cabannes, Stefano Vigogna

https://openreview.net/forum?id=BklIgOO76D

Keywords: Statistical learning, breaking the curse, kernel methods

Compressor summary: This paper studies how smoothness helps overcome the curse of dimensionality in statistical learning and explores the impact of constants and transient periods on generalization error.


A normative theory of social conflict

Sergey A. Shuvaev, Evgeny M Amelchenko, Dmitry Smagin, Natalia Kudryavtseva, Grigori Enikolopov, Alexei A. Koulakov

https://openreview.net/forum?id=BkQM8huiIc

Keywords: neuroscience, decision-making, normative modeling, game theory, Bayesian methods, POMDP, inverse rational control, belief, theory of mind

Compressor summary: The study uses neural data from mice to develop a Theory of Mind model explaining how they form beliefs about strength and behavior in social conflict situations.


Participatory Personalization in Classification

Hailey James, Chirag Nagpal, Katherine A Heller, Berk Ustun

https://openreview.net/forum?id=Bj1QSgiBPP

Keywords: healthcare, algorithmic fairness, data privacy, classification, interpretability

Compressor summary: The authors propose participatory systems, which allow individuals to opt into personalization of prediction models using their group attributes, improving both performance and privacy.


Zero-shot causal learning

Hamed Nilforoshan, Michael Moor, Yusuf H Roohani, Yining Chen, Anja Šurina, Michihiro Yasunaga, Sara Oblak, Jure Leskovec

https://openreview.net/forum?id=BfQJrIiOZC

Keywords: causal inference, CATE, CATE estimation, causal machine learning, causal ML, heterogenous treatment effects, causality, potential outcomes, treatment effect

Compressor summary: CaML is a framework that predicts personalized effects of novel interventions using meta-learning and individual features, outperforming existing methods in medical and cell-line datasets.


MG-ViT: A Multi-Granularity Method for Compact and Efficient Vision Transformers

Yu Zhang, Yepeng Liu, Duoqian Miao, Qi Zhang, Yiwei Shi, Liang Hu

https://openreview.net/forum?id=Bf6WFWNCUP

Keywords: Efficient AI, Vision Transformer, Image Classification, Multi-Granularity, Three-Way Decisions

Compressor summary: The paper introduces a multi-granularity strategy for compressing Vision Transformers that reduces computational cost without sacrificing performance on classification, detection, and segmentation tasks.


The Contextual Lasso: Sparse Linear Models via Deep Neural Networks

Ryan Thompson, Amir Dezfouli, Robert Kohn

https://openreview.net/forum?id=BdvCo8RVlx

Keywords: feature selection, sparsity, sparse regression, varying coefficients, deep learning

Compressor summary: The contextual lasso is a new method for creating interpretable sparse linear models that adapt to different contexts using a deep neural network and a novel lasso regularizer.


LANCE: Stress-testing Visual Models by Generating Language-guided Counterfactual Images

Viraj Uday Prabhu, Sriram Yenamandra, Prithvijit Chattopadhyay, Judy Hoffman

https://openreview.net/forum?id=BbIxB4xnbq

Keywords: image classification, robustness, guided diffusion models, counterfactuals

Compressor summary: The paper introduces an algorithm to create realistic test images based on language input to stress-test visual models and reveal hidden biases.


Waypoint Transformer: Reinforcement Learning via Supervised Learning with Intermediate Targets

Anirudhan Badrinath, Yannis Flet-Berliac, Allen Nie, Emma Brunskill

https://openreview.net/forum?id=BYywOFbRFz

Keywords: offline reinforcement learning, reinforcement learning via supervised learning, behavioral cloning

Compressor summary: The Waypoint Transformer is a new reinforcement learning method that improves offline learning by using automatically-generated waypoints to guide the agent's trajectory.


Offline Multi-Agent Reinforcement Learning with Implicit Global-to-Local Value Regularization

Xiangsen Wang, Haoran Xu, Yinan Zheng, Xianyuan Zhan

https://openreview.net/forum?id=BXQtgwA2n0

Keywords: Offline reinforcement learning; multi-agent reinforcement learning; multi-agent cooperation

Compressor summary: OMIGA is a new offline multi-agent RL algorithm with implicit global-to-local value regularization that outperforms existing methods on various MuJoCo and StarCraft II tasks.


Mip-Grid: Anti-aliased Grid Representations for Neural Radiance Fields

Seungtae Nam, Daniel Rho, Jong Hwan Ko, Eunbyung Park

https://openreview.net/forum?id=BW6nZf7TnK

Keywords: Novel view synthesis, Neural radiance fields

Compressor summary: mip-Grid is a novel approach that improves grid-based neural radiance field methods by integrating anti-aliasing techniques, achieving fast training speed and high rendering quality.


From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Models to Pre-trained Machine Reader

Weiwen Xu, Xin Li, Wenxuan Zhang, Meng Zhou, Wai Lam, Luo Si, Lidong Bing

https://openreview.net/forum?id=BVN9Kgvwzv

Keywords: Machine Reading Comprehension, Pre-training, Natural Language Understanding

Compressor summary: PMR is a novel method that improves machine reading comprehension models by retrofitting pre-trained masked language models using Wikipedia data and a Wiki Anchor Extraction task, leading to better performance, explainability, and versatility in various extraction and classification tasks.


Greedy Pruning with Group Lasso Provably Generalizes for Matrix Sensing

Nived Rajaraman, Fnu Devvrit, Aryan Mokhtari, Kannan Ramchandran

https://openreview.net/forum?id=BTRcVP7ZJn

Keywords: Greedy Pruning; Matrix Sensing; Lasso regularization

Compressor summary: The paper investigates the theory behind pruning and fine-tuning methods for reducing model complexity in overparameterized matrix sensing problems, and shows that pruning columns with low norms leads to good generalization.


BQ-NCO: Bisimulation Quotienting for Efficient Neural Combinatorial Optimization

Darko Drakulic, Sofia Michel, Florian Mai, Arnaud Sors, Jean-Marc Andreoli

https://openreview.net/forum?id=BRqlkTDvvm

Keywords: Combinatorial Optimization, Markov Decision Processes, Bisimulation, Policy Learning, Out-of-Distribution Generalization, Routing Problems, TSP, CVRP, KP.

Compressor summary: The paper proposes a method for solving combinatorial optimization problems using Markov decision processes and bisimulation quotienting to exploit symmetries, achieving state-of-the-art results and better out-of-distribution generalization.


Passive learning of active causal strategies in agents and language models

Andrew Kyle Lampinen, Stephanie C.Y. Chan, Ishita Dasgupta, Andrew Joo Hun Nam, Jane X Wang

https://openreview.net/forum?id=BRpi8YAfac

Keywords: passive; causal; offline; agency; language models

Compressor summary: The paragraph discusses how passively-trained agents can learn generalizable strategies for determining and using causal structures by intervening at test time, and how natural language explanations can further enhance their performance in complex environments.


Optimal privacy guarantees for a relaxed threat model: Addressing sub-optimal adversaries in differentially private machine learning

Georgios Kaissis, Alexander Ziller, Stefan Kolek, Anneliese Riess, Daniel Rueckert

https://openreview.net/forum?id=BRSgVw85Mc

Keywords: Differential Privacy, Membership Inference Attack, Hypothesis Testing, Data Reconstruction Attack, Security

Compressor summary: The paper studies how private machine learning models can be attacked by less powerful adversaries with partial data, and suggests ways to set noise levels based on this risk.


Identifiable Contrastive Learning with Automatic Feature Importance Discovery

Qi Zhang, Yifei Wang, Yisen Wang

https://openreview.net/forum?id=BQA7wR2KBF

Keywords: Self-supervised Learning, Contrastive Learning, Identifiability, Representation Learning

Compressor summary: The paper introduces triCL, a contrastive learning method that uses a 3-factor contrast to learn more interpretable and identifiable features by using a learnable diagonal matrix $S$ that captures the importance of each feature.


Uncovering and Quantifying Social Biases in Code Generation

Yan Liu, Xiaokang Chen, Yan Gao, Zhe Su, Fengji Zhang, Daoguang Zan, Jian-Guang Lou, Pin-Yu Chen, Tsung-Yi Ho

https://openreview.net/forum?id=BOP5McdqGy

Keywords: Social bias, code fairness

Compressor summary: This paper investigates the social bias problem in automatic code generation tools like Copilot and proposes a method to construct code prompts that reveal and measure these biases.


Fast Model DeBias with Machine Unlearning

Ruizhe Chen, Jianfei Yang, Huimin Xiong, Jianhong Bai, Tianxiang Hu, Jin Hao, YANG FENG, Joey Tianyi Zhou, Jian Wu, Zuozhu Liu

https://openreview.net/forum?id=BL9Pc7xsdX

Keywords: Model Debias, Bias Mitigation, Machine Unlearning, Counterfactual Fairness

Compressor summary: The paper proposes FMD, an efficient method for identifying, evaluating, and removing biases in deep neural networks using counterfactuals and machine unlearning.


Outlier-Robust Gromov-Wasserstein for Graph Data

Lemin Kong, Jiajin Li, Jianheng Tang, Anthony Man-Cho So

https://openreview.net/forum?id=BKAFLUcpBS

Keywords: Gromov Wasserstein, Robust Optimization, Nonconvex Optimization

Compressor summary: The paper introduces a new robust version of Gromov-Wasserstein distance called RGW that minimizes the impact of outliers in graph learning tasks using efficient algorithms and validates its effectiveness on real data.


Retaining Beneficial Information from Detrimental Data for Neural Network Repair

Long-Kai Huang, Peilin Zhao, Junzhou Huang, Sinno Pan

https://openreview.net/forum?id=BJ1vOqh3hJ

Keywords: Model Repair; Fine-tuning

Compressor summary: The paper proposes a new approach to repair deep learning models by identifying and separating harmful data from beneficial data using a clean set and enhancing the model's performance.


Recommender Systems with Generative Retrieval

Shashank Rajput, Nikhil Mehta, Anima Singh, Raghunandan Hulikal Keshavan, Trung Vu, Lukasz Heldt, Lichan Hong, Yi Tay, Vinh Q. Tran, Jonah Samost, Maciej Kula, Ed H. Chi, Maheswaran Sathiamoorthy

https://openreview.net/forum?id=BJ0fQUU32w

Keywords: Recommender Systems, Generative Retrieval, Vector Quantization

Compressor summary: The paper proposes a generative retrieval approach using semantically meaningful Semantic IDs for items, which improves recommender system performance and generalization.


Flow: Per-instance Personalized Federated Learning

Kunjal Panchal, Sunav Choudhary, Nisarg Parikh, Lijun Zhang, Hui Guan

https://openreview.net/forum?id=BI031mw7iS

Keywords: federated learning, personalization, statistical heterogeneity, dynamic routing

Compressor summary: Flow is a per-instance personalization FL algorithm that adapts dynamic models to each client's data distributions and instances, improving clients' accuracy.


OKRidge: Scalable Optimal k-Sparse Ridge Regression

Jiachang Liu, Sam Rosen, Chudi Zhong, Cynthia Rudin

https://openreview.net/forum?id=BHxsP5fSHv

Keywords: Sparse Ridge Regression, Dynamical Systems

Compressor summary: The authors propose OKRidge, a fast algorithm for identifying sparse governing equations in nonlinear dynamical systems using a novel lower bound calculation and beam search warm-starting.


An Improved Relaxation for Oracle-Efficient Adversarial Contextual Bandits

Kiarash Banihashem, MohammadTaghi Hajiaghayi, Suho Shin, Max Springer

https://openreview.net/forum?id=BHZsJ2sTkG

Keywords: contextual bandits, adversarial bandits, oracle-efficient online learning

Compressor summary: The paper presents a new algorithm for adversarial contextual bandits that improves existing regret bounds and makes fewer calls to an offline optimization oracle per round.


ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings

Shibo Hao, Tianyang Liu, Zhen Wang, Zhiting Hu

https://openreview.net/forum?id=BHXsb69bSx

Keywords: large language model, tool learning

Compressor summary: The authors propose ToolkenGPT, which enables large language models to master tools by using vector embeddings of tools that are plugged into the LLM head and executed during text generation.


Regularity as Intrinsic Reward for Free Play

Cansu Sancaktar, Justus Piater, Georg Martius

https://openreview.net/forum?id=BHHrX3CRE1

Keywords: Intrinsic Motivation, Reinforcement Learning, Model-based Planning, Regularity, Manipulation, Zero-shot Generalization, Unsupervised Exploration

Compressor summary: We propose regularity as a novel reward signal for intrinsically-motivated reinforcement learning. Taking inspiration from child development, we postulate that striving for structure and order helps guide exploration towards a subspace of tasks that are not favored by naive uncertainty-based intrinsic rewards. Our generalized formulation of Regularity as Intrinsic Reward (RaIR) allows us to operationalize it within model-based reinforcement learning. In a synthetic environment, we showcase the plethora of structured patterns that can emerge from pursuing this regularity objective. We also demonstrate the strength of our method in a multi-object robotic manipulation environment. We incorporate RaIR into free play and use it to complement the model’s epistemic uncertainty as an intrinsic reward. Doing so, we witness the autonomous construction of towers and other regular structures during free play, which leads to a substantial improvement in zero-shot downstream task performance on assembly tasks.


Large Language Models Are Latent Variable Models: Explaining and Finding Good Demonstrations for In-Context Learning

Xinyi Wang, Wanrong Zhu, Michael Saxon, Mark Steyvers, William Yang Wang

https://openreview.net/forum?id=BGvkwZEGt7

Keywords: Large language models, Bayesian explanation, in-context learning

Compressor summary: The study examines in-context learning using Bayesian methods and proposes an algorithm to select optimal demonstrations for few-shot text classification, improving performance on real-world datasets.


Predicting mutational effects on protein-protein binding via a side-chain diffusion probabilistic model

Shiwei Liu, Tian Zhu, Milong Ren, Yu Chungong, Dongbo Bu, Haicang Zhang

https://openreview.net/forum?id=BGP5Vjt93A

Keywords: Riemannian Diffusion Probabilistic Model, Mutation, Protein-protein binding

Compressor summary: SidechainDiff is a new method that uses unlabelled experimental data to predict how mutations affect protein-protein binding by learning side-chain conformations and structural context representations.


Score-based Source Separation with Applications to Digital Communication Signals

Tejas Jayashankar, Gary C.F. Lee, Alejandro Lancho, Amir Weiss, Yury Polyanskiy, Gregory Wornell

https://openreview.net/forum?id=BFGQQKicuu

Keywords: Diffusion models, score-based models, source separation, digital communications, maximum a posteriori (MAP) estimation, alpha-posterior, Gaussian smoothing, score distillation sampling, radio frequency systems, interference mitigation

Compressor summary: The authors propose a new method for separating mixed signals using diffusion-based generative models that improves bit error rate in radio-frequency systems and has applications beyond conditional sampling.


HyP-NeRF: Learning Improved NeRF Priors using a HyperNetwork

Bipasha Sen, Gaurav Singh, Aditya Agarwal, Rohith Agaram, Madhava Krishna, Srinath Sridhar

https://openreview.net/forum?id=BExDjNDYkN

Keywords: neural radiance field, hypernetwork, multi-hash encoding, NeRF

Compressor summary: HyP-NeRF is a method for learning generalizable NeRF priors using hypernetworks that improves quality and enables multiple downstream tasks.


TensorNet: Cartesian Tensor Representations for Efficient Learning of Molecular Potentials

Guillem Simeon, Gianni De Fabritiis

https://openreview.net/forum?id=BEHlPdBZ2e

Keywords: Neural network interatomic potentials, Equivariant graph neural network, Message passing neural network

Compressor summary: TensorNet is a novel neural network architecture that uses Cartesian tensor representations to efficiently model molecular systems with reduced computational cost and improved performance.


Multi-Object Representation Learning via Feature Connectivity and Object-Centric Regularization

Alex Foo, Wynne Hsu, Mong-Li Lee

https://openreview.net/forum?id=BDno5qWEFh

Keywords: Object-Centric Learning, Multi-Object Representation Learning

Compressor summary: The paragraph describes a new method for discovering object-centered representations from images that improves robustness, sample efficiency, and interpretability, and outperforms existing methods on various types of real-world images.


Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models

Alvin Heng, Harold Soh

https://openreview.net/forum?id=BC1IJdsuYB

Keywords: generative models, forgetting

Compressor summary: Selective Amnesia is a technique that allows controlling the forgetting of specific concepts in deep generative models to prevent misuse for harmful or inappropriate content.


Subspace Identification for Multi-Source Domain Adaptation

Zijian Li, Ruichu Cai, Guangyi Chen, Boyang Sun, Zhifeng Hao, Kun Zhang

https://openreview.net/forum?id=BACQLWQW8u

Keywords: Domain Adaptation, Identification

Compressor summary: The paper proposes a new method for transferring knowledge between multiple labeled source domains to an unlabeled target domain using subspace identification and variational inference, which works better than existing methods.


Sampling from Structured Log-Concave Distributions via a Soft-Threshold Dikin Walk

Oren Mangoubi, Nisheeth K Vishnoi

https://openreview.net/forum?id=BA7NHAzbpO

Keywords: Logconcave sampling, Dikin walk, Markov chain Monte Carlo, interior point methods

Compressor summary: The paper presents a new Dikin walk algorithm to sample from a log-concave distribution defined by a Lipschitz or smooth convex function on a bounded polytope, with improved efficiency and acceptance ratio compared to previous works.


CamoPatch: An Evolutionary Strategy for Generating Camoflauged Adversarial Patches

Phoenix Neale Williams, Ke Li

https://openreview.net/forum?id=B94G0MXWQX

Keywords: Evolutionary Strategy, Adversarial Attack, Adversarial Patches, Computational Art, Computer Vision

Compressor summary: Deep neural networks (DNNs) have demonstrated vulnerabilities to adversarial examples, which raises concerns about their reliability in safety-critical applications. While the majority of existing methods generate adversarial examples by making small modifications to the entire image, recent research has proposed a practical alternative known as adversarial patches. Adversarial patches have shown to be highly effective in causing DNNs to misclassify by distorting a localized area (patch) of the image. However, existing methods often produce clearly visible distortions since they do not consider the visibility of the patch. To address this, we propose a novel method for constructing adversarial patches that approximates the appearance of the area it covers. We achieve this by using a set of semi-transparent, RGB-valued circles, drawing inspiration from the computational art community. We utilize an evolutionary strategy to optimize the properties of each shape, and employ a simulated annealing approach to optimize the patch's location. Our approach achieves better or comparable performance to state-of-the-art methods on ImageNet DNN classifiers while achieving a lower $l_2$ distance from the original image. By minimizing the visibility of the patch, this work further highlights the vulnerabilities of DNNs to adversarial patches.


Optimization and Bayes: A Trade-off for Overparameterized Neural Networks

Zhengmian Hu, Heng Huang

https://openreview.net/forum?id=B7QkdEnjL9

Keywords: Bayesian learning, Generalization

Compressor summary: The paper introduces TansBL, which combines ERM and Bayesian learning for neural networks, balancing generalization and computational complexity.


An Alternative to Variance: Gini Deviation for Risk-averse Policy Gradient

Yudong Luo, Guiliang Liu, Pascal Poupart, Yangchen Pan

https://openreview.net/forum?id=B7QRV4XXiK

Keywords: risk-averse RL, mean-variance RL

Compressor summary: The paper proposes a new risk measure for Reinforcement Learning, called Gini deviation, which addresses the limitations of variance-based methods and improves policy learning.


POMDP Planning for Object Search in Partially Unknown Environment

Yongbo Chen, Hanna Kurniawati

https://openreview.net/forum?id=B6qZdrGRpm

Keywords: robotics, Partially Observable Markov Decision Process (POMDP), object search

Compressor summary: The paper proposes a method for mobile robots to search for objects in complex environments using a POMDP formulation, a perception module, and a planning algorithm that improves performance over baseline approaches.


A Batch-to-Online Transformation under Random-Order Model

Jing Dong, Yuichi Yoshida

https://openreview.net/forum?id=B6HSIgvyJ3

Keywords: online learning, random model setting

Compressor summary: The framework transforms offline approximation algorithms into online ones with low $\epsilon$-approximate regret and applies to various problems like clustering, matrix approximation, and regression.


Universal Gradient Descent Ascent Method for Nonconvex-Nonconcave Minimax Optimization

Taoli Zheng, Linglingzhi Zhu, Anthony Man-Cho So, Jose Blanchet, Jiajin Li

https://openreview.net/forum?id=B6FihisDBl

Keywords: Nonconvex-Nonconcave Minimax Optimization, Limit Cycle

Compressor summary: The paper proposes a new optimization algorithm, DS-GDA, that works for various nonconvex-nonconcave problems and achieves convergence with better complexity than existing methods.


Communication-Efficient Federated Bilevel Optimization with Global and Local Lower Level Problems

Junyi Li, Feihu Huang, Heng Huang

https://openreview.net/forum?id=B5XwENgy0T

Keywords: Federated Learning, Bilevel Optimization

Compressor summary: The paper proposes FedBiOAcc, a communication-efficient algorithm for Federated Bilevel Optimization problems with good theoretical and empirical performance.


Online Nonstochastic Model-Free Reinforcement Learning

Udaya Ghai, Arushi Gupta, Wenhan Xia, Karan Singh, Elad Hazan

https://openreview.net/forum?id=B5LpWAaBVA

Keywords: Control, Reinforcement Learning, Online Learning, Regret Minimization, Bandit Linear Control

Compressor summary: The authors propose a novel policy class based on disturbance signals and introduce efficient algorithms for optimizing them, improving robustness in model-free reinforcement learning agents facing adversarial or dynamic environments.


Optimal Time Complexities of Parallel Stochastic Optimization Methods Under a Fixed Computation Model

Alexander Tyurin, Peter Richtárik

https://openreview.net/forum?id=B4xF1wfQnF

Keywords: nonconvex optimization, convex optimization, parallel methods, asynchronous methods, lower bounds

Compressor summary: The paper introduces a new protocol for parallel optimization methods with an unbiased stochastic gradient oracle, establishes minimax complexities, and reveals implications for asynchronous optimization methods.


Grammar Prompting for Domain-Specific Language Generation with Large Language Models

Bailin Wang, Zi Wang, Xuezhi Wang, Yuan Cao, Rif A. Saurous, Yoon Kim

https://openreview.net/forum?id=B4tkwuzeiY

Keywords: semantic parsing, large language models, PDDL, AI planning, molecule generation, data efficiency, grammar-based learning

Compressor summary: Grammar prompting helps large language models generate outputs for structured languages by using external knowledge and domain-specific constraints from a grammar in BNF during in-context learning.


Fast Approximation of Similarity Graphs with Kernel Density Estimation

Peter Macgregor, He Sun

https://openreview.net/forum?id=B4G87Bq5wA

Keywords: similarity graphs, spectral clustering

Compressor summary: The paper proposes a new algorithm to efficiently construct a sparse similarity graph that preserves cluster structure for data points in high dimensions using kernel density estimation.


Recurrent Temporal Revision Graph Networks

YIZHOU CHEN, Anxiang Zeng, Qingtao Yu, Kerui Zhang, Cao Yuanpeng, Kangle Wu, Guangda Huzhang, Han Yu, Zhiming Zhou

https://openreview.net/forum?id=B3UDx1rNOy

Keywords: temporal graph, temporal graph network, temporal graph model expressiveness, continuous-time dynamic graph

Compressor summary: The paragraph discusses a novel framework for temporal neighbor aggregation that uses recurrent neural networks to integrate information from all historical neighbors, improving accuracy and expressiveness in graph networks.


Counterfactually Comparing Abstaining Classifiers

Yo Joong Choe, Aditya Gangrade, Aaditya Ramdas

https://openreview.net/forum?id=B2DEcj4a7i

Keywords: abstaining classifiers, black-box model evaluation, causal inference, missing data

Compressor summary: The paper proposes a method to evaluate abstaining classifiers by treating their abstentions as missing data and estimating their counterfactual performance.


Parameterizing Context: Unleashing the Power of Parameter-Efficient Fine-Tuning and In-Context Tuning for Continual Table Semantic Parsing

Yongrui Chen, Shenyu Zhang, Guilin Qi, Xinnan Guo

https://openreview.net/forum?id=B01uiWhjpc

Keywords: semantic parsing, continual learning, few-shot learning

Compressor summary: This paper introduces a novel method combining parameter-efficient fine-tuning and in-context tuning for training a continual table semantic parser that overcomes challenges like catastrophic forgetting and limited supervision.


Partial Counterfactual Identification of Continuous Outcomes with a Curvature Sensitivity Model

Valentyn Melnychuk, Dennis Frauen, Stefan Feuerriegel

https://openreview.net/forum?id=AygwZzdCM0

Keywords: causal inference, counterfactual inference, partial identification, sensitivity model, normalizing flows, causal machine learning

Compressor summary: The paper proposes a novel sensitivity model and deep generative model to relax assumptions and achieve partial counterfactual identification of continuous outcomes in structural causal models.


Model-Free Reinforcement Learning with the Decision-Estimation Coefficient

Dylan J Foster, Noah Golowich, Jian Qian, Alexander Rakhlin, Ayush Sekhari

https://openreview.net/forum?id=Ay3WvSrtpO

Keywords: Decision making, learning theory, bandits, reinforcement learning theory, online learning, decision-estimation coefficient

Compressor summary: The paper proposes a method to improve interactive decision making by combining a meta-algorithm with a specialized estimation technique, achieving better guarantees and accommodating more lenient notions of estimation error.


ProteinNPT: Improving protein property prediction and design with non-parametric transformers

Pascal Notin, Ruben Weitzman, Debora Susan Marks, Yarin Gal

https://openreview.net/forum?id=AwzbQVuDBk

Keywords: Non-parametric transformers, protein design, protein property prediction, fitness prediction, Bayesian optimization, ProteinGym

Compressor summary: ProteinNPT is a non-parametric transformer variant for protein sequence optimization that outperforms existing methods and enables iterative protein design in multi-task settings with label scarcity.


Extraction and Recovery of Spatio-Temporal Structure in Latent Dynamics Alignment with Diffusion Model

Yule Wang, Zijing Wu, Chengrui Li, Anqi Wu

https://openreview.net/forum?id=AuXd54odxm

Keywords: Neural distribution alignment, Diffusion model, Neuroscience, Neural decoding

Compressor summary: The proposed ERDiff method aligns neural signals across domains by using a diffusion model to preserve spatio-temporal structures in latent dynamics, resulting in better alignment quality and neural decoding performance.


Calibrate and Boost Logical Expressiveness of GNN Over Multi-Relational and Temporal Graphs

Yeyuan Chen, Dingmin Wang

https://openreview.net/forum?id=AtHJ7TLheF

Keywords: Knowledge Graphs, First-Order Logic, Temporal Knowledge Graph, Graph Neural Networks

Compressor summary: The paper analyzes the expressiveness of Graph Neural Networks (GNNs) for Boolean node classification over multi-relational graphs, proposes a graph transformation technique to improve GNNs' capabilities, and tests their approach empirically.


FeCAM: Exploiting the Heterogeneity of Class Distributions in Exemplar-Free Continual Learning

Dipam Goswami, Yuyang Liu, Bartłomiej Twardowski, Joost van de Weijer

https://openreview.net/forum?id=Asx5eDqFZl

Keywords: Continual Learning, Class-Incremental Learning

Compressor summary: The paper proposes a prototypical network approach for class-incremental learning that uses anisotropic Mahalanobis distance and feature covariance modeling, achieving state-of-the-art results without updating the backbone network.


Neural Lyapunov Control for Discrete-Time Systems

Junlin Wu, Andrew Clark, Yiannis Kantaros, Yevgeniy Vorobeychik

https://openreview.net/forum?id=ArRycLMoUg

Keywords: nonlinear systems, Lyapunov stability, neural Lyapunov control

Compressor summary: The authors propose a novel method for learning neural Lyapunov control in discrete-time systems using mixed-integer linear programming, sublevel sets computation, and gradient-based counterexample finding, achieving faster and better results on four benchmarks than existing approaches.


Resetting the Optimizer in Deep RL: An Empirical Study

Kavosh Asadi, Rasool Fakoor, Shoham Sabach

https://openreview.net/forum?id=AnFUgNC3Yc

Keywords: Deep Reinforcement Learning, Rainbow Algorithm, Atari benchmark, Adam Optimizer

Compressor summary: The text discusses how resetting internal parameters of optimization algorithms, such as Adam, can improve the performance of deep reinforcement learning on Atari games.


Causal Context Connects Counterfactual Fairness to Robust Prediction and Group Fairness

Jacy Reese Anthis, Victor Veitch

https://openreview.net/forum?id=AmwgBjXqc3

Keywords: causal graphs, causality, counterfactual fairness, domain generalization, fairness, robustness, machine learning, artificial intelligence

Compressor summary: The paper proposes a causal context approach to relate counterfactual fairness, robust prediction, and group fairness in data-generating processes and shows how counterfactual fairness is equivalent to different group fairness metrics in various scenarios.


You Only Condense Once: Two Rules for Pruning Condensed Datasets

Yang He, Lingao Xiao, Joey Tianyi Zhou

https://openreview.net/forum?id=AlTyimRsLf

Keywords: Dataset Condensation, Dataset Pruning

Compressor summary: YOCO is a method to adjust the size of a condensed dataset for efficient on-device training using simple rules and improving accuracy.


Phase diagram of early training dynamics in deep neural networks: effect of the learning rate, depth, and width

Dayal Singh Kalra, Maissam Barkeshli

https://openreview.net/forum?id=Al9yglQGKj

Keywords: Optimization dynamics, Phase diagrams, learning rate transition, Catapult effect

Compressor summary: The paper studies how the learning rate, depth, and width of deep neural networks affect their optimization dynamics and the sharpness of the loss landscape during training with stochastic gradient descent.


Towards Anytime Classification in Early-Exit Architectures by Enforcing Conditional Monotonicity

Metod Jazbec, James Urquhart Allingham, Dan Zhang, Eric Nalisnick

https://openreview.net/forum?id=Akslsk891N

Keywords: anytime algorithms, early-exit neural networks, conditional monotonicity, anytime uncertainty

Compressor summary: The paper proposes a modification to early-exit neural networks that allows them to provide better predictions as computation time increases, enabling anytime predictive modeling with these architectures.


Strategic Data Sharing between Competitors

Nikita Tsoy, Nikola Konstantinov

https://openreview.net/forum?id=AkK3S2spZs

Keywords: Incentives, collaborative learning, federated learning, game theory, competition, oligopolistic markets, strategic behavior, Nash equilibrium

Compressor summary: The paper presents a framework to analyze how data sharing affects machine learning model quality and profitability in collaborative learning settings, and finds that market conditions and task difficulty influence collaboration incentives.


A Deep Instance Generative Framework for MILP Solvers Under Limited Data Availability

Zijie Geng, Xijun Li, Jie Wang, Xiao Li, Yongdong Zhang, Feng Wu

https://openreview.net/forum?id=AiEipk1X0c

Keywords: Learning to Optimize, Machine Learning for Combinatorial Optimization, Mixed-Integer Linear Programming, Graph Generation

Compressor summary: G2MILP is a deep generative framework that creates realistic and novel mixed-integer linear programs (MILPs) from bipartite graphs without prior expert knowledge, improving MILP solver performance under limited data availability.


Optimal Excess Risk Bounds for Empirical Risk Minimization on $p$-Norm Linear Regression

Ayoub El Hanchi, Murat A Erdogdu

https://openreview.net/forum?id=Ah2Q8mLH96

Keywords: Excess risk bounds, Linear regression, Lp-norm, Fast rates

Compressor summary: The paper investigates how many samples are needed for different p-norms in linear regression and proves bounds on the performance of empirical risk minimization.


Multi-task Representation Learning for Pure Exploration in Bilinear Bandits

Subhojyoti Mukherjee, Qiaomin Xie, Josiah P. Hanna, Robert D Nowak

https://openreview.net/forum?id=AfC8PVQZ9z

Keywords: Linear Bandits, Experimental design, Pure Exploration, Representation Learning

Compressor summary: The paper proposes GOBLIN, a multi-task representation learning algorithm for bilinear bandits that reduces exploration time and improves sample complexity by sharing a common low-dimensional linear representation across tasks.


To Repeat or Not To Repeat: Insights from Scaling LLM under Token-Crisis

Fuzhao Xue, Yao Fu, Wangchunshu Zhou, Zangwei Zheng, Yang You

https://openreview.net/forum?id=Af5GvIj3T5

Keywords: Large Language Model, Transformer Scaling, Foundation Model Pre-training

Compressor summary: The study investigates how repeating pre-training data affects large language models, causing overfitting and multi-epoch degradation, and explores factors contributing to it, as well as regularization techniques and hyper-parameter tuning methods.


No-Regret Online Prediction with Strategic Experts

Omid Sadeghi, Maryam Fazel

https://openreview.net/forum?id=AesN5bYnJr

Keywords: incentive-compatible, online prediction with expert advice, forecasting

Compressor summary: The authors study a generalization of online binary prediction with expert advice where experts act strategically and aim to maximize their influence, and they design algorithms that encourage truthful reporting and achieve sublinear regret.


A Unified Solution for Privacy and Communication Efficiency in Vertical Federated Learning

Ganyu Wang, Bin Gu, Qingsong Zhang, Xiang Li, Boyu Wang, Charles Ling

https://openreview.net/forum?id=AYiRHZirD2

Keywords: Vertical Federated Learning, Zeroth Order Optimization, Communication Efficiency, Privacy

Compressor summary: The paper proposes a cascaded hybrid optimization approach for Vertical Federated Learning that improves convergence and privacy protection by using a zeroth-order gradient on the output layer and a first-order gradient elsewhere, with experimental results showing its advantages over existing methods.


Two Heads are Better Than One: A Simple Exploration Framework for Efficient Multi-Agent Reinforcement Learning

Jiahui Li, Kun Kuang, Baoxiang Wang, Xingchen Li, Fei Wu, Jun Xiao, Long Chen

https://openreview.net/forum?id=AYLlZMmUbo

Keywords: multi-agent reinforcement learning, influence-based exploration

Compressor summary: COIN is an exploration method for cooperative multi-agent reinforcement learning that combines curiosity-based and influence-based intrinsic rewards to effectively explore large and complex state spaces.


Every Parameter Matters: Ensuring the Convergence of Federated Learning with Dynamic Heterogeneous Models Reduction

Hanhan Zhou, Tian Lan, Guru Prasadh Venkataramani, Wenbo Ding

https://openreview.net/forum?id=AWpWaub6nf

Keywords: Federated Learning, Optimization, Deep Learning

Compressor summary: The paper proposes a framework for heterogeneous FL algorithms with online model extraction and proves their convergence under certain conditions for IID and non-IID data, while introducing the concept of minimum coverage index to improve efficiency.


Enhancing User Intent Capture in Session-Based Recommendation with Attribute Patterns

Xin Liu, Zheng Li, Yifan Gao, Jingfeng Yang, Tianyu Cao, Zhengyang Wang, Bing Yin, Yangqiu Song

https://openreview.net/forum?id=AV3iZlDrzF

Keywords: session-based recommendation, representation learning, pattern mining

Compressor summary: FAPAT is a new method for session-based recommendation in E-commerce that uses attribute transition graphs and frequent attribute patterns to capture user intents and outperforms existing methods by 4.5% on average.


A Reduction-based Framework for Sequential Decision Making with Delayed Feedback

Yunchang Yang, Han Zhong, Tianhao Wu, Bin Liu, Liwei Wang, Simon Shaolei Du

https://openreview.net/forum?id=AT6NaLPwy0

Keywords: sequential decision making, delay, reinforcement learning

Compressor summary: The paper presents a framework to improve sample efficiency in sequential decision making with stochastic delays using different multi-batched algorithms.


Dense and Aligned Captions (DAC) Promote Compositional Reasoning in VL Models

Sivan Doveh, Assaf Arbelle, Sivan Harary, Roei Herzig, Donghyun Kim, Paola Cascante-Bonilla, Amit Alfassy, Rameswar Panda, Raja Giryes, Rogerio Feris, Shimon Ullman, Leonid Karlinsky

https://openreview.net/forum?id=ARrwf7Ev2T

Keywords: computer vision, deep learning, vision and language models

Compressor summary: The paper identifies two factors that limit VL models' compositional reasoning performance and proposes a fine-tuning approach to improve it on a standard paired VL dataset, leading to significant performance increase.


Differentiable Sampling of Categorical Distributions Using the CatLog-Derivative Trick

Lennert De Smet, Emanuele Sansone, Pedro Zuidberg Dos Martires

https://openreview.net/forum?id=AQyqxXctsN

Keywords: gradient estimation, categorical random variables, probability theory, discrete distributions

Compressor summary: The CatLog-Derivative trick improves upon the Log-Derivative trick by accounting for the discrete nature of categorical distributions, leading to a more efficient and unbiased gradient estimator called IndeCateR.


When Do Transformers Shine in RL? Decoupling Memory from Credit Assignment

Tianwei Ni, Michel Ma, Benjamin Eysenbach, Pierre-Luc Bacon

https://openreview.net/forum?id=APGXBNkt6h

Keywords: Memory-based RL, Transformers, Credit Assignment, Online RL, Model-free RL

Compressor summary: The paper investigates why Transformer architecture works well for reinforcement learning problems involving long-term dependencies, finding that it enhances memory capability but not credit assignment.


Linguistic Binding in Diffusion Models: Enhancing Attribute Correspondence through Attention Map Alignment

Royi Rassin, Eran Hirsch, Daniel Glickman, Shauli Ravfogel, Yoav Goldberg, Gal Chechik

https://openreview.net/forum?id=AOKU4nRw1W

Keywords: syntax, diffusion, stable diffusion, attribute, attention

Compressor summary: SynGen improves text-to-image generation by using syntax to guide cross-attention maps and maintain correct associations between entities and their visual attributes.


3D-Aware Visual Question Answering about Parts, Poses and Occlusions

Xingrui Wang, Wufei Ma, Zhuowan Li, Adam Kortylewski, Alan Yuille

https://openreview.net/forum?id=AMIJEupsNq

Keywords: VQA, reasoning, 3D scene understanding, analysis-by-synthesis, neural modular network, neuro-symbolic reasoning

Compressor summary: The paragraph introduces the task of 3D-aware VQA, which involves answering questions about the 3D structure of visual scenes, and presents a new dataset (Super-CLEVR-3D) and a novel model (PO3D-VQA) to address this challenge.


Boosting Adversarial Transferability by Achieving Flat Local Maxima

Zhijin Ge, Xiaosen Wang, Hongying Liu, Fanhua Shang, Yuanyuan Liu

https://openreview.net/forum?id=AKAMNDe2Sw

Keywords: Adversarial attack, Adversarial transferability, Black-box Attack

Compressor summary: The authors propose a method to generate adversarial examples that transfer better by optimizing the gradient norm and using random sampling and averaging of gradients, achieving improved transferability on both normal and adversarial models.


Fantastic Robustness Measures: The Secrets of Robust Generalization

Hoki Kim, Jinseong Park, Yujin Choi, Jaewook Lee

https://openreview.net/forum?id=AGVBqJuL0T

Keywords: Adversarial Robustness, Generalization, Measures

Compressor summary: The study analyzes how different measures of robust generalization relate to adversarial examples in various settings using over 1,300 models trained on CIFAR-10 and more than 100 models from RobustBench.


Active Bipartite Ranking

James Cheshire, Vincent Laurent, Stephan Clémençon

https://openreview.net/forum?id=AGMVzMGcGP

Keywords: bipartite ranking, multi armed bandits, active learning

Compressor summary: The paper presents an active learning framework for bipartite ranking and proposes a new algorithm called active-rank that minimizes the distance between the ROC curve of the ranking function and the optimal one.


DIFFER:Decomposing Individual Reward for Fair Experience Replay in Multi-Agent Reinforcement Learning

Xunhan Hu, Jian Zhao, Wengang Zhou, Ruili Feng, Houqiang Li

https://openreview.net/forum?id=AG9A7Ae9r3

Keywords: Experience Replay; Reinforcement Learning; Multi-Agent System

Compressor summary: The paper proposes DIFFER, a framework to decompose individual rewards in cooperative multi-agent reinforcement learning for more efficient and fair learning.


Training Energy-Based Normalizing Flow with Score-Matching Objectives

Chen-Hao Chao, Wei-Fang Sun, Yen-Chang Hsu, Zsolt Kira, Chun-Yi Lee

https://openreview.net/forum?id=AALLvnv95q

Keywords: flow-based models, score-matching methods

Compressor summary: The paper introduces EBFlow, a flow-based modeling method that connects flow- and energy-based models, optimizes with score-matching for faster training and better performance than existing methods.


Universal Online Learning with Gradient Variations: A Multi-layer Online Ensemble Approach

Yu-Hu Yan, Peng Zhao, Zhi-Hua Zhou

https://openreview.net/forum?id=AA1xrgAP5z

Keywords: online learning

Compressor summary: The paper proposes an online convex optimization method that adapts to different function types and can achieve problem-dependent guarantees, while using only one gradient query per round.


NAS-X: Neural Adaptive Smoothing via Twisting

Dieterich Lawson, Michael Y. Li, Scott Linderman

https://openreview.net/forum?id=A9mHph8GJk

Keywords: sequence models, probabilistic inference, reweighted wake-sleep, sequential monte carlo, smoothing, mechanistic models

Compressor summary: NAS-X is a method for approximating inferences and model learning in complex sequential latent variable models using neural adaptive smoothing and twisting, which improves performance over previous methods in various applications, including neuronal dynamics.


AGD: an Auto-switchable Optimizer using Stepwise Gradient Difference for Preconditioning Matrix

Yun Yue, Zhiling Ye, Jiadi Jiang, Yongchao Liu, Ke Zhang

https://openreview.net/forum?id=A954O4tDmU

Keywords: adaptive optimizer, gradient difference, auto switch, AGD

Compressor summary: AGD is a new optimizer that combines a dynamic preconditioning matrix and an auto-switching function between SGD and adaptive optimizers, improving generalization performance on NLP, CV, and RecSys datasets.


On the Importance of Feature Separability in Predicting Out-Of-Distribution Error

RENCHUNZI XIE, Hongxin Wei, Lei Feng, Yuzhou Cao, Bo An

https://openreview.net/forum?id=A86JTXllHa

Keywords: Machine Learning, Uncertainty Estimation

Compressor summary: The paper proposes a dataset-level score based on feature dispersion to estimate test accuracy under distribution shift, showing that inter-class dispersion is strongly correlated with model accuracy on out-of-distribution data.


Image Captioners Are Scalable Vision Learners Too

Michael Tschannen, Manoj Kumar, Andreas Peter Steiner, Xiaohua Zhai, Neil Houlsby, Lucas Beyer

https://openreview.net/forum?id=A7feCufBhL

Keywords: contrastive learning, CLIP, CapPa, Cap, vision-language, image captioning, visual representation learning, weakly supervised learning, VLM, multimodal learning, VQA, image classification

Compressor summary: The paper compares image captioning and contrastive pretraining for vision backbones, finding that captioning alone produces competitive or better results on various tasks.


On Convergence of Polynomial Approximations to the Gaussian Mixture Entropy

Caleb Dahlke, Jason Pacheco

https://openreview.net/forum?id=A7ESFTMJWs

Keywords: entropy, Gaussian mixture model, uncertainty quantification, approximate inference

Compressor summary: The paper proposes a new Taylor series for Gaussian mixture model entropy that converges to the true value and improves accuracy over previous methods.


One-2-3-45: Any Single Image to 3D Mesh in 45 Seconds without Per-Shape Optimization

Minghua Liu, Chao Xu, Haian Jin, Linghao Chen, Mukund Varma T, Zexiang Xu, Hao Su

https://openreview.net/forum?id=A6X9y8n4sT

Keywords: single image reconstruction, 3d generation, mesh reconstruction, diffusion models

Compressor summary: The authors propose a novel method that generates a 360-degree 3D textured mesh from a single image using a view-conditioned 2D diffusion model and an SDF-based neural surface reconstruction method, achieving better geometry, faster runtime, and supporting the text-to-3D task.


Generalized Semi-Supervised Learning via Self-Supervised Feature Adaptation

Jiachen Liang, RuiBing Hou, Hong Chang, Bingpeng Ma, Shiguang Shan, Xilin CHEN

https://openreview.net/forum?id=A6PRwRjI8V

Keywords: semi-supervised learning, self-supervised learning

Compressor summary: The paper proposes a new semi-supervised learning setting where unlabeled data has a different feature distribution than labeled data, and introduces SSFA, a framework that adapts the model features to improve pseudo-label quality.


Keep Various Trajectories: Promoting Exploration of Ensemble Policies in Continuous Control

Chao Li, Chen GONG, Qiang He, Xinwen Hou

https://openreview.net/forum?id=A6JDQDv7Nt

Keywords: Reinforcement Learning, Ensemble Exploration, Control Tasks

Compressor summary: TEEN is a new ensemble RL algorithm that enhances diversity and performance of sub-policies in complex sequential decision-making problems.


Effective Bayesian Heteroscedastic Regression with Deep Neural Networks

Alexander Immer, Emanuele Palumbo, Alexander Marx, Julia E Vogt

https://openreview.net/forum?id=A6EquH0enk

Keywords: Heteroscedastic Regression, Marginal Likelihood, Bayesian Neural Networks, Uncertainty Estimaton, Model Selection, Laplace Approximation

Compressor summary: The authors propose an efficient method to handle uncertainties in complex regression problems using neural networks and Gaussian processes, which improves generalization and requires no hyperparameter tuning.


Combinatorial Group Testing with Selfish Agents

Giorgos Chionas, Dariusz Rafal Kowalski, Piotr Krysta

https://openreview.net/forum?id=A5yMv7XPuA

Keywords: Combinatorial Group Testing, Adversarial Equilibrium, Contention Resolution, selfish agents, learning time, adaptive learning algorithms

Compressor summary: The paper presents a game-theoretic approach to Combinatorial Group Testing with selfish agents and shows how adaptive strategies can achieve near-optimal learning time depending on whether the number of active agents is known or not.


Maximize to Explore: One Objective Function Fusing Estimation, Planning, and Exploration

Zhihan Liu, Miao Lu, Wei Xiong, Han Zhong, Hao Hu, Shenao Zhang, Sirui Zheng, Zhuoran Yang, Zhaoran Wang

https://openreview.net/forum?id=A57UMlUJdc

Keywords: reinforcement learning; online learning; game

Compressor summary: MEX is an easy-to-implement RL framework that balances exploration and exploitation automatically and outperforms baselines in various MuJoCo environments.


Koopa: Learning Non-stationary Time Series Dynamics with Koopman Predictors

Yong Liu, Chenyu Li, Jianmin Wang, Mingsheng Long

https://openreview.net/forum?id=A4zzxu82a7

Keywords: Time series forecasting, Deep learning

Compressor summary: Koopa is a novel Koopman forecaster that uses modern theory to handle non-stationary time series with complex dynamics, achieving competitive performance while reducing training time and memory.


Oracle Complexity of Single-Loop Switching Subgradient Methods for Non-Smooth Weakly Convex Functional Constrained Optimization

Yankun Huang, Qihang Lin

https://openreview.net/forum?id=A383wMho4h

Keywords: Constrained optimization, first-order method, non-smooth optimization, non-convex optimization

Compressor summary: The paper analyzes the oracle complexity of the switching subgradient method for solving non-convex constrained optimization problems with weakly convex or convex constraint functions, and shows that it outperforms double-loop methods in terms of simplicity and efficiency.


Tuning Multi-mode Token-level Prompt Alignment across Modalities

Dongsheng Wang, Miaoge Li, Xinyang Liu, MingSheng Xu, Bo Chen, Hanwang Zhang

https://openreview.net/forum?id=A253n2EXCd

Keywords: Multi-modal prompt learning; Optimal transport

Compressor summary: The text describes a new method for improving vision-language models by tuning multiple prompts across modalities using optimal transportation, which enhances semantic alignment and leads to better image recognition performance.


Accelerating Molecular Graph Neural Networks via Knowledge Distillation

Filip Ekström Kelvinius, Dimitar Georgiev, Artur Toshev, Johannes Gasteiger

https://openreview.net/forum?id=A18PgVSUgf

Keywords: GNN, graph neural networks, knowledge distillation, molecules, molecular simulations

Compressor summary: The paper explores using knowledge distillation to accelerate molecular graph neural networks without sacrificing predictive accuracy in energy and force prediction tasks.


k-Median Clustering via Metric Embedding: Towards Better Initialization with Differential Privacy

Chenglin Fan, Ping Li, Xiaoyun Li

https://openreview.net/forum?id=9zV2OXCrVF

Keywords: privacy, clustering

Compressor summary: The paper introduces a new initialization method for k-median problem in metric spaces called HST, which produces better and more efficient initial centers than existing methods, and can be extended to generate private initial centers with differential privacy.


Three Iterations of (d − 1)-WL Test Distinguish Non Isometric Clouds of d-dimensional Points

Valentino delle Rose, Alexander Kozachinskiy, Cristobal Rojas, Mircea Petrache, Pablo Barcelo

https://openreview.net/forum?id=9yhYcjsdab

Keywords: euclidean graphs, point clouds, WL test, graph neural networks

Compressor summary: The Weisfeiler-Lehman test can distinguish point clouds in Euclidean space using a limited number of iterations and is complete for three dimensions or higher.


MultiFusion: Fusing Pre-Trained Models for Multi-Lingual, Multi-Modal Image Generation

Marco Bellagente, Manuel Brack, Hannah Benita Teufel, Felix Friedrich, Björn Deiseroth, Constantin Eichenberg, Andrew Dai, Robert John Nicholas Baldock, Souradeep Nanda, Koen Oostermeijer, Andres Felipe Cruz-Salinas, Patrick Schramowski, Kristian Kersting, Samuel Weinbach

https://openreview.net/forum?id=9ych3krqP0

Keywords: diffusion, image generation, multimodal

Compressor summary: MultiFusion is a system that enables complex and nuanced image generation from multiple modalities and languages by fusing pre-trained models without extensive training.


Probabilistic Inference in Reinforcement Learning Done Right

Jean Tarbouriech, Tor Lattimore, Brendan O'Donoghue

https://openreview.net/forum?id=9yQ2aaArDn

Keywords: Reinforcement learning, Bayesian inference, Exploration

Compressor summary: The paper proposes VAPOR, a new variational Bayesian approach for Reinforcement learning that efficiently approximates the posterior probability of state-action optimality and leads to better exploration policies.


Diversify Your Vision Datasets with Automatic Diffusion-based Augmentation

Lisa Dunlap, Alyssa Umino, Han Zhang, Jiezhi Yang, Joseph E. Gonzalez, Trevor Darrell

https://openreview.net/forum?id=9wrYfqdrwk

Keywords: data augmentation, diffusion, vision and language

Compressor summary: ALIA uses natural language descriptions and large vision models to generate diverse and visually consistent image augmentations for fine-grained classification tasks, improving performance on domain generalization and contextual bias.


Unbiased Compression Saves Communication in Distributed Optimization: When and How Much?

Yutong He, Xinmeng Huang, Kun Yuan

https://openreview.net/forum?id=9v6gpFTfCM

Keywords: Communication Compression, Distributed Optimization, Unbiased Compression, Optimal Complexity

Compressor summary: The paper investigates when unbiased communication compression reduces total communication cost in distributed optimization, showing that it can achieve this under certain conditions, and provides theoretical and empirical evidence for its effectiveness.


Dynamic Regret of Adversarial Linear Mixture MDPs

Long-Fei Li, Peng Zhao, Zhi-Hua Zhou

https://openreview.net/forum?id=9tUjsRLjf2

Keywords: dynamic regret, adversarial MDPs, linear mixture MDPs, policy optimization

Compressor summary: The paper proposes a new algorithm for reinforcement learning in adversarial MDPs with unknown transition kernels, achieving near-optimal dynamic regret.


Gradient-Free Kernel Stein Discrepancy

Matthew A Fisher, Chris J. Oates

https://openreview.net/forum?id=9rmwPAjk9O

Keywords: Bayesian, discrepancy, kernel, sampling, Stein's method

Compressor summary: The paper proposes non-canonical Stein discrepancies, which don't need derivatives, for posterior approximation in complex models, with convergence guarantees and applications in sampling and variational inference.


Efficient Adversarial Attacks on Online Multi-agent Reinforcement Learning

Guanlin Liu, Lifeng Lai

https://openreview.net/forum?id=9qlJGjO7bA

Keywords: adversarial attacks; multi agent reinforcement learning;

Compressor summary: The paper investigates how adversarial attacks affect multi-agent reinforcement learning and proposes a mixed attack strategy with both action poisoning and reward poisoning that can efficiently target MARL agents.


Generalized Logit Adjustment: Calibrating Fine-tuned Models by Removing Label Bias in Foundation Models

Beier Zhu, Kaihua Tang, Qianru Sun, Hanwang Zhang

https://openreview.net/forum?id=9qG6cMGUWk

Keywords: Foundation Model, Class Bias, Generalized Logit Adjustment

Compressor summary: The paper proposes a method called Generalized Logit Adjustment (GLA) to reduce biases in foundation models like CLIP and improve their performance on various tasks.


NeRF-IBVS: Visual Servo Based on NeRF for Visual Localization and Navigation

Yuanze Wang, Yichao Yan, Dianxi Shi, Wenhan Zhu, Jianqiang Xia, Tan Jeff, Songchang Jin, KE GAO, XIAOBO LI, Xiaokang Yang

https://openreview.net/forum?id=9pLaDXX8m3

Keywords: NeRF, Image-Based Visual Servoing (IBVS), visual localization, visual navigation

Compressor summary: The paper proposes a novel visual localization method that uses few posed images with coarse pseudo-3D labels from NeRF to train a coordinate regression network, estimate a coarse pose with PNP, optimize the pose with image-based visual servo, and provide navigation prior for navigation without custom markers or depth sensors.


Geometric Transformer with Interatomic Positional Encoding

Yusong Wang, Shaoning Li, Tong Wang, Bin Shao, Nanning Zheng, Tie-Yan Liu

https://openreview.net/forum?id=9o6KQrklrE

Keywords: Geometric Deep Learning, Molecular Modeling, Positional Encoding

Compressor summary: The paper proposes Geoformer, a novel geometric Transformer that uses Interatomic Positional Encoding to effectively model molecular structures for various property prediction and outperforms existing algorithms on benchmark datasets.


Beyond Invariance: Test-Time Label-Shift Adaptation for Addressing "Spurious" Correlations

Qingyao Sun, Kevin Patrick Murphy, Sayna Ebrahimi, Alexander D'Amour

https://openreview.net/forum?id=9mJXDcr17V

Keywords: Distribution shift, Spurious correlation, Group robustness

Compressor summary: The paper proposes a method (TTLSA) to correct for label shifts in test data that includes additional meta-data labels, and shows improved performance on various datasets.


Multi-body SE(3) Equivariance for Unsupervised Rigid Segmentation and Motion Estimation

Jia-Xing Zhong, Ta-Ying Cheng, Yuhang He, Kai Lu, Kaichen Zhou, Andrew Markham, Niki Trigoni

https://openreview.net/forum?id=9lygTqLdWn

Keywords: Dynamic Point Cloud Analytics, Multi-body Motion

Compressor summary: The authors propose a novel method for segmenting and estimating motion in 3D scenes without category information, using an unsupervised training strategy and lightweight architecture.


RangePerception: Taming LiDAR Range View for Efficient and Accurate 3D Object Detection

Yeqi BAI, Ben Fei, Youquan Liu, Tao MA, Yuenan Hou, Botian Shi, Yikang LI

https://openreview.net/forum?id=9kFQEJSyCM

Keywords: 3D Detection, Autonomous Driving

Compressor summary: RangePerception is an efficient and accurate RV-based 3D object detection framework that addresses two challenges in existing methods, achieving higher performance and speed than BEV-based methods.


Hypervolume Maximization: A Geometric View of Pareto Set Learning

Xiaoyuan Zhang, Xi Lin, Bo Xue, Yifan Chen, Qingfu Zhang

https://openreview.net/forum?id=9ieV1hnuva

Keywords: multiobjective optimization;multitask learning;hypervolume maximization;Pareto set learning

Compressor summary: The paper proposes a new neural network-based method for modeling the entire Pareto set in multiobjective optimization and shows its effectiveness on several problems.


Rewrite Caption Semantics: Bridging Semantic Gaps for Language-Supervised Semantic Segmentation

Yun Xing, Jian Kang, Aoran Xiao, Jiahao Nie, Ling Shao, Shijian Lu

https://openreview.net/forum?id=9iafshF7s3

Keywords: language-supervised semantic segmentation, vision-language pre-training

Compressor summary: CoCu is a method that uses CLIP to find and incorporate missing visual concepts from image-text pairs into pre-training, improving zero-shot recognition and language-supervised segmentation performance.


(Almost) Provable Error Bounds Under Distribution Shift via Disagreement Discrepancy

Elan Rosenfeld, Saurabh Garg

https://openreview.net/forum?id=9i8MD9btc8

Keywords: accuracy estimation, error bounds, distribution shift, unsupervised domain adaptation

Compressor summary: The paper proposes a new bound on the error of deep neural networks under distribution shift using unlabeled test data, which is simpler, more accurate, and easier to evaluate than previous methods.


Django: Detecting Trojans in Object Detection Models via Gaussian Focus Calibration

Guangyu Shen, Siyuan Cheng, Guanhong Tao, Kaiyuan Zhang, Yingqi Liu, Shengwei An, Shiqing Ma, Xiangyu Zhang

https://openreview.net/forum?id=9fb975Au9G

Keywords: backdoor detection; object detection;

Compressor summary: The paper proposes Django, a backdoor detection framework for object detection models that uses dynamic Gaussian weighting to prioritize vulnerable boxes and improve trigger inversion efficiency.


DPM-Solver-v3: Improved Diffusion ODE Solver with Empirical Model Statistics

Kaiwen Zheng, Cheng Lu, Jianfei Chen, Jun Zhu

https://openreview.net/forum?id=9fWKExmKa0

Keywords: diffusion models, fast sampling, ODE solver

Compressor summary: The paper proposes a new fast ODE solver for DPMs that improves sampling efficiency and sample quality, especially with fewer function evaluations or large guidance scales.


Inner Product-based Neural Network Similarity

Wei Chen, Zichen Miao, Qiang Qiu

https://openreview.net/forum?id=9eneYFIGKq

Keywords: Neural Network Similarity, Filter Subspace

Compressor summary: The paper proposes a new method to quickly compare neural network models by simplifying their representation to filter subspace distance, which is efficient and robust.


Reinforcement-Enhanced Autoregressive Feature Transformation: Gradient-steered Search in Continuous Space for Postfix Expressions

Dongjie Wang, Meng Xiao, Min Wu, pengfei wang, Yuanchun Zhou, Yanjie Fu

https://openreview.net/forum?id=9dp35y5C0p

Keywords: Feature Transformation, Autoregressive Generation, Continuous Space Optimization

Compressor summary: The authors propose a new method to optimize feature transformation as a continuous space optimization task, and demonstrate its effectiveness and robustness through experiments and case studies.


Diffusion Representation for Asymmetric Kernels via Magnetic Transform

Mingzhen He, FAN He, Ruikai Yang, Xiaolin Huang

https://openreview.net/forum?id=9cQzO3rXgR

Keywords: Asymmetric kernels, diffusion maps, magnetic transform, dimension reduction

Compressor summary: The paper introduces MagDM, a new diffusion map technique that can handle asymmetric data using the magnetic transform, and shows its effectiveness on synthetic and real-world datasets.


Learning threshold neurons via edge of stability

Kwangjun Ahn, Sebastien Bubeck, Sinho Chewi, Yin Tat Lee, Felipe Suarez, Yi Zhang

https://openreview.net/forum?id=9cQ6kToLnJ

Keywords: Gradient descent, edge of stability, generalization

Compressor summary: The paper analyzes how large learning rates affect two-layer neural networks and shows that using a high learning rate can improve generalization by enabling the training of threshold-like neurons.


Learning Space-Time Continuous Latent Neural PDEs from Partially Observed States

Valerii Iakovlev, Markus Heinonen, Harri Lähdesmäki

https://openreview.net/forum?id=9cF6RUwMe7

Keywords: neural, PDEs, neural PDEs, partial observations, space time continuous

Compressor summary: The paper presents a new neural network model that learns partial differential equations from noisy and partial observations without relying on regular grids, achieving high performance and data efficiency.


Hierarchical Decomposition of Prompt-Based Continual Learning: Rethinking Obscured Sub-optimality

Liyuan Wang, Jingyi Xie, Xingxing Zhang, Mingyi Huang, Hang Su, Jun Zhu

https://openreview.net/forum?id=9XieH21Tlf

Keywords: Continual Learning, Catastrophic Forgetting, Pre-training, Prompt Tuning

Compressor summary: The paper proposes Hierarchical Decomposition (HiDe-)Prompt, an approach that improves the performance of continual learning with self-supervised pre-training by optimizing hierarchical components with task-specific prompts and representation statistics.


Large Language Models for Automated Data Science: Introducing CAAFE for Context-Aware Automated Feature Engineering

Noah Hollmann, Samuel Müller, Frank Hutter

https://openreview.net/forum?id=9WSxQZ9mG7

Keywords: AutoML, AutoDS, Automated Feature Engineering, LLM Code Generation, Tabular Data, Feature Engineering, Automated Data Science, Automated Machine Learning

Compressor summary: The authors present CAAFE, a feature engineering method that uses large language models to generate semantically meaningful features for tabular datasets based on their descriptions, improving performance and interpretability in automated machine learning systems.


Brain Diffusion for Visual Exploration: Cortical Discovery using Large Scale Generative Models

Andrew Luo, Margaret Marie Henderson, Leila Wehbe, Michael J. Tarr

https://openreview.net/forum?id=9VqMaSjf7U

Keywords: neuroscience, brain, fmri, generative models, diffusion models, image synthesis, visual cortex

Compressor summary: The authors present BrainDiVE, a data-driven approach to explore the fine-grained functional organization of the brain's higher visual cortex by synthesizing images that activate specific brain regions, revealing novel subdivisions and functional differences within these regions.


LuminAIRe: Illumination-Aware Conditional Image Repainting for Lighting-Realistic Generation

Jiajun Tang, Haofeng Zhong, Shuchen Weng, Boxin Shi

https://openreview.net/forum?id=9UxUTGCteW

Keywords: Illumination, Image Generation, Conditional Image Repainting

Compressor summary: The paper introduces a new task (LuminAIRe) and a dataset (Car-LuminAIRe) to improve conditional image repainting methods by using estimated 3D geometry and physically-based illumination rendering.


Diffused Task-Agnostic Milestone Planner

Mineui Hong, Minjae Kang, Songhwai Oh

https://openreview.net/forum?id=9Tx2znbyTm

Keywords: Multi-task decision-making, Offline reinforcement learning, Planning, Diffusion model

Compressor summary: The paper proposes a diffusion-based generative sequence model to plan milestones in a latent space for long-term planning, vision-based control, and multi-task decision-making, achieving superior results on various tasks.


Mean-field Langevin dynamics: Time-space discretization, stochastic gradient, and variance reduction

Taiji Suzuki, Denny Wu, Atsushi Nitanda

https://openreview.net/forum?id=9STYRIVx6u

Keywords: mean-field regime, interacting particle system, propagation of chaos, Neural network optimization, MMD minimization, kernel stein discrepancy

Compressor summary: The paper proposes a framework to analyze the MFLD method for training two-layer neural networks and shows its convergence rate and applicability to various learning problems and gradient estimators.


Intervention Generalization: A View from Factor Graph Models

Gecia Bravo-Hermsdorff, David Watson, Jialin Yu, Jakob Zeitler, Ricardo Silva

https://openreview.net/forum?id=9S8oVumknA

Keywords: Causality, experimental design

Compressor summary: The paper proposes a method to generalize causal inference from past experiments to novel conditions using factor graph models that abstract away unmeasured confounding and feedback mechanisms.


Test-time Training for Matching-based Video Object Segmentation

Juliette Bertrand, Giorgos Kordopatis-Zilos, Yannis Kalantidis, Giorgos Tolias

https://openreview.net/forum?id=9QsdPQlWiE

Keywords: VOS, video object segmentation, test-time training, test-time adaptation

Compressor summary: The paper proposes test-time training strategies for video object segmentation tasks under distribution shifts, such as video corruptions and style transfers, and introduces a new test set with extreme distribution shifts.


Robust Learning with Progressive Data Expansion Against Spurious Correlation

Yihe Deng, Yu Yang, Baharan Mirzasoleiman, Quanquan Gu

https://openreview.net/forum?id=9QEVJ9qm46

Keywords: spurious correlation, robustness, robust learning

Compressor summary: The paper analyzes how two-layer convolutional neural networks can be influenced by non-generalizable features and proposes a new algorithm called PDE that improves robustness and performance on various tasks.


Incentives in Federated Learning: Equilibria, Dynamics, and Mechanisms for Welfare Maximization

Aniket Murhekar, Zhuowen Yuan, Bhaskar Ray Chaudhury, Bo Li, Ruta Mehta

https://openreview.net/forum?id=9OqezkNxnX

Keywords: Federated learning, Nash equilibrium, Mechanism design, Welfare maximization

Compressor summary: The paper proposes a budget-balanced mechanism to improve the welfare of agents participating in collaborative federated learning by ensuring optimal trade-offs between learning payoff and data sharing costs.


Loss Decoupling for Task-Agnostic Continual Learning

Yan-Shuo Liang, Wu-Jun Li

https://openreview.net/forum?id=9Oi3YxIBSa

Keywords: Continual Learning, stability, plasticity

Compressor summary: LODE is a method for task-agnostic continual learning that separates the objectives of distinguishing new classes from old and between different new classes, achieving better stability and plasticity than existing replay-based methods.


Cross-modal Prompts: Adapting Large Pre-trained Models for Audio-Visual Downstream Tasks

Haoyi Duan, Yan Xia, Mingze Zhou, Li Tang, Jieming Zhu, Zhou Zhao

https://openreview.net/forum?id=9MwidIH4ea

Keywords: audio-visual, multi-modal prompt, clip, cross-modal attention

Compressor summary: The paper proposes a novel DG-SCT attention mechanism that uses audio and visual modalities as soft prompts to dynamically adjust the parameters of pre-trained models for multi-modal tasks, achieving state-of-the-art results and few-shot/zero-shot performance.


Optimal Transport-Guided Conditional Score-Based Diffusion Model

Xiang Gu, Liwei Yang, Jian Sun, Zongben Xu

https://openreview.net/forum?id=9Muli2zoFn

Keywords: optimal transport, diffusion probabilistic model, conditional score-based model, unpaired super-resolution, image-to-image translation

Compressor summary: OTCS is a novel model for conditional generation of target data using unpaired or partially paired data based on optimal transport, and it achieves effective data transport and image translation.


The Behavior and Convergence of Local Bayesian Optimization

Kaiwen Wu, Kyurae Kim, Roman Garnett, Jacob R. Gardner

https://openreview.net/forum?id=9KtX12YmA7

Keywords: Bayesian optimization, convergence rates

Compressor summary: The text discusses the benefits of using local optimization strategies in Bayesian optimization for high-dimensional problems, and provides a rigorous analysis of a recent algorithm's performance and convergence rates.


Provable Advantage of Curriculum Learning on Parity Targets with Mixed Inputs

Emmanuel Abbe, Elisabetta Cornacchia, Aryo Lotfi

https://openreview.net/forum?id=9Ihu0VBOTq

Keywords: curriculum learning, parities, time complexity, sample complexity, neural networks, SGD

Compressor summary: The text explains that curriculum learning, starting with simpler tasks before more complex ones, can help neural networks learn parities faster than traditional methods, especially when the data is a mix of sparse and dense inputs.


Spike-driven Transformer

Man Yao, JiaKui Hu, Zhaokun Zhou, Li Yuan, Yonghong Tian, Bo XU, Guoqi Li

https://openreview.net/forum?id=9FmolyOHi5

Keywords: Spiking Neural Networks; Transformer; Neuromorphic Computing; Event-driven; Linear Attention

Compressor summary: The paper proposes a spike-driven Transformer that uses event-driven and binary spike communication to achieve energy efficiency and state-of-the-art performance in the spiking neural network domain.


Should We Learn Most Likely Functions or Parameters?

Shikai Qiu, Tim G. J. Rudner, Sanyam Kapoor, Andrew Gordon Wilson

https://openreview.net/forum?id=9EndFTDiqh

Keywords: Function-Space Modeling, Maximum A Posteriori Estimation, Generalization

Compressor summary: The paper explores how estimating the most likely function implied by a model and data can improve predictive performance, but warns of potential pathological solutions when using neural networks.


Scale-teaching: Robust Multi-scale Training for Time Series Classification with Noisy Labels

Zhen Liu, Peitian Ma, Dongliang Chen, Wenbin Pei, Qianli Ma

https://openreview.net/forum?id=9D0fELXbrg

Keywords: time series classification, deep neural networks, noisy labels

Compressor summary: The paper proposes a deep learning approach called Scale-teaching to handle noisy labels in time series data by using multiple DNNs, cross-scale fusion, and label propagation.


Accelerating Motion Planning via Optimal Transport

An Thai Le, Georgia Chalvatzaki, Armin Biess, Jan Peters

https://openreview.net/forum?id=9B9J8X23LK

Keywords: Motion Planning, Trajectory Optimization, Optimal Transport

Compressor summary: MPOT is a gradient-free method that optimizes smooth trajectories over nonlinear costs using the Sinkhorn Step and regular polytopes, outperforming other motion planners in various problems.


AmadeusGPT: a natural language interface for interactive animal behavioral analysis

Shaokai Ye, Jessy Lauer, Mu Zhou, Alexander Mathis, Mackenzie W Mathis

https://openreview.net/forum?id=9AcG3Tsyoq

Keywords: ChatGPT, GPT3.5, GPT4, behavioral analysis, LLMs, human-AI interaction, behavioral neuroscience

Compressor summary: The paragraph introduces AmadeusGPT, a natural language interface that converts descriptions of animal behaviors into machine code using large-language models and a dual-memory mechanism, improving the efficiency and accuracy of behavior analysis.


Scenario Diffusion: Controllable Driving Scenario Generation With Diffusion

Ethan Pronovost, Meghana Reddy Ganesina, Noureldin Hendy, Zeyu Wang, Andres Morales, Kai Wang, Nicholas Roy

https://openreview.net/forum?id=99MHSB98yZ

Keywords: Deep Learning, (Other) Applications, (Other) Machine Learning Topics

Compressor summary: Scenario Diffusion generates diverse and controllable synthetic traffic scenarios for autonomous vehicles using latent diffusion, object detection, and trajectory regression.


Generalization bounds for neural ordinary differential equations and deep residual networks

Pierre Marion

https://openreview.net/forum?id=992vogTP1L

Keywords: residual neural networks, neural ODEs, generalization bound

Compressor summary: The paper develops a generalization bound for continuous-in-time parameterized ODEs, which includes time-dependent neural ODEs and deep residual networks, by using a Lipschitz-based argument that relates the magnitude of weight matrix differences to generalization.


Accelerating Reinforcement Learning with Value-Conditional State Entropy Exploration

Dongyoung Kim, Jinwoo Shin, Pieter Abbeel, Younggyo Seo

https://openreview.net/forum?id=97E3YXvcFM

Keywords: Reinforcement Learning, State Entropy, Exploration

Compressor summary: The paper proposes a new exploration technique for reinforcement learning that considers the value of states when maximizing state entropy to prevent bias and improve performance in various tasks.


Towards Data-Algorithm Dependent Generalization: a Case Study on Overparameterized Linear Regression

Jing Xu, Jiaye Teng, Yang Yuan, Andrew C Yao

https://openreview.net/forum?id=966yOmwk6d

Keywords: data-algorithm dependent generalization analysis, overparameterized linear regression

Compressor summary: The paper proposes a new concept called data-algorithm compatibility to analyze generalization in overparameterized models by considering the entire training trajectory instead of just the final model.


Generalizable One-shot 3D Neural Head Avatar

Xueting Li, Shalini De Mello, Sifei Liu, Koki Nagano, Umar Iqbal, Jan Kautz

https://openreview.net/forum?id=95q46MpBGZ

Keywords: Neural Radiance Field, Portrait Reconstruction and Animation

Compressor summary: The authors present a method that can reconstruct and animate realistic 3D head avatars from single-view images, generalizing to different identities and capturing details beyond the face. The method uses three branches with tri-planes and volumetric rendering for high fidelity results.


Distance-Restricted Folklore Weisfeiler-Leman GNNs with Provable Cycle Counting Power

Junru Zhou, Jiarui Feng, Xiyuan Wang, Muhan Zhang

https://openreview.net/forum?id=94rKFkcm56

Keywords: Cycle counting, graph neural networks

Compressor summary: The paper introduces $d$-Distance-Restricted FWL(2) GNNs, a novel class of graph neural networks that can efficiently count certain graph substructures like cycles by exploiting graph sparsity and avoiding expensive preprocessing steps.


UltraRE: Enhancing RecEraser for Recommendation Unlearning via Error Decomposition

Yuyuan Li, Chaochao Chen, Yizhao Zhang, Weiming Liu, Lingjuan Lyu, Xiaolin Zheng, Dan Meng, Jun Wang

https://openreview.net/forum?id=93NLxUojvc

Keywords: recommendation unlearning, machine unlearning, recommender systems, ensemble learning

Compressor summary: UltraRE is a new framework that enhances RecEraser's recommendation unlearning by addressing redundancy, relevance, and combination losses.


Self-supervised Object-Centric Learning for Videos

Görkay Aydemir, Weidi Xie, Fatma Guney

https://openreview.net/forum?id=919tWtJPXe

Keywords: Unsupervised Object Discovery, Unsupervised Video Object Segmentation, Object-Centric Learning, Unsupervised Video Multi Object Segmentation

Compressor summary: The paper presents an unsupervised method for segmenting multiple objects in real-world video sequences using object-centric learning and a masking strategy.


Distribution-Free Statistical Dispersion Control for Societal Applications

Zhun Deng, Thomas P Zollo, Jake Snell, Toniann Pitassi, Richard Zemel

https://openreview.net/forum?id=917crxqJdA

Keywords: societal dispersion, distribution-free uncertainty quantification

Compressor summary: The paragraph discusses the importance of controlling the dispersion of loss distribution in machine learning models for high-stakes applications and proposes a new framework to handle various statistical functionals.


GUST: Combinatorial Generalization by Unsupervised Grouping with Neuronal Coherence

Hao Zheng, Hui Lin, Rong Zhao

https://openreview.net/forum?id=90O5cvFZkZ

Keywords: neuronal coherence, combinatorial generalization, perceptual grouping, unsupervised learning

Compressor summary: GUST is an iterative neural network architecture that mimics neuronal coherence in the human brain to learn how to group sensory information effectively and create symbolic representations of scenes.


Training neural operators to preserve invariant measures of chaotic attractors

Ruoxi Jiang, Peter Y. Lu, Elena Orlova, Rebecca Willett

https://openreview.net/forum?id=8xx0pyMOW1

Keywords: Neural operators, contrastive learning, optimal transport, chaotic attractors, invariant measures

Compressor summary: The paper proposes new methods for training neural networks to forecast chaotic systems over long time horizons while preserving their statistical properties using optimal transport or contrastive learning.


TMT-VIS: Taxonomy-aware Multi-dataset Joint Training for Video Instance Segmentation

rongkun Zheng, Lu Qi, Xi Chen, Yi Wang, Kun Wang, Yu Qiao, Hengshuang Zhao

https://openreview.net/forum?id=8xTOtxinMH

Keywords: taxonomy-aware, multiple-datasets, video instance segementation

Compressor summary: The paper proposes TMT-VIS, a model that uses taxonomy information to improve video instance segmentation performance on multiple datasets.


OpenMask3D: Open-Vocabulary 3D Instance Segmentation

Ayça Takmaz, Elisabetta Fedele, Robert Sumner, Marc Pollefeys, Federico Tombari, Francis Engelmann

https://openreview.net/forum?id=8vuDHCxrmy

Keywords: open-world, open-vocabulary, 3D vision, point cloud, instance segmentation, 3D instance segmentation

Compressor summary: OpenMask3D is a method for open-vocabulary 3D instance segmentation that uses multi-view fusion of CLIP-based image embeddings to separate multiple objects in scenes.


Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts

Eduard Tulchinskii, Kristian Kuznetsov, Kushnareva Laida, Daniil Cherniavskii, Sergey Nikolenko, Evgeny Burnaev, Serguei Barannikov, Irina Piontkovskaya

https://openreview.net/forum?id=8uOZ0kNji6

Keywords: generated texts detection, intrinsic dimension, TDA, Persistent Homology, ChatGPT

Compressor summary: The authors propose a new method to distinguish between natural and AI-generated texts based on the intrinsic dimensionality of text embeddings, which is lower for AI-generated texts and stable across different languages and domains.


Constant Approximation for Individual Preference Stable Clustering

Anders Aamand, Justin Y. Chen, Allen Liu, Sandeep Silwal, Pattara Sukprasert, Ali Vakilian, Fred Zhang

https://openreview.net/forum?id=8rDbUoYc0p

Keywords: clustering, fairness, approximation algorithms

Compressor summary: The paper introduces IP stability as a new clustering objective, shows that an O(1)-IP stable clustering always exists for general metrics, and provides efficient algorithms for various generalizations of IP stability.


One-for-All: Bridge the Gap Between Heterogeneous Architectures in Knowledge Distillation

Zhiwei Hao, Jianyuan Guo, Kai Han, Yehui Tang, Han Hu, Yunhe Wang, Chang Xu

https://openreview.net/forum?id=8qePPvL1VY

Keywords: knowledge distillation, feature distillation, heterogeneous architectures

Compressor summary: The paper proposes OFA-KD, a framework for distilling knowledge between heterogeneous models that improves performance by aligning features and adaptively enhancing the target.


FLSL: Feature-level Self-supervised Learning

Qing Su, Anton Netchaev, Hai Li, Shihao Ji

https://openreview.net/forum?id=8pOBo5NgTQ

Keywords: Transformer, ViT, Dense Prediction, Self-supervised Learning, Mean Shift, Self-attention, Representation learning

Compressor summary: The paper proposes a new SSL method, FLSL, that uses transformer for joint embedding and clustering, which improves dense prediction tasks like object detection and segmentation.


Self-Supervised Learning of Representations for Space Generates Multi-Modular Grid Cells

Rylan Schaeffer, Mikail Khona, Tzuhsuan Ma, Cristobal Eyzaguirre, Sanmi Koyejo, Ila R Fiete

https://openreview.net/forum?id=8ox2vrQiTF

Keywords: self-supervised learning, unsupervised learning, grid cells, neuroscience, systems neuroscience, representation learning

Compressor summary: The authors propose a new self-supervised learning framework to synthesize multi-periodic grid cells in deep recurrent neural networks, drawing from dynamical systems, coding theory, function optimization and supervised deep learning approaches.


Sparse Deep Learning for Time Series Data: Theory and Applications

Mingxuan Zhang, Yan Sun, Faming Liang

https://openreview.net/forum?id=8niGwlkLAX

Keywords: Sparse Deep Learning, Uncertainty Quantification, Model Compression, Variable Selection, Dependent Data

Compressor summary: The paper studies sparse deep learning with dependent data, showing that sparse recurrent neural networks can be consistently estimated and outperform existing methods in uncertainty quantification and model compression for time series data.


Towards Stable Backdoor Purification through Feature Shift Tuning

Rui Min, Zeyu Qin, Li Shen, Minhao Cheng

https://openreview.net/forum?id=8muKbaAgsh

Keywords: Backdoor Defense, Model-tuning

Compressor summary: The paper proposes Feature Shift Tuning (FST), a method to disentangle backdoor and clean features for tuning-based backdoor purification, which performs well against diverse attack scenarios with low tuning costs.


Investigating how ReLU-networks encode symmetries

Georg Bökman, Fredrik Kahl

https://openreview.net/forum?id=8lbFwpebeu

Keywords: loss landscape, network merging, linear mode connectivity, equivariance, group convolutional neural network, permutation, group, symmetry, invariance, weight space ensembling

Compressor summary: The authors study how neural networks' equivariance properties affect their layers and propose a conjecture related to the recent permutation conjecture, while providing experiments to support their findings.


When Does Optimizing a Proper Loss Yield Calibration?

Jarosław Błasiok, Parikshit Gopalan, Lunjia Hu, Preetum Nakkiran

https://openreview.net/forum?id=8kyIChWsAG

Keywords: calibration, deep learning, theory, optimization

Compressor summary: This work studies under what conditions optimizing a proper loss function over a restricted family of predictors yields calibrated models, and provides a rigorous answer by introducing the concept of local optimality.


Closing the Computational-Statistical Gap in Best Arm Identification for Combinatorial Semi-bandits

Ruo-Chun Tzeng, Po-An Wang, Alexandre Proutiere, Chi-Jen Lu

https://openreview.net/forum?id=8jg8z3ASiw

Keywords: best-arm identification; combinatorial semi-bandit; no-regret learning;

Compressor summary: Perturbed Frank-Wolfe Sampling (P-FWS) is an algorithm that efficiently identifies the best arm in combinatorial semi-bandits by solving an optimization problem with a single iteration of the Frank-Wolfe algorithm and leveraging structural properties.


A Riemannian Exponential Augmented Lagrangian Method for Computing the Projection Robust Wasserstein Distance

Bo Jiang, Ya-Feng Liu

https://openreview.net/forum?id=8hKCNVqrlf

Keywords: Barzilai-Borwein method, exponential augmented Lagrangian, inexact gradient, Stiefel manifold, Sinkhorn iteration, Wasserstein distance

Compressor summary: The paper proposes a new method (REALM) to compute the projection robust Wasserstein distance more efficiently and accurately than existing methods, by using an inexact Riemannian Barzilai-Borwein method with Sinkhorn iteration.


Uniform-in-Time Wasserstein Stability Bounds for (Noisy) Stochastic Gradient Descent

Lingjiong Zhu, Mert Gurbuzbalaban, Anant Raj, Umut Simsekli

https://openreview.net/forum?id=8fLatmFQgF

Keywords: Algorithmic stability, SGD, Wasserstein distance

Compressor summary: The paper introduces a unified proof technique for deriving Wasserstein stability bounds for stochastic optimization algorithms like SGD, showing that ergodicity is crucial for time-uniform bounds and extending previous results to more general cases.


Accurate Interpolation for Scattered Data through Hierarchical Residual Refinement

Shizhe Ding, Boyang Xia, Dongbo Bu

https://openreview.net/forum?id=8d9wVXri89

Keywords: Interpolation algorithm, scattered data, deep learning, residual learning

Compressor summary: HINT is a novel neural network-based interpolation method that uses residuals on observed points to guide target function estimation and hierarchical local constraints in correlation modeling to improve interpolation accuracy.


Provably Efficient Offline Reinforcement Learning in Regular Decision Processes

Roberto Cipollone, Anders Jonsson, Alessandro Ronca, Mohammad Sadegh Talebi

https://openreview.net/forum?id=8bQc7oRnjm

Keywords: Reinforcement Learning, Offline Reinforcement Learning, Regular Decision Processes, Sample complexity, Automata

Compressor summary: The paper introduces RegORL, an algorithm that combines automata learning and offline RL to learn near-optimal policies in unknown episodic RDPs using pre-collected data without further exploration.


Convex and Non-convex Optimization Under Generalized Smoothness

Haochuan Li, Jian Qian, Yi Tian, Alexander Rakhlin, Ali Jadbabaie

https://openreview.net/forum?id=8aunGrXdkl

Keywords: Optimization, Convergence, Generalized smoothness

Compressor summary: The paper proposes a generalized non-uniform smoothness condition for optimization problems, which enables convergence rates for various methods in both convex and non-convex settings without gradient clipping or heavy-tailed noise restrictions.


No Change, No Gain: Empowering Graph Neural Networks with Expected Model Change Maximization for Active Learning

Zixing Song, Yifei Zhang, Irwin King

https://openreview.net/forum?id=8aDG51pxFc

Keywords: Graph Neural Networks, Expected Model Change Maximization

Compressor summary: The paper proposes an active learning method for graph neural networks that uses the Expected Model Change Maximization principle to improve prediction performance on unlabeled data.


On the Convergence of Encoder-only Shallow Transformers

Yongtao Wu, Fanghui Liu, Grigorios Chrysos, Volkan Cevher

https://openreview.net/forum?id=8ZveVHfmIE

Keywords: Transformer, convergence, scaling, initialization, over-parameterization

Compressor summary: The paper develops a global convergence theory for shallow Transformers with encoder-only architecture under realistic settings and analyzes the impact of architectures, initialization, scaling, and softmax on their performance.


A Unified Discretization Framework for Differential Equation Approach with Lyapunov Arguments for Convex Optimization

Kansei Ushiyama, Shun Sato, Takayasu Matsuo

https://openreview.net/forum?id=8YN62t19AW

Keywords: Convex optimization, Numerical analysis, Ordinary differential equations, Convergence estimate

Compressor summary: This paper introduces a concept called weak discrete gradient (wDG) that allows transitioning from continuous to discrete optimization methods in the differential equation approach for convex optimization, simplifying analysis and enabling faster convergence rates.


Mitigating the Effect of Incidental Correlations on Part-based Learning

Gaurav Bhatt, Deepayan Das, Leonid Sigal, Vineeth N. Balasubramanian

https://openreview.net/forum?id=8Xn3D9OtqI

Keywords: part-based learning, interpretability, few-shot learning, vision transformers

Compressor summary: The study proposes two regularization methods to improve interpretability and generalization of part-based representations in intelligent systems by reducing incidental background correlations.


Near-Optimal $k$-Clustering in the Sliding Window Model

David Woodruff, Peilin Zhong, Samson Zhou

https://openreview.net/forum?id=8XRMbNAP6Z

Keywords: clustering, streaming algorithms, sliding window model

Compressor summary: The paper presents the first algorithm that achieves near-optimal clustering in the sliding window model and develops an online coreset data structure for this purpose.


MixFormerV2: Efficient Fully Transformer Tracking

Yutao Cui, Tianhui Song, Gangshan Wu, Limin Wang

https://openreview.net/forum?id=8WvYAycmDJ

Keywords: Efficient Tracking, Fully Transformer, Distillation, Model Pruning

Compressor summary: The paper proposes MixFormerV2, a transformer-based tracker that uses four special prediction tokens and a new distillation-based model reduction paradigm to achieve high accuracy and efficiency on both GPU and CPU platforms.


The Grand Illusion: The Myth of Software Portability and Implications for ML Progress.

Fraser Mince, Dzung Dinh, Jonas Kgomo, Neil Thompson, Sara Hooker

https://openreview.net/forum?id=8VTbfVfAfI

Keywords: hardware, software, meta study, portability

Compressor summary: The lack of portability in popular ML frameworks across different hardware types hinders exploratory research and innovation in machine learning due to significant function loss and performance slowdown.


Learning Shared Safety Constraints from Multi-task Demonstrations

Konwoo Kim, Gokul Swamy, Zuxin Liu, Ding Zhao, Sanjiban Choudhury, Steven Wu

https://openreview.net/forum?id=8U31BCquNF

Keywords: constraints, inverse reinforcement learning, safe reinforcement learning

Compressor summary: The paper proposes a method to learn safety constraints from expert demonstrations for robotic tasks, using inverse reinforcement learning and diverse demonstrations to avoid overly conservative constraints.


Semantic segmentation of sparse irregular point clouds for leaf/wood discrimination

Yuchen BAI, Jean-Baptiste Durand, Grégoire Laurent Vincent, Florence Forbes

https://openreview.net/forum?id=8SUtvEZCF2

Keywords: UAV, Deep Learning, Semantic Segmentation, Lidar, Class Imbalance, Point Cloud

Compressor summary: Lidar technology helps monitor forests and track climate change, but identifying leaf points from wood points in UAV data is challenging; a new neural network model based on Pointnet ++ architecture uses point geometry only to overcome these issues.


Real-World Image Super-Resolution as Multi-Task Learning

Wenlong Zhang, Xiaohui Li, Guangyuan SHI, Xiangyu Chen, Yu Qiao, Xiaoyun Zhang, Xiao-Ming Wu, Chao Dong

https://openreview.net/forum?id=8SCz56sUGP

Keywords: Image super-resolution

Compressor summary: The paper proposes a multi-task learning approach to improve real-world image super-resolution by grouping similar degradation tasks together and training them separately, which reduces task competition and enhances performance.


Data-driven Optimal Filtering for Linear Systems with Unknown Noise Covariances

Shahriar Talebi, Amirhossein Taghvaei, Mehran Mesbahi

https://openreview.net/forum?id=8S9Fbee743

Keywords: Optimal filtering, data-driven control, stochastic optimization, learning

Compressor summary: The paper studies how to learn the best filtering policy for a linear system with unknown noise matrices using noisy data, and proposes a method that minimizes prediction error, has good convergence properties, and scales well with problem size.


New Bounds for Hyperparameter Tuning of Regression Problems Across Instances

Nina Balcan, Anh Tuan Nguyen, Dravyansh Sharma

https://openreview.net/forum?id=8QGukmdAbh

Keywords: Elastic Net, logistic regression, data-driven algorithm design, learning theory, regularization

Compressor summary: This paper studies how many samples are needed to tune regularization parameters in linear and logistic regressions with different constraints, and provides new upper and lower bounds for their validation loss functions.


Rethinking Gauss-Newton for learning over-parameterized models

Michael Arbel, Romain Menegaux, Pierre Wolinski

https://openreview.net/forum?id=8Oukmqfek2

Keywords: implicit bias, gauss newton

Compressor summary: The paper analyzes how Gauss Newton's method optimizes over-parameterized networks, finding that it converges faster than gradient descent but has an implicit bias that affects generalization.


Reinforcement Learning for Fine-tuning Text-to-Image Diffusion Models

Ying Fan, Olivia Watkins, Yuqing Du, Hao Liu, Moonkyung Ryu, Craig Boutilier, Pieter Abbeel, Mohammad Ghavamzadeh, Kangwook Lee, Kimin Lee

https://openreview.net/forum?id=8OTPepXzeh

Keywords: Diffusion models, RLHF

Compressor summary: The authors propose using online reinforcement learning to fine-tune text-to-image diffusion models, achieving better alignment and image quality than supervised fine-tuning methods.


Few-Shot Class-Incremental Learning via Training-Free Prototype Calibration

Qi-Wei Wang, Da-Wei Zhou, Yi-Kai Zhang, De-Chuan Zhan, Han-Jia Ye

https://openreview.net/forum?id=8NAxGDdf7H

Keywords: Few-Shot Class-Incremental Learning, Continual Learning, Class-Incremental Learning

Compressor summary: The paper introduces a new strategy (TEEN) to improve few-shot class-incremental learning by enhancing the discriminability of new classes using weighted prototypes, and shows its effectiveness on standard benchmarks and few-shot learning tasks.


Bootstrapping Vision-Language Learning with Decoupled Language Pre-training

Yiren Jian, Chongyang Gao, Soroush Vosoughi

https://openreview.net/forum?id=8Kch0ILfQH

Keywords: vision-language pretraining, multi-modal learning, uni-modal auxiliary learning

Compressor summary: The Prompt-Transformer (P-Former) is a model that predicts optimal prompts for language models to align with visual features, improving performance and reducing the need for image-text pairs in vision-language pre-training.


3D Indoor Instance Segmentation in an Open-World

Mohamed El Amine Boudjoghra, Salwa K. Al Khatib, Jean Lahoud, Hisham Cholakkal, Rao Muhammad Anwer, Salman Khan, Fahad Khan

https://openreview.net/forum?id=8JsbdJjRvY

Keywords: open-world, 3d instance segmentation

Compressor summary: The paper presents a new method for 3D indoor instance segmentation that can handle unknown classes and improve its performance using pseudo-labels and realistic open-world scenarios.


DAC-DETR: Divide the Attention Layers and Conquer

Zhengdong Hu, Yifan Sun, Jingdong Wang, Yi Yang

https://openreview.net/forum?id=8JMexYVcXB

Keywords: deep learning, computer vision, object detection, transformer

Compressor summary: The paper presents a method called Divide-And-Conquer DETR (DAC-DETR) that improves object detection by separating cross-attention and self-attention layers in the decoder, leading to better performance on popular benchmarks.


Systematic Visual Reasoning through Object-Centric Relational Abstraction

Taylor Whittington Webb, Shanka Subhra Mondal, Jonathan Cohen

https://openreview.net/forum?id=8JCZe7QrPy

Keywords: relational reasoning, object-centric representations, abstract rule learning, relational inductive biases, systematic generalization

Compressor summary: The Object-Centric Relational Abraction (OCRA) model combines object-centric and relational abstraction approaches to achieve strong systematic generalization in complex visual tasks.


Exploring Loss Functions for Time-based Training Strategy in Spiking Neural Networks

Yaoyu Zhu, Wei Fang, Xiaodong Xie, Tiejun Huang, Zhaofei Yu

https://openreview.net/forum?id=8IvW2k5VeA

Keywords: Spiking neural networks, Spike encoding, Time-based training

Compressor summary: The paper proposes an enhanced counting loss for training spiking neural networks (SNNs) with time-based schemes, which better utilizes temporal information and improves performance on most datasets.


Efficient Subgame Refinement for Extensive-form Games

Zhenxing Ge, Zheng Xu, Tianyu Ding, Wenbin Li, Yang Gao

https://openreview.net/forum?id=8HzOyg1ngp

Keywords: Subgame solving, extensive-form game, imperfect information

Compressor summary: The GS2 framework uses a generation function to identify important subgame nodes, reducing their size and improving performance in large imperfect information games.


Pre-training Contextualized World Models with In-the-wild Videos for Reinforcement Learning

Jialong Wu, Haoyu Ma, Chaoyi Deng, Mingsheng Long

https://openreview.net/forum?id=8GuEVzAUQS

Keywords: Model-based reinforcement learning, world model, pre-training

Compressor summary: The paper proposes Contextualized World Models to pre-train world models with in-the-wild videos for efficient learning of visual control tasks in reinforcement learning.


Conservative State Value Estimation for Offline Reinforcement Learning

Liting Chen, Jie Yan, Zhengdao Shao, Lu Wang, Qingwei Lin, Saravan Rajmohan, Thomas Moscibroda, Dongmei Zhang

https://openreview.net/forum?id=8GSCaoFot9

Keywords: Offline Reinforcement Learning

Compressor summary: The paper proposes a new approach called Conservative State Value Estimation (CSVE) for offline reinforcement learning, which improves state value estimation by penalizing out-of-distribution states and performs well in classic continual control tasks of D4RL.


Differentially Private Statistical Inference through $\beta$-Divergence One Posterior Sampling

Jack Jewson, Sahra Ghalebikesabi, Christopher C. Holmes

https://openreview.net/forum?id=8FbuHeVU7D

Keywords: differential privacy, beta-divergence, posterior sampling, generalised Bayesian inference

Compressor summary: $\beta$D-Bayes is a new method that allows more accurate and private estimation for various types of statistical models without altering the original data or model.


Exploiting Connections between Lipschitz Structures for Certifiably Robust Deep Equilibrium Models

Aaron J Havens, Alexandre Araujo, Siddharth Garg, Farshad Khorrami, Bin Hu

https://openreview.net/forum?id=8F3Lutda7R

Keywords: Deep equilibrium models, Lipschitz networks, certifiable robustness

Compressor summary: The paper explores how different deep equilibrium models can be reparameterized as Lipschitz-bounded equilibrium networks, which could improve their certified robustness.


Analyzing Generalization of Neural Networks through Loss Path Kernels

Yilan Chen, Wei Huang, Hao Wang, Charlotte Loh, Akash Srivastava, Lam M. Nguyen, Tsui-Wei Weng

https://openreview.net/forum?id=8Ba7VJ7xiM

Keywords: generalization, deep learning theory, neural tangent kernel, neural architecture search

Compressor summary: This paper introduces a new kernel called loss path kernel, which measures data similarity using loss gradient agreement along gradient flow paths, and shows its application in deriving tight generalization bounds for neural networks and guiding neural architecture search.


Label-efficient Segmentation via Affinity Propagation

Wentong Li, Yuqian Yuan, Song Wang, Wenyu Liu, Dongqi Tang, Jian liu, Jianke Zhu, Lei Zhang

https://openreview.net/forum?id=8BPzLxF9p5

Keywords: Computer Vision, Segmentation, Weakly-supervised Learning

Compressor summary: The paper proposes an affinity propagation method that uses local and global pairwise affinity terms to generate accurate soft pseudo labels for weakly-supervised segmentation tasks, reducing the need for pixel-wise labeling and improving performance on three types of tasks.


Towards Automated Circuit Discovery for Mechanistic Interpretability

Arthur Conmy, Augustine N. Mavor-Parker, Aengus Lynch, Stefan Heimersheim, Adrià Garriga-Alonso

https://openreview.net/forum?id=89ia77nZ8u

Keywords: Mechanistic Interpretability, Pruning, Science of Deep Learning, AI Safety

Compressor summary: The paper presents a systematic method to interpret transformer models and automates one step of the process using novel algorithms, demonstrating their effectiveness on GPT-2 Small.


Tame a Wild Camera: In-the-Wild Monocular Camera Calibration

Shengjie Zhu, Abhinav Kumar, Masa Hu, Xiaoming Liu

https://openreview.net/forum?id=898RcRYWCg

Keywords: Monocular Camera Calibration; Camera Pose Estimation; Image Editing

Compressor summary: The paper proposes a method to calibrate intrinsic camera parameters using monocular 3D priors from depthmaps and surface normals, which can improve 3D sensing tasks and applications.


Out-of-distribution Detection Learning with Unreliable Out-of-distribution Sources

Haotian Zheng, Qizhou Wang, Zhen Fang, Xiaobo Xia, Feng Liu, Tongliang Liu, Bo Han

https://openreview.net/forum?id=87Qnneer8l

Keywords: out-of-distribution detection

Compressor summary: The paper proposes ATOL, a data generation-based learning method for OOD detection that uses an auxiliary task to relieve mistaken OOD generation and improve reliability.


Embracing the chaos: analysis and diagnosis of numerical instability in variational flows

Zuheng Xu, Trevor Campbell

https://openreview.net/forum?id=87Nu9SagB7

Keywords: variational flow, numerical instability, shadowing property

Compressor summary: The paper examines how numerical instability affects variational flows' reliability in sampling, density evaluation, and ELBO estimation, and proposes a diagnostic procedure to validate results from unstable flows.


Geometry-Informed Neural Operator for Large-Scale 3D PDEs

Zongyi Li, Nikola Borislavov Kovachki, Chris Choy, Boyi Li, Jean Kossaifi, Shourya Prakash Otta, Mohammad Amin Nabian, Maximilian Stadler, Christian Hundt, Kamyar Azizzadenesheli, Anima Anandkumar

https://openreview.net/forum?id=86dXbqT5Ua

Keywords: partial differential equation, computational fluid dynamics, neural operator

Compressor summary: GINO is an efficient method to learn solutions for large-scale PDEs with varying geometries, achieving significant speed-up and accuracy improvements over existing methods on 3D fluid simulation.


Nearest Neighbour with Bandit Feedback

Stephen Pasteris, Chris Hicks, Vasilios Mavroudis

https://openreview.net/forum?id=86ADcKOHAw

Keywords: Nearest Neighbours, Contextual Bandits

Compressor summary: The paper presents an efficient algorithm for contextual bandits that adapts the nearest neighbour rule and handles adversarial settings with minimal assumptions, providing regret bounds and application to online classification.


Convolutions Die Hard: Open-Vocabulary Segmentation with Single Frozen Convolutional CLIP

Qihang Yu, Ju He, Xueqing Deng, Xiaohui Shen, Liang-Chieh Chen

https://openreview.net/forum?id=83LJRUzXWj

Keywords: open-vocabulary panoptic segmentation, panoptic segmentation, vision and language

Compressor summary: The authors propose a single-stage framework for open-vocabulary segmentation using a frozen convolutional CLIP backbone that simplifies the existing two-stage pipeline, achieves better accuracy-cost trade-off, and outperforms previous methods on various benchmarks.


Expanding Small-Scale Datasets with Guided Imagination

Yifan Zhang, Daquan Zhou, Bryan Hooi, Kai Wang, Jiashi Feng

https://openreview.net/forum?id=82HeVCqsfh

Keywords: Dataset Expansion, Guided Imagination

Compressor summary: The paper proposes a Guided Imagination Framework that uses generative models to create new labeled data for dataset expansion, improving model accuracy on various image and medical datasets.


One-step differentiation of iterative algorithms

Jerome Bolte, Edouard Pauwels, Samuel Vaiter

https://openreview.net/forum?id=81snFfE3vR

Keywords: automatic differentiation, implicit differentiation, super-linear algorithms, bilevel optimization.

Compressor summary: One-step differentiation is an easy and efficient method for derivative estimation that works well for fast optimization algorithms and has theoretical and practical implications in bilevel optimization.


Fast Scalable and Accurate Discovery of DAGs Using the Best Order Score Search and Grow Shrink Trees

Bryan Andrews, Joseph Ramsey, Ruben Sanchez Romero, Jazmin Camchong, Erich Kummerfeld

https://openreview.net/forum?id=80g3Yqlo1a

Keywords: Causal Discovery, Directed Acyclic Graphs, DAGs, fMRI, Graphical Models, High Dimension, Densely Connected

Compressor summary: BOSS and GSTs are efficient algorithms for learning directed acyclic graphs (DAGs) from highly connected variables, achieving high accuracy and fast execution time in various conditions, and applied to fMRI data analysis.


A Simple Solution for Offline Imitation from Observations and Examples with Possibly Incomplete Trajectories

Kai Yan, Alex Schwing, Yu-Xiong Wang

https://openreview.net/forum?id=805CW5w2CY

Keywords: offline Imitation learning, learning from observations, positive-unlabeled learning

Compressor summary: TAILO is a novel imitation learning method that uses a discriminator to scale weights for weighted behavior cloning based on expert states, improving robustness and performance, especially when learning from incomplete trajectories.


LD2: Scalable Heterophilous Graph Neural Network with Decoupled Embeddings

Ningyi Liao, Siqiang Luo, Xiang Li, Jieming Shi

https://openreview.net/forum?id=7zkFc9TGKz

Keywords: Graph neural networks, Scalability, Heterophilous Graphs, Non-Homophily

Compressor summary: LD2 is a scalable heterophilous GNN that simplifies the learning process by decoupling graph propagation and generating expressive embeddings prior to training, achieving optimal time complexity and memory footprint, and improving speed and efficiency on large-scale graphs.


The Target-Charging Technique for Privacy Analysis across Interactive Computations

Edith Cohen, Xin Lyu

https://openreview.net/forum?id=7yjsYrajlt

Keywords: Differential Privacy; Adaptive Composition; Sparse Vector Technique

Compressor summary: TCT is a framework that improves privacy in interactive settings by charging a small overhead for computations that hit their targets, while keeping most others essentially free.


A Theory of Multimodal Learning

Zhou Lu

https://openreview.net/forum?id=7xlrdSOm3g

Keywords: Multimodal Learning

Compressor summary: The paper presents a theory explaining how multimodal machine learning can generalize better than unimodal learning, especially when there is diversity and connection among different sensory inputs.


HyTrel: Hypergraph-enhanced Tabular Data Representation Learning

Pei Chen, Soumajyoti Sarkar, Leonard Lausen, Balasubramaniam Srinivasan, Sheng Zha, Ruihong Huang, George Karypis

https://openreview.net/forum?id=7vqlzODS28

Keywords: Tabular Language Model, Tabular Representation Learning, Pretraining, Tabular Data, Table, Hypergraph

Compressor summary: HyTrel is a tabular language model that uses hypergraphs to capture structural properties of tabular data and improve performance on downstream tasks.


Federated Learning with Manifold Regularization and Normalized Update Reaggregation

Xuming An, Li Shen, Han Hu, Yong Luo

https://openreview.net/forum?id=7uPnuoYqac

Keywords: federated learning; manifold regularization; update reaggregation

Compressor summary: FedMRUR is a novel federated learning framework that uses hyperbolic graph manifolds and a new optimizer to reduce model inconsistency and improve convergence speed, achieving state-of-the-art results.


DP-Mix: Mixup-based Data Augmentation for Differentially Private Learning

Wenxuan Bao, Francesco Pittaluga, Vijay Kumar b g, Vincent Bindschaedler

https://openreview.net/forum?id=7rm3OcASkg

Keywords: differential privacy, deep learning, data augmentation

Compressor summary: The paper proposes two new data augmentation techniques for differentially private learning that improve performance and are compatible with privacy constraints.


ExPT: Synthetic Pretraining for Few-Shot Experimental Design

Tung Nguyen, Sudhanshu Agrawal, Aditya Grover

https://openreview.net/forum?id=7qfkImn0dL

Keywords: experimental design, few-shot, black-box optimization, synthetic pretraining, in-context learning, transformer

Compressor summary: The paper introduces Experiment Pretrained Transformers (ExPT), a foundation model for few-shot experimental design that uses synthetic pretraining and in-context learning to generate optimal input designs with limited labeled data.


MeGraph: Capturing Long-Range Interactions by Alternating Local and Hierarchical Aggregation on Multi-Scaled Graph Hierarchy

Honghua Dong, Jiawei Xu, Yu Yang, Rui Zhao, Shiwen Wu, Chun Yuan, Xiu Li, Chris J. Maddison, Lei Han

https://openreview.net/forum?id=7p5YWe8GqG

Keywords: Long-Range Interactions, Hierachical Structure, Multi-Scale, Graph Pooling, Graph Neural Networks(GNNs)

Compressor summary: MeGraph is a model that combines local and hierarchical information in a multi-scale graph hierarchy to better capture long-range interactions and outperforms existing methods on various benchmarks.


Dis-inhibitory neuronal circuits can control the sign of synaptic plasticity

Julian Rossbroich, Friedemann Zenke

https://openreview.net/forum?id=7otRtfrRqo

Keywords: Credit assignment, hebbian plasticity, inhibitory microcircuits, bio-plausible learning

Compressor summary: The paragraph discusses a microcircuit model that shows how error signals can be encoded in top-down synapses and influence Hebbian learning, resolving the discrepancy between functional models and experimental observations.


Energy-Efficient Scheduling with Predictions

Eric Balkanski, Noemie Perivier, Clifford Stein, Hao-Ting Wei

https://openreview.net/forum?id=7ntySBR3Ey

Keywords: Scheduling, algorithms with predictions, speed scaling, energy minimization

Compressor summary: The paper proposes a learning-augmented algorithmic framework for energy-efficient scheduling that improves performance when predictions are accurate and maintains bounded competitive ratios regardless of prediction error, as shown by empirical results on real and synthetic datasets.


AMAG: Additive, Multiplicative and Adaptive Graph Neural Network For Forecasting Neuron Activity

Jingyuan Li, Leo Scholl, Trung Le, Pavithra Rajeswaran, Amy L Orsborn, Eli Shlizerman

https://openreview.net/forum?id=7ntI4kcoqG

Keywords: Neuroscience and Cognitive Science, Neural Activity Forecasting, Graph Neural Network

Compressor summary: The text describes a model (AMAG) that uses deep learning and neural network interactions to forecast neural activity, capturing temporal causality and improving performance on synthetic and real neural data from macaques.


A Sublinear-Time Spectral Clustering Oracle with Improved Preprocessing Time

Ranran Shen, Pan Peng

https://openreview.net/forum?id=7nXaoclHed

Keywords: Sublinear-time algorithms, Spectral Clustering, Graph Clustering, Random Walks

Compressor summary: The paper proposes an efficient spectral clustering algorithm for graphs with strong clusterability, allowing sublinear-time preprocessing and queries, while handling some noise and deviation from ideal conditions.


Structure from Duplicates: Neural Inverse Graphics from a Pile of Objects

Tianhang Cheng, Wei-Chiu Ma, Kaiyu Guan, Antonio Torralba, Shenlong Wang

https://openreview.net/forum?id=7irm2VJARb

Keywords: 3d reconstruction, inverse rendering, pose estimation, single view reconstruction, nerf, duplicates

Compressor summary: The paragraph introduces Structure from Duplicates (SfD), a novel inverse graphics framework that reconstructs geometry, material, and illumination from a single image containing multiple identical objects, using them as a robust prior for single-image inverse graphics and proposing an in-plane rotation-robust Structure from Motion formulation for joint 6-DoF object pose estimation.


A Theory of Link Prediction via Relational Weisfeiler-Leman on Knowledge Graphs

Xingyue Huang, Miguel Romero Orth, Ismail Ilkan Ceylan, Pablo Barcelo

https://openreview.net/forum?id=7hLlZNrkt5

Keywords: graph neural networks, knowledge graphs, expressivity, logical characterization

Compressor summary: The authors analyze and compare different graph neural network models for link prediction on knowledge graphs and provide a logical characterization of their expressive power based on relational Weisfeiler-Leman algorithms.


Physics-Driven ML-Based Modelling for Correcting Inverse Estimation

Ruiyuan Kang, Tingting Mu, Panos Liatsis, Dimitrios Kyritsis

https://openreview.net/forum?id=7h1YaSGaHS

Keywords: Failure detection, Physical evaluation, Network-based optimization, Generative model, Hybrid surrogate model

Compressor summary: The authors propose GEESE, a method to detect and correct failed machine learning estimations in engineering problems using simulations and optimization, outperforming existing approaches.


Unsupervised Optical Flow Estimation with Dynamic Timing Representation for Spike Camera

Lujie Xia, Ziluo Ding, Rui Zhao, Jiyuan Zhang, Lei Ma, Zhaofei Yu, Tiejun Huang, Ruiqin Xiong

https://openreview.net/forum?id=7gbjsgcN5p

Keywords: Optical flow, unsupervised learning, spike camera

Compressor summary: The authors propose a dynamic timing representation for spike streams, an unsupervised learning method for optical flow estimation, and a synthetic validation dataset called SSES for evaluating their approach in autonomous driving scenarios.


Multi-Agent Learning with Heterogeneous Linear Contextual Bandits

Anh Do, Thanh Nguyen-Tang, Raman Arora

https://openreview.net/forum?id=7f6vH3mmhr

Keywords: Multi-agent, Bandits, Cooperative

Compressor summary: The paper proposes H-LINUCB, a novel distributed learning algorithm for linear contextual bandits, which optimally cooperates to minimize group regret when agents have knowledge of task similarity or dissimilarity.


Information Maximizing Curriculum: A Curriculum-Based Approach for Learning Versatile Skills

Denis Blessing, Onur Celik, Xiaogang Jia, Moritz Reuss, Maximilian Xiling Li, Rudolf Lioutikov, Gerhard Neumann

https://openreview.net/forum?id=7eW6NzSE4g

Keywords: Imitation Learning, Verstile Skill Learning, Curriculum Learning

Compressor summary: The paper proposes a curriculum-based imitation learning method that uses weighted data and mixture of experts policy with maximum entropy objective to learn versatile policies from multimodal human demonstrations.


Random-Access Infinite Context Length for Transformers

Amirkeivan Mohtashami, Martin Jaggi

https://openreview.net/forum?id=7eHn64wOVy

Keywords: large language models, memory, context length

Compressor summary: The paper proposes a new attention mechanism called landmark attention that allows Transformers to handle longer contexts without compromising flexibility or memory requirements.


Certification of Distributional Individual Fairness

Matthew Robert Wicker, Vihari Piratla, Adrian Weller

https://openreview.net/forum?id=7cnMLZvTy9

Keywords: Fairness, Individual Fairness, Deep Learning, Certification, Trustworthy ML

Compressor summary: The authors propose a novel method for certifying individual fairness in neural networks that is more computationally efficient and scalable than previous approaches, allowing them to analyze larger models and handle distributional shifts.


Beyond Geometry: Comparing the Temporal Structure of Computation in Neural Circuits with Dynamical Similarity Analysis

Mitchell Ostrow, Adam Joseph Eisen, Leo Kozachkov, Ila R Fiete

https://openreview.net/forum?id=7blSUMwe7R

Keywords: Computational Neuroscience, Neural Data Analysis, Statistical Shape Metrics, Representational Similarity Analysis, Recurrent Neural Networks, Dynamical Systems

Compressor summary: The paper introduces Dynamical Similarity Analysis (DSA), a novel method to compare recurrent neural networks based on their dynamics, which can distinguish different computations and learning rules that are not possible with standard geometric approaches.


Bias in Evaluation Processes: An Optimization-Based Model

L. Elisa Celis, Amit Kumar, Anay Mehrotra, Nisheeth K Vishnoi

https://openreview.net/forum?id=7b4oobeB4w

Keywords: bias, evaluation, maximum entropy, selection

Compressor summary: The paragraph discusses a model that explains how biases arise in evaluation processes, such as hiring or admissions, due to resource-information trade-offs and risk-averseness, and shows how this model can be used to study and reduce these biases.


Generalized equivalences between subsampling and ridge regularization

Pratik Patil, Jin-Hong Du

https://openreview.net/forum?id=7aoVQkNmQ6

Keywords: subsampling, ridge regularization, asymptotic equivalences, proportional asymptotics

Compressor summary: The text shows that subsampling and ridge regularization have similar effects on ensemble ridge estimators, and provides methods to find the equivalent parameters for both.


Information Theoretic Lower Bounds for Information Theoretic Upper Bounds

Roi Livni

https://openreview.net/forum?id=7anW5TWbCJ

Keywords: Learning Theory

Compressor summary: The paper explores how mutual information relates to generalization in stochastic convex optimization and finds that existing information-theoretic bounds do not fully capture the performance of some learning algorithms.


Bounding the Invertibility of Privacy-preserving Instance Encoding using Fisher Information

Kiwan Maeng, Chuan Guo, Sanjay Kariyappa, G. Edward Suh

https://openreview.net/forum?id=7ZQiucQu2u

Keywords: privacy, instance encoding, split learning

Compressor summary: The paper introduces a new measure for evaluating the privacy of instance encoding based on Fisher information, which can bound its invertibility both theoretically and empirically.


Joint processing of linguistic properties in brains and language models

SUBBA REDDY OOTA, Manish Gupta, Mariya Toneva

https://openreview.net/forum?id=7WeCyYy9TL

Keywords: Linguistic properties, fMRI, probing tasks, cognitive neuroscience, language models, NLP

Compressor summary: The correspondence between human brain processing and language models is affected by specific linguistic properties, particularly syntactic ones, which are crucial for maintaining alignment with fMRI recordings.


Masked Image Residual Learning for Scaling Deeper Vision Transformers

Guoxi Huang, Hongtao Fu, Adrian G. Bors

https://openreview.net/forum?id=7WTA298wts

Keywords: self-supervised learning, vision transformer, masked image modeling

Compressor summary: MIRL is a self-supervised learning framework that improves the training of deeper ViTs by teaching them to recover masked image residuals, leading to better accuracy and generalization on downstream tasks.


Generative Modeling through the Semi-dual Formulation of Unbalanced Optimal Transport

Jaemoo Choi, Jaewoong Choi, Myungjoo Kang

https://openreview.net/forum?id=7WQt1J13ex

Keywords: Optimal Transport, Generative modeling, Generative adversarial network

Compressor summary: The paper proposes a new generative model based on Unbalanced Optimal Transport, which improves robustness, stability, and convergence compared to traditional Optimal Transport methods, achieving lower FID scores on image generation tasks.


A State Representation for Diminishing Rewards

Ted Moskovitz, Samo Hromadka, Ahmed Touati, Diana L Borsa, Maneesh Sahani

https://openreview.net/forum?id=7Uix1eQZ8z

Keywords: reinforcement learning, successor features, successor representation, neuroscience

Compressor summary: The paper introduces a new state representation, the $\lambda$ representation (λR), that helps agents adapt to shifting reward priorities in multitask reinforcement learning and natural behaviors like foraging.


On student-teacher deviations in distillation: does it pay to disobey?

Vaishnavh Nagarajan, Aditya Krishna Menon, Srinadh Bhojanapalli, Hossein Mobahi, Sanjiv Kumar

https://openreview.net/forum?id=7UdVPRmpif

Keywords: knowledge distillation, regularization, understanding, underfitting, theory

Compressor summary: The paper investigates why students trained with knowledge distillation may overconfident but still outperform their teachers, and shows that this is due to an exaggerated bias of gradient descent that improves generalization.


Order Matters in the Presence of Dataset Imbalance for Multilingual Learning

Dami Choi, Derrick Xin, Hamid Dadkhahi, Justin Gilmer, Ankush Garg, Orhan Firat, Chih-Kuan Yeh, Andrew M. Dai, Behrooz Ghorbani

https://openreview.net/forum?id=7RMGI4slcb

Keywords: Multitask Optimization, Multilingual, Pre-training, Language Models, Language Sampling, Low Resource Languages, Overfitting

Compressor summary: This paper investigates how pre-training on high-resource tasks and fine-tuning on a mix of high/low-resource tasks improves multi-task learning with data imbalance, and demonstrates its effectiveness in neural machine translation and multi-lingual language modeling.


Tempo Adaptation in Non-stationary Reinforcement Learning

Hyunin Lee, Yuhao Ding, Jongmin Lee, Ming Jin, Javad Lavaei, Somayeh Sojoudi

https://openreview.net/forum?id=7R8noSP4vL

Keywords: Non-stationary RL, Reinforcement Learning

Compressor summary: The text proposes a method to synchronize time between an agent and a changing environment in reinforcement learning, improving policy performance.


Statistical and Computational Trade-off in Multi-Agent Multi-Armed Bandits

Filippo Vannella, Alexandre Proutiere, Jaeseong Jeong

https://openreview.net/forum?id=7PJ6LaIOO4

Keywords: Multi-Agent Multi-Armed Bandits, Multi-Armed Bandits, Regret Minimization

Compressor summary: The paper presents a regret minimization algorithm for MAMABs with factor graph rewards, which approximates a lower bound using Mean Field techniques and shows its performance in radio communications networks.


Dual Self-Awareness Value Decomposition Framework without Individual Global Max for Cooperative MARL

Zhiwei Xu, Bin Zhang, Dapeng Li, Guangchong Zhou, Zeren Zhang, Guoliang Fan

https://openreview.net/forum?id=7LtzqnfuOs

Keywords: Multi-Agent Reinforcement Learning, Individual Global Max

Compressor summary: The paper introduces a new value decomposition method for multi-agent reinforcement learning that rejects the Individual Global Max principle and uses dual self-awareness to solve the credit assignment problem and improve exploration.


Bounding training data reconstruction in DP-SGD

Jamie Hayes, Borja Balle, Saeed Mahloujifar

https://openreview.net/forum?id=7LZ4tZrYlx

Keywords: Differential privacy, reconstruction

Compressor summary: The text discusses how privacy levels in DP-SGD affect the risk of training data reconstruction and suggests that less noise may improve utility while emphasizing that the standard privacy guarantee may not reflect the actual risk.


Representation Equivalent Neural Operators: a Framework for Alias-free Operator Learning

Francesca Bartolucci, Emmanuel de Bezenac, Bogdan Raonic, Roberto Molinaro, Siddhartha Mishra, Rima Alaifari

https://openreview.net/forum?id=7LSEkvEGCM

Keywords: Operator Learning, Neural Operators, PDEs, Frame theory, Sampling theory

Compressor summary: The paper proposes ReNO, a framework to address issues in neural operator learning by measuring inconsistency between neural operators and their discrete representations using the concept of operator aliasing.


Adaptive Contextual Perception: How To Generalize To New Backgrounds and Ambiguous Objects

Zhuofan Ying, Peter Hase, Mohit Bansal

https://openreview.net/forum?id=7JuReDmGSL

Keywords: Computer vision, out-of-distribution generalization, representational geometry

Compressor summary: The paper investigates how biological vision systems adapt to context for out-of-distribution generalization and proposes new augmentation methods based on representational geometry analysis and causal intervention.


Long-Term Fairness with Unknown Dynamics

Tongxin Yin, Reilly Raab, Mingyan Liu, Yang Liu

https://openreview.net/forum?id=7INd5Yu9ET

Keywords: Long-term Fairness, Dynamics, Reinforcement Learning

Compressor summary: The paper proposes an online reinforcement learning approach for achieving long-term fairness in policy decisions affecting human populations, with algorithmic solutions that adapt to unknown dynamics and prove probabilistic bounds on loss and fairness violations.


AbDiffuser: full-atom generation of in-vitro functioning antibodies

Karolis Martinkus, Jan Ludwiczak, WEI-CHING LIANG, Julien Lafrance-Vanasse, Isidro Hotzel, Arvind Rajpal, Yan Wu, Kyunghyun Cho, Richard Bonneau, Vladimir Gligorijevic, Andreas Loukas

https://openreview.net/forum?id=7GyYpomkEa

Keywords: antibody generation, diffusion, equivariance

Compressor summary: AbDiffuser is a new diffusion model that generates antibody structures and sequences by using novel protein representation, physics-based constraints, and strong diffusion priors, achieving high in silico and in vitro performance.


Change point detection and inference in multivariate non-parametric models under mixing conditions

Carlos Misael Madrid Padilla, Haotian Xu, Daren Wang, OSCAR HERNAN MADRID PADILLA, Yi Yu

https://openreview.net/forum?id=7Fb2lCwS76

Keywords: Multivariate; Nonparametric; Change point inference; short range dependence; Long-run variance; Confidence interval.

Compressor summary: The paper analyzes how to find changes in the underlying distributions of multivariate time series with smooth densities and unknown change points.


Exact Optimality of Communication-Privacy-Utility Tradeoffs in Distributed Mean Estimation

Berivan Isik, Wei-Ning Chen, Ayfer Ozgur, Tsachy Weissman, Albert No

https://openreview.net/forum?id=7ETbK9lQd7

Keywords: distributed mean estimation, privacy, compression, communication, federated analytics.

Compressor summary: The paper explores how to estimate a population mean with privacy guarantees and shared randomness, proposing an exact-optimal mechanism based on a rotationally symmetric codebook and $k$-closest encoding.


AVIS: Autonomous Visual Information Seeking with Large Language Model Agent

Ziniu Hu, Ahmet Iscen, Chen Sun, Kai-Wei Chang, Yizhou Sun, David A Ross, Cordelia Schmid, Alireza Fathi

https://openreview.net/forum?id=7EMphtUgCI

Keywords: large language model, visual question answering, dynamic decision making, Tool augmented LLM

Compressor summary: AVIS is a framework that uses an LLM to strategize the use of external tools and analyze their outputs to answer complex visual questions by following a user study-informed sequence of actions.


Defending against Data-Free Model Extraction by Distributionally Robust Defensive Training

Zhenyi Wang, Li Shen, Tongliang Liu, Tiehang Duan, Yanjun Zhu, Donglin Zhan, David Doermann, Mingchen Gao

https://openreview.net/forum?id=7DZAVpOoAK

Keywords: Data-Free Model Extraction; Defense

Compressor summary: MeCo is a defense method against data-free model cloning that randomizes inputs to confuse attackers and protect the original model's utility.


Learning Energy-based Model via Dual-MCMC Teaching

Jiali Cui, Tian Han

https://openreview.net/forum?id=7962B4nXX7

Keywords: Energy-based model, MCMC, Joint-training, Generator model

Compressor summary: The paper proposes a joint learning framework for energy-based models that uses a complementary generator model to improve MCMC sampling and avoid bias in learning.


Language Model Tokenizers Introduce Unfairness Between Languages

Aleksandar Petrov, Emanuele La Malfa, Philip Torr, Adel Bibi

https://openreview.net/forum?id=78yDLKi95p

Keywords: LLM, language model, tokenizer, multilingual, language, fairness

Compressor summary: The paper argues that current language models are not fair to all languages and suggests using multilingually fair subword tokenizers for better performance.


IDEA: An Invariant Perspective for Efficient Domain Adaptive Image Retrieval

Haixin Wang, Hao Wu, Jinan Sun, Shikun Zhang, Chong Chen, Xian-Sheng Hua, Xiao Luo

https://openreview.net/forum?id=77i6itptQW

Keywords: domain adaption, binary descriptor, causal inference

Compressor summary: The paper proposes an IDEA model that learns to generate discriminative hash codes using causal and non-causal features, while minimizing the impact of non-causal effects on domain invariance for cross-domain retrieval.


Fragment-based Pretraining and Finetuning on Molecular Graphs

Kha-Dinh Luong, Ambuj Singh

https://openreview.net/forum?id=77Nq1KjmLl

Keywords: Self-supervised Learning, Graph Neural Network, Molecule

Compressor summary: The authors propose GraphFP, a pretraining method for GNNs that leverages fragment-level information and improves performance on molecular property prediction tasks.


Hierarchical clustering with dot products recovers hidden tree structure

Annie Gray, Alexander Modell, Patrick Rubin-Delanchy, Nick Whiteley

https://openreview.net/forum?id=75v88kyyko

Keywords: agglomerative clustering, generative model, graphical model, hierarchical clustering, high-dimensional data

Compressor summary: The paper proposes a simple variant of agglomerative clustering that better recovers hierarchical structure in data using maximum average dot product and shows its advantages over other methods.


No-Regret Learning in Dynamic Competition with Reference Effects Under Logit Demand

Mengzi Amy Guo, Donghao Ying, Javad Lavaei, Zuo-Jun Shen

https://openreview.net/forum?id=75Mxzfoeq7

Keywords: no-regret learning, price competition, reference effect, last-iterate convergence

Compressor summary: The paper proposes an algorithm for two firms to learn a stable equilibrium in a dynamic price competition with opaque information using online projected gradient ascent and stationary Nash equilibrium concepts.


Smoothing the Landscape Boosts the Signal for SGD: Optimal Sample Complexity for Learning Single Index Models

Alex Damian, Eshaan Nichani, Rong Ge, Jason D. Lee

https://openreview.net/forum?id=73XPopmbXH

Keywords: statistical learning, learning theory, single index model, gradient descent, stochastic gradient descent

Compressor summary: The paper studies how many samples are needed to learn a single index model with respect to an isotropic Gaussian distribution using online SGD, and shows that $n \gtrsim d^{k^\star/2}$ samples suffice, closing the gap between upper and lower bounds.


Understanding, Predicting and Better Resolving Q-Value Divergence in Offline-RL

Yang Yue, Rui Lu, Bingyi Kang, Shiji Song, Gao Huang

https://openreview.net/forum?id=71P7ugOGCV

Keywords: Offline RL, Theory

Compressor summary: The authors study the cause of Q-value estimation divergence in offline RL, propose a new metric to measure it, and introduce LayerNorm as a solution to improve extrapolation behavior and achieve better performance.


VPGTrans: Transfer Visual Prompt Generator across LLMs

Ao Zhang, Hao Fei, Yuan Yao, Wei Ji, Li Li, Zhiyuan Liu, Tat-Seng Chua

https://openreview.net/forum?id=716PvHoDct

Keywords: Visual Prompt Generator, Efficient Transfer, Multimodality

Compressor summary: This paper explores transferring visual prompt generators (VPG) between multimodal large language models (MLLMs) to reduce training costs and presents a two-stage transfer framework called VPGTrans that achieves significant efficiency improvements.


Causes and Effects of Unanticipated Numerical Deviations in Neural Network Inference Frameworks

Alexander Schlögl, Nora Hofer, Rainer Böhme

https://openreview.net/forum?id=6zyFgr1b8Q

Keywords: machine learning, security, reproducibility, forensics

Compressor summary: The study investigates numerical deviations in CNN inference results across different platforms due to hardware-specific optimizations, such as SIMD use on CPUs and convolution algorithms on GPUs.


Weitzman's Rule for Pandora's Box with Correlations

Evangelia Gergatsouli, Christos Tzamos

https://openreview.net/forum?id=6wBkT2ndDu

Keywords: pandora's box, stochastic optimization, discrete optimization, learning from samples, algorithms under uncertainty

Compressor summary: The paper studies Pandora's Box problem with correlated value distributions and shows that Weitzman's rule works for both independent and correlated cases, improving approximation guarantees and simplifying the algorithm.


PUe: Biased Positive-Unlabeled Learning Enhancement by Causal Inference

Xutao Wang, Hanting Chen, Tianyu Guo, Yunhe Wang

https://openreview.net/forum?id=6vtZIoxZoJ

Keywords: PU learning, causal inference, semi-supervised learning

Compressor summary: The paper proposes a PU learning enhancement algorithm using causal inference to improve accuracy on non-uniform label distribution datasets.


Efficient Uncertainty Quantification and Reduction for Over-Parameterized Neural Networks

Ziyi Huang, Henry Lam, Haofeng Zhang

https://openreview.net/forum?id=6vnwhzRinw

Keywords: frequentist uncertainty, epistemic uncertainty, procedural variability, confidence intervals, batching, cheap bootstrap

Compressor summary: The paragraph discusses a new method to quantify and remove uncertainty in deep learning models using a single auxiliary network and light-computation resampling methods.


Explaining Predictive Uncertainty with Information Theoretic Shapley Values

David Watson, Joshua O'Hara, Niek Tax, Richard Mudd, Ido Guy

https://openreview.net/forum?id=6rabAZhCRS

Keywords: Explainable AI, interpretable ML, feature attributions, information theory, Shapley values

Compressor summary: The authors propose a method to explain predictive uncertainty using Shapley values, which are related to information theory and conditional independence testing, and have applications in various tasks.


Where are we in the search for an Artificial Visual Cortex for Embodied Intelligence?

Arjun Majumdar, Karmesh Yadav, Sergio Arnaud, Yecheng Jason Ma, Claire Chen, Sneha Silwal, Aryan Jain, Vincent-Pierre Berges, Tingfan Wu, Jay Vakil, Pieter Abbeel, Jitendra Malik, Dhruv Batra, Yixin Lin, Oleksandr Maksymets, Aravind Rajeswaran, Franziska Meier

https://openreview.net/forum?id=6qLzQeFGio

Keywords: representation learning, pre-training, foundation models, embodied AI, reinforcement learning, imitation learning

Compressor summary: The paper evaluates various pre-trained visual representations for Embodied AI tasks using CortexBench, a new benchmark, and shows that adapting a large model to specific tasks improves performance.


Bayesian Optimization with Cost-varying Variable Subsets

Sebastian Shenghong Tay, Chuan-Sheng Foo, Daisuke Urano, Richalynn Leong, Bryan Kian Hsiang Low

https://openreview.net/forum?id=6oiux75UDj

Keywords: Bayesian optimization, Gaussian processes

Compressor summary: The paper proposes a novel algorithm for Bayesian optimization with cost-varying variable subsets, which balances informativeness and costs, and shows its superior performance over baselines.


Learning the Efficient Frontier

Philippe Chatigny, Ivan Sergienko, Ryan Ferguson, Jordan Weir, Maxime Bergeron

https://openreview.net/forum?id=6lnoUqFd5R

Keywords: Efficient Frontier, Convex Optimization, Resource Allocation, Constrainted Optimization, Finance

Compressor summary: NeuralEF is a framework that uses neural networks to quickly and accurately predict the optimal portfolio for a given risk level in the efficient frontier problem, considering heterogeneous constraints and variable inputs.


Mask Propagation for Efficient Video Semantic Segmentation

Yuetian Weng, Mingfei Han, Haoyu He, Mingjie Li, Lina Yao, Xiaojun Chang, Bohan Zhuang

https://openreview.net/forum?id=6ljXBlojde

Keywords: Video Semantic Segmentation; Inference Efficiency

Compressor summary: The paper proposes MPVSS, an efficient mask propagation framework for video semantic segmentation that reduces computational costs by reusing predictions from key frames and warping them to non-key frames using learned queries.


Language Model Alignment with Elastic Reset

Michael Noukhovitch, Samuel Lavoie, Florian Strub, Aaron Courville

https://openreview.net/forum?id=6lgugutkin

Keywords: reinforcement learning from human feedback (rlhf), language

Compressor summary: Elastic Reset is a new algorithm that reduces reward hacking and language drift in fine-tuning language models with reinforcement learning by periodically resetting the online model to an exponentially moving average of itself.


Towards A Richer 2D Understanding of Hands at Scale

Tianyi Cheng, Dandan Shan, Ayda Sultan Hassen, Richard Ely Locke Higgins, David Fouhey

https://openreview.net/forum?id=6ldTxwhgtP

Keywords: human-object interaction; hand object detection; hand detection

Compressor summary: The new model produces detailed outputs about hand interactions, using large-scale annotations from multiple datasets to improve AI understanding of hands in contact with objects.


AI for Interpretable Chemistry: Predicting Radical Mechanistic Pathways via Contrastive Learning

Mohammadamin Tavakoli, Pierre Baldi, Ann Marie Carlton, Yinting Chiu, Alexander Shmakov, David Van Vranken

https://openreview.net/forum?id=6kRQTPEVip

Keywords: Chemistry, Reactions, Contrastive, Radical, Graph

Compressor summary: The new reaction predictor system, RMechRP, uses contrastive learning and mechanistic pathways to provide accurate and interpretable predictions of radical reactions, overcoming limitations of existing deep learning-based methods that rely on US patents data.


Adversarially Robust Distributed Count Tracking via Partial Differential Privacy

Zhongzheng Xiong, Xiaoyi Zhu, Zengfeng Huang

https://openreview.net/forum?id=6kINNTYQcm

Keywords: Distributed Tracking, Adaptive Robustness, Differential Privacy, Generalization

Compressor summary: The paper studies how to track a function of items received by multiple sites communicating with a central server, considering adaptive adversaries and extending the differential privacy framework for optimal communication.


Synthetic Experience Replay

Cong Lu, Philip J. Ball, Yee Whye Teh, Jack Parker-Holder

https://openreview.net/forum?id=6jNQ1AY1Uf

Keywords: Reinforcement Learning, Diffusion Models, Synthetic Data, Sample-Efficient RL

Compressor summary: The authors propose Synthetic Experience Replay (SynthER), a method that uses generative models to upsample an agent's collected experience, improving the performance of deep reinforcement learning agents in both offline and online settings with limited data.


The Exact Sample Complexity Gain from Invariances for Kernel Regression

Behrooz Tahmasebi, Stefanie Jegelka

https://openreview.net/forum?id=6iouUxI45W

Keywords: invariances, manifolds, sample complexity

Compressor summary: The paper studies how encoding invariances into models reduces sample complexity for kernel ridge regression on compact manifolds with group action-invariant target functions, and presents a minimax optimal rate using differential geometry.


Statistical Analysis of Quantum State Learning Process in Quantum Neural Networks

Hao-Kai Zhang, Chenghong Zhu, Mingrui Jing, Xin Wang

https://openreview.net/forum?id=6gcY0MGNhj

Keywords: quantum neural networks, quantum state learning, quantum computing, quantum machine learning, quantum optimization

Compressor summary: The paper develops a no-go theorem for learning an unknown quantum state with quantum neural networks (QNNs) and shows that local minima become more difficult to avoid as the qubit count increases, limiting the learnability and scalability of QNNs.


When is Agnostic Reinforcement Learning Statistically Tractable?

Zeyu Jia, Gene Li, Alexander Rakhlin, Ayush Sekhari, Nathan Srebro

https://openreview.net/forum?id=6gWpJ0IExE

Keywords: Agnostic Reinforcement Learning, Sample Complexity, Learning Theory, Complexity Measure

Compressor summary: The paper studies the number of interactions needed to learn an approximately optimal policy for unknown MDPs using a new complexity measure called spanning capacity, and introduces a new algorithm called POPLER that enables efficient online RL with certain conditions.


Optimal and Fair Encouragement Policy Evaluation and Learning

Angela Zhou

https://openreview.net/forum?id=6fuZs3ibGA

Keywords: causal inference, fairness in machine learning, algorithmic fairness, criminal justice, policy learning, off-policy evaluation

Compressor summary: The paragraph discusses how to optimize treatment recommendations for individuals who may not follow them, considering factors like who benefits most and fairness, using a new online learning algorithm.


Conformalized matrix completion

Yu Gui, Rina Barber, Cong Ma

https://openreview.net/forum?id=6f320HfMeS

Keywords: matrix completion, conformal inference, uncertainty quantification

Compressor summary: The paper proposes a new method, conformalized matrix completion (cmc), which predicts missing entries in data matrices without requiring a low-rank assumption and provides valid prediction intervals using conformal prediction framework.


An Efficient End-to-End Training Approach for Zero-Shot Human-AI Coordination

Xue Yan, Jiaxian Guo, Xingzhou Lou, Jun Wang, Haifeng Zhang, Yali Du

https://openreview.net/forum?id=6ePsuwXUwf

Keywords: Zero-Shot Coordination, Human-AI coordination, Training Efficiency, Partner Modeling

Compressor summary: E3T is an efficient and diverse agent for zero-shot human-AI coordination that uses a mixture of ego policy and random policy, partner modeling, and end-to-end training to improve collaboration efficiency and performance.


Balancing Risk and Reward: A Batched-Bandit Strategy for Automated Phased Release

Yufan Li, Jialiang Mao, Iavor Bojinov

https://openreview.net/forum?id=6e86TccKyQ

Keywords: bandit algorithms, online learning, causality, Bayesian inference

Compressor summary: The paper proposes an algorithm that automatically determines the release percentage of new products or updates in phased releases, balancing risk and ramp-up speed.


Offline Imitation Learning with Variational Counterfactual Reasoning

Zexu Sun, Bowei He, Jinxin Liu, Xu Chen, Chen Ma, Shuai Zhang

https://openreview.net/forum?id=6d9Yxttb3w

Keywords: offline imitaion learning, counterfactual reasoning, data augmentation

Compressor summary: OILCA is a framework for offline imitation learning that uses counterfactual data augmentation to generate expert data and improve the agent's generalization ability.


Globally injective and bijective neural operators

Takashi Furuya, Michael Anthony Puthawala, Matti Lassas, Maarten V. de Hoop

https://openreview.net/forum?id=6cc69ArD3O

Keywords: Deep Learning, Operator Learning, Functional Analysis, Injectivity, Bijectivity, Universal approximation

Compressor summary: This paper investigates when neural networks can learn injective and surjective operators between function spaces and their applications in uncertainty quantification and inverse problems.


Strategic Classification under Unknown Personalized Manipulation

Han Shao, Avrim Blum, Omar Montasser

https://openreview.net/forum?id=6cJKcIxPck

Keywords: strategic classification, mistake bound in online learning, PAC learning

Compressor summary: The study analyzes how strategic feature manipulation affects learning in classification problems with different levels of information available to the learner, focusing on ball and non-ball manipulations.


HAP: Structure-Aware Masked Image Modeling for Human-Centric Perception

Junkun Yuan, Xinyu Zhang, Hao Zhou, Jian Wang, Zhongwei Qiu, Zhiyin Shao, Shaofeng Zhang, Sifan Long, Kun Kuang, Kun Yao, Junyu Han, Errui Ding, Lanfen Lin, Fei Wu, Jingdong Wang

https://openreview.net/forum?id=6XPPfZkhKi

Keywords: human centric perception, masked image modeling, structural-aware pre-training

Compressor summary: The paper introduces HAP, a pre-training method for human-centric perception tasks that uses masked image modeling with human structure priors to improve performance on various benchmarks.


DELTA: Diverse Client Sampling for Fasting Federated Learning

Lin Wang, Yongxin Guo, Tao Lin, Xiaoying Tang

https://openreview.net/forum?id=6XC5iKqRVm

Keywords: federated learning, client sampling

Compressor summary: DELTA is an unbiased sampling scheme for Federated Learning that considers client diversity and local variance to improve convergence and reduce variance caused by partial client participation.


Adversarial Model for Offline Reinforcement Learning

Mohak Bhardwaj, Tengyang Xie, Byron Boots, Nan Jiang, Ching-An Cheng

https://openreview.net/forum?id=6UCMa0Qgej

Keywords: model based, offline, reinforcement learning, adversarial training

Compressor summary: ARMOR is a novel model-based offline RL framework that learns policies to improve upon any reference policy, optimizing for worst-case performance and being robust to hyperparameters.


Preference-grounded Token-level Guidance for Language Model Fine-tuning

Shentao Yang, Shujian Zhang, Congying Xia, Yihao Feng, Caiming Xiong, Mingyuan Zhou

https://openreview.net/forum?id=6SRE9GZ9s6

Keywords: Preference Learning, Training Guidance Learning, Language Model Fine-tuning, Text Sequence Generation

Compressor summary: The paper proposes a method to align language models with preferences by iterating between grounding preferences into token-level training and improving the LM with learned guidance, using a framework that extends pairwise-preference learning and two minimalist learning objectives.


Robust Lipschitz Bandits to Adversarial Corruptions

Yue Kang, Cho-Jui Hsieh, Thomas Chun Man Lee

https://openreview.net/forum?id=6RiqluMFNz

Keywords: bandits

Compressor summary: The paper introduces robust Lipschitz bandit algorithms that can handle adaptive adversaries corrupting stochastic rewards up to a budget $C$, achieving sub-linear regret even when $C$ is unknown, and provides lower bounds and experiments.


ComSL: A Composite Speech-Language Model for End-to-End Speech-to-Text Translation

Chenyang Le, Yao Qian, Long Zhou, Shujie LIU, Yanmin Qian, Michael Zeng, Xuedong Huang

https://openreview.net/forum?id=6Qx7G1xrAk

Keywords: end-to-end speech to text translation, cross-modality learning, joint speech and language training

Compressor summary: ComSL is a data-efficient speech-language model that combines cross-modality learning and multi-task learning to achieve state-of-the-art performance in speech-to-text translation across 21 languages.


Probabilistic Weight Fixing: Large-scale training of neural network weight uncertainties for quantisation.

Chris Subia-Waud, Srinandan Dasmahapatra

https://openreview.net/forum?id=6Odmtoek02

Keywords: Quantization, compression, bayesian neural networks, accelerators

Compressor summary: The paper proposes a probabilistic method to cluster neural network weights based on their position-specific uncertainty, improving compression and accuracy in large models.


TempME: Towards the Explainability of Temporal Graph Neural Networks via Motif Discovery

Jialin Chen, Zhitao Ying

https://openreview.net/forum?id=6OOgw4boZI

Keywords: Explainability, Temporal Graph Neural Network

Compressor summary: TempME is a novel approach that uncovers the most pivotal temporal motifs guiding the prediction of temporal graph neural networks, improving their explainability and trustworthiness.


Decision-Aware Actor-Critic with Function Approximation and Theoretical Guarantees

Sharan Vaswani, Amirreza Kazemi, Reza Babanezhad Harikandeh, Nicolas Le Roux

https://openreview.net/forum?id=6MQ5cheYDZ

Keywords: Decision-aware reinforcement learning, Actor-Critic algorithm, Off-policy updates, General function approximation, Theoretical guarantees

Compressor summary: The paper proposes a joint objective for training actor and critic in reinforcement learning that improves policy improvement and uses surrogate functions similar to TRPO and PPO.


Polyhedron Attention Module: Learning Adaptive-order Interactions

Tan Zhu, Fei Dou, Xinyu Wang, Jin Lu, Jinbo Bi

https://openreview.net/forum?id=6JrckqCxtl

Keywords: Feature interaction modeling, model interpretation framework, adptive-order interaction, piece-wise polynomial

Compressor summary: PAM is a new module that allows neural networks to model nonlinear feature interactions adaptively using polyhedrons, improving predictive performance and interpretability.


Learning World Models with Identifiable Factorization

Yu-Ren Liu, Biwei Huang, Zhengmao Zhu, Honglong Tian, Mingming Gong, Yang Yu, Kun Zhang

https://openreview.net/forum?id=6JJq5TW9Mc

Keywords: Model-based Reinforcement Learning; Causal Representation Learning;

Compressor summary: IFactor is a framework that models four categories of latent state variables to efficiently represent information in high-dimensional, noisy, and non-stationary environments for reinforcement learning.


Hierarchical Randomized Smoothing

Yan Scholten, Jan Schuchardt, Aleksandar Bojchevski, Stephan Günnemann

https://openreview.net/forum?id=6IhNHKyuJO

Keywords: Adversarial Robustness, Robustness Certification, Randomized Smoothing, Graph Neural Networks

Compressor summary: Hierarchical randomized smoothing adds targeted noise to subsets of entities in complex objects, improving robustness while maintaining accuracy in classification tasks.


Certified Minimax Unlearning with Generalization Rates and Deletion Capacity

Jiaqi Liu, Jian Lou, Zhan Qin, Kui Ren

https://openreview.net/forum?id=6H8Md75kAw

Keywords: machine unlearning, machin learning privacy, minimax learning, certified removal

Compressor summary: The paper proposes a new certified machine unlearning algorithm for minimax models that uses a total-Hessian-based complete Newton update and Gaussian mechanism, and achieves better generalization and deletion capacity than existing methods.


Revisiting Scalarization in Multi-Task Learning: A Theoretical Perspective

Yuzheng Hu, Ruicheng Xian, Qilong Wu, Qiuling Fan, Lang Yin, Han Zhao

https://openreview.net/forum?id=6EqUpqMnwl

Keywords: multi-task learning, scalarization, Pareto front

Compressor summary: The paper investigates whether scalarization can fully explore the Pareto front for linear multi-task learning models and finds that it is inherently incapable of doing so, especially for balanced trade-offs between tasks.


LinkerNet: Fragment Poses and Linker Co-Design with 3D Equivariant Diffusion

Jiaqi Guan, Xingang Peng, PeiQi Jiang, Yunan Luo, Jian Peng, Jianzhu Ma

https://openreview.net/forum?id=6EaLIw3W7c

Keywords: Linker design, generative models

Compressor summary: The paper proposes a new approach to design stable drug-candidate molecules for targeted protein degradation by jointly learning fragment poses and linker structures using a 3D equivariant diffusion model.


DDF-HO: Hand-Held Object Reconstruction via Conditional Directed Distance Field

Chenyangguang Zhang, Yan Di, Ruida Zhang, Guangyao Zhai, Fabian Manhardt, Federico Tombari, Xiangyang Ji

https://openreview.net/forum?id=6EDHfVHicP

Keywords: hand-held object reconstruction, directed distance field, human-object interaction

Compressor summary: The proposed DDF-HO method uses Directed Distance Fields to represent shapes and capture hand-object interactions more effectively than existing Signed Distance Field methods, achieving better performance on synthetic and real-world datasets.


In-Context Learning Unlocked for Diffusion Models

Zhendong Wang, Yifan Jiang, Yadong Lu, yelong shen, Pengcheng He, Weizhu Chen, Zhangyang Wang, Mingyuan Zhou

https://openreview.net/forum?id=6BZS2EAkns

Keywords: diffusion models, in-context learning

Compressor summary: Prompt Diffusion is a framework that enables in-context learning in diffusion-based generative models using vision-language prompts and a trained diffusion model, achieving high-quality results on various tasks and unseen vision tasks.


Causal Fairness for Outcome Control

Drago Plecko, Elias Bareinboim

https://openreview.net/forum?id=6AAbWSF6Qg

Keywords: Fair Machine Learning, Causal Inference, Decision-Making

Compressor summary: The paper studies outcome control, where automated systems optimize an outcome variable while ensuring fairness and equity by considering sensitive attributes like gender, race, and religion.


Logarithmic-Regret Quantum Learning Algorithms for Zero-Sum Games

Minbo Gao, Zhengfeng Ji, Tongyang Li, Qisheng Wang

https://openreview.net/forum?id=69dAz94zPv

Keywords: Online learning, quantum computing, zero-sum games, linear programming, optimistic multiplicative weight update

Compressor summary: The paper presents a quantum algorithm that finds an approximate Nash equilibrium for zero-sum games quickly and with succinct outputs, using a novel quantum multi-sampling technique.


Counterfactual Memorization in Neural Language Models

Chiyuan Zhang, Daphne Ippolito, Katherine Lee, Matthew Jagielski, Florian Tramèr, Nicholas Carlini

https://openreview.net/forum?id=67o9UQgTD0

Keywords: Memorization, Language Models

Compressor summary: The paper proposes a method to study and quantify language model memorization by examining how predictions change when certain documents are removed from training data.


Revisit the Power of Vanilla Knowledge Distillation: from Small Scale to Large Scale

Zhiwei Hao, Jianyuan Guo, Kai Han, Han Hu, Chang Xu, Yunhe Wang

https://openreview.net/forum?id=67MTWzhEOn

Keywords: knowledge distillation, small-data pitfall, vanilla kd

Compressor summary: The paper argues that using larger datasets and stronger data augmentation techniques can improve knowledge distillation approaches for limited-capacity architectures, and demonstrates this with state-of-the-art results on ImageNet.


A Competitive Algorithm for Agnostic Active Learning

Yihan Zhou, Eric Price

https://openreview.net/forum?id=66XhNDahk6

Keywords: active learning, binary classification, competitive ratio

Compressor summary: The paper proposes a new agnostic active learning algorithm that is competitive with the optimal algorithm and has a logarithmic overhead in the query complexity, and shows that improving this is NP-hard.


Computational Complexity of Learning Neural Networks: Smoothness and Degeneracy

Amit Daniely, Nathan Srebro, Gal Vardi

https://openreview.net/forum?id=65aDEXIhih

Keywords: Learning neural networks, Computational complexity, Hardness of learning, Smoothed analysis, Degenerate weights

Compressor summary: The paper investigates if assumptions about Gaussian inputs and non-degenerate weights are enough to learn deeper neural networks efficiently, and finds that they are not for depth-$3$ ReLU networks.


No-Regret Learning with Unbounded Losses: The Case of Logarithmic Pooling

Eric Neyman, Tim Roughgarden

https://openreview.net/forum?id=639RkUOmW8

Keywords: Logarithmic pooling, online learning, no-regret learning, calibrated experts, online mirror descent, prediction with expert advice

Compressor summary: The paper proposes a novel algorithm for learning expert weights in an online adversarial setting, where experts' forecasts are consistent and calibrated, to minimize log loss using logarithmic pooling and achieve low regret.


Counterfactual Conservative Q Learning for Offline Multi-agent Reinforcement Learning

Jianzhun Shao, Yun Qu, Chen Chen, Hongchang Zhang, Xiangyang Ji

https://openreview.net/forum?id=62zmO4mv8X

Keywords: multi-agent reinforcement learning, offline reinforcement learning

Compressor summary: The paper proposes CounterFactual Conservative Q-Learning (CFCQL), a new multi-agent offline RL algorithm that estimates conservative values for each agent separately, improving performance and reducing the impact of out-of-distribution actions and value overestimation.


MosaicBERT: A Bidirectional Encoder Optimized for Fast Pretraining

Jacob Portes, Alexander R Trott, Sam Havens, DANIEL KING, Abhinav Venigalla, Moin Nadeem, Nikhil Sardana, Daya Khudia, Jonathan Frankle

https://openreview.net/forum?id=5zipcfLC2Z

Keywords: BERT, Pretraining, Efficiency, FlashAttention, ALiBi

Compressor summary: MosaicBERT is a fast and efficient BERT-style encoder architecture that allows for low-cost pretraining, enabling researchers to customize their own NLP models.


No Representation Rules Them All in Category Discovery

Sagar Vaze, Andrea Vedaldi, Andrew Zisserman

https://openreview.net/forum?id=5ytypAqAsR

Keywords: Category discovery, semi-supervised learning, self-supervised learning, classification

Compressor summary: The paper introduces Clevr-4, a synthetic dataset for Generalized Category Discovery, and proposes a novel method called $\mu$GCD that leverages consistent findings from representation learning to outperform existing approaches on both synthetic and real data.


Many-body Approximation for Non-negative Tensors

Kazu Ghalamkari, Mahito Sugiyama, Yoshinobu Kawahara

https://openreview.net/forum?id=5yedZXV7wt

Keywords: Tensor decomposition, Energy based model, Tensor networks

Compressor summary: The paper introduces many-body approximation, a new method for decomposing non-negative tensors that avoids problems with traditional methods by using energy-based modeling and global optimization.


Amortized Reparametrization: Efficient and Scalable Variational Inference for Latent SDEs

Kevin Course, Prasanth B. Nair

https://openreview.net/forum?id=5yZiP9fZNv

Keywords: variational inference, differential equations, dynamical systems, neural ordinary differential equations, latent stochastic differential equations

Compressor summary: The paragraph describes a novel method for inferring latent stochastic differential equations with low computational cost and memory usage, using amortization and a reparametrization technique.


StreamNet: Memory-Efficient Streaming Tiny Deep Learning Inference on the Microcontroller

Hong Sheng Zheng, Yu-Yuan Liu, Chen-Fong Hsu, Tsung Tai Yeh

https://openreview.net/forum?id=5t5u8PQa2T

Keywords: TinyML models, edge AIs, Microcontroller

Compressor summary: StreamNet is a method that improves the performance and memory efficiency of deploying TinyML models on low-power MCUs using 1D and 2D streaming processing and smart parameter selection.


Module-wise Training of Neural Networks via the Minimizing Movement Scheme

Skander Karkar, Ibrahim Ayed, Emmanuel de Bezenac, patrick gallinari

https://openreview.net/forum?id=5sV53leJCv

Keywords: Deep learning, greedy layerwise training, memory, optimal transport

Compressor summary: The paper introduces TRGL, a module-wise regularization method for neural networks that improves accuracy and reduces memory usage in limited memory settings.


Composing Parameter-Efficient Modules with Arithmetic Operation

Jinghan Zhang, Shiqi Chen, Junteng Liu, Junxian He

https://openreview.net/forum?id=5r3e27I9Gy

Keywords: Parameter-efficient fine-tuning, module composition

Compressor summary: The paper proposes a method to compose different parameter-efficient modules using linear arithmetic operations in the weight space for various tasks, improving their performance.


Pairwise Causality Guided Transformers for Event Sequences

Xiao Shou, Debarun Bhattacharjya, Tian Gao, Dharmashankar Subramanian, Oktie Hassanzadeh, Kristin Bennett

https://openreview.net/forum?id=5q8xovQF7r

Keywords: temporal event sequences, causal inference, transformer, causal knowledge graph

Compressor summary: The paper proposes a new method to improve transformer-based models' performance in predicting temporal event sequences using pairwise causal knowledge, and shows its effectiveness in several experiments and applications.


Compression with Bayesian Implicit Neural Representations

Zongyu Guo, Gergely Flamich, Jiajun He, Zhibo Chen, José Miguel Hernández-Lobato

https://openreview.net/forum?id=5otj6QKUMI

Keywords: Neural Compression, Implicit Neural Representation, Relative Entropy Coding, Bayesian Neural Network

Compressor summary: The authors propose a new method to compress data by overfitting neural networks to their functional representation and encoding the weights using relative entropy coding, which improves rate-distortion performance and allows adjusting trade-offs for different network architectures.


Cross-Scale MAE: A Tale of Multiscale Exploitation in Remote Sensing

Maofeng Tang, Andrei Liviu Cozma, Konstantinos Georgiou, Hairong Qi

https://openreview.net/forum?id=5oEVdOd6TV

Keywords: Remote Sensting, Self-Supervised Learning

Compressor summary: The paper introduces Cross-Scale MAE, a self-supervised model that learns consistent and meaningful representations for remote sensing image analysis using scale augmentation, cross-scale consistency constraints, contrastive and generative losses, and xFormers library.


LinGCN: Structural Linearized Graph Convolutional Network for Homomorphically Encrypted Inference

Hongwu Peng, Ran Ran, Yukui Luo, Jiahui Zhao, Shaoyi Huang, Kiran Thorat, Tong Geng, Chenghong Wang, Xiaolin Xu, Wujie Wen, Caiwen Ding

https://openreview.net/forum?id=5loV5tVzsY

Keywords: Privacy-Preserving Machine Learning, efficient private inference, machine learning as a service, homomorphic encryption, non-linear pruning, ST-GCN

Compressor summary: LinGCN is a framework that optimizes the performance of homomorphic encryption-based Graph Convolution Networks by reducing multiplication depth and improving latency, accuracy, and scalability for encrypted inference in applications like personal healthcare and financial systems.


Evolving Standardization for Continual Domain Generalization over Temporal Drift

Mixue Xie, Shuang Li, Longhui Yuan, Chi Harold Liu, Zehui Dai

https://openreview.net/forum?id=5hVXbiEGXB

Keywords: domain generalization, sequential learning, temporal drift, feature standardization

Compressor summary: EvoS is a method for continual domain generalization over temporal drift that uses multi-scale attention to learn evolving feature distribution patterns and standardize features with generated statistics, enabling models to adapt to gradually shifting data distributions in real-world applications.


A Cross-Moment Approach for Causal Effect Estimation

Yaroslav Kivva, Saber Salehkaleybar, Negar Kiyavash

https://openreview.net/forum?id=5gz7npbQ6Z

Keywords: Causal inference, Difference-in-Difference, Structural causal models, Potential outcome, Proxy learning

Compressor summary: The paper proposes a new method to estimate the causal effect of a treatment on an outcome using cross moments, when only one proxy variable is available, and shows its advantages over existing methods that require restrictive assumptions or two proxy variables.


Quantification of Uncertainty with Adversarial Models

Kajetan Schweighofer, Lukas Aichberger, Mykyta Ielanskyi, Günter Klambauer, Sepp Hochreiter

https://openreview.net/forum?id=5eu00pcLWa

Keywords: uncertainty, uncertainty quantification, predictive uncertainty, epistemic uncertainty, out of distribution, mc dropout, deep ensembles, sg-mcmc, adversarial model, adversarial model search, imagenet

Compressor summary: QUAM is a new method to estimate epistemic uncertainty better than current methods by focusing on regions where the product of divergence function and posterior is large, which corresponds to adversarial models.


Replicability in Reinforcement Learning

Amin Karbasi, Grigoris Velegkas, Lin Yang, Felix Zhou

https://openreview.net/forum?id=5cPz5hrjy6

Keywords: Theory, Reinforcement Learning Theory, Statistical Learning Theory, Reproducibility, Replicability

Compressor summary: The authors study replicability in reinforcement learning algorithms and provide efficient algorithms with theoretical guarantees for various settings.


The Goldilocks of Pragmatic Understanding: Fine-Tuning Strategy Matters for Implicature Resolution by LLMs

Laura Eline Ruis, Akbir Khan, Stella Biderman, Sara Hooker, Tim Rocktäschel, Edward Grefenstette

https://openreview.net/forum?id=5bWW9Eop7l

Keywords: large language models, pragmatics, natural language processing, communication, conversation, implicature, language model fine-tuning

Compressor summary: The paragraph discusses a study on whether large language models can understand implicatures, a type of inference based on beliefs and knowledge, and finds that instruction-tuned models perform better than others in this task.


Approximately Equivariant Graph Networks

Ningyuan Teresa Huang, Ron Levie, Soledad Villar

https://openreview.net/forum?id=5aeyKAZr0L

Keywords: graph neural networks, equivariant machine learning, symmetry, generalization, statistical learning

Compressor summary: The paper explores how active symmetries in GNNs affect their performance on various tasks and proposes a bias-variance formula to choose an appropriate symmetry group for each task.


Self-Adaptive Motion Tracking against On-body Displacement of Flexible Sensors

Chengxu Zuo, Jiawei Fang, Shihui Guo, Yipeng Qin

https://openreview.net/forum?id=5ZMBiS1uMq

Keywords: motion tracking, flexible sensor, on-body displacement, deep learning, domain adaptation

Compressor summary: The authors propose a self-adaptive motion tracking network to deal with the challenges posed by flexible sensors' displacement on human body during different sessions, using a light-weight learnable layer, a Fourier-encoded LSTM network, and a novel sequence discrepancy loss.


Understanding Contrastive Learning via Distributionally Robust Optimization

Junkang Wu, Jiawei Chen, Jiancan Wu, Wentao Shi, Xiang Wang, Xiangnan He

https://openreview.net/forum?id=5XshcizH9w

Keywords: contrastive learning, distributionally robust optimization, mutual information

Compressor summary: The study shows that contrastive learning is robust to sampling bias, explains its connection to distributionally robust optimization and mutual information, and proposes a new loss function to improve performance and convergence.


Tree of Thoughts: Deliberate Problem Solving with Large Language Models

Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L. Griffiths, Yuan Cao, Karthik R Narasimhan

https://openreview.net/forum?id=5Xc1ecxO1h

Keywords: large language model, general problem solving, heuristic search, reasoning, planning, decision making

Compressor summary: The Tree of Thoughts framework improves language models' problem-solving abilities by enabling them to explore, plan, and evaluate multiple reasoning paths in tasks requiring non-trivial planning or search.


Characteristic Circuits

Zhongjie Yu, Martin Trapp, Kristian Kersting

https://openreview.net/forum?id=5W7cXno10k

Keywords: Characteristic Circuit, Characteristic Function, Probabilistic Circuit, Heterogeneous Data, Density Estimation

Compressor summary: Characteristic circuits are a new type of probabilistic model that can learn high-dimensional distributions over different types of data and perform efficient inference, even when no closed-form density function is available.


Replicable Clustering

Hossein Esfandiari, Amin Karbasi, Vahab Mirrokni, Grigoris Velegkas, Felix Zhou

https://openreview.net/forum?id=5VQFAvUHcd

Keywords: Theory, Clustering Theory, Statistical Learning Theory, Reproducibility, Replicability

Compressor summary: The authors propose replicable algorithms for statistical clustering problems like $k$-medians, $k$-means, and $k$-centers by using approximation routines for their combinatorial versions in a black-box manner.


Evaluating the Robustness of Interpretability Methods through Explanation Invariance and Equivariance

Jonathan Crabbé, Mihaela van der Schaar

https://openreview.net/forum?id=5UwnKSgY6u

Keywords: interpretability, explainability, robustness, invariance, equivariance, geometric deep learning

Compressor summary: The paper proposes a framework to measure and improve the robustness of interpretability methods for neural networks that are invariant under specific symmetry groups.


Dynamic Prompt Learning: Addressing Cross-Attention Leakage for Text-Based Image Editing

Kai Wang, Fei Yang, Shiqi Yang, Muhammad Atif Butt, Joost van de Weijer

https://openreview.net/forum?id=5UXXhVI08r

Keywords: Diffusion Models; Text-guided Image Edit; Textual Inversion; Localization

Compressor summary: Dynamic Prompt Learning (DPL) improves text-to-image generation by focusing cross-attention maps on correct nouns in the prompt, enabling fine-grained image editing without unwanted changes.


Uni3DETR: Unified 3D Detection Transformer

Zhenyu Wang, Ya-Li Li, Xi Chen, Hengshuang Zhao, Shengjin Wang

https://openreview.net/forum?id=5UOYGfobhC

Keywords: 3d object detection, unified object detection, point clouds

Compressor summary: Uni3DETR is a unified 3D detector that works well for both indoor and outdoor scenes by using a detection transformer with point-voxel interaction and a mixture of query points.


Variational Inference with Gaussian Score Matching

Chirag Modi, Robert M. Gower, Charles Margossian, Yuling Yao, David Blei, Lawrence K. Saul

https://openreview.net/forum?id=5TTV5IZnLL

Keywords: Variational Inference, score matching, KL projection, polyak stepsize

Compressor summary: The authors propose a new variational inference method called Gaussian score matching VI (GSM-VI), which uses the principle of score matching and is faster and equally or more accurate than black box variational inference (BBVI) for various Bayesian inference problems.


Inconsistency, Instability, and Generalization Gap of Deep Neural Network Training

Rie Johnson, Tong Zhang

https://openreview.net/forum?id=5SIz31OGFV

Keywords: Deep neural network training, Generalization gap, Empirical study

Compressor summary: The paper analyzes how inconsistency and instability of model outputs affect the generalization gap in deep neural networks, and shows that reducing inconsistency improves performance.


Smoothed Analysis of Sequential Probability Assignment

Alankrita Bhatt, Nika Haghtalab, Abhishek Shetty

https://openreview.net/forum?id=5R9bZlpZKj

Keywords: Online learning, Log loss, Information theory, Smoothed Analysis, Beyond worst case analysis, Oracle Efficient Online Learning

Compressor summary: The study explores smoothed analysis for sequential probability assignment problem with contexts, leading to optimal fast rates and efficient algorithms for minimax learning.


Debias Coarsely, Sample Conditionally: Statistical Downscaling through Optimal Transport and Probabilistic Diffusion Models

Zhong Yi Wan, Ricardo Baptista, Anudhyan Boral, Yi-Fan Chen, John Anderson, Fei Sha, Leonardo Zepeda-Nunez

https://openreview.net/forum?id=5NxJuc0T1P

Keywords: optimal transport, probabilistic diffusion models, statistical downscaling

Compressor summary: The paper presents a two-stage probabilistic framework that uses unpaired data to downscale low-resolution numerical simulations of fluid flow to high-resolution data, while preserving the physical statistics and addressing the core difficulties in weather and climate modeling.


Global Optimality in Bivariate Gradient-based DAG Learning

Chang Deng, Kevin Bello, Pradeep Kumar Ravikumar, Bryon Aragam

https://openreview.net/forum?id=5MG5C5aS6m

Keywords: global optimization, nonconvex optimization, graphical models, directed acyclic graphs, structure learning

Compressor summary: The paper shows how a simple optimization method can find the best solution for learning an acyclic graphical model from data without getting stuck in spurious solutions.


Fast Bellman Updates for Wasserstein Distributionally Robust MDPs

Zhuodong Yu, Ling Dai, Shaohang Xu, Siyang Gao, Chin Pang Ho

https://openreview.net/forum?id=5La4Y8BnQw

Keywords: Markov decision processes, distributionally robust optimization

Compressor summary: The paper introduces a fast algorithm for solving distributionally robust MDPs with Wasserstein ambiguity sets, which reduces computational cost and improves performance over existing methods.


GeoTMI: Predicting Quantum Chemical Property with Easy-to-Obtain Geometry via Positional Denoising

Hyeonsu Kim, Jeheon Woo, SEONGHWAN KIM, Seokhyun Moon, Jun Hyeong Kim, Woo Youn Kim

https://openreview.net/forum?id=5JcKKRX2iH

Keywords: Mutual information, Easy-to-obtain geometry, Denoising, 3D Graph neural network, OC20

Compressor summary: GeoTMI is a new framework that uses denoising to predict quantum chemical properties accurately using easy-to-obtain corrupted geometries.


Computational Guarantees for Doubly Entropic Wasserstein Barycenters

Tomas Vaskevicius, Lénaïc Chizat

https://openreview.net/forum?id=5HahZRA0fy

Keywords: Wasserstein barycenters, entropic penalization, optimal transport, Sinkhorn's algorithm

Compressor summary: The paper presents an algorithm for computing doubly regularized Wasserstein barycenters with convergence guarantees and extends it to handle discrete point clouds using Monte Carlo sampling.


PAC Learning Linear Thresholds from Label Proportions

Anand Paresh Brahmbhatt, Rishi Saket, Aravindan Raghuveer

https://openreview.net/forum?id=5Gw9YkJkFF

Keywords: PAC learning, Learning from label proportions, Linear thresholds

Compressor summary: The paper presents efficient algorithms for learning linear threshold functions from label proportions using Gaussian distributions and shows their effectiveness through experiments.


Strong and Precise Modulation of Human Percepts via Robustified ANNs

Guy Gaziv, Michael J. Lee, James J. DiCarlo

https://openreview.net/forum?id=5GmTI4LNqX

Keywords: Vision, Object Recognition, Human, Primate, Ventral Stream, Adversarial Examples, Behavior Modulation, Behavioral Alignment

Compressor summary: The paragraph discusses how robustified artificial neural networks (ANNs) can disrupt and manipulate human visual perception by finding low-norm image perturbations, while human percepts remain stable in the same regime.


Diffusion Model for Graph Inverse Problems: Towards Effective Source Localization on Complex Networks

Xin Yan, Hui Fang, Qiang He

https://openreview.net/forum?id=5Fr8Nwi5KF

Keywords: Diffusion Model, Graph Inverse Problems, Source Localization, Information Diffusion

Compressor summary: The paper proposes a probabilistic model called DDMSL to locate the sources and reconstruct the paths of information diffusion over complex networks, using Markov chains and a reversible residual network.


Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective

Zeyuan Yin, Eric Xing, Zhiqiang Shen

https://openreview.net/forum?id=5Fgdk3hZpb

Keywords: dataset condensation and distillation, ImageNet Scale

Compressor summary: The SRe$^2$L framework creates smaller synthetic data to train models efficiently on different dataset scales and resolutions, achieving state-of-the-art results with fast speed and less memory consumption.


Provable Guarantees for Neural Networks via Gradient Feature Learning

Zhenmei Shi, Junyi Wei, Yingyu Liang

https://openreview.net/forum?id=5F04bU79eK

Keywords: neural networks, gradient descent, feature learning, provable guarantees, theoretical analysis

Compressor summary: This paper presents a unified analysis framework for two-layer neural networks trained by gradient descent, which explains their ability to learn features from gradients and applies to various problems, revealing new insights into network learning.


Adaptive Normalization for Non-stationary Time Series Forecasting: A Temporal Slice Perspective

Zhiding Liu, Mingyue Cheng, Zhi Li, Zhenya Huang, Qi Liu, Yanhu Xie, Enhong Chen

https://openreview.net/forum?id=5BqDSw8r5j

Keywords: Time series forecasting, deep learning, normalization

Compressor summary: The paragraph describes a new method called SAN that improves time series forecasting by adaptively normalizing and denormalizing data at the slice level, accounting for non-stationary changes in statistical properties.


A Spectral Theory of Neural Prediction and Alignment

Abdulkadir Canatar, Jenelle Feather, Albert Wakhloo, SueYeon Chung

https://openreview.net/forum?id=5B1ZK60jWn

Keywords: computational neuroscience, neural manifolds, neuro-AI, statistical physics, representational geometry

Compressor summary: The authors analyze how different deep neural networks predict visual cortical activity using a theoretical framework that relates generalization error to spectral properties, and introduce geometrical measures to interpret the neural prediction error.


Common Ground in Cooperative Communication

Xiaoran Hao, Yash Jhaveri, Patrick Shafto

https://openreview.net/forum?id=5AMa9fiyJq

Keywords: Cooperative Communication, Common Ground, Bayesian Theory

Compressor summary: The paragraph discusses a new theory of cooperative communication that accounts for different levels of common ground, connects it to variational autoencoding, and tests it with simulations.


Efficient Meta Neural Heuristic for Multi-Objective Combinatorial Optimization

Jinbiao Chen, Jiahai Wang, Zizhen Zhang, Zhiguang Cao, Te Ye, Siyuan Chen

https://openreview.net/forum?id=593fc38lhN

Keywords: neural heuristic, meta learning, deep reinforcement learning, multi-objective combinatorial optimization

Compressor summary: The proposed efficient meta neural heuristic (EMNH) uses a multi-task model for parallel learning and a hierarchical method for fine-tuning to solve MOCOPs faster and better than existing neural heuristics.


Bayesian Metric Learning for Uncertainty Quantification in Image Retrieval

Frederik Rahbæk Warburg, Marco Miani, Silas Brack, Søren Hauberg

https://openreview.net/forum?id=58XMiu8kot

Keywords: Laplace approximation, metric learning, uncertainty quantification, weight posterior, bayesian

Compressor summary: The paper introduces a Bayesian encoder for metric learning using the Laplace Approximation, which improves uncertainty, out-of-distribution detection, and predictive performance.


Reward-Directed Conditional Diffusion: Provable Distribution Estimation and Reward Improvement

Hui Yuan, Kaixuan Huang, Chengzhuo Ni, Minshuo Chen, Mengdi Wang

https://openreview.net/forum?id=58HwnnEdtF

Keywords: Theory, Diffusion Model, Reward Optimization, Low-dimensional Data, Distribution estimation

Compressor summary: The authors propose a method to generate samples with desired properties using a reward function and a conditional diffusion model, and show its effectiveness in theory and practice.


Language Models are Weak Learners

Hariharan Manikandan, Yiding Jiang, J Zico Kolter

https://openreview.net/forum?id=559NJBfN20

Keywords: language model, prompting, tabular data, summarization, boosting, adaboost

Compressor summary: The paper demonstrates that large language models can act as weak learners in a boosting algorithm for tabular data, achieving better results than traditional tree-based methods and few-shot learning.


GeoPhy: Differentiable Phylogenetic Inference via Geometric Gradients of Tree Topologies

Takahiro Mimori, Michiaki Hamada

https://openreview.net/forum?id=54z8M7NTbJ

Keywords: phylogenetic inference, variational inference, control variates, hyperbolic space

Compressor summary: GeoPhy is a new method for phylogenetic inference that uses continuous geometric spaces to represent tree topologies and improve variational Bayesian models without restricting possible tree candidates.


Quantifying the Cost of Learning in Queueing Systems

Daniel Freund, Thodoris Lykouris, Wentao Weng

https://openreview.net/forum?id=54hYifmQZU

Keywords: bandits, learning, queueing systems, optimal control

Compressor summary: The paper introduces the Cost of Learning in Queueing (CLQ), a metric to measure how much parameter uncertainty affects queue length in queueing systems, and develops a unified analysis framework for CLQ that combines Lyapunov and bandit analysis.


Towards a fuller understanding of neurons with Clustered Compositional Explanations

Biagio La Rosa, Leilani H. Gilpin, Roberto Capobianco

https://openreview.net/forum?id=51PLYhMFWz

Keywords: compositional explanations, network dissection, explainable artificial intelligence, interpretability

Compressor summary: Clustered Compositional Explanations is a method for broadening the spectrum of neuron behavior approximated by logical formulas of concepts, by combining existing techniques with clustering and a novel search heuristic.


Recovering Simultaneously Structured Data via Non-Convex Iteratively Reweighted Least Squares

Christian Kümmerle, Johannes Maly

https://openreview.net/forum?id=50hs53Zb3w

Keywords: low-rank models, sparsity, iteratively reweighted least squares, non-convex optimization, quadratic convergence, simultaneously structured data

Compressor summary: The paper introduces an algorithm (IRLS) for recovering data with multiple structures (row-sparsity and low-rank) from linear observations, which has better performance and sample complexity than existing methods.


A Unified Detection Framework for Inference-Stage Backdoor Defenses

Xun Xian, Ganghua Wang, Jayanth Srinivasa, Ashish Kundu, Xuan Bi, Mingyi Hong, Jie Ding

https://openreview.net/forum?id=4zWEyYGGfI

Keywords: Backdoor attacks, Backdoor Defense, Security for AI

Compressor summary: The study proposes a framework for detecting backdoor attacks in machine learning models that guarantees few false positives and maximizes the accuracy of identifying poisoned samples.


On Proper Learnability between Average- and Worst-case Robustness

Vinod Raman, UNIQUE SUBEDI, Ambuj Tewari

https://openreview.net/forum?id=4yXnnCK3r9

Keywords: Adversarial Robustness, PAC Learning

Compressor summary: The authors study how relaxing the robust loss in adversarial learning can improve sample complexity and learnability, while showing that some existing relaxations still require more than finite VC dimension for proper learning.


Attention as Implicit Structural Inference

Ryan Singh, Christopher Buckley

https://openreview.net/forum?id=4xckZu4MPG

Keywords: Attention, Structural Inference, Variational Inference, Predictive Coding, Graphical Models

Compressor summary: The paragraph discusses how attention mechanisms in Transformers can be viewed as inference over adjacency structures in graphical models, unifying different attentional architectures and suggesting potential modifications and generalizations.


Med-UniC: Unifying Cross-Lingual Medical Vision-Language Pre-Training by Diminishing Bias

Zhongwei Wan, Che Liu, Mi Zhang, Jie Fu, Benyou Wang, Sibo Cheng, Lei Ma, César Quilodrán-Casas, Rossella Arcucci

https://openreview.net/forum?id=4vpsQdRBlK

Keywords: Medical Vision Langauge Pretraining, Cross-lingual, Language bias

Compressor summary: This paper introduces Med-UniC, a framework to integrate medical data from English and Spanish, using Cross-Lingual Text Alignment Regularization (CTR) to reduce community bias and improve performance in various medical vision-language tasks.


Unsupervised Behavior Extraction via Random Intent Priors

Hao Hu, Yiqin Yang, Jianing Ye, Ziqing Mai, Chongjie Zhang

https://openreview.net/forum?id=4vGVQVz5KG

Keywords: offline RL, reward-free, behavior extraction

Compressor summary: UBER is an unsupervised method that uses diverse pseudo-rewards from random neural networks to extract useful behaviors from offline reward-free datasets and improve online RL performance.


HiBug: On Human-Interpretable Model Debug

Muxi Chen, YU LI, Qiang Xu

https://openreview.net/forum?id=4sDHLxKb1L

Keywords: model debugging, error slice discovery

Compressor summary: HiBug is a framework that uses pre-trained models like chatGPT to automatically discover and explain model bugs in computer vision tasks by identifying underperforming data slices and suggesting human-understandable attributes.


Censored Sampling of Diffusion Models Using 3 Minutes of Human Feedback

TaeHo Yoon, Kibeom Myoung, Keon Lee, Jaewoong Cho, Albert No, Ernest K. Ryu

https://openreview.net/forum?id=4qG2RKuZaA

Keywords: Generative models, Diffusion probabilistic models, Controlled generation, Human Feedback, RLHF

Compressor summary: The paper proposes censored image generation using a pre-traired diffusion model and minimal human feedback, which is very efficient in preventing undesirable images.


Energy-Based Models for Anomaly Detection: A Manifold Diffusion Recovery Approach

Sangwoong Yoon, Young-Uk Jin, Yung-Kyun Noh, Frank C. Park

https://openreview.net/forum?id=4nSDDokpfK

Keywords: Energy-based Models, Anomaly Detection, Generative Models, Out-of-Distribution Detection, Recovery Likelihood

Compressor summary: The proposed algorithm MPDR trains energy-based models for anomaly detection by perturbing data points along low-dimensional manifolds and generating near-manifold negative samples.


Stable Vectorization of Multiparameter Persistent Homology using Signed Barcodes as Measures

David Loiseaux, Luis Scoccola, Mathieu Carrière, Magnus Bakke Botnan, Steve Oudot

https://openreview.net/forum?id=4mwORQjAim

Keywords: topological data analysis, multiparameter persistent homology, kernel methods, optimal transport

Compressor summary: The paper introduces a new method for stable vectorization of multiparameter persistent homology descriptors using signed Radon measures and demonstrates its effectiveness on different data sets.


Online Pricing for Multi-User Multi-Item Markets

Yigit Efe Erginbas, Thomas Courtade, Kannan Ramchandran, Soham Rajesh Phade

https://openreview.net/forum?id=4mXYJzoPhf

Keywords: revenue, price, offer, online

Compressor summary: The study proposes online algorithms to price items efficiently and learn user valuations from feedback, and compares their performance under three user valuation models.


Multi-Modal Inverse Constrained Reinforcement Learning from a Mixture of Demonstrations

Guanren Qiao, Guiliang Liu, Pascal Poupart, zhiqiang xu

https://openreview.net/forum?id=4mPiqh4pLb

Keywords: Inverse Constrained Reinforcement Learning, Learning from Demonstrations, Muti-Modal Learning

Compressor summary: MMICRL is a new algorithm that can learn multiple constraints from diverse expert agents by using flow-based density estimation and contrastive learning, leading to better imitation policies and behavior diversity.


Operator Learning with Neural Fields: Tackling PDEs on General Geometries

Louis Serrano, Lise Le Boudec, Armand Kassaï Koupaï, Thomas X Wang, Yuan Yin, Jean-Noël Vittaut, patrick gallinari

https://openreview.net/forum?id=4jEjq5nhg1

Keywords: PDEs, Physics, Operator Learning, Deep Learning, Spatiotemporal

Compressor summary: CORAL is a new method that uses coordinate-based networks to solve PDEs on any spatial sampling and geometry, achieving robust performance across multiple resolutions and problem domains.


Train 'n Trade: Foundations of Parameter Markets

Tzu-Heng Huang, Harit Vishwakarma, Frederic Sala

https://openreview.net/forum?id=4iV26fZPUD

Keywords: Parameter Market, Pricing, Efficient Model Training

Compressor summary: The authors propose a framework for trading model parameters as commodities, which could improve large-scale model training and create value for agents.


Data-Dependent Bounds for Online Portfolio Selection Without Lipschitzness and Smoothness

Chung-En Tsai, Ying-Ting Lin, Yen-Huan Li

https://openreview.net/forum?id=4iTAUsyisM

Keywords: Online portfolio selection, small-loss bound, gradual-variation bound, second-order bound, optimistic FTRL with self-concordant regularizers

Compressor summary: The paper presents new regret bounds for online portfolio selection that depend on the data and apply to non-Lipschitz, non-smooth losses, using novel smoothness characterizations and analysis techniques.


Embroid: Unsupervised Prediction Smoothing Can Improve Few-Shot Classification

Neel Guha, Mayee F Chen, Kush Bhatia, Azalia Mirhoseini, Frederic Sala, Christopher Re

https://openreview.net/forum?id=4iMpwAlza1

Keywords: language models, prompting, embeddings, weak supervision

Compressor summary: Embroid is a method that improves language models' prompt-based learning by using consistency between predictions for similar samples to generate corrected predictions without additional labeled data.


AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback

Yann Dubois, Xuechen Li, Rohan Taori, Tianyi Zhang, Ishaan Gulrajani, Jimmy Ba, Carlos Guestrin, Percy Liang, Tatsunori Hashimoto

https://openreview.net/forum?id=4hturzLcKX

Keywords: Instruction-Following, Reinforcement Learning from Human Feedback, Artificial General Intelligence, Large Language Models

Compressor summary: AlpacaFarm is a simulator for training large language models with human feedback, which is cheaper, more trustworthy, and easier to evaluate than real-world data.


Make the U in UDA Matter: Invariant Consistency Learning for Unsupervised Domain Adaptation

Zhongqi Yue, Qianru Sun, Hanwang Zhang

https://openreview.net/forum?id=4hYIxI8ds0

Keywords: unsupervised domain adaptation, transfer learning

Compressor summary: The paper proposes ICON, a method for unsupervised domain adaptation that removes spurious correlations between features by learning an invariant classifier that is consistent with labels in the source domain and clusters in the target domain.


Recovering from Out-of-sample States via Inverse Dynamics in Offline Reinforcement Learning

Ke Jiang, Jia-Yu Yao, Xiaoyang Tan

https://openreview.net/forum?id=4gLWjSaw4o

Keywords: Offline reinforcement learning, state distributional shift, state recovery, inverse dynamics model

Compressor summary: The paper proposes a method to improve offline reinforcement learning by using an inverse dynamics model to guide actions that recover the state distribution, leading to better performance on benchmark tasks.


Importance Weighted Actor-Critic for Optimal Conservative Offline Reinforcement Learning

Hanlin Zhu, Paria Rashidinejad, Jiantao Jiao

https://openreview.net/forum?id=4e0NJbkkd8

Keywords: offline RL, actor-critic, l_2 single-policy concentrability, average bellman error

Compressor summary: A-Crab is an offline reinforcement learning algorithm that uses importance sampling and actor-critic framework, achieving optimal convergence rates and outperforming existing methods in complex environments with insufficient data coverage.


Likelihood Ratio Confidence Sets for Sequential Decision Making

Nicolas Emmenegger, Mojmir Mutny, Andreas Krause

https://openreview.net/forum?id=4anryczeED

Keywords: confidence sets, uncertainty quantification, bandits, active learning, testing

Compressor summary: The paper proposes a new likelihood-based method to construct valid uncertainty estimates for sequential decision-making algorithms that works well for problems with well-specified likelihoods and can handle different noise distributions and estimators.


What Makes Data Suitable for a Locally Connected Neural Network? A Necessary and Sufficient Condition Based on Quantum Entanglement.

Yotam Alexander, Nimrod De La Vega, Noam Razin, Nadav Cohen

https://openreview.net/forum?id=4aIpgq1nuI

Keywords: Deep Learning, Locally Connected Neural Networks, Data Distributions, Quantum Entanglement, Tensor Networks

Compressor summary: The authors apply quantum physics concepts to study the suitability of locally connected neural networks for different data distributions, and propose a preprocessing method based on low quantum entanglement.


Breaking the Communication-Privacy-Accuracy Tradeoff with $f$-Differential Privacy

Richeng Jin, Zhonggen Su, Caijun Zhong, Zhaoyang Zhang, Tony Quek, Huaiyu Dai

https://openreview.net/forum?id=4ZaPpVDjGQ

Keywords: Differential privacy, federated data analytics, discrete valued-mechanism, distributed mean estimation

Compressor summary: The paper explores how discrete-valued mechanisms with finite output space can provide local differential privacy guarantees for federated data analytics while also improving communication efficiency and accuracy.


Learning Sample Difficulty from Pre-trained Models for Reliable Prediction

Peng Cui, Dan Zhang, Zhijie Deng, Yinpeng Dong, Jun Zhu

https://openreview.net/forum?id=4WPhXYMK6N

Keywords: uncertainty calibration, sample difficulty, reliable prediction

Compressor summary: The authors propose a method to use large pre-trained models to improve the reliability and calibration of downstream models by incorporating sample difficulty into entropy regularization, leading to better accuracy and uncertainty estimation.


Structured State Space Models for In-Context Reinforcement Learning

Chris Lu, Yannick Schroecker, Albert Gu, Emilio Parisotto, Jakob Nicolaus Foerster, Satinder Singh, Feryal Behbahani

https://openreview.net/forum?id=4W9FVg1j6I

Keywords: Reinforcement Learning, Meta-Learning, State Space Models

Compressor summary: The authors propose a modified S4 model for reinforcement learning tasks, which initializes and resets the hidden state in parallel, resulting in faster speeds and better performance than Transformers and RNNs on various tasks.


Adaptive Selective Sampling for Online Prediction with Experts

Rui M. Castro, Fredrik Hellström, Tim van Erven

https://openreview.net/forum?id=4VAF3d5jNg

Keywords: Online learning, prediction with experts, selective sampling, active learning

Compressor summary: The text describes a method for online prediction with expert advice that uses selective sampling to reduce label queries while still achieving optimal regret guarantees in some cases.


Training-free Diffusion Model Adaptation for Variable-Sized Text-to-Image Synthesis

Zhiyu Jin, Xuli Shen, Bin Li, Xiangyang Xue

https://openreview.net/forum?id=4ULTSBBY4U

Keywords: Text-to-Image Synthesis, Variable-Sized Image Synthesis, Entropy

Compressor summary: The paper proposes a scaling factor for diffusion models to handle various image sizes and aspect ratios, improving visual fidelity, quality, and alignment without extra training.


MuSe-GNN: Learning Unified Gene Representation From Multimodal Biological Graph Data

Tianyu Liu, Yuge Wang, Zhitao Ying, Hongyu Zhao

https://openreview.net/forum?id=4UCktT9XZx

Keywords: Multimodal Learning; Representation Learning; Graph Neural Network; Similarity Learning; Contrastive Learning; Computational Biology and Bioinformatics; Single-cell genomics

Compressor summary: The novel Multimodal Similarity Learning Graph Neural Network model combines multimodal machine learning and deep graph neural networks to learn gene representations from diverse data, enabling better analysis of gene functions and related biomedical phenomena.


Reversible and irreversible bracket-based dynamics for deep graph neural networks

Anthony Gruber, Kookjin Lee, Nathaniel Trask

https://openreview.net/forum?id=4SoTUaTK8N

Keywords: graph neural networks, structure preserving machine learning, neural ordinary differential equations, hamiltonian dynamics, metriplectic dynamics

Compressor summary: This paper introduces new graph neural networks (GNNs) based on physics principles that conserve energy or generate dissipation, and explains their performance in relation to reversibility and irreversibility.


Reference-Based POMDPs

Edward Kim, Yohan Karunanayake, Hanna Kurniawati

https://openreview.net/forum?id=4Sn2vUs0zA

Keywords: POMDP, planning under uncertainty, long horizon

Compressor summary: The paper proposes a modified POMDP problem called Reference-Based POMDP, which balances expected reward and similarity to a reference policy, and shows that it improves long-horizon navigation problems compared to the standard POMDP approach.


Cal-DETR: Calibrated Detection Transformer

Muhammad Akhtar Munir, Salman Khan, Muhammad Haris Khan, Mohsen Ali, Fahad Khan

https://openreview.net/forum?id=4SkPTD6XNP

Keywords: Model Calibration, Object Detection, Detection Transformers, Uncertainty

Compressor summary: The paragraph discusses a new mechanism, Cal-DETR, for calibrating uncertainty in DNN-based object detectors using train-time approaches, which improves their performance and reliability in safety-critical applications.


Binary Classification with Confidence Difference

Wei Wang, Lei Feng, Yuchen Jiang, Gang Niu, Min-Ling Zhang, Masashi Sugiyama

https://openreview.net/forum?id=4RoD1o7yq6

Keywords: Weakly supervised learning, binary classification, unbiased risk estimator

Compressor summary: The paper proposes a new weakly supervised binary classification method called ConfDiff classification, which uses confidence differences between unlabeled data pairs instead of pointwise labeling confidence, and shows that it improves model performance and mitigates overfitting issues.


CRoSS: Diffusion Model Makes Controllable, Robust and Secure Image Steganography

Jiwen Yu, Xuanyu Zhang, Youmin Xu, Jian Zhang

https://openreview.net/forum?id=4R2Y5B12jm

Keywords: Diffusion models, image steganography, Stable Diffusion, coverless steganography

Compressor summary: The authors propose a new image steganography method, CRoSS, that uses diffusion models to improve security and natural robustness without additional training.


High Precision Causal Model Evaluation with Conditional Randomization

Chao Ma, Cheng Zhang

https://openreview.net/forum?id=4PkBhz18in

Keywords: causality, causal inference, causal model evaluation

Compressor summary: The pairs estimator is a new method to improve causal model evaluation in real-world experiments by reducing variance and achieving near-RCT performance.


Generalization in the Face of Adaptivity: A Bayesian Perspective

Moshe Shenfeld, Katrina Ligett

https://openreview.net/forum?id=4L9g1jUDtO

Keywords: Differential Privacy, Adaptive Data Analysis

Compressor summary: The paper proposes a novel approach to prevent overfitting in adaptive data analysis by using noise-addition algorithms and introducing a new data-dependent stability notion.


Object-centric Learning with Cyclic Walks between Parts and Whole

Ziyu Wang, Mike Zheng Shou, Mengmi Zhang

https://openreview.net/forum?id=4L3RfWnDzL

Keywords: object representation learning, slot attention, object-centric, contrastive random walks

Compressor summary: The authors propose a method to learn object-centric representations from scenes using cyclic walks between perceptual features and object entities, which enables reasoning abilities in both humans and machines.


FIRAL: An Active Learning Algorithm for Multinomial Logistic Regression

Youguang Chen, George Biros

https://openreview.net/forum?id=4L2OlXhiTM

Keywords: statistical learning, active learning, logistic regression, regret minimization

Compressor summary: The paper studies how to use pool-based active learning with multinomial logistic regression for multiclass classification, proves theoretical bounds on excess risk using Fisher Information Ratio (FIR), proposes a regret minimization algorithm based on FIR, and shows experimental results that beat other methods on synthetic and real datasets.


Direction-oriented Multi-objective Learning: Simple and Provable Stochastic Algorithms

Peiyao Xiao, Hao Ban, Kaiyi Ji

https://openreview.net/forum?id=4Ks8RPcXd9

Keywords: Multi-objective optimization, multi-task leaning, stochastic algorithms, convergence and complexity, Pareto stationarity

Compressor summary: The paper proposes SDMGrad, a stochastic optimization method that improves sample complexity and convergence for multi-objective machine learning problems with multiple criteria and tasks.


When Does Confidence-Based Cascade Deferral Suffice?

Wittawat Jitkrittum, Neha Gupta, Aditya Krishna Menon, Harikrishna Narasimhan, Ankit Singh Rawat, Sanjiv Kumar

https://openreview.net/forum?id=4KZhZJSPYU

Keywords: cascades, deferral rules, adaptive computation, model confidence

Compressor summary: The paper investigates when confidence-based deferral in cascade classifiers may fail and proposes alternative deferral strategies for such scenarios.


On the Variance, Admissibility, and Stability of Empirical Risk Minimization

Gil Kur, Eli Putterman, Alexander Rakhlin

https://openreview.net/forum?id=4KV2xLeqPN

Keywords: empirical risk minimization, bias-variance decomposition, admissibility

Compressor summary: The paper proves that ERM's suboptimality is due to its bias, not variance, and provides proofs and extensions for various settings and models, as well as stability results and a discussion on the irregular loss landscape of ERM.


Effectively Learning Initiation Sets in Hierarchical Reinforcement Learning

Akhil Bagaria, Ben M Abbatematteo, Omer Gottesman, Matt Corsaro, Sreehari Rammohan, George Konidaris

https://openreview.net/forum?id=4JCVw8oMlf

Keywords: hierarchical reinforcment learning

Compressor summary: The paragraph discusses the challenges of learning options in hierarchical reinforcement learning, especially the issue of learning initiation sets, and proposes a new method that uses off-policy value estimation and classification to improve performance on various tasks.


Neural approximation of Wasserstein distance via a universal architecture for symmetric and factorwise group invariant functions

Samantha Chen, Yusu Wang

https://openreview.net/forum?id=4JB42GBxGs

Keywords: neural networks, Wasserstein distance, universal approximation, optimal transport

Compressor summary: The paper proposes a neural network architecture that approximates Wasserstein distance between point sets with low complexity, independent of the input size, and shows its superior performance over other models.


Learning to Receive Help: Intervention-Aware Concept Embedding Models

Mateo Espinosa Zarlenga, Katherine M. Collins, Krishnamurthy Dj Dvijotham, Adrian Weller, Zohreh Shams, Mateja Jamnik

https://openreview.net/forum?id=4ImZxqmT1K

Keywords: Explainable Artificial Intelligence, Concept Bottleneck Models, Concept-based Explainability, Interpretability, XAI, Concept Interventions

Compressor summary: IntCEMs are a new type of neural architecture that allows users to improve the model's performance by correcting mispredicted concepts during training, based on an learned intervention policy.


Removing Hidden Confounding in Recommendation: A Unified Multi-Task Learning Approach

Haoxuan Li, Kunhan Wu, Chunyuan Zheng, Yanghao Xiao, Hao Wang, Zhi Geng, Fuli Feng, Xiangnan He, Peng Wu

https://openreview.net/forum?id=4IWJZjbRFj

Keywords: Debiased recommender system, Multi-task learning, Causal inference

Compressor summary: The paper analyzes the limitations of existing debiasing methods for recommender systems and proposes a new multi-task learning approach to address hidden confounding.


Block-Coordinate Methods and Restarting for Solving Extensive-Form Games

Darshan Chakrabarti, Jelena Diakonikolas, Christian Kroer

https://openreview.net/forum?id=4AmJVaJ78I

Keywords: extensive-form games, first-order methods, coordinate descent

Compressor summary: The paper presents a new cyclic coordinate-descent method for solving sequential games that exploits the recursive structure of dilated regularizers, achieving fast convergence and sometimes outperforming state-of-the-art algorithms.


On the Exploitability of Instruction Tuning

Manli Shu, Jiongxiao Wang, Chen Zhu, Jonas Geiping, Chaowei Xiao, Tom Goldstein

https://openreview.net/forum?id=4AQ4Fnemox

Keywords: Trustworthy machine learning, Large language models, Supervised fine-tuning, instruction tuning

Compressor summary: The paper introduces AutoPoison, a tool that can change the behavior of large language models by injecting specific examples into their training data.


Direct Diffusion Bridge using Data Consistency for Inverse Problems

Hyungjin Chung, Jeongsol Kim, Jong Chul Ye

https://openreview.net/forum?id=497CevPdOg

Keywords: Diffusion models, Inverse problems, Diffusion bridge

Compressor summary: The paper proposes a new method called data Consistent DDB that improves the performance of Direct Diffusion Bridges for solving inverse problems by ensuring data consistency and is open-sourced.


Lightweight Vision Transformer with Bidirectional Interaction

Qihang Fan, Huaibo Huang, Xiaoqiang Zhou, Ran He

https://openreview.net/forum?id=492Hfmgejy

Keywords: Vision Transformer, Lightweight Vision Backbone, Convolution Neural Network

Compressor summary: The paper proposes FASA, a mechanism for vision transformers that adaptively extracts local and global information and their bidirectional interaction, leading to improved performance and efficiency in multiple vision tasks.


Zeroth-Order Methods for Nondifferentiable, Nonconvex, and Hierarchical Federated Optimization

Yuyang Qiu, Uday Shanbhag, Farzad Yousefian

https://openreview.net/forum?id=46x3zvYCyQ

Keywords: Federated Learning, Nonsmooth Optimization, Nonconvex Optimization, Bilevel Optimization

Compressor summary: The paper proposes a framework for communication-efficient decentralized training that handles nonconvex optimization, bilevel optimization, and minimax problems in federated learning using randomized smoothing and implicit zeroth-order methods.


Unsupervised Protein-Ligand Binding Energy Prediction via Neural Euler's Rotation Equation

Wengong Jin, Siranush Sarkizova, Xun Chen, Nir Hacohen, Caroline Uhler

https://openreview.net/forum?id=46gYakmj4e

Keywords: Energy-based Models, Denoising Score Matching, Equivariant Neural Networks

Compressor summary: The paper proposes a new generative modeling approach for predicting binding affinity of antibodies, using a rotation prediction network that models forces between protein and ligand atoms.


Accelerated On-Device Forward Neural Network Training with Module-Wise Descending Asynchronism

Xiaohan Zhao, Hualin Zhang, Zhouyuan Huo, Bin Gu

https://openreview.net/forum?id=45RBLZBJid

Keywords: asynchronous algorithm, one-device learning, forward gradient descent, directional derivative, forward algorithms

Compressor summary: The paper introduces AsyncFGD, an asynchronous framework for on-device learning that reduces memory usage and improves hardware efficiency by decoupling dependencies and utilizing stale parameters.


A Unified Framework for U-Net Design and Analysis

Christopher Williams, Fabian Falck, George Deligiannidis, Christopher C. Holmes, Arnaud Doucet, Saifuddin Syed

https://openreview.net/forum?id=43ruO2fMjq

Keywords: U-Net, ResNet, Multi-ResNet, Generalised U-Net, Wavelets, Diffusion models, Generative modelling, PDE Modelling, Image Segmentation

Compressor summary: The paper presents a framework for analysing and designing general U-Net architectures, proposes Multi-ResNets with a simplified encoder and novel U-Net designs that incorporate function constraints and data geometry, and shows their improved performance in various tasks.


The Pick-to-Learn Algorithm: Empowering Compression for Tight Generalization Bounds and Improved Post-training Performance

Dario Paccagnan, Marco Campi, Simone Garatti

https://openreview.net/forum?id=40L3viVWQN

Keywords: Statistical learning theory, Compression theory, Generalization bounds

Compressor summary: The authors propose Pick-to-Learn, a meta-algorithm that improves compression properties and generalization bounds for learning algorithms on MNIST and a synthetic regression problem.


Uncoupled and Convergent Learning in Two-Player Zero-Sum Markov Games with Bandit Feedback

Yang Cai, Haipeng Luo, Chen-Yu Wei, Weiqiang Zheng

https://openreview.net/forum?id=3xSwxlB0fd

Keywords: two-player zero-sum Markov game, last-iterate convergence, path convergence, learning in games

Compressor summary: The paper proposes an efficient and rational algorithm for learning in two-player zero-sum Markov games with various feedback types and convergence rates, outperforming previous methods in terms of assumptions and synchronization.


Estimating Noise Correlations Across Continuous Conditions With Wishart Processes

Amin Nejatbakhsh, Isabel Garon, Alex H Williams

https://openreview.net/forum?id=3ucmcMzCXD

Keywords: Noise Correlations, Wishart Process, Variational Inference

Compressor summary: The authors propose a new method to estimate the covariance structure of neural noise across naturalistic behaviors using Wishart process models, which can provide information on signal fidelity and noise correlations in unseen conditions.


The Quantization Model of Neural Scaling

Eric J Michaud, Ziming Liu, Uzay Girit, Max Tegmark

https://openreview.net/forum?id=3tbTw2ga8K

Keywords: scaling laws, emergence, language models, science of deep learning

Compressor summary: The Quantization Model explains neural scaling laws by describing how network abilities are divided into discrete chunks and learned in order of decreasing use frequency.


Projection Regret: Reducing Background Bias for Novelty Detection via Diffusion Models

Sungik Choi, Hankook Lee, Honglak Lee, Moontae Lee

https://openreview.net/forum?id=3qHlPqzjM1

Keywords: Novelty detection, out-of-distribution detection, consistency models, diffusion models, score-based generative models

Compressor summary: The paper proposes Projection Regret (PR), a novelty detection method for diffusion models that mitigates background bias and outperforms existing methods.


ReHLine: Regularized Composite ReLU-ReHU Loss Minimization with Linear Computation and Linear Convergence

Ben Dai, Yixuan Qiu

https://openreview.net/forum?id=3pEBW2UPAD

Keywords: coordinate descent, linear convergence, primal-dual methods, empirical risk minimization, linear constraints, quantile regression

Compressor summary: ReHLine is a novel algorithm for minimizing regularized ERMs with convex piecewise linear-quadratic loss functions and optional linear constraints, which handles diverse problems and has provable linear convergence rate and linear computational complexity.


DisDiff: Unsupervised Disentanglement of Diffusion Probabilistic Models

Tao Yang, Yuwang Wang, Yan Lu, Nanning Zheng

https://openreview.net/forum?id=3ofe0lpwQP

Keywords: Diffusion Probabilistic Model, Disentangled representation

Compressor summary: The authors propose a new task called disentanglement of diffusion probabilistic models (DPMs), which aims to discover and represent inherent factors behind observations, and introduce an unsupervised approach named DisDiff that achieves this task for the first time.


RECESS Vaccine for Federated Learning: Proactive Defense Against Model Poisoning Attacks

Haonan Yan, Wenjing Zhang, Qian Chen, Xiaoguang Li, Wenhai Sun, HUI LI, Xiaodong Lin

https://openreview.net/forum?id=3n8PNUdvSg

Keywords: Federated Learning, Model Poisoning Attacks, Proactive Detection, Robust Aggregation, Benign Outlier Identification

Compressor summary: RECESS is a novel defense for federated learning against model poisoning attacks that uses trust scoring based aggregation and proactive querying to detect and protect against malicious clients.


Beyond probability partitions: Calibrating neural networks with semantic aware grouping

Jia-Qi Yang, De-Chuan Zhan, Le Gan

https://openreview.net/forum?id=3kitbpEZZO

Keywords: Uncertainty calibration, Deep neural networks

Compressor summary: The paper introduces Partitioned Calibration Error (PCE) as a generalized definition of calibration error in deep networks, showing that accurate models should be calibrated across any data partition, and proposes a method to learn semantic-aware partitioning functions for improved calibration.


PERFOGRAPH: A Numerical Aware Program Graph Representation for Performance Optimization and Program Analysis

Ali TehraniJamsaz, Quazi Ishtiaque Mahmud, Le Chen, Nesreen K. Ahmed, Ali Jannesari

https://openreview.net/forum?id=3jAsfo8x8k

Keywords: program representation, graph representation, program analysis, graph neural networks, performance optimization

Compressor summary: PERFOGRAPH is a novel graph-based program representation that captures numerical information and aggregate data structure, improving machine learning methods' ability to reason about programs and achieving state-of-the-art results in various applications.


Learning Interpretable Low-dimensional Representation via Physical Symmetry

Xuanjie Liu, Daniel Chin, Yichen Huang, Gus Xia

https://openreview.net/forum?id=3iSj4l8ZGT

Keywords: Physics Symmetry, Time series data, Self-supervised Learning, Representation Augmentation

Compressor summary: The study proposes using physical symmetry as a self-consistency constraint for learning interpretable music representations, leading to a linear pitch factor and representation augmentation.


Differentiable and Stable Long-Range Tracking of Multiple Posterior Modes

Ali Younis, Erik B. Sudderth

https://openreview.net/forum?id=3gxiOEf2D6

Keywords: particle, filter, mixture, belief propagation, nonparametric, deep learning, generative, discriminative, graphical model, multiple modes, mutli-modal

Compressor summary: The paragraph describes a novel particle filter method that uses deep neural networks to learn nonparametric representations of uncertainty in latent object states from arbitrary observations, improving tracking and localization performance.


QuantSR: Accurate Low-bit Quantization for Efficient Image Super-Resolution

Haotong Qin, Yulun Zhang, Yifu Ding, Yifan liu, Xianglong Liu, Martin Danelljan, Fisher Yu

https://openreview.net/forum?id=3gamyee9Yh

Keywords: Super Resolution, Model Quantization, Deep Learning

Compressor summary: The paper proposes a novel quantized image super-resolution network called QuantSR that achieves accurate and efficient processing under low-bit quantization using the Redistribution-driven Learnable Quantizer (RLQ) and Depth-dynamic Quantized Architecture (DQA).


Mitigating Over-smoothing in Transformers via Regularized Nonlocal Functionals

Tam Minh Nguyen, Tan Minh Nguyen, Richard Baraniuk

https://openreview.net/forum?id=3fd776zKmo

Keywords: transformers, self-attention, total variation, nonlocal functionals, over-smoothing

Compressor summary: NeuTRENO is a new type of transformer model that reduces token uniformity by penalizing the difference between input and output tokens in self-attention layers.


SoTTA: Robust Test-Time Adaptation on Noisy Data Streams

Taesik Gong, Yewon Kim, Taeckyung Lee, Sorn Chottananurak, Sung-Ju Lee

https://openreview.net/forum?id=3bdXag2rUd

Keywords: test-time adaptation, domain adaptation, deep learning, machine learning

Compressor summary: The paper introduces SoTTA, a novel test-time adaptation algorithm that is robust to noisy data by using high-confidence uniform-class sampling and entropy-sharpness minimization.


A Data-Free Approach to Mitigate Catastrophic Forgetting in Federated Class Incremental Learning for Vision Tasks

Sara Babakniya, Zalan Fabian, Chaoyang He, Mahdi Soltanolkotabi, Salman Avestimehr

https://openreview.net/forum?id=3b9sqxCW1x

Keywords: federated learning, class incremental learning, generative models, data-free, continual learning

Compressor summary: The paper proposes a framework for federated class incremental learning that uses generative models to synthesize data from past distributions and train the model without requesting data or storing old information from clients, addressing privacy and resource limitations in federated learning.


Volume Feature Rendering for Fast Neural Radiance Field Reconstruction

Kang Han, Wei Xiang, Lu Yu

https://openreview.net/forum?id=3aVZhMfsyz

Keywords: neural rendering, volume rendering, view synthesis, 3D reconstruction

Compressor summary: NeRFs use neural networks to generate realistic images from multiple views, but this can be slow due to many color network evaluations. This paper proposes a new method called VFR that combines queries into one vector and uses a larger color network for better quality with less training time.


Deep learning with kernels through RKHM and the Perron-Frobenius operator

Yuka Hashimoto, Masahiro Ikeda, Hachem Kadri

https://openreview.net/forum?id=3ZrGmenVM2

Keywords: kernel method. generalization bound. C*-algebra. Perron-Frobenius operator and Koopman operator.

Compressor summary: The paper introduces deep RKHM, a deep learning framework for kernel methods using reproducing kernel Hilbert module and Perron-Frobenius operator, with a new generalization bound and insights on benign overfitting.


ReTR: Modeling Rendering Via Transformer for Generalizable Neural Surface Reconstruction

Yixun Liang, Hao He, Ying-Cong Chen

https://openreview.net/forum?id=3ZICE99e6n

Keywords: 3D vision, 3D reconstruction, Generalizable Neural Surface Reconstruction

Compressor summary: The paper introduces ReTR, a novel framework that uses transformer architecture to improve neural surface reconstruction by redesigning the rendering process with a learnable meta-ray token and cross-attention mechanism, resulting in more accurate and confident surface assessment.


Learning Large Graph Property Prediction via Graph Segment Training

Kaidi Cao, Phitchaya Mangpo Phothilimthana, Sami Abu-El-Haija, Dustin Zelle, Yanqi Zhou, Charith Mendis, Jure Leskovec, Bryan Perozzi

https://openreview.net/forum?id=3YDukx2cpr

Keywords: Graph Neural Networks, Graph Property Prediction

Compressor summary: The Graph Segment Training framework uses a divide-and-conquer approach to learn large graph property prediction with constant memory, and introduces techniques to handle input distribution shift and stale embeddings.


A General Framework for Equivariant Neural Networks on Reductive Lie Groups

Ilyes Batatia, Mario Geiger, Jose M Munoz, Tess Smidt, Lior Silberman, Christoph Ortner

https://openreview.net/forum?id=3XStpETaO8

Keywords: equivariance, point clouds, machine learning, particle physics

Compressor summary: The paper introduces a new neural network architecture that respects symmetries in various scientific fields, such as high energy physics and computer vision, using reductive Lie groups.


Birth of a Transformer: A Memory Viewpoint

Alberto Bietti, Vivien Cabannes, Diane Bouchacourt, Herve Jegou, Leon Bottou

https://openreview.net/forum?id=3X2EbBLNsk

Keywords: transformers, language models, deep learning theory, interpretability

Compressor summary: The paragraph discusses how transformers learn from global and context-specific information and the importance of understanding their internal mechanisms for reliability.


Learning via Wasserstein-Based High Probability Generalisation Bounds

Paul Viallard, Maxime Haddouche, Umut Simsekli, Benjamin Guedj

https://openreview.net/forum?id=3Wrolscjbx

Keywords: Wasserstein, PAC-Bayes, Generalisation Bound, Algorithm

Compressor summary: The paper proposes new generalization bounds for PAC-Bayesian learning using Wasserstein distance, which are stronger than previous ones and can be used in structural risk minimization.


Why Does Sharpness-Aware Minimization Generalize Better Than SGD?

Zixiang Chen, Junkai Zhang, Yiwen Kou, Xiangning Chen, Cho-Jui Hsieh, Quanquan Gu

https://openreview.net/forum?id=3WAnGWLpSQ

Keywords: Sharpness Aware Algorithm, Deep Learning Theory

Compressor summary: The paper explores why Sharpness-Aware Minimization (SAM) works better than Stochastic Gradient Descent (SGD) for certain neural network settings, especially in preventing overfitting and noise learning.


DP-HyPO: An Adaptive Private Framework for Hyperparameter Optimization

Hua Wang, Sheng Gao, Huanyu Zhang, Weijie J Su, Milan Shen

https://openreview.net/forum?id=3Py8A1j5N3

Keywords: Differential Privacy, Hyperparameter Tuning, Deep Learning

Compressor summary: DP-HyPO is a novel framework for private and adaptive hyperparameter optimization that bridges the gap between private and non-private methods, providing a thorough differential privacy analysis and showing its effectiveness on various real-world datasets.


From Pixels to UI Actions: Learning to Follow Instructions via Graphical User Interfaces

Peter Shaw, Mandar Joshi, James Cohan, Jonathan Berant, Panupong Pasupat, Hexiang Hu, Urvashi Khandelwal, Kenton Lee, Kristina Toutanova

https://openreview.net/forum?id=3PjCt4kmRx

Keywords: instruction following, web tasks, user interface tasks, vision and language, representation learning, reinforcement learning, imitation learning, tree search, language grounding, web agents, computer control

Compressor summary: The paper proposes pixel-based agents that use keyboard and mouse actions to interact with GUIs and shows they can perform better than humans on GUI instruction following tasks.


Undirected Probabilistic Model for Tensor Decomposition

Zerui Tao, Toshihisa Tanaka, Qibin Zhao

https://openreview.net/forum?id=3NWWgB2SuF

Keywords: Tensor decomposition, tensor completion, probabilistic methods

Compressor summary: The paper introduces a flexible tensor decomposition framework using energy-based models and neural networks to learn underlying structures and distributions without prior assumptions, and proposes a variational objective for efficient training.


CAMEL: Communicative Agents for "Mind" Exploration of Large Language Model Society

Guohao Li, Hasan Abed Al Kader Hammoud, Hani Itani, Dmitrii Khizbullin, Bernard Ghanem

https://openreview.net/forum?id=3IyL2XWDkG

Keywords: Communicative Agents, Large Language Models, AI Society, Role-Playing, Society of Mind

Compressor summary: This paper introduces role-playing, a framework for enabling autonomous cooperation among chat-based agents using inception prompting, and demonstrates its potential for studying multi-agent systems.


DFRD: Data-Free Robustness Distillation for Heterogeneous Federated Learning

Kangyang Luo, Shuai Wang, Yexuan Fu, Xiang Li, Yunshi Lan, Ming Gao

https://openreview.net/forum?id=3H9QH1v6U9

Keywords: Federated Learning, Data Heterogeneity, Model Heterogeneity, Data-Free Distillation

Compressor summary: DFRD is a new federated learning method that uses data-free knowledge distillation and a conditional generator to learn a robust global model from privately decentralized clients.


Post Hoc Explanations of Language Models Can Improve Language Models

Satyapriya Krishna, Jiaqi Ma, Dylan Z Slack, Asma Ghandeharioun, Sameer Singh, Himabindu Lakkaraju

https://openreview.net/forum?id=3H37XciUEv

Keywords: Machine Learning Explainability, Large Language Models

Compressor summary: AMPLIFY is a framework that uses post hoc explanations to generate rationales for large language models, improving their performance on complex tasks without human involvement.


The Pursuit of Human Labeling: A New Perspective on Unsupervised Learning

Artyom Gadetsky, Maria Brbic

https://openreview.net/forum?id=3GpIeVYw8X

Keywords: unsupervised learning, deep learning, generalization, self-supervised learning, clustering

Compressor summary: HUME is a simple framework that can infer human labels on a dataset without supervision, using linear classifiers on top of pretrained representations, and achieves state-of-the-art performance on several image classification benchmarks.


Addressing Negative Transfer in Diffusion Models

Hyojun Go, Jinyoung Kim, Yunsung Lee, Seunghyun Lee, Shinhyeok Oh, Hyeongdon Moon, Seungtaek Choi

https://openreview.net/forum?id=3G2ec833mW

Keywords: Diffusion Models, Multi-Task Learning

Compressor summary: The paper analyzes diffusion training in multi-task learning, identifies challenges like negative transfer and task affinity, and proposes interval clustering to improve it.


Neural Frailty Machine: Beyond proportional hazard assumption in neural survival regressions

Ruofan Wu, Jiawei Qiao, Mingzhe Wu, Wen Yu, Ming Zheng, Tengfei LIU, Tianyi Zhang, Weiqiang Wang

https://openreview.net/forum?id=3Fc9gnR0fa

Keywords: Survival Analysis, Theory, Semiparametric statistics

Compressor summary: The paper introduces neural frailty machine (NFM), a neural model for survival regressions that extends the proportional hazard assumption and uses multiplicative frailty, with theoretical guarantees and empirical results showing its effectiveness.


GLIME: General, Stable and Local LIME Explanation

Zeren Tan, Yang Tian, Jian Li

https://openreview.net/forum?id=3FJaFElIVN

Keywords: Explanation, LIME, Stability, Local fidelity, Interpretability

Compressor summary: The paper proposes \textsc{Glime}, a framework that improves the stability and local fidelity of explanations for black-box machine learning models by modifying LIME's sampling approach and formulation.


Posthoc privacy guarantees for collaborative inference with modified Propose-Test-Release

Abhishek Singh, Praneeth Vepakomma, Vivek Sharma, Ramesh Raskar

https://openreview.net/forum?id=3DMDNwd7ND

Keywords: privacy, deep learning, neural networks, adversarial learning, reconstruction guarantees, collaborative inference, MLaaS

Compressor summary: The paper proposes a new framework to provide formal privacy guarantees for cloud-based machine learning inference using local sensitivity and extends an existing method for neural network queries.


HotBEV: Hardware-oriented Transformer-based Multi-View 3D Detector for BEV Perception

Peiyan Dong, Zhenglun Kong, Xin Meng, Pinrui Yu, Yifan Gong, Geng Yuan, Hao Tang, Yanzhi Wang

https://openreview.net/forum?id=3Cj67k38st

Keywords: Multi-view 3D detection, Hardware efficiency, Autonomous driving

Compressor summary: This paper proposes a latency-aware model design for bird's-eye-view detection in autonomous driving systems, using efficient building operators and a hardware-oriented backbone to achieve significant speedups while maintaining accuracy.


On the Last-iterate Convergence in Time-varying Zero-sum Games: Extra Gradient Succeeds where Optimism Fails

Yi Feng, Hu Fu, Qun Hu, Ping Li, Ioannis Panageas, bo peng, Xiao Wang

https://openreview.net/forum?id=3CJOaJugMG

Keywords: zero sum game, time-varying game, optimistic gradient, extra gradient, momentum method

Compressor summary: The paper investigates how OGDA and EG perform in time-varying bilinear zero-sum games and finds that they have different last-iterate behaviors.


Schema-learning and rebinding as mechanisms of in-context learning and emergence

Sivaramakrishnan Swaminathan, Antoine Dedieu, Rajkumar Vasudeva Raju, Murray Shanahan, Miguel Lazaro-Gredilla, Dileep George

https://openreview.net/forum?id=3AreDQZ8eO

Keywords: mechanistic interpretability, in-context learning, emergence, large language models

Compressor summary: This paper proposes an interpretable alternative method for in-context learning using causal graphs and shows how it works at different levels of complexity.


Time-uniform confidence bands for the CDF under nonstationarity

Paul Mineiro, Steven R Howard

https://openreview.net/forum?id=39cFjnRpYm

Keywords: off-policy evaluation, anytime-valid

Compressor summary: The paper proposes bounds on the distribution of a sequence of real-valued random variables and their importance-weighted extension for counterfactual distributions in randomized experiments.


Im-Promptu: In-Context Composition from Image Prompts

Bhishma Dedhia, Michael Chang, Jake Snell, Thomas L. Griffiths, Niraj Jha

https://openreview.net/forum?id=38o372YoYt

Keywords: in-context learning, compositionality, generative models

Compressor summary: This paper explores how analogical reasoning can help in-context visual learners generalize to new tasks and domains, and introduces a meta-learning framework called Im-Promptu that trains agents with different levels of compositional granularity.


Fairness Aware Counterfactuals for Subgroups

Loukas Kavouras, Konstantinos Tsopelas, Giorgos Giannopoulos, Dimitris Sacharidis, Eleni Psaroudaki, Nikolaos Theologitis, Dimitrios Rontogiannis, Dimitris Fotakis, Ioannis Emiris

https://openreview.net/forum?id=38dQv3OwN3

Keywords: subgroup fairness, recourse, counterfactual explanations

Compressor summary: FACTS is a framework to audit subgroup fairness using counterfactual explanations that consider individual and group-level difficulties in achieving desired outcomes while being efficient and explainable.


Explore to Generalize in Zero-Shot RL

Ev Zisselman, Itai Lavie, Daniel Soudry, Aviv Tamar

https://openreview.net/forum?id=37cADkATD0

Keywords: Reinforcement Learning, Generalization, State Space Maximum Entropy Exploration

Compressor summary: ExpGen is a reinforcement learning algorithm that uses exploration to achieve zero-shot generalization on unseen tasks, outperforming previous invariance-based methods and achieving state-of-the-art results on ProcGen challenge domains.


Representational Strengths and Limitations of Transformers

Clayton Sanford, Daniel Hsu, Matus Telgarsky

https://openreview.net/forum?id=36DxONZ9bA

Keywords: self-attention, approximation theory, communication complexity

Compressor summary: The text discusses the benefits and drawbacks of attention layers in transformers, focusing on their representation power and complexity in different scenarios, using proof techniques that highlight communication complexity and sparse averaging as a key attention task.


Equivariant Spatio-Temporal Attentive Graph Networks to Simulate Physical Dynamics

Liming Wu, Zhichao Hou, Jirui Yuan, Yu Rong, Wenbing Huang

https://openreview.net/forum?id=35nFSbEBks

Keywords: Equivariance, Spatio-Temporal GNNs, Physical Dynamics

Compressor summary: The paper introduces ESTAG, an equivariant spatio-temporal graph neural network that uses a novel Equivariant Discrete Fourier Transform and forward attention to simulate non-Markovian physical systems.


Idempotent Learned Image Compression with Right-Inverse

Yanghao Li, Tongda Xu, Yan Wang, Jingjing Liu, Ya-Qin Zhang

https://openreview.net/forum?id=35dOU92OJM

Keywords: learned image compression, idempotent compression, right-inverse

Compressor summary: The paper proposes a new idempotent image compression codec using blocked convolution and null-space enhancement, which achieves state-of-the-art performance and has less quality decay after re-compression.


Evolving Connectivity for Recurrent Spiking Neural Networks

Guan Wang, Yuhao Sun, Sijie Cheng, Sen Song

https://openreview.net/forum?id=30o4ARmfC3

Keywords: neuromorphic computing, spiking neural networks, evolutionary algorithms, inference-only approach, hardware-friendly, robotic locomotion tasks

Compressor summary: The evolving connectivity (EC) framework trains recurrent spiking neural networks without gradients by optimizing connection probability distributions with natural evolution strategies, achieving comparable performance to deep neural networks and outperforming gradient-trained RSNNs on robotic locomotion tasks.


Mutual Information Regularized Offline Reinforcement Learning

Xiao Ma, Bingyi Kang, Zhongwen Xu, Min Lin, Shuicheng YAN

https://openreview.net/forum?id=2z8noau98f

Keywords: Mutual Information, Offline Reinforcement Learning

Compressor summary: MISA is a novel framework for offline RL that maximizes mutual information between states and actions to improve policy evaluation and constrain the policy improvement direction.


A Diffusion-Model of Joint Interactive Navigation

Matthew Niedoba, Jonathan Wilder Lavington, Yunpeng Liu, Vasileios Lioutas, Justice Sefas, Xiaoxuan Liang, Dylan Green, Setareh Dabiri, Berend Zwartsenberg, Adam Scibior, Frank Wood

https://openreview.net/forum?id=2yXExAl0FW

Keywords: Diffusion Models, Trajecotry Forecasting, Autonomous Vehicles, Motion Forecasting, Simulation

Compressor summary: The paper introduces DJINN, a diffusion method that generates realistic traffic scenarios for autonomous vehicle simulations using past, present, or future state observations.


Compressed Video Prompt Tuning

Bing Li, Jiaxin Chen, Xiuguo Bao, Di Huang

https://openreview.net/forum?id=2vADOf3K00

Keywords: Compressed video, Action Recognition, Prompt Tuning

Compressor summary: The paper introduces Compressed Video Prompt Tuning (CVPT), a novel approach to adapt pre-trained raw video models to compressed video understanding tasks, using conditional prompts and cross-modal complementary blocks to improve efficiency and accuracy.


Training on Foveated Images Improves Robustness to Adversarial Attacks

Muhammad A Shah, Aqsa Kashaf, Bhiksha Raj

https://openreview.net/forum?id=2tfG9QaFA7

Keywords: adversarial robustness, computer vision, biologically-inspired, retina, blurring

Compressor summary: The authors propose a method called RBlur that simulates peripheral vision blurring and color desaturation to improve the robustness of deep neural networks against various types of image corruptions, including adversarial attacks.


What Do Deep Saliency Models Learn about Visual Attention?

Shi Chen, Ming Jiang, Qi Zhao

https://openreview.net/forum?id=2rq4LwwjfE

Keywords: Saliency prediction, human attention, low-level vision

Compressor summary: The paper introduces a method to analyze how deep saliency models predict human visual attention by breaking down their implicit features into interpretable components related to semantics and applying it to different scenarios.


Double Auctions with Two-sided Bandit Feedback

Soumya Basu, Abishek Sankararaman

https://openreview.net/forum?id=2nTpPxJ5Bs

Keywords: Double Auction, Markets, Bandits, Regret

Compressor summary: The paper studies how buyers and sellers can efficiently learn their valuations and discover prices in Double Auction markets with bandit feedback, and shows that certain regrets are unachievable.


Density of States Prediction of Crystalline Materials via Prompt-guided Multi-Modal Transformer

Namkyeong Lee, Heewoong Noh, Sungwon Kim, Dongmin Hyun, Gyoung S. Na, Chanyoung Park

https://openreview.net/forum?id=2lWh1G1W1I

Keywords: ML4Materials, AI4Science, Graph Neural Networks

Compressor summary: The paper proposes a multi-modal transformer that predicts density of states (DOS) in crystalline materials by considering both energy levels and the relationships between atoms, and shows its effectiveness on two types of DOS.


Memory-Efficient Fine-Tuning of Compressed Large Language Models via sub-4-bit Integer Quantization

Jeonghoon Kim, Jung Hyun Lee, Sungdong Kim, Joonsuk Park, Kang Min Yoo, Se Jung Kwon, Dongsoo Lee

https://openreview.net/forum?id=2jUKhUrBxP

Keywords: Large Language Models, Parameter-Efficient Fine-Tuning, Neural Network Quantization

Compressor summary: PEQA is a method that combines parameter-efficient fine-tuning and quantization to reduce memory demands and computational costs of large language models while maintaining or improving their performance.


A Combinatorial Algorithm for Approximating the Optimal Transport in the Parallel and MPC Settings

Nathaniel Lahn, Sharath Raghvendra, Kaiyi Zhang

https://openreview.net/forum?id=2izFpGERjU

Keywords: Optimal Transport, Combinatorial Optimization

Compressor summary: The paper introduces a parallel combinatorial algorithm to compute an additive epsilon-approximation of the optimal transport distance with improved complexity and speed compared to existing methods.


Neuro-symbolic Learning Yielding Logical Constraints

Zenan Li, Yunpeng Huang, Zhaoyu Li, Yuan Yao, Jingwei Xu, Taolue Chen, Xiaoxing Ma, Jian Lu

https://openreview.net/forum?id=2ioRi2uwLR

Keywords: Neuro-symbolic learning, logical constraint learning, symbol grounding, difference-of-convex relaxation

Compressor summary: The paper presents a framework that integrates neural network training, symbol grounding, and logical constraint synthesis for end-to-end learning of neuro-symbolic systems, using techniques to relax and maintain logical constraints.


Estimating Propensity for Causality-based Recommendation without Exposure Data

Zhongzhou Liu, Yuan Fang, Min Wu

https://openreview.net/forum?id=2hhIDEHhkk

Keywords: recommendation systems, causal effect, propensity score, propensity estimation

Compressor summary: The paper proposes PropCare, a framework that estimates exposure and propensity from interaction data for causality-based recommendation systems without additional input.


What is the Inductive Bias of Flatness Regularization? A Study of Deep Matrix Factorization Models

Khashayar Gatmiry, Zhiyuan Li, Tengyu Ma, Sashank J. Reddi, Stefanie Jegelka, Ching-Yao Chuang

https://openreview.net/forum?id=2hQ7MBQApp

Keywords: Sharpness minimization, Deep learning, Matrix factorization, Deep linear networks, Implicit bias, SGD, Trace of Hessian regularizer

Compressor summary: This paper investigates how minimizing the sharpness of the loss function improves generalization in deep linear networks trained with certain measurements, and shows that it is equivalent to minimizing a matrix norm.


Mode Connectivity in Auction Design

Christoph Hertrich, Yixin Tao, László A. Végh

https://openreview.net/forum?id=2gn9WFlqJ4

Keywords: Differentiable Economics, Mechanism Design, Neural Network Theory, Mode Connectivity, RochetNet

Compressor summary: The paper analyzes mode connectivity of neural networks for optimal auction design and provides the first theoretical justification for their empirical success in this area.


TopP&R: Robust Support Estimation Approach for Evaluating Fidelity and Diversity in Generative Models

Pum Jun Kim, Yoojin Jang, Jisu Kim, Jaejun Yoo

https://openreview.net/forum?id=2gUCMr6fDY

Keywords: GAN, Evaluation, Support Estimation

Compressor summary: TopP&R is a new and better way to evaluate generative models by measuring the significance and confidence of features in samples, unlike existing metrics that rely on unreliable support estimates.


Effective Targeted Attacks for Adversarial Self-Supervised Learning

Minseon Kim, Hyeonjeong Ha, Sooel Son, Sung Ju Hwang

https://openreview.net/forum?id=2f0dlMZlNb

Keywords: Adversarial self supervised learning, targeted attack, self supervised learning, contrastive learning, positive mining

Compressor summary: The paragraph discusses a new method for improving adversarial robustness in unsupervised learning by using positive mining for targeted attacks instead of general adversarial attacks.


Foundation Model is Efficient Multimodal Multitask Model Selector

Fanqing Meng, Wenqi Shao, zhanglin peng, Chonghe Jiang, Kaipeng Zhang, Yu Qiao, Ping Luo

https://openreview.net/forum?id=2ep5PXEZiw

Keywords: transfer learning, model selection, foundation model

Compressor summary: The paper proposes EMMS, a fast and effective method to estimate the transferability of pre-trained neural networks on multiple multi-modal tasks without fine-tuning them, using large-scale foundation models and unified label embeddings.


Learning Invariant Representations with a Nonparametric Nadaraya-Watson Head

Alan Q. Wang, Minh Nguyen, Mert R. Sabuncu

https://openreview.net/forum?id=2ePf1sBgLU

Keywords: Invariant representations, causality, domain generalization

Compressor summary: The paper proposes a nonparametric method using Nadaraya-Watson heads to learn invariant representations across different environments, which can improve domain generalization in machine learning models.


Probabilistic Exponential Integrators

Nathanael Bosch, Philipp Hennig, Filip Tronarp

https://openreview.net/forum?id=2dx5MNs2Ip

Keywords: Probabilistic numerics, differential equations, exponential integrators, Kalman filters, Gaussian processes

Compressor summary: The paper presents a new class of probabilistic integrators for dynamical systems that improve stability and efficiency in stiff problems by including fast, linear dynamics in the prior, and generalizing them to non-linear systems using piece-wise semi-linearity.


Bilevel Coreset Selection in Continual Learning: A New Formulation and Algorithm

Jie Hao, Kaiyi Ji, Mingrui Liu

https://openreview.net/forum?id=2dtU9ZbgSN

Keywords: Coreset Selection, Continual Learning, Bilevel Optimization

Compressor summary: The paper proposes a new bilevel formulation for coreset selection in rehearsal-based continual learning, using a novel regularizer and an efficient optimization algorithm that outperforms baselines.


Efficient Online Clustering with Moving Costs

Dimitris Christou, EFSTRATIOS PANTELEIMON SKOULAKIS, Volkan Cevher

https://openreview.net/forum?id=2doqt9r0r0

Keywords: Online Learning, Regret Analysis, Clustering, k-Median

Compressor summary: The paper proposes an online learning algorithm for clustering facilities with moving costs that achieves logarithmic regret and beats the best fixed solution's average cost by a factor of poly-log(n).


Robust Second-Order Nonconvex Optimization and Its Application to Low Rank Matrix Sensing

Shuyao Li, Yu Cheng, Ilias Diakonikolas, Jelena Diakonikolas, Rong Ge, Stephen Wright

https://openreview.net/forum?id=2ccH4zjKVs

Keywords: low rank matrix sensing, non-convex optimization, high-dimensional robust statistics, second-order optimization, statistical query model

Compressor summary: The paper proposes a general framework to find approximate second-order stationary points in nonconvex optimization with dimension-independent accuracy guarantees, and applies it to low rank matrix sensing with robustness to data corruption.


Setting the Trap: Capturing and Defeating Backdoors in Pretrained Language Models through Honeypots

Ruixiang Tang, Jiayi Yuan, Yiming Li, Zirui Liu, Rui Chen, Xia Hu

https://openreview.net/forum?id=2cYxNWNzk3

Keywords: Backdoor Defense, Honeypot

Compressor summary: The study proposes an honeypot module to absorb backdoor information and prevent backdoor attacks during fine-tuning of pretrained language models, achieving substantial reduction in attack success rate.


Towards Accelerated Model Training via Bayesian Data Selection

Zhijie Deng, Peng Cui, Jun Zhu

https://openreview.net/forum?id=2bRG4Hj8qd

Keywords: data selection, training acceleration, probabilistic modeling, Bayesian methods

Compressor summary: The paper proposes an efficient algorithm that selects data based on its impact on generalization loss, using Bayesian treatment and zero-shot predictors, and shows improved training efficiency on noisy and imbalanced datasets compared to existing methods.


Enhancing CLIP with CLIP: Exploring Pseudolabeling for Limited-Label Prompt Tuning

Cristina Menghini, Andrew Delworth, Stephen Bach

https://openreview.net/forum?id=2b9aY2NgXE

Keywords: vision-language models, prompt-tuning, pseudolabels, self-training

Compressor summary: The study explores using zero-shot pseudolabels and prompt tuning strategies to enhance CLIP's performance on image classification tasks without requiring additional labeled data.


Implicit Convolutional Kernels for Steerable CNNs

Maksim Zhdanov, Nico Hoffmann, Gabriele Cesa

https://openreview.net/forum?id=2YtdxqvdjX

Keywords: equivariance; group convolutions; implicit kernels; physical simulations

Compressor summary: The paper introduces a flexible way to implement Steerable CNNs using MLPs to parameterize G-steerable kernels, which generalizes to different groups G and performs well on diverse tasks.


CLIP-OGD: An Experimental Design for Adaptive Neyman Allocation in Sequential Experiments

Jessica Dai, Paula Gradu, Christopher Harshaw

https://openreview.net/forum?id=2Xqvk2KVAq

Keywords: causal inference, randomized experiments, online optimization

Compressor summary: The paper studies how to create adaptive designs for causal inference that are almost as efficient as non-adaptive ones and proposes new performance measures and an example design that performs well.


Structured Federated Learning through Clustered Additive Modeling

Jie Ma, Tianyi Zhou, Guodong Long, Jing Jiang, Chengqi Zhang

https://openreview.net/forum?id=2XT3UpOv48

Keywords: Federated Learning

Compressor summary: The paper proposes Clustered Additive Modeling (CAM) for heterogeneous federated learning, which combines global and cluster models to capture shared features and prevent "clustering collapse."


Scaling laws for language encoding models in fMRI

Richard Antonello, Aditya Vaidya, Alexander Huth

https://openreview.net/forum?id=2W4LxJbgec

Keywords: Encoding Models, Language Models, Neuroscience, Scaling Laws

Compressor summary: Larger language models like OPT and LLaMA can better predict brain responses to natural language using fMRI, with performance scaling logarithmically with model size and nearing a theoretical maximum for certain brain areas.


Revisiting Implicit Differentiation for Learning Problems in Optimal Control

Ming Xu, Timothy L Molloy, Stephen Gould

https://openreview.net/forum?id=2URr3mkagy

Keywords: implicit differentiation, bi-level optimization; constrained learning and control; safe learning for control

Compressor summary: The paper presents a new method for solving non-convex optimal control problems using the implicit function theorem, which improves scalability, numerical stability, and parallelization compared to previous methods.


On the Connection between Pre-training Data Diversity and Fine-tuning Robustness

Vivek Ramanujan, Thao Nguyen, Sewoong Oh, Ali Farhadi, Ludwig Schmidt

https://openreview.net/forum?id=2SScUiWUbn

Keywords: robustness, out-of-distribution shifts, finetuning, pretraining

Compressor summary: The main factor affecting the robustness of fine-tuned models after pre-training is data quantity, while other factors like label space, image diversity, and domain have limited impact.


DoWG Unleashed: An Efficient Universal Parameter-Free Gradient Descent Method

Ahmed Khaled, Konstantin Mishchenko, Chi Jin

https://openreview.net/forum?id=2RQhgx1WLA

Keywords: normalized gradient descent, gradient descent, adagrad, adaptive optimization, parameter-free, smooth optimization, convex optimization, edge of stability

Compressor summary: DoWG is a new gradient-based optimizer that works efficiently and universally without parameters and achieves this by using a distance-based weighted average of gradients.


Group Robust Classification Without Any Group Information

Christos Tsirigotis, Joao Monteiro, Pau Rodriguez, David Vazquez, Aaron Courville

https://openreview.net/forum?id=2OcNWFHFpk

Keywords: out-of-distribution generalization, robustness, fairness, spurious correlations, systematic generalization, model selection

Compressor summary: The paper proposes a new bias-unsupervised method to improve group robustness in ERM by using pretrained self-supervised models and logit adjustment training loss, addressing the limitations of existing approaches that depend on group information or bias labels.


CoLLAT: On Adding Fine-grained Audio Understanding to Language Models using Token-Level Locked-Language Tuning

Amila Silva, Spencer Whitehead, Chris Lengerich, Hugh James Leather

https://openreview.net/forum?id=2NncD8AaFK

Keywords: Audio Understanding, Contrastive Learning, Audio-Language Grounding

Compressor summary: The proposed method $CoLLAT$ enhances audio understanding and provides fine-grained audio grounding using a novel pretraining objective, achieving state-of-the-art results in various downstream tasks.


Segment Anything in 3D with NeRFs

Jiazhong Cen, Zanwei Zhou, Jiemin Fang, chen yang, Wei Shen, Lingxi Xie, Dongsheng Jiang, XIAOPENG ZHANG, Qi Tian

https://openreview.net/forum?id=2NkGfA66Ne

Keywords: Segmentation, NeRF, 3D segmentation

Compressor summary: SA3D is an efficient method that uses SAM and NeRF to segment 3D objects from 2D images with minimal manual input.


Gaussian Membership Inference Privacy

Tobias Leemann, Martin Pawelczyk, Gjergji Kasneci

https://openreview.net/forum?id=2NUFe4TZMS

Keywords: Privacy, Membership Inference Attacks

Compressor summary: The paper introduces $f$-MIP, a novel privacy notion for machine learning models that considers realistic adversaries and offers better utility, and proposes $\mu$-GMIP, an enhancement of $f$-MIP with added noise to gradient updates.


Unifying Predictions of Deterministic and Stochastic Physics in Mesh-reduced Space with Sequential Flow Generative Model

Luning Sun, Xu Han, Han Gao, Jian-Xun Wang, Liping Liu

https://openreview.net/forum?id=2JtwuJtoa0

Keywords: AI4Science, Fluid Dynamics, Generative Models, Graph Neural Network

Compressor summary: The paper proposes a new model that combines generative and sequential networks to accurately predict both deterministic and stochastic dynamical systems, using an autoencoder and a conditional normalizing flow model.


Constructing Non-isotropic Gaussian Diffusion Model Using Isotropic Gaussian Diffusion Model for Image Editing

Xi Yu, Xiang Gu, Haozhi Liu, Jian Sun

https://openreview.net/forum?id=2Ibp83esmb

Keywords: score-based diffusion model, non-isotropic Gaussian diffusion model, image editing

Compressor summary: The Non-isotropic Gaussian Diffusion Model (NGDM) is a novel technique for image editing that uses different noises and diffusion times for different pixels, achieving high quality results while preserving the source image.


A General Theory of Correct, Incorrect, and Extrinsic Equivariance

Dian Wang, Xupeng Zhu, Jung Yeon Park, Mingxi Jia, Guanang Su, Robert Platt, Robin Walters

https://openreview.net/forum?id=2FMJtNDLeE

Keywords: Equivariance, Deep Learning, Error Bound, Symmetry

Compressor summary: The paper presents a theory to analyze and quantify various types of equivariance in neural networks when the ground truth function is only partially symmetric, and studies its impact on model error.


ViSt3D: Video Stylization with 3D CNN

Ayush Pande, Gaurav Sharma

https://openreview.net/forum?id=2EiqizElGO

Keywords: Video style transfer

Compressor summary: The paper presents a new approach to video stylization using 3D CNN that disentangles motion and appearance, stylizes the appearance, and then adds back the motion, as well as introducing a new dataset for training and testing video stylization networks.


Improved Bayesian Regret Bounds for Thompson Sampling in Reinforcement Learning

Ahmadreza Moradipari, Mohammad Pedramfar, Modjtaba Shokrian Zini, Vaneet Aggarwal

https://openreview.net/forum?id=2EVTB1idyR

Keywords: Thompson Sampling, Reinforcement Learning, Bayesian Regret

Compressor summary: The paper shows new Bayesian regret bounds for Thompson Sampling in various reinforcement learning scenarios, with improved upper bounds for the information ratio.


Any-to-Any Generation via Composable Diffusion

Zineng Tang, Ziyi Yang, Chenguang Zhu, Michael Zeng, Mohit Bansal

https://openreview.net/forum?id=2EDqbSCnmF

Keywords: Generative AI, Diffusion Model, Multimodal Generation, Audio-Video Generation

Compressor summary: CoDi is a novel generative model that can create any combination of language, image, video, or audio from any input modalities, using a unique composable generation strategy and multimodal space alignment.


Learning to Compress Prompts with Gist Tokens

Jesse Mu, Xiang Lisa Li, Noah Goodman

https://openreview.net/forum?id=2DtxPCL3T5

Keywords: language models, instruction finetuning, prompt compression, distillation, context distillation, prompting, soft prompting, efficiency

Compressor summary: The paper proposes gisting, a method that trains language models to compress prompts into smaller sets of tokens for compute efficiency and other benefits.


Navigating Data Heterogeneity in Federated Learning: A Semi-Supervised Approach for Object Detection

Taehyeon Kim, Eric Lin, Junu Lee, Christian Lau, Vaikkunth Mugunthan

https://openreview.net/forum?id=2D7ou48q0E

Keywords: Federated Learning, Semi-Supervised Learning, Object Detection

Compressor summary: The paper introduces FedSTO, a semi-supervised federated object detection framework for autonomous driving that uses selective training and orthogonality regularization to handle data shift and limited labeled data.


Discriminative Calibration: Check Bayesian Computation from Simulations and Flexible Classifier

Yuling Yao, Justin Domke

https://openreview.net/forum?id=2Cmdh5z6ph

Keywords: simulation based calibration, simulation based inference, Bayesian computation, diagnostics, classifier two-sample test, likelihood-free

Compressor summary: The paper proposes a classification approach that learns test statistics from data to improve Bayesian calibration accuracy and provide interpretable divergence measures, which can be used for simulation-based or traditional inference methods.


Characterizing the Optimal $0-1$ Loss for Multi-class Classification with a Test-time Attacker

Sihui Dai, Wenxin Ding, Arjun Nitin Bhagoji, Daniel Cullina, Haitao Zheng, Ben Y. Zhao, Prateek Mittal

https://openreview.net/forum?id=2CRaOpEKWh

Keywords: adversarial robustness, graph theory, fundamental bounds

Compressor summary: The paper develops a framework to find information-theoretic lower bounds on robust loss for multi-class classifiers under test-time attacks and compares them with state-of-the-art methods.


DOSE: Diffusion Dropout with Adaptive Prior for Speech Enhancement

Wenxin Tai, Yue Lei, Fan Zhou, Goce Trajcevski, Ting Zhong

https://openreview.net/forum?id=2C2WZfCfo9

Keywords: speech enhancement, diffusion models, adaptive prior, dropout, generalization

Compressor summary: The paper proposes DOSE, a model-agnostic method that uses two techniques to incorporate condition information into DDPMs for speech enhancement, improving speech quality and stability.


ISP: Multi-Layered Garment Draping with Implicit Sewing Patterns

Ren Li, Benoît Guillard, Pascal Fua

https://openreview.net/forum?id=2BrHBj1Puu

Keywords: garment modeling, draping, deformation, human body modeling

Compressor summary: The paper introduces a new model for draping garments on human body models that can handle multi-layered clothing and various poses, using a combination of 2D and 3D parameterization.


Solving Inverse Physics Problems with Score Matching

Benjamin Holzschuh, Simona Vegetti, Nils Thuerey

https://openreview.net/forum?id=2BpoGPSDCR

Keywords: inverse problems, diffusion models, learned corrections, score matching

Compressor summary: The authors propose a method to solve inverse problems in physics using diffusion models that combines an approximate inverse simulator and a learned correction function, achieving high accuracy and temporal stability.


Learning Functional Transduction

Mathieu Chalvidal, Thomas Serre, Rufin VanRullen

https://openreview.net/forum?id=2BFZ8cPIf6

Keywords: Meta-learning, Neural Operators, Kernel methods, In-context learning

Compressor summary: The paper proposes a hybrid approach that combines transductive and inductive methods using vector-valued Reproducing Kernel Banach Spaces to create a meta-learned neural approximator called Transducer, which can efficiently capture new functional relationships and model physical systems with little data.


Sharpness-Aware Minimization Leads to Low-Rank Features

Maksym Andriushchenko, Dara Bahri, Hossein Mobahi, Nicolas Flammarion

https://openreview.net/forum?id=29WbraPk8U

Keywords: sharpness-aware minimization, low-rank features, understanding feature learning

Compressor summary: Sharpness-aware minimization (SAM) reduces feature rank by pruning activations across various neural network architectures and objectives.


Improving Graph Matching with Positional Reconstruction Encoder-Decoder Network

Yixiao Zhou, Ruiqi Jia, Hongxiang Lin, Hefeng Quan, Yumeng Zhao, Xiaoqing Lyu

https://openreview.net/forum?id=28RTu9MOT6

Keywords: Graph Matching, Positional Encoding

Compressor summary: The paper introduces a new graph matching method, PREGM, that leverages spatial information of keypoints to establish correspondence between keypoint sets in images using a positional reconstruction encoder-decoder (PR-EnDec).


Going beyond persistent homology using persistent homology

Johanna Emilia Immonen, Amauri H Souza, Vikas Garg

https://openreview.net/forum?id=27TdrEvqLD

Keywords: graph representation learning, topological deep learning, persistent homology, graph neural networks

Compressor summary: The text introduces a new concept of color-separating sets to improve the representation of graphs using persistent homology and proposes RePHINE, a model that combines vertex- and edge-level information, achieving better results than existing methods on graph classification tasks.


Noether Embedding: Efficient Learning of Temporal Regularities

Chi Gao, Zidong Zhou, Luping Shi

https://openreview.net/forum?id=27CRbwewyb

Keywords: Schema Learning, Temporal Regularity, Event Embedding

Compressor summary: The authors propose Noether Embedding (NE), an efficient method to learn temporal regularities from event embeddings, which outperforms existing methods in detecting and querying valid temporal patterns.


General Munchausen Reinforcement Learning with Tsallis Kullback-Leibler Divergence

Lingwei Zhu, Zheng Chen, Matthew Kyle Schlegel, Martha White

https://openreview.net/forum?id=26qqUHi9XF

Keywords: reinforcement learning, entropy regularization, Tsallis KL divergence

Compressor summary: The paper proposes a new policy optimization approach in reinforcement learning using Tsallis KL divergence, which generalizes the standard KL divergence and can improve performance in Atari games.


Beyond Normal: On the Evaluation of Mutual Information Estimators

Paweł Czyż, Frederic Grabowski, Julia E Vogt, Niko Beerenwinkel, Alexander Marx

https://openreview.net/forum?id=25vRtG56YH

Keywords: Mutual Information, Invariance, Benchmark, Geometric Machine Learning

Compressor summary: The paper introduces a diverse family of distributions with known ground-truth mutual information and a language-independent benchmarking platform for evaluating mutual information estimators in various settings.


CAPro: Webly Supervised Learning with Cross-modality Aligned Prototypes

Yulei Qin, Xingyu Chen, Yunhang Shen, Chaoyou Fu, Yun Gu, Ke Li, Xing Sun, Rongrong Ji

https://openreview.net/forum?id=25HiFHPcXg

Keywords: webly supervised learning, representation learning, visual-semantic alignment, collective bootstrapping

Compressor summary: CAPro is a method that combats label noise in webly supervised learning by using textual prototypes, visual feature enhancement, and collective bootstrapping to learn semantically aligned visual representations.


DiffuseBot: Breeding Soft Robots With Physics-Augmented Generative Diffusion Models

Tsun-Hsuan Wang, Juntian Zheng, Pingchuan Ma, Yilun Du, Byungchul Kim, Andrew Everett Spielberg, Joshua B. Tenenbaum, Chuang Gan, Daniela Rus

https://openreview.net/forum?id=1zo4iioUEs

Keywords: soft robot, diffusion model, co-design

Compressor summary: The paper introduces DiffuseBot, a model that generates soft robot morphologies for various tasks by combining physics simulation with diffusion models.


Comparing Causal Frameworks: Potential Outcomes, Structural Models, Graphs, and Abstractions

Duligur Ibeling, Thomas Icard

https://openreview.net/forum?id=1zKRwh5Rl2

Keywords: potential outcomes framework, structural causal model, causal inference, logic, probability, graphical causal models, causal abstraction, causal machine learning

Compressor summary: The paper clarifies how Rubin causal models and structural causal models relate and shows that all Rubin models can be represented by structural models, despite some differences in their principles.


Computing a human-like reaction time metric from stable recurrent vision models

Lore Goetschalckx, Lakshmi Narasimhan Govindarajan, Alekh Karkada Ashok, Aarit Ahuja, David Sheinberg, Thomas Serre

https://openreview.net/forum?id=1xPsn2gCOe

Keywords: alignment, RNNs, reaction times, equilibrium dynamics, perceptual grouping, decision making

Compressor summary: The authors present a method to measure reaction times in recurrent vision models, which helps align them with human visual decision-making patterns across different tasks.


Hierarchical VAEs provide a normative account of motion processing in the primate brain

Hadi Vafaii, Jacob L. Yates, Daniel A. Butts

https://openreview.net/forum?id=1wOkHN9JK8

Keywords: NeuroAI, VAE, Dorsal stream, Hierarchical Bayesian Inference

Compressor summary: The paper evaluates a new hierarchical VAE model for motion perception tasks and shows how it aligns with brain function in interpreting causal relationships between stimuli and neuronal responses.


Rethinking Bias Mitigation: Fairer Architectures Make for Fairer Face Recognition

Samuel Dooley, Rhea Sanjay Sukthanker, John P Dickerson, Colin White, Frank Hutter, Micah Goldblum

https://openreview.net/forum?id=1vzF4zWQ1E

Keywords: Bias Mitigation, Fairness, Facial Recognition

Compressor summary: The paragraph discusses how biases in face recognition systems are inherent to neural network architectures, and presents a new method for finding fairer architecture designs that outperforms existing methods on accuracy and fairness.


Unexpected Improvements to Expected Improvement for Bayesian Optimization

Sebastian Ament, Sam Daulton, David Eriksson, Maximilian Balandat, Eytan Bakshy

https://openreview.net/forum?id=1vyAG6j9PE

Keywords: Bayesian Optimization, Gaussian Process, Multi-Objective Optimization

Compressor summary: LogEI is a new acquisition function family that is easier to optimize numerically than EI and its variants, improving their performance in Bayesian optimization.


Boosting with Tempered Exponential Measures

Richard Nock, Ehsan Amid, Manfred K Warmuth

https://openreview.net/forum?id=1vvsIJtnnr

Keywords: Boosting, optimization, exponential families

Compressor summary: The paper introduces $t$-AdaBoost, a generalization of AdaBoost that uses tempered exponential measures to improve convergence rates and bound leveraging coefficients for decision tree induction.


Explaining V1 Properties with a Biologically Constrained Deep Learning Architecture

Galen Pogoncheff, Jacob Granley, Michael Beyeler

https://openreview.net/forum?id=1uirUsR9E7

Keywords: NeuroAI, Neuroscience, Visual Stream, Convolutional Neural Networks, Biologically inspired deep learning

Compressor summary: The authors improve CNNs by incorporating neuroscience-derived components to better explain V1 neural activity and properties, advancing NeuroAI research.


Topological Obstructions and How to Avoid Them

Babak Esmaeili, Robin Walters, Heiko Zimmermann, Jan-Willem van de Meent

https://openreview.net/forum?id=1tviRBNxI9

Keywords: representation learning, variational autoencoders, homeomorphism, topological, equivariant, lie groups, normalizing flows

Compressor summary: The paper studies the challenges of training models with geometric latent spaces and proposes a new flow-based model that improves interpretability and generalization by mapping data to multimodal distributions over geometric spaces.


Decorate3D: Text-Driven High-Quality Texture Generation for Mesh Decoration in the Wild

Yanhui Guo, Xinxin Zuo, Peng Dai, Juwei Lu, Xiaolin Wu, Li Cheng, Youliang Yan, Songcen Xu, Xiaofei Wu

https://openreview.net/forum?id=1recIOnzOF

Keywords: Texture Generation, Text-Driven, 3D-Consistent Editing, Neural Radiance Field

Compressor summary: Decorate3D is a method that uses neural networks to create and edit 3D objects from images, allowing users to edit or generate textures with high quality.


Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation

Jiawei Liu, Chunqiu Steven Xia, Yuyao Wang, LINGMING ZHANG

https://openreview.net/forum?id=1qvx610Cu7

Keywords: LLM4Code, ChatGPT, Automated Test Generation

Compressor summary: EvalPlus is a framework that uses automatic test input generation to rigorously evaluate the functional correctness of code synthesized by large language models, revealing previously undetected errors and improving programming benchmarks.


Energy Discrepancies: A Score-Independent Loss for Energy-Based Models

Tobias Schröder, Zijing Ou, Jen Ning Lim, Yingzhen Li, Sebastian Josef Vollmer, Andrew Duncan

https://openreview.net/forum?id=1qFnxhdbxg

Keywords: Energy-based models, statistical discrepancy, latent-variable model, density estimation

Compressor summary: Energy Discrepancy is a new loss function for training energy-based models that is faster, more accurate, and has theoretical guarantees compared to existing methods.


Large Language Models are Visual Reasoning Coordinators

Liangyu Chen, Bo Li, Sheng Shen, Jingkang Yang, Chunyuan Li, Kurt Keutzer, Trevor Darrell, Ziwei Liu

https://openreview.net/forum?id=1q0feiJ2i4

Keywords: visual reasoning, large language models

Compressor summary: The paper introduces Cola, a method that uses a large language model (LLM) to coordinate multiple vision-language models (VLMs) for better visual reasoning by enabling natural language communication between them.


Uncertainty-Aware Instance Reweighting for Off-Policy Learning

Xiaoying Zhang, Junpu Chen, Hongning Wang, Hong Xie, Yang Liu, John C.S. Lui, Hang Li

https://openreview.net/forum?id=1pWNhmbllE

Keywords: off-policy learning, uncertainty

Compressor summary: The paragraph discusses off-policy learning, a procedure used in applications like search engines and recommender systems, and proposes a new estimator (UIPS) that models uncertainty to improve the learning process.


Alternating Updates for Efficient Transformers

Cenk Baykal, Dylan J Cutler, Nishanth Dikkala, Nikhil Ghosh, Rina Panigrahy, Xin Wang

https://openreview.net/forum?id=1p6teT6F73

Keywords: efficiency, efficient transformers

Compressor summary: AltUp is a method that increases a model's capacity by widening token embeddings without increasing latency much, and it can be combined with other techniques for even better results.


Online Convex Optimization with Unbounded Memory

Raunak Kumar, Sarah Dean, Robert Kleinberg

https://openreview.net/forum?id=1osmdAfD4P

Keywords: online learning, online convex optimization, online linear control

Compressor summary: This paper introduces Online Convex Optimization with Unbounded Memory (OCO-UM), a generalization of OCO that accounts for long-term dependence on past decisions, and provides upper and lower bounds on policy regret.


Energy Guided Diffusion for Generating Neurally Exciting Images

Paweł A. Pierzchlewicz, Konstantin Friedrich Willeke, Arne Nix, Pavithra Elumalai, Kelli Restivo, Tori Shinn, Cate Nealley, Gabrielle Rodriguez, Saumil Patel, Katrin Franke, Andreas S. Tolias, Fabian H. Sinz

https://openreview.net/forum?id=1moStpWGUj

Keywords: most exciting inputs, diffusion models, energy guidance, attention, macaque V4

Compressor summary: The authors propose a new method called Energy Guidance (EGG) for generating most exciting inputs (MEIs) that improve predictions of neuronal activity in macaque V4, reduce computational costs, and can be used for other tasks like studying invariances.


Adjustable Robust Reinforcement Learning for Online 3D Bin Packing

Yuxin Pan, Yize Chen, Fangzhen Lin

https://openreview.net/forum?id=1mdTYi1jAW

Keywords: online 3D bin packing problem, combinatorial optimization problem, reinforcement learning

Compressor summary: The paragraph discusses a new reinforcement learning framework, AR2L, which balances average and worst-case performance for solving online 3D bin packing problems, using a permutation-based attacker to evaluate its robustness.


Exploring the Optimal Choice for Generative Processes in Diffusion Models: Ordinary vs Stochastic Differential Equations

Yu Cao, Jingrun Chen, Yixin Luo, Xiang ZHOU

https://openreview.net/forum?id=1mJQq6zYaE

Keywords: diffusion models; stochastic differential equations; score-based generative models; asymptotic analysis

Compressor summary: The paper analyzes when ODE or SDE models are more suitable for computer vision tasks by studying error accumulation in two limiting scenarios and comparing numerical results with different data distributions.


Laplacian Canonization: A Minimalist Approach to Sign and Basis Invariant Spectral Embedding

George Ma, Yifei Wang, Yisen Wang

https://openreview.net/forum?id=1mAYtdoYw6

Keywords: Graph Neural Networks, Positional Encoding, Spectral Embedding, Laplacian Eigenvectors

Compressor summary: The paper proposes Laplacian Canonization, a minimal and efficient pre-processing method for graph embedding that enhances the invariance properties of spectral embedding techniques.


Exponentially Convergent Algorithms for Supervised Matrix Factorization

Joowon Lee, Hanbaek Lyu, Weixin Yao

https://openreview.net/forum?id=1kgK0r8PGg

Keywords: Supervised matrix factorization, multi-objective optimization, global convergence, linear convergence, statistical estimation

Compressor summary: The paper proposes a new method and efficient algorithm for supervised matrix factorization that improves feature extraction and classification tasks in high-dimensional data, such as identifying cancer-related gene groups.


Implicit Transfer Operator Learning: Multiple Time-Resolution Models for Molecular Dynamics

Mathias Schreiner, Ole Winther, Simon Olsson

https://openreview.net/forum?id=1kZx7JiuA2

Keywords: AI4Science, Molecular Dynamics, equivariant neural networks, stochastic dynamics

Compressor summary: Implicit Transfer Operator Learning is a framework to learn surrogates for molecular dynamics simulations with multiple time-resolutions, enabling faster and more accurate modeling of molecular systems using coarse representations.


Neural Sculpting: Uncovering hierarchically modular task structure in neural networks through pruning and network analysis

Shreyas Malakarjun Patil, Loizos Michael, Constantine Dovrolis

https://openreview.net/forum?id=1jhmWkZGy6

Keywords: Neural networks, Hierarchical modularity, Pruning, Sparsity

Compressor summary: The paragraph discusses a method to identify the hierarchical structure of sub-functions in a task using deep neural networks, focusing on the domain of Boolean functions and two vision tasks from the MNIST dataset.


Towards Efficient Image Compression Without Autoregressive Models

Muhammad Salman Ali, Yeongwoong Kim, Maryam Qamar, Sung-Chang Lim, Donghyun Kim, Chaoning Zhang, Sung-Ho Bae, Hui Yong Kim

https://openreview.net/forum?id=1ihGy9vAIg

Keywords: Image Compression, Correlation

Compressor summary: The paper proposes a novel method to improve learned image compression by introducing a correlation loss that reduces discrepancies between latent features and the assumed distribution, achieving similar performance with significantly less computational complexity.


Find What You Want: Learning Demand-conditioned Object Attribute Space for Demand-driven Navigation

Hongcheng Wang, Andy Guan Hong Chen, Xiaoqi Li, Mingdong Wu, Hao Dong

https://openreview.net/forum?id=1hZwxBgQ3G

Keywords: Visual Navigation, Demand-Driven Navigation

Compressor summary: The paper introduces Demand-driven Navigation (DDN), a method that allows agents to navigate using the user's demand, not just object names, and leverages common sense knowledge and visual features for better navigation performance.


Momentum Provably Improves Error Feedback!

Ilyas Fatkhullin, Alexander Tyurin, Peter Richtárik

https://openreview.net/forum?id=1h92PmnKov

Keywords: Heavy-ball momentum, Polyak momentum, Error feedback, Federated Learning, Distributed Optimization, Stochastic optimization, Nonconvex optimization

Compressor summary: The paper proposes a simple fix that improves error feedback algorithms for distributed machine learning by applying Polyak's momentum to the latest version, EF21, and shows theoretical and practical benefits.


Reliable learning in challenging environments

Nina Balcan, Steve Hanneke, Rattana Pukdee, Dravyansh Sharma

https://openreview.net/forum?id=1h7Uh9zUXc

Keywords: Reliable machine learning, adversarial robustness, distribution shift, theory

Compressor summary: The paper proposes a reliable learner with optimal guarantees in challenging test-time environments like adversarial attacks and distribution shifts, and demonstrates strong performance on examples.


Exact Verification of ReLU Neural Control Barrier Functions

Hongchao Zhang, Junlin Wu, Yevgeniy Vorobeychik, Andrew Clark

https://openreview.net/forum?id=1h2TAUEfc4

Keywords: Safety, Neural Barrier Function, Verification

Compressor summary: The paper proposes a method to verify the safety of neural control barrier functions with ReLU activation for nonlinear systems using piecewise linear segments and interval bound propagation.


Learning Unseen Modality Interaction

Yunhua Zhang, Hazel Doughty, Cees G. M. Snoek

https://openreview.net/forum?id=1g0A9kE8Id

Keywords: Multimodal Learning

Compressor summary: The paper proposes a method to learn from different modalities without assuming all combinations are available during training and shows its effectiveness on various tasks like video classification, robot state regression, and multimedia retrieval.


A Theoretical Analysis of Optimistic Proximal Policy Optimization in Linear Markov Decision Processes

Han Zhong, Tong Zhang

https://openreview.net/forum?id=1bTG4sJ7tN

Keywords: policy optimization, adversarial lienar MDPs, RL theory

Compressor summary: The paper proposes an optimistic PPO variant for linear MDPs and proves a state-of-the-art regret bound, while introducing novel algorithm design and analysis techniques.


Incentivized Communication for Federated Bandits

Zhepei Wei, Chuanhao Li, Haifeng Xu, Hongning Wang

https://openreview.net/forum?id=1aQivXgZKj

Keywords: contextual bandit, federated learning, incentive mechanism

Compressor summary: The paper introduces a new federated bandit learning problem that considers self-interested clients and proposes a near-optimal incentivized communication protocol called Inc-FedUCB.


Unified Lower Bounds for Interactive High-dimensional Estimation under Information Constraints

Jayadev Acharya, Clement Louis Canonne, Ziteng Sun, Himanshu Tyagi

https://openreview.net/forum?id=1ZzG6td0el

Keywords: statistical estimation; interactivity; local differential privacy; communication constraint

Compressor summary: The paper develops a framework to derive lower bounds on distributed parameter estimation under various constraints and for different types of distributions using interactive protocols and shows its versatility and effectiveness.


Convolutional State Space Models for Long-Range Spatiotemporal Modeling

Jimmy T.H. Smith, Shalini De Mello, Jan Kautz, Scott Linderman, Wonmin Byeon

https://openreview.net/forum?id=1ZvEtnrHS1

Keywords: spatiotemporal modeling, ConvLSTM, RNN, state spaces, SSM, S4, S5, long-range dependencies, video prediction

Compressor summary: ConvS5 is a fast and efficient model that combines ConvLSTM tensor modeling with state space methods for long-range spatiotemporal sequence generation.


Langevin Quasi-Monte Carlo

Sifan Liu

https://openreview.net/forum?id=1YEF6TA8Di

Keywords: Completely uniformly distributed; log-concave sampling; low-discrepancy; MCMC;

Compressor summary: LMC can reduce estimation error by using quasi-random samples from CUD sequences for generating Gaussian perturbations.


Tanh Works Better with Asymmetry

Dongjin Kim, Woojeong Kim, Suhyun Kim

https://openreview.net/forum?id=1WpmOipyYI

Keywords: Batch Normalization, Activation Functions, Saturation, Sparsity

Compressor summary: Swapping the order of Batch Normalization and bounded activation functions like Tanh improves performance due to increased asymmetry, sparsity, and a modified Tanh function with consistent asymmetry.


Robust Knowledge Transfer in Tiered Reinforcement Learning

Jiawei Huang, Niao He

https://openreview.net/forum?id=1WMdoiVMov

Keywords: Reinforcement Learning Theory, Transfer RL, Tiered RL

Compressor summary: The paper explores how to transfer knowledge between different tasks in parallel reinforcement learning, focusing on a condition called Optimal Value Dominance and proposing novel algorithms for near-optimal performance.


Adaptive Topological Feature via Persistent Homology: Filtration Learning for Point Clouds

Naoki Nishikawa, Yuichi Ike, Kenji Yamanishi

https://openreview.net/forum?id=1TJaITmK2Q

Keywords: point cloud, persistence homology, isometry-invariant networks, filtration learning

Compressor summary: The paper proposes a framework that learns an adaptive filtration for point clouds using neural networks and shows its effectiveness in enhancing machine learning accuracy for various applications.


CommonScenes: Generating Commonsense 3D Indoor Scenes with Scene Graphs

Guangyao Zhai, Evin Pinar Örnek, Shun-Cheng Wu, Yan Di, Federico Tombari, Nassir Navab, Benjamin Busam

https://openreview.net/forum?id=1SF2tiopYJ

Keywords: Scene Graph, Scene Synthesis, Diffusion Model, Graph Convolution Network

Compressor summary: CommonScenes is a model that converts scene graphs into realistic 3D scenes with controllable objects, using variational auto-encoder and latent diffusion, and improves generation consistency, quality, and diversity compared to other methods.


Single-Stage Visual Query Localization in Egocentric Videos

Hanwen Jiang, Santhosh Kumar Ramakrishnan, Kristen Grauman

https://openreview.net/forum?id=1SAzP7W43j

Keywords: Visual Query Localization, Egocentric Video, Spatial-Temporal Correspondence, Episodic Memory

Compressor summary: VQLoC is a single-stage framework that uses joint query-to-frame and frame-to-frame correspondences to perform visual query Localization on long-form egocentric videos faster and more accurately than prior methods.


Gaussian Partial Information Decomposition: Bias Correction and Application to High-dimensional Data

Praveen Venkatesh, Corbett Bennett, Sam Gale, Tamina K. Ramirez, Greggory Heller, Severine Durand, Shawn R Olsen, Stefan Mihalas

https://openreview.net/forum?id=1PnSOKQKvq

Keywords: partial information decomposition, estimation, bias, inter-area interaction, neuroscience

Compressor summary: The paper presents a new method for efficiently computing partial information decompositions on multivariate Gaussian distributions and evaluates its performance on simulated and real mouse brain data.


Budgeting Counterfactual for Offline RL

Yao Liu, Pratik Chaudhari, Rasool Fakoor

https://openreview.net/forum?id=1MUxtSBUox

Keywords: reinforcement learning, offline reinforcement learning, counterfactual reasoning

Compressor summary: The paper proposes a method for offline reinforcement learning that uses dynamic programming to limit counterfactual reasoning and extrapolation errors, achieving better performance than existing approaches on D4RL benchmarks.


Re-Think and Re-Design Graph Neural Networks in Spaces of Continuous Graph Diffusion Functionals

Tingting Dan, Jiaqi Ding, Ziquan Wei, Shahar Z Kovalsky, Minjeong Kim, Won Hwa Kim, Guorong Wu

https://openreview.net/forum?id=1M8nDkUU9b

Keywords: graph neural networks (GNNs), total variation (TV), Euler–Lagrange equation, calculus of variations, over-smoothing, min-max optimization

Compressor summary: The paper presents a general framework for enhancing graph neural networks (GNNs) based on continuous diffusion and variational analysis, which improves their ability to model long-range dependencies and global patterns in graphs, as well as overcomes the over-smoothing problem. The paper also introduces a novel GAN to predict spreading flows in graphs using neural transport equation.


Blurred-Dilated Method for Adversarial Attacks

Yang Deng, Weibin Wu, Jianping Zhang, Zibin Zheng

https://openreview.net/forum?id=1JlAV2paGu

Keywords: Transferable adversarial example

Compressor summary: The paper proposes a new method called Blurred-Dilated (BD) that modifies the source model to generate adversarial examples for DNNs in black-box settings, improving transferability and effectiveness against other target models.


Banana: Banach Fixed-Point Network for Pointcloud Segmentation with Inter-Part Equivariance

Congyue Deng, Jiahui Lei, Bokui Shen, Kostas Daniilidis, Leonidas Guibas

https://openreview.net/forum?id=1IOU2329Za

Keywords: 3D deep learning, equivariant network, pointcloud segmentation, multi-body system

Compressor summary: Banana is a Banach fixed-point network that enables equivariant segmentation by co-evolving part assignment and per-part SE(3)-equivariance in complex systems.


Learning and Collusion in Multi-unit Auctions

Simina Branzei, Mahsa Derakhshan, Negin Golrezaei, Yanjun Han

https://openreview.net/forum?id=1HKJ3lPz6m

Keywords: multi-unit auctions, repeated auctions, online learning, collusion, games and learning, lower bounds, multiplicative weight updates, bandit learning

Compressor summary: The paper studies repeated multi-unit auctions with uniform pricing for CO2 emissions licenses, and analyzes their efficiency, regret, and collusion-proneness.


On Computing Pairwise Statistics with Local Differential Privacy

Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Adam Sealfon

https://openreview.net/forum?id=1GxKVprbwM

Keywords: differential privacy, local differential privacy, pairwise statistics

Compressor summary: The authors develop methods to compute pairwise statistics of user inputs while preserving differential privacy in a local model setting.


Adapting Neural Link Predictors for Data-Efficient Complex Query Answering

Erik Arakelyan, Pasquale Minervini, Daniel Daza, Michael Cochez, Isabelle Augenstein

https://openreview.net/forum?id=1G7CBp8o7L

Keywords: complex query answering, neural link prediction, knowledge graph embeddings, knowledge graphs, relational learning, adapters

Compressor summary: CQD$^{\mathcal{A}}$ is a score adaptation model that improves the accuracy of complex query answering on incomplete knowledge graphs by recalibrating neural link prediction scores with minimal parameter increase and supporting negation reasoning.


How a Student becomes a Teacher: learning and forgetting through Spectral methods

Lorenzo Giambagli, Lorenzo Buffoni, Lorenzo Chicchi, Duccio Fanelli

https://openreview.net/forum?id=1FVmMlifl7

Keywords: Network Slimming, Spectral Analysis, Node Pruning, Teacher-Student

Compressor summary: The paper proposes a new optimization method for student networks in machine learning that allows identifying and isolating an invariant subnetwork that matches the complexity of the teacher network, without degrading performance.


Large language models transition from integrating across position-yoked, exponential windows to structure-yoked, power-law windows

David Skrill, Samuel Victor Norman-Haignere

https://openreview.net/forum?id=1EYKYJeZtR

Keywords: language modeling, temporal integration, transformers, timescales, model interpretation

Compressor summary: The paragraph discusses how language models integrate linguistic information over time and how their integration patterns resemble those of human brains, with structure-dependent exponential and power-law windows across different layers.


HeadSculpt: Crafting 3D Head Avatars with Text

Xiao Han, Yukang Cao, Kai Han, Xiatian Zhu, Jiankang Deng, Yi-Zhe Song, Tao Xiang, Kwan-Yee K. Wong

https://openreview.net/forum?id=1DmP6ySKYq

Keywords: 3D generative model, head avatar, diffusion models, neural rendering

Compressor summary: The paragraph discusses a new method called HeadSculpt that can create and edit high-quality 3D head avatars from textual descriptions by incorporating 3D awareness and fine-grained editing capabilities.


Cascading Contextual Assortment Bandits

Hyunjun Choi, Rajan Udwani, Min-hwan Oh

https://openreview.net/forum?id=1DTCoyAFiV

Keywords: cascade bandit, assortment bandit, upper confidence bound, exploration and exploitation, combinatorial optimization

Compressor summary: The text introduces a new combinatorial bandit model called cascading contextual assortment bandit and two UCB-based algorithms for it that improve existing regret bounds and dependence on problem constants.


Should I Stop or Should I Go: Early Stopping with Heterogeneous Populations

Hammaad Adam, Fan Yin, Mary Hu, Neil Tenenholtz, Lorin Crawford, Lester Mackey, Allison Koenecke

https://openreview.net/forum?id=1CpVHL10fh

Keywords: Randomized experiments, heterogeneous effects, causal machine learning, fairness, sequential testing, clinical trials, A/B testing

Compressor summary: The paper introduces CLASH, a machine learning method to stop randomized experiments early when the treatment harms different groups of participants unevenly.


PoET: A generative model of protein families as sequences-of-sequences

Timothy Fei Truong Jr, Tristan Bepler

https://openreview.net/forum?id=1CJ8D7P8RZ

Keywords: protein fitness prediction, transformer, retrieval, language model, MSA, generative model, protein engineering

Compressor summary: PoET is a generative model that learns to create related proteins from any family by using a Transformer layer and attention mechanism, improving variant function prediction and sequence generation.


De novo Drug Design using Reinforcement Learning with Multiple GPT Agents

Xiuyuan Hu, Guoqing Liu, Yang Zhao, Hao Zhang

https://openreview.net/forum?id=1B6YKnHYBb

Keywords: De novo drug design, Molecular generation, Multi-agent reinforcement learning, GPT

Compressor summary: The paper introduces MolRL-MGPT, a reinforcement learning algorithm with multiple GPT agents for diverse molecular generation in drug design, and shows its effectiveness in designing SARS-CoV-2 inhibitors.


Task-aware Distributed Source Coding under Dynamic Bandwidth

Po-han Li, Sravan Kumar Ankireddy, Ruihan Zhao, Hossein Nourkhiz Mahjoub, Ehsan Moradi Pari, ufuk topcu, Sandeep P. Chinchali, Hyeji Kim

https://openreview.net/forum?id=1A4ZqTmnye

Keywords: Data Compression, Distributed Source Coding, Semantic Communication, Multi-sensor Networks, Bandwidth Allocation, Information Theory

Compressor summary: The authors propose a novel distributed compression framework, NDPCA, that learns low-rank task representations and flexibly adapts to varying available bandwidth in multi-sensor networks, improving performance on tasks like robotic arm manipulation and object detection.


Instructing Goal-Conditioned Reinforcement Learning Agents with Temporal Logic Objectives

Wenjie Qiu, Wensen Mao, He Zhu

https://openreview.net/forum?id=19AgWnmyoV

Keywords: Goal-Conditioned Reinforcement Learning, Linear Temporal Logic

Compressor summary: The paper proposes a new method for teaching reinforcement learning agents to follow complex temporal logic instructions without extra training, using a simple technique that works with any regular expression.


Deciphering Spatio-Temporal Graph Forecasting: A Causal Lens and Treatment

Yutong Xia, Yuxuan Liang, Haomin Wen, Xu Liu, Kun Wang, Zhengyang Zhou, Roger Zimmermann

https://openreview.net/forum?id=17Zkztjlgt

Keywords: Spatio-temporal forecasting

Compressor summary: CaST is a novel framework that uses causal treatments to improve Spatio-Temporal Graph Neural Networks' forecasting performance and address temporal out-of-distribution issues and dynamic spatial causation.


Towards Understanding the Dynamics of Gaussian-Stein Variational Gradient Descent

Tianle Liu, Promit Ghosal, Krishna Balasubramanian, Natesh S. Pillai

https://openreview.net/forum?id=14ZM7FfPx8

Keywords: Stein variational gradient descent, Gaussian variational inference, Rates of Convergence

Compressor summary: SVGD is a nonparametric sampling algorithm whose Gaussian variant has theoretical properties and practical advantages, including convergence to the target distribution in certain cases.


Accelerating Monte Carlo Tree Search with Probability Tree State Abstraction

Yangqing Fu, Ming Sun, Buqing Nie, Yue Gao

https://openreview.net/forum?id=0zeLTZAqaJ

Keywords: reinforcement learning, mento carlo tree search, state abstraction

Compressor summary: The PTSA algorithm improves the efficiency of MCTS-based algorithms like AlphaGo and MuZero by reducing the search space and providing theoretical guarantees, leading to faster training on various tasks.


Fine-Grained Theoretical Analysis of Federated Zeroth-Order Optimization

Jun Chen, Hong Chen, Bin Gu, Hao Deng

https://openreview.net/forum?id=0ycX03sMAT

Keywords: Federated zeroth-order optimization, stability analysis, theoretical guarantee, non-convex optimization, sub-Weibull distribution

Compressor summary: The paper develops a theory for FedZO, a black-box optimization method that combines zeroth-order optimization and federated learning, by analyzing its stability, generalization error bound, and convergence rate in both synchronous and asynchronous settings.


On the Convergence of No-Regret Learning Dynamics in Time-Varying Games

Ioannis Anagnostides, Ioannis Panageas, Gabriele Farina, Tuomas Sandholm

https://openreview.net/forum?id=0x2Ou3xHbH

Keywords: no-regret learning, optimistic gradient descent, time-varying games, dynamic regret

Compressor summary: The paper studies how optimistic gradient descent learns in time-varying multiagent settings and provides convergence bounds for equilibrium gaps and second-order variations.


Lovász Principle for Unsupervised Graph Representation Learning

Ziheng Sun, Chris Ding, Jicong Fan

https://openreview.net/forum?id=0vdEHDwamk

Keywords: Lovász Number, graph-level representation learning, unsupervised learning, semi-supervised learning

Compressor summary: The paper introduces a new graph representation method, Lovász principle, based on graph theory concepts that can be improved with neural networks and outperforms existing methods in some tasks.


The Adversarial Consistency of Surrogate Risks for Binary Classification

Natalie Frank, Jonathan Niles-Weed

https://openreview.net/forum?id=0uARg5G04K

Keywords: Adversarial learning, surrogate risks, optimal transport

Compressor summary: The paper investigates which surrogate losses can replace the original loss function without changing the minimizing sequences for robust binary classification under adversarial attacks.


Two-Stage Predict+Optimize for MILPs with Unknown Parameters in Constraints

Xinyi HU, Jasper C.H. Lee, Jimmy H.M. Lee

https://openreview.net/forum?id=0tnhFpyWjb

Keywords: Constraint optimization, Predict+Optimize

Compressor summary: The paper introduces Two-Stage Predict+Optimize, a new framework for end-to-end training of supervised learning models to handle unknown parameters in both optimization objectives and constraints, with improved prediction performance over existing methods.


Collaborative Score Distillation for Consistent Visual Editing

Subin Kim, Kyungmin Lee, June Suk Choi, Jongheon Jeong, Kihyuk Sohn, Jinwoo Shin

https://openreview.net/forum?id=0tEjORCGFD

Keywords: Score Distillation Sampling, Diffusion model, Editing

Compressor summary: Collaborative Score Distillation (CSD) is a novel method that uses Stein Variational Gradient Descent to achieve consistent visual synthesis across multiple images for various editing tasks.


Transformer-based Planning for Symbolic Regression

Parshin Shojaee, Kazem Meidani, Amir Barati Farimani, Chandan K. Reddy

https://openreview.net/forum?id=0rVXQEeFEL

Keywords: Symbolic Regression, Transformers, Planning, Deep Learning

Compressor summary: TPSR is a new method for symbolic regression that combines transformers with Monte Carlo Tree Search, improving equation accuracy, complexity, extrapolation, and robustness.


Convex-Concave Zero-Sum Stochastic Stackelberg Games

Denizalp Goktas, Arjun Prakash, Amy Greenwald

https://openreview.net/forum?id=0rEJx5QAxt

Keywords: Stackelberg games, Equilibrium Computation, Policy Gradient

Compressor summary: The paper proposes policy gradient methods for solving zero-sum stochastic Stackelberg games with noisy gradients, proves convergence to Stackelberg equilibrium for convex-concave games, and shows that this approach leads to safer and more effective solutions for reach-avoid problems.


Graph Contrastive Learning with Stable and Scalable Spectral Encoding

Deyu Bo, Yuan Fang, Yang Liu, Chuan Shi

https://openreview.net/forum?id=0kz5RmHxmE

Keywords: Graph Contrastive Learning, Spectral Embedding

Compressor summary: EigenMLP is a spectral encoder for graph contrastive learning that is stable, scalable, and invariant to transformations, while Sp$^{2}$GCL fuses spatial and spectral views for better representation learning.


Model Sparsity Can Simplify Machine Unlearning

Jinghan Jia, Jiancheng Liu, Parikshit Ram, Yuguang Yao, Gaowen Liu, Yang Liu, Pranay Sharma, Sijia Liu

https://openreview.net/forum?id=0jZH883i34

Keywords: Machine unlearning, model pruning

Compressor summary: The text introduces model sparsification via weight pruning as a novel approach to improve machine unlearning efficiency and performance in various scenarios, such as defending against backdoor attacks and enhancing transfer learning.


Described Object Detection: Liberating Object Detection with Flexible Expressions

Chi Xie, Zhao Zhang, Yixuan Wu, Feng Zhu, Rui Zhao, Shuang Liang

https://openreview.net/forum?id=0hwq2vOHT4

Keywords: open-vocabulary object detection, referring expression comprehension, multi-modal detection

Compressor summary: The paper introduces a new task called Described Object Detection (DOD) that extends Open-Vocabulary object Detection (OVD) and Referring Expression Comprehension (REC) by using more flexible language expressions, and presents a new dataset ($D^3$) to evaluate it.


Rethinking Semi-Supervised Imbalanced Node Classification from Bias-Variance Decomposition

Liang Yan, Gengchen Wei, Chen Yang, Shengzhong Zhang, Zengfeng Huang

https://openreview.net/forum?id=0gvtoxhvMY

Keywords: Imbalanced Node Classification, Bias-Variance Decomposition, Graph Neural Networks

Compressor summary: The paper proposes a method to handle class imbalance in graph neural networks using data augmentation, variance decomposition, and regularization, and shows its effectiveness on several benchmark datasets.


Modeling Dynamics over Meshes with Gauge Equivariant Nonlinear Message Passing

Jung Yeon Park, Lawson L.S. Wong, Robin Walters

https://openreview.net/forum?id=0eXniewIvr

Keywords: message passing, dynamics, mesh, symmetry, equivariance

Compressor summary: The paragraph discusses a new gauge equivariant architecture using nonlinear message passing for modeling surface PDEs on meshes that leverages surface geometry and outperforms existing methods for complex and nonlinear dynamics.


A Finite-Particle Convergence Rate for Stein Variational Gradient Descent

Jiaxin Shi, Lester Mackey

https://openreview.net/forum?id=0eRDQQK2TW

Keywords: Stein Variational Gradient Descent, SVGD, variational inference, sampling, optimization, Stein's method

Compressor summary: The paper studies SVGD's convergence rate and shows it has an order of ${1/}{\sqrt{\log\log n}}$ when the target distribution is sub-Gaussian with a Lipschitz score.


SAMoSSA: Multivariate Singular Spectrum Analysis with Stochastic Autoregressive Noise

Abdullah Omar Alomar, Munther A. Dahleh, Sean Mann, Devavrat Shah

https://openreview.net/forum?id=0e4eiXoUn5

Keywords: Time series, System Identification, Singular Spectrum Analysis

Compressor summary: The paper proposes a two-stage algorithm, SAMoSSA, that combines multivariate Singular Spectrum Analysis and Autoregressive models to estimate non-stationary trends and stationary noise in time series data, and provides theoretical guarantees for its forecasting performance.


Score-based Generative Models with Lévy Processes

Eunbi Yoon, Keehun Park, Sungwoong Kim, Sungbin Lim

https://openreview.net/forum?id=0Wp3VHX0Gm

Keywords: Generative Model, Score-based Method, Lévy processes

Compressor summary: The Lévy-Itō Model (LIM) is a novel score-based generative model that uses heavy-tailed Lévy processes to overcome the limitations of Brownian motion, achieving faster and more diverse sampling with high fidelity on various image datasets.


No-Regret Online Reinforcement Learning with Adversarial Losses and Transitions

Tiancheng Jin, Junyan Liu, Chloé Rouyer, William Chang, Chen-Yu Wei, Haipeng Luo

https://openreview.net/forum?id=0WLMVDdvDF

Keywords: reinforcement Learning, best of both worlds, MDP, robust RL, adversarial corruption

Compressor summary: The paper presents algorithms that can learn in adversarial Markov Decision Processes with smoothly increasing regret depending on the degree of maliciousness of the adversary, while handling adversarial transitions and losses.


Improved Frequency Estimation Algorithms with and without Predictions

Anders Aamand, Justin Y. Chen, Huy Nguyen, Sandeep Silwal, Ali Vakilian

https://openreview.net/forum?id=0VcvYQ3uPh

Keywords: learning-augmented algorithms, algorithms with predictions, data-driven algorithms, sublinear, streaming, frequency estimation, sketching

Compressor summary: The paper proposes a novel frequency estimation algorithm that outperforms existing methods without predictions and further improves with heavy-hitter predictions.


A Fast and Accurate Estimator for Large Scale Linear Model via Data Averaging

Rui Wang, Yanyan Ouyang, Panpan Yu, Wangli Xu

https://openreview.net/forum?id=0Tq1RGJBid

Keywords: Big data, Data averaging, Order statistic, Sampling method, Sketching method.

Compressor summary: This paper studies how to estimate linear models with large data sets and varying dimensions, using sketching techniques and a new averaging-based method that improves upon existing methods in terms of speed and accuracy.


Efficient Diffusion Policies For Offline Reinforcement Learning

Bingyi Kang, Xiao Ma, Chao Du, Tianyu Pang, Shuicheng YAN

https://openreview.net/forum?id=0P6uJtndWu

Keywords: Offline Reinforcement Learning, Diffusion Models

Compressor summary: Efficient diffusion policy (EDP) improves offline reinforcement learning by addressing the limitations of diffusion models and enabling compatibility with maximum likelihood-based RL algorithms, achieving state-of-the-art results on D4RL benchmark tasks.


Pruning vs Quantization: Which is Better?

Andrey Kuzmin, Markus Nagel, Mart Van Baalen, Arash Behboodi, Tijmen Blankevoort

https://openreview.net/forum?id=0OU1ZXXxs5

Keywords: Neural network quantization, neural network pruning, magnitude pruning, post-training quantization, quantization-aware training

Compressor summary: This paper compares neural network quantization and pruning techniques for compressing deep neural networks and shows that quantization usually performs better than pruning, except in rare cases where extreme compression is needed.


Improved Communication Efficiency in Federated Natural Policy Gradient via ADMM-based Gradient Updates

Guangchen Lan, Han Wang, James Anderson, Christopher Brinton, Vaneet Aggarwal

https://openreview.net/forum?id=0ORqsMY6OL

Keywords: reinforcement learning, federated learning

Compressor summary: FedNPG-ADMM is a framework that reduces communication complexity in federated reinforcement learning by using the alternating direction method of multipliers (ADMM) to approximate global policy gradient directions efficiently.


SaVeNet: A Scalable Vector Network for Enhanced Molecular Representation Learning

Sarp Aykent, Tian Xia

https://openreview.net/forum?id=0OImBCFsdf

Keywords: geometric deep learning, molecule property prediction, geometric representation learning

Compressor summary: The SaVeNet framework is a more efficient and effective way to learn geometric features in molecules, overcoming challenges like computational inefficiency and limited generalizability.


Gaussian Mixture Solvers for Diffusion Models

Hanzhong Allan Guo, Cheng Lu, Fan Bao, Tianyu Pang, Shuicheng YAN, Chao Du, Chongxuan Li

https://openreview.net/forum?id=0NuseeBuB4

Keywords: Diffusion models, SDE-based solver, Gaussian mixture, Stroke-based synthesis

Compressor summary: The paragraph introduces a new method called Gaussian Mixture Solvers (GMS) for improving sampling efficiency and quality in diffusion models, especially for image generation and stroke-based synthesis tasks.


ScaleLong: Towards More Stable Training of Diffusion Model via Scaling Network Long Skip Connection

Zhongzhan Huang, Pan Zhou, Shuicheng YAN, Liang Lin

https://openreview.net/forum?id=0N73P8pH2l

Keywords: Diffusion Model, Stable Training, Network architectures

Compressor summary: The authors analyze the instability of UNet in diffusion models due to its long skip connects and propose a coefficient scaling framework that improves stability and robustness, leading to faster training.


Universal Prompt Tuning for Graph Neural Networks

Taoran Fang, Yunchao Mercer Zhang, Yang Yang, Chunping Wang, Lei CHEN

https://openreview.net/forum?id=0LmWBhIYLi

Keywords: graph neural networks, prompt tuning

Compressor summary: The paper introduces Graph Prompt Feature (GPF), a universal prompt-based tuning method for graph neural networks that works under any pre-training strategy and outperforms fine-tuning and specialized methods in various scenarios.


Dynamic Non-monotone Submodular Maximization

Kiarash Banihashem, Leyla Biabani, Samira Goudarzi, MohammadTaghi Hajiaghayi, Peyman Jabbarzade, Morteza Monemizadeh

https://openreview.net/forum?id=0K1ZTfHZ0N

Keywords: Non-monotone submodular maximization, dynamic algorithm, oracle query, video summarization

Compressor summary: The paper presents the first dynamic algorithms to solve non-monotone submodular maximization under cardinality constraints and demonstrates their effectiveness on real-world data sets.


Multi-Agent Meta-Reinforcement Learning: Sharper Convergence Rates with Task Similarity

Weichao Mao, Haoran Qiu, Chen Wang, Hubertus Franke, Zbigniew Kalbarczyk, Ravi Iyer, Tamer Basar

https://openreview.net/forum?id=0Iw2dLh8uq

Keywords: Reinforcement learning, game theory, multi-agent systems, meta-learning

Compressor summary: This paper explores how meta-learning can improve multi-agent reinforcement learning in various game settings and provides theoretical and empirical evidence for its benefits.


Efficient Potential-based Exploration in Reinforcement Learning using Inverse Dynamic Bisimulation Metric

YIMING WANG, Ming Yang, Renzhi Dong, Binbin Sun, Furui Liu, Leong Hou U

https://openreview.net/forum?id=0FhKURbTyF

Keywords: Reinforcement learning, reward shaping, potential-based exploration, inverse dynamic bisimulation metric

Compressor summary: LIBERTY is a potential-based exploration bonus for deep RL that uses bisimulation metric to measure state discrepancy, enhancing agent's discovery of novel states and improving training efficiency.


AR-Diffusion: Auto-Regressive Diffusion Model for Text Generation

Tong Wu, Zhihao Fan, Xiao Liu, Hai-Tao Zheng, Yeyun Gong, yelong shen, Jian Jiao, Juntao Li, zhongyu wei, Jian Guo, Nan Duan, Weizhu Chen

https://openreview.net/forum?id=0EG6qUQ4xE

Keywords: text generation, diffusion model, auto-regression, sequential dependency

Compressor summary: AR-Diffusion is a new diffusion model for text generation that uses dynamic denoising steps to account for the sequential dependency of natural language and achieves faster and better results than existing models on various tasks.


Adaptive Uncertainty Estimation via High-Dimensional Testing on Latent Representations

Tsai Hor Chan, Kin Wai Lau, Jiajun Shen, Guosheng Yin, Lequan Yu

https://openreview.net/forum?id=0DpKUzl1Se

Keywords: Bayesian deep learning, high-dimensional testing, uncertainty estimation, out-of-distribution detection

Compressor summary: The proposed framework uses data-adaptive high-dimensional hypothesis testing for uncertainty estimation in deep neural networks, without retraining the feature encoder, and improves performance on OOD detection and task-specific prediction.


Partial Label Learning with Dissimilarity Propagation guided Candidate Label Shrinkage

Yuheng Jia, Fuchao Yang, Yongqiang Dong

https://openreview.net/forum?id=0CbmvZPBGB

Keywords: partial label learning, dissimilarity propagation, candidate label shrinkage

Compressor summary: The paper proposes a method to disambiguate candidate labels in partial label learning using adversarial similarity and dissimilarity matrices, which improves performance and has theoretical guarantees.


Gaussian Process Probes (GPP) for Uncertainty-Aware Probing

Zi Wang, Alexander Ku, Jason Michael Baldridge, Thomas L. Griffiths, Been Kim

https://openreview.net/forum?id=0BwB03qA5T

Keywords: Interpretability, probing, Bayesian, Gaussian process, transparency

Compressor summary: Gaussian process probes (GPP) is a framework that allows measuring uncertainty about concepts represented by models using a distribution over classifiers without needing access to training data or model details, which can be applied to any pre-trained model and helps understand and evaluate their capabilities.


Stochastic Multi-armed Bandits: Optimal Trade-off among Optimality, Consistency, and Tail Risk

David Simchi-Levi, Zeyu Zheng, Feng Zhu

https://openreview.net/forum?id=0BfQT652sC

Keywords: multi-armed bandit, worst-case optimality, instance-dependent consistency, light-tailed risk

Compressor summary: The paper studies how to design policies for the stochastic multi-armed bandit problem with three desired properties: worst-case optimality, instance-dependent consistency, and light-tailed risk, and proposes a novel policy that achieves optimal trade-offs among them.


Task Arithmetic in the Tangent Space: Improved Editing of Pre-Trained Models

Guillermo Ortiz-Jimenez, Alessandro Favero, Pascal Frossard

https://openreview.net/forum?id=0A9f2jZDGW

Keywords: model editing, transfer learning, neural tangent kernel, vision-language pre-training, deep learning science

Compressor summary: Task arithmetic improves vision-language model performance by leveraging weight disentanglement and linearizing the model's tangent space.


Online Ad Procurement in Non-stationary Autobidding Worlds

Jason Cheuk Nam Liang, Haihao Lu, Baoyu Zhou

https://openreview.net/forum?id=09bZyE9tfp

Keywords: autobidding, online advertising, bandit online convex optimization, constrained optimization

Compressor summary: The paragraph describes an online learning framework for optimizing ad platform lever decisions in a realistic bandit feedback environment with non-stationary procurement outcomes, and presents a primal-dual algorithm that achieves low regret across various scenarios.


On the Powerfulness of Textual Outlier Exposure for Visual OoD Detection

Sangha Park, Jisoo Mok, Dahuin Jung, Saehyung Lee, Sungroh Yoon

https://openreview.net/forum?id=090ORrOAPL

Keywords: Out-of-distribution detection

Compressor summary: The paper explores using textual outliers instead of visual ones for Out-of-Distribution (OoD) detection in neural networks, introducing new methods to generate them and showing their competitive performance on benchmarks.


Exposing flaws of generative model evaluation metrics and their unfair treatment of diffusion models

George Stein, Jesse C. Cresswell, Rasa Hosseinzadeh, Yi Sui, Brendan Leigh Ross, Valentin Villecroze, Zhaoyan Liu, Anthony L. Caterini, Eric Taylor, Gabriel Loaiza-Ganem

https://openreview.net/forum?id=08zf7kTOoh

Keywords: generative models, generative model evaluation, self-supervised learning, representation learning, metrics

Compressor summary: The authors systematically evaluate generative models using human perception of image realism and 17 modern metrics, finding that existing metrics are flawed and propose improvements and a new dataset and library for future work.


Knowledge Diffusion for Distillation

Tao Huang, Yuan Zhang, Mingkai Zheng, Shan You, Fei Wang, Chen Qian, Chang Xu

https://openreview.net/forum?id=08hStXdT1s

Keywords: knowledge distillation, diffusion models

Compressor summary: The paper proposes DiffKD, a novel knowledge distillation method that uses diffusion models to denoise and match features between teacher and student models, achieving state-of-the-art performance on various tasks.


Exploiting hidden structures in non-convex games for convergence to Nash equilibrium

Iosif Sakos, Emmanouil-Vasileios Vlatakis-Gkaragkounis, Panayotis Mertikopoulos, Georgios Piliouras

https://openreview.net/forum?id=05P1U0jk8r

Keywords: Nash Equilibrium, Games, Gradient, Non-monotone VI, Natural Gradient, Precondition

Compressor summary: The paper introduces PHGD, a first-order method that exploits latent convex structures in non-cooperative games to achieve convergence under minimal assumptions.


Robustness Guarantees for Adversarially Trained Neural Networks

Poorya Mianjy, Raman Arora

https://openreview.net/forum?id=02Uc0G2Cym

Keywords: Adversarial training, neural networks, robustness, guarantees

Compressor summary: The paper proposes a method to train robust neural networks against adversarial attacks by maximizing a lower bound on the loss function using projected gradient descent and provides convergence guarantees and empirical evidence.


Complexity Matters: Rethinking the Latent Space for Generative Modeling

Tianyang Hu, Fei Chen, Haonan Wang, Jiawei Li, Wenjia Wang, Jiacheng Sun, Zhenguo Li

https://openreview.net/forum?id=00EKYYu3fD

Keywords: generative model, latent space, distance between distributions, generative adversarial network, vqgan

Compressor summary: The study proposes a novel method to find the optimal latent space for generative models by considering model complexity and improves the performance of various models with less complex ones.


Finite Population Regression Adjustment and Non-asymptotic Guarantees for Treatment Effect Estimation

Mehrdad Ghadiri, David Arbour, Tung Mai, Cameron N Musco, Anup Rao

https://openreview.net/forum?id=009LK0vLcY

Keywords: regression adjustment; treatment effect estimation; average treatment effect

Compressor summary: The paper explores regression adjustment in finite populations, using randomized numerical linear algebra to select subsets of subjects for experiments and providing non-asymptotic accuracy bounds for estimating sample mean, individual treatment effects, and average treatment effect.


Collaborative Alignment of NLP Models

Fereshte Khani, Marco Tulio Ribeiro

https://openreview.net/forum?id=f39Q3JyoIi

Keywords: alignment, collaborative alignment, debugging, nlp, interference, multi-user interaction

Compressor summary: CoAligment is a framework that helps multiple users align NLP models with their values by learning local and global models and generating instances to resolve disagreements.


Wide Neural Networks as Gaussian Processes: Lessons from Deep Equilibrium Models

Tianxiang Gao, Xiaokai Huo, Hailiang Liu, Hongyang Gao

https://openreview.net/forum?id=Z2he2Y0MoH

Keywords: Gradient descent, deep equilibrium model, Gaussian processes, kernel methods, NNGP, NTK

Compressor summary: The paper analyzes deep equilibrium models (DEQs), infinite-depth neural networks that converge to Gaussian processes, and shows they have non-degenerate kernels, enabling training and generalization.


Auditing for Human Expertise

Rohan Alur, Loren Laine, Darrick K Li, Manish Raghavan, Devavrat Shah, Dennis Shung

https://openreview.net/forum?id=VEpU9rFaQr

Keywords: hypothesis testing, human-AI complementarity, machine learning for healthcare

Compressor summary: The authors propose a statistical test to determine if human experts add value in high-stakes prediction tasks beyond what an algorithm can capture, using the example of admissions decisions for patients with acute gastrointestinal bleeding.


Sub-optimality of the Naive Mean Field approximation for proportional high-dimensional Linear Regression

Jiaze Qiu

https://openreview.net/forum?id=OXhymu6MeN

Keywords: Variational Bayes; Naive Mean Field; Gaussian comparison inequalities; High-dimensional statistics; Proportional asymptotic.

Compressor summary: The paper derives asymptotic characterizations for the NMF approximation in high-dimensional linear regression and shows its limitations in practice.


Diffusion-Based Adversarial Sample Generation for Improved Stealthiness and Controllability

Haotian Xue, Alexandre Araujo, Bin Hu, Yongxin Chen

https://openreview.net/forum?id=JTwxylP6U9

Keywords: Robustness, Adversarial Samples, Diffusion Model

Compressor summary: The paper proposes a new method called Diff-PGD for creating realistic adversarial samples that stay close to natural images while being effective against neural networks.