This page contains summaries of all Neurips 2023 accepted papers generated by the compressor, my personal LLM-based project.
Hao Sun, Boris van Breugel, Jonathan Crabbé, Nabeel Seedat, Mihaela van der Schaar
https://openreview.net/forum?id=zyhxRc9bew
Keywords: Uncertainty Explaination, Uncertainty Quantification, Interpretability
Compressor summary: This paper introduces a framework to classify uncertain examples from machine learning models using the confusion density matrix, which helps compare different uncertainty quantification methods.
Aahlad Manas Puli, Lily H Zhang, Yoav Wald, Rajesh Ranganath
https://openreview.net/forum?id=zyZkaqNnpa
Keywords: shortcut learning, spurious correlations, perfect stable feature, perception tasks, implicit bias in optimization, improving inductive biases
Compressor summary: The paragraph discusses how default-ERM models tend to use shortcuts in perception tasks due to their preference for maximizing margins, and proposes alternative loss functions called margin control (MARG-CTRL) that encourage uniform-margin solutions and mitigate shortcut learning.
Xilie Xu, Jingfeng Zhang, Feng Liu, Masashi Sugiyama, Mohan Kankanhalli
https://openreview.net/forum?id=zuXyQsXVLF
Keywords: robust pre-training, adversarial contrastive learning
Compressor summary: Adversarial contrastive learning (ACL) enhances standard contrastive learning by using adversarial data and introduces adversarial invariant regularization (AIR) to improve robustness against style factors, as shown by causal reasoning and empirical results.
Michael Scholkemper, Michael T Schaub
https://openreview.net/forum?id=ztDxO15N7f
Keywords: Role Extraction, Graph Learning, Node Embeddings, Weisfeiler Lehman, Equitable Partition
Compressor summary: The paragraph discusses a new method for identifying structural roles of nodes in complex networks, based on graph-isomorphism tests and the Weisfeiler-Leman algorithm, and presents a benchmark to test the approach.
Yuanyuan Wang, Xi Geng, Wei Huang, Biwei Huang, Mingming Gong
https://openreview.net/forum?id=zsOOqjaj2z
Keywords: Linear SDE, Identification, Causal inference
Compressor summary: The paper proposes identifiability conditions for determining the generator of linear SDEs from its solution process distribution, which is useful for causal inference, and provides sufficient and necessary conditions for additive noise and sufficient ones for multiplicative noise, along with geometric interpretations and simulations.
Jifan Zhang, Shuai Shao, saurabh verma, Robert D Nowak
https://openreview.net/forum?id=zrUEHZ6s9C
Keywords: Deep Learning, Active Learning
Compressor summary: TAILOR is a meta algorithm that adaptively selects among various active learning strategies to efficiently reduce labeling efforts for deep learning applications using class-balanced rewards.
Yu Wang, Zhun Zhong, Pengchong Qiao, Xuxin Cheng, Xiawu Zheng, Chang Liu, Nicu Sebe, Rongrong Ji, Jie Chen
https://openreview.net/forum?id=zrLxHYvIFL
Keywords: open-world semi-supervised learning; novel class discovery;
Compressor summary: The paper proposes TIDA, a framework for open-world semi-supervised learning that leverages multi-granularity taxonomic context priors to enhance representation learning and improve pseudo label quality.
Donglin Xia, Xiao Wang, Nian Liu, Chuan Shi
https://openreview.net/forum?id=zrCmeqV3Sz
Keywords: Graph neural networks, network representation learning, deep learning
Compressor summary: The paper proposes CIT, a mechanism that improves GNNs' generalization ability by transferring cluster information and preserving node diversity when the test graph structure differs from the training graph structure.
Filippo Maria Bianchi, Veronica Lachi
https://openreview.net/forum?id=zqyVjCjhYD
Keywords: Graph Neural Networks, Graph pooling, Expressive power
Compressor summary: This paper studies how graph pooling affects the expressiveness of Graph Neural Networks (GNNs) and provides a criterion for choosing or designing pooling operators based on their ability to preserve the GNN's message-passing power.
Shaokui Wei, Mingda Zhang, Hongyuan Zha, Baoyuan Wu
https://openreview.net/forum?id=zqOcW3R9rd
Keywords: Backdoor Attack, Trustworthy AI, Backdoor Learning
Compressor summary: The paper proposes a method called Shared Adversarial Unlearning (SAU) to purify backdoored machine learning models using a small clean dataset and adversarial training techniques, achieving state-of-the-art performance.
Yan Dai, Kwangjun Ahn, Suvrit Sra
https://openreview.net/forum?id=zq4vFneRiA
Keywords: Sharpness-Aware Minimization, Normalization, Deep Learning Theory
Compressor summary: The paragraph discusses a paper that explores how normalization, a key component of the SAM optimizer, improves deep neural network performance by stabilizing the algorithm and enabling it to drift along a manifold of minima.
Jiaqi Wang, Xingyi Yang, Suhan Cui, Liwei Che, Lingjuan Lyu, Dongkuan Xu, Fenglong Ma
https://openreview.net/forum?id=zpVCITHknd
Keywords: Federated Learning
Compressor summary: The paper introduces pFedHR, a framework for personalized federated learning that handles model heterogeneity by reassembling diverse models using server-side optimization and minimizing the impact of different data distributions.
Sehyun Hwang, Sohyun Lee, Hoyoung Kim, Minhyeon Oh, Jungseul Ok, Suha Kwak
https://openreview.net/forum?id=znudaK78u8
Keywords: semantic segmentation; active learning; partial label learning
Compressor summary: The paper presents a new active learning method for semantic segmentation that uses multi-class labeling of local image regions and disambiguates partial labels with two loss functions and pseudo labels, achieving better performance and annotation efficiency.
Jaeyeon Kim, Asuman E. Ozdaglar, Chanwoo Park, Ernest K. Ryu
https://openreview.net/forum?id=znY173SCxu
Keywords: Convex Optimization, Acceleration, First-Order methods
Compressor summary: The paper introduces H-duality, a novel concept in convex optimization that reveals symmetries between different first-order optimization methods and leads to new efficient techniques for minimizing function values and gradient magnitudes.
Shiqiang Zhang, Juan S Campos, Christian Wolfgang Feldmann, David Walz, Frederik Sandfort, Miriam Mathea, Calvin Tsay, Ruth Misener
https://openreview.net/forum?id=znW5jNIOED
Keywords: Mixed-integer optimization, Graph neural network, Symmetry-breaking, Molecular design
Compressor summary: The paper proposes methods to optimize machine learning models constrained by trained graph neural networks (GNNs) using symmetry-breaking constraints and applies them to molecular design.
Quan Xiao, Songtao Lu, Tianyi Chen
https://openreview.net/forum?id=zn5ihqknGj
Keywords: Bilevel optimization, nonconvex constrained optimization, convergence analysis
Compressor summary: This paper introduces a new method (GALET) for solving bilevel optimization problems with nonconvex lower-level objectives and shows that it has the same convergence rate as gradient descent.
Xiaobin Rui, Zhixiao Wang, Jiayu Zhao, Lichao Sun, Wei Chen
https://openreview.net/forum?id=zmWNe1V6jg
Keywords: influence maximization, approximation algorithm, social fairness
Compressor summary: The paper proposes an efficient algorithm for maximizing welfare fairness in community structures using a weighted maximum coverage problem approach.
Saber Sheybani, Himanshu Hansaria, Justin Newell Wood, Linda B. Smith, Zoran Tiganj
https://openreview.net/forum?id=zkfyOkBVpz
Keywords: Curriculum learning, Self-supervised learning, Slow changes, Infant development
Compressor summary: Infant visual experiences affect the development of their visual system, and starting with data from younger infants improves learning outcomes for AI models trained on their visual inputs.
Leonidas Tsepenekas, Ivan Brugere, Freddy Lecue, Daniele Magazzeni
https://openreview.net/forum?id=zjpjsJeVJZ
Keywords: individual fairness; similarity learning; active learning
Compressor summary: The paper proposes an efficient sampling framework to learn similarity functions between different groups using limited expert feedback, and provides theoretical and empirical evidence for its effectiveness.
Çağlar Hızlı, S. T. John, Anne Tuulikki Juuti, Tuure Tapani Saarinen, Kirsi Hannele Pietiläinen, Pekka Marttinen
https://openreview.net/forum?id=zfHCKDzzC8
Keywords: Machine learning for healthcare, Causal mediation, Gaussian process, Point Process
Compressor summary: The authors propose a new method to analyze the causal effects of an intervention on outcomes and mediators in dynamic systems using temporal point processes.
Ruoyu Li, Qing Li, Yu Zhang, Dan Zhao, Yong Jiang, Yong Yang
https://openreview.net/forum?id=zfCNwRQ569
Keywords: unsupervised anomaly detection, global explanation, rule extraction
Compressor summary: The paper proposes a post-hoc method to explain black-box unsupervised anomaly detection models using distribution decomposition rules and boundary inference rules, making them more interpretable and trustworthy.
Thomas Steinke, Alexander Knop
https://openreview.net/forum?id=zdli6OxpWd
Keywords: differential privacy, user-level privacy, person-level privacy, sensitivity
Compressor summary: The authors propose a method to count distinct elements in a dataset with person-level differential privacy, by approximating the problem with a max-flow solution and optimizing the sensitivity bound.
Qinghua Liu, Gellért Weisz, András György, Chi Jin, Csaba Szepesvari
https://openreview.net/forum?id=zaQ7wV9NOg
Keywords: Theory of reinforcement learning, policy optimization
Compressor summary: This paper introduces Optimistic NPG, a simple and efficient policy optimization framework for online RL with linear MDPs, which improves over existing algorithms in terms of computation and exploration.
Licong Lin, Mufang Ying, Suvrojit Ghosh, Koulik Khamaru, Cun-Hui Zhang
https://openreview.net/forum?id=zXckveawHa
Keywords: Adaptive linear regression, bandit algorithms, high dimensional statistics, statistical inference
Compressor summary: The paper explores how adaptive data collection affects estimation and inference in high-dimensional linear models and proposes a new estimator for single coordinate inference with better performance than OLS.
Yihan Wang, Jatin Chauhan, Wei Wang, Cho-Jui Hsieh
https://openreview.net/forum?id=zWxKYyW9ik
Keywords: prompt-tuning; language model; expressive power
Compressor summary: The paper analyzes the theoretical and practical aspects of prompt tuning for transformer-based models, showing its universality for approximating sequence-to-sequence functions and its limitations for limited-depth models and non-invertible datasets.
Nils Sturma, Chandler Squires, Mathias Drton, Caroline Uhler
https://openreview.net/forum?id=zW1uVN6Mbv
Keywords: linear structural equation models, causality, representation learning, independent component analysis, structure identifiability, multiple views, graphical model
Compressor summary: The paper proposes a method to identify a shared causal representation of unpaired data from multiple domains using linear models.
Yunho Jin, Chun-Feng Wu, David Brooks, Gu-Yeon Wei
https://openreview.net/forum?id=zUYfbdNl1m
Keywords: Throughput, GPU utilization, Sequence length prediction
Compressor summary: The paper proposes $S^3$, a system that predicts the output sequence length of large language models to increase memory and GPU utilization, and handle mispredictions.
Mingyuan Zhou, Tianqi Chen, Zhendong Wang, Huangjie Zheng
https://openreview.net/forum?id=zTSlm4nmlH
Keywords: Diffusion models, KL-divergence upper bounds, multiplicative transitions, scaled and shifted beta distributions
Compressor summary: Beta diffusion is a new method for generating data within limited ranges using multiplicative transitions based on beta distributions and optimized with KL-divergence upper bounds, which outperform negative ELBOs in experiments with synthetic and natural images.
Xudong XU, Dejan Markovic, Jacob Sandakly, Todd Keebler, Steven Krenn, Alexander Richard
https://openreview.net/forum?id=zQTi3pziFp
Keywords: sound field, spatial audio, virtual humans, human body, body modeling
Compressor summary: The authors present a system that generates realistic 3D spatial audio for human bodies using input from microphones and body pose, and introduce a new multimodal dataset for this task.
Qianyi Chen, Bo Li, LU DENG, Yong Wang
https://openreview.net/forum?id=zQOYGDc9pu
Keywords: AB test, interference, causal inference, optimization, social network
Compressor summary: The paper proposes a new method for designing randomized network experiments that balances bias and variance by optimizing the covariance matrix of the treatment assignment vector using projected gradient descent.
Jongheon Jeong, Jinwoo Shin
https://openreview.net/forum?id=zQ4yraDiRe
Keywords: adversarial robustness, certified robustness, randomized smoothing, denoised smoothing, diffusion models
Compressor summary: The paper presents a method called multi-scale smoothing that improves the trade-off between robustness and accuracy in denoised smoothing by selectively applying randomized smoothing among multiple noise scales, and proposes diffusion fine-tuning to enhance the performance of diffusion models.
Anastasios Nikolas Angelopoulos, Emmanuel Candes, Ryan Tibshirani
https://openreview.net/forum?id=zPYeYv6YYs
Keywords: conformal prediction, time series, uncertainty quantification, distribution shift
Compressor summary: The paper presents uncertainty quantification algorithms for time series prediction that use conformal prediction and control theory, improve coverage on COVID-19 forecasts, and provide an extendable codebase.
Jianing Li, Vardan Papyan
https://openreview.net/forum?id=zOCIKYVaF5
Keywords: Deep Learning, Residual Networks, Neural Networks, Generalization, Spectral Analysis
Compressor summary: This paper investigates Residual Alignment in ResNet architectures and shows how it aligns intermediate representations linearly across layers and affects generalization.
Alistair White, Niki Kilbertus, Maximilian Gelbrecht, Niklas Boers
https://openreview.net/forum?id=zO2dAQfvHf
Keywords: neural differential equations, neural ordinary differential equations, constraints, conservation laws, stabilization, dynamical systems, dynamics, scientific machine learning, physics-informed machine learning
Compressor summary: SNDEs are a method to enforce constraints on neural differential equations by adding a stabilization term, which makes them compatible with various models and improves performance.
Sungwon Kim, Kevin J. Shih, Rohan Badlani, Joao Felipe Santos, Evelina Bakhturina, Mikyas T. Desta, Rafael Valle, Sungroh Yoon, Bryan Catanzaro
https://openreview.net/forum?id=zNA7u7wtIN
Keywords: text-to-speech, zero-shot TTS, flow matching generative model
Compressor summary: P-Flow is a fast and data-efficient zero-shot text-to-speech model that uses speech prompts for speaker adaptation and achieves high quality and pronunciation with less training data and faster sampling speed than existing models.
Bo Liu, Yihao Feng, Peter Stone, qiang liu
https://openreview.net/forum?id=zMeemcUeXL
Keywords: multitask learning, multitask optimization, conflicting gradients, knowledge transfer
Compressor summary: The paper introduces FAMO, a method that balances task losses in multitask learning using minimal space and time, outperforming existing techniques.
Manu Srinath Halvagal, Axel Laborieux, Friedemann Zenke
https://openreview.net/forum?id=zMNUNd9zs1
Keywords: Self-supervised learning, Non-contrastive learning, Learning dynamics
Compressor summary: Non-contrastive self-supervised learning (SSL) methods use different losses to stabilize learning and avoid representational collapse, and a new family of loss functions called IsoLoss can further improve their performance.
Cornelius Brand, Robert Ganian, Mathis Rocton
https://openreview.net/forum?id=zIEaOZ0saA
Keywords: neural network training, computational complexity, ReLU networks, Linear networks
Compressor summary: The article presents new algorithms that improve the efficiency of training neural networks with linear or ReLU activation functions.
Tao Shen, Yifan Cui
https://openreview.net/forum?id=zGdH4tKtOW
Keywords: Optimal treatment regimes, Policy-making, Proximal causal inference, Unmeasured confounding, Value function
Compressor summary: The paragraph discusses a new framework for causal inference using proxy variables, and proposes an optimal individualized treatment regime based on outcome and treatment confounding bridges, with theoretical guarantees and applications shown.
Lorenzo Beretta, Vincent Cohen-Addad, Silvio Lattanzi, Nikos Parotsidis
https://openreview.net/forum?id=zGRWp7yRqd
Keywords: Clustering, k-means, approximation algorithms
Compressor summary: The paper proposes an improved $k$-means clustering algorithm by combining $k$-means++ sampling with larger local-search neighborhoods, achieving a 9 + ε approximation ratio and practical efficiency.
Momchil Peychev, Mark Niklas Mueller, Marc Fischer, Martin Vechev
https://openreview.net/forum?id=zEoP4vzFKy
Keywords: ImageNet, evaluation, error classification, error analysis
Compressor summary: The paragraph discusses how automated error classification can help evaluate computer vision models and finds that top-1 accuracy is still useful for measuring model performance despite its limitations.
Christopher A. Choquette-Choo, Arun Ganesh, Ryan McKenna, Hugh Brendan McMahan, J Keith Rush, Abhradeep Guha Thakurta, Zheng Xu
https://openreview.net/forum?id=zEm6hF97Pz
Keywords: Machine Learning, Differential Privacy, Optimization, Private Machine Learning, Federated Learning, Privacy Amplification, Matrix Factorization
Compressor summary: The paper presents a method to improve privacy-utility tradeoffs for machine learning using matrix factorization with banded matrices, which can be applied in both centralized and federated settings.
Lin Guan, Karthik Valmeekam, Sarath Sreedharan, Subbarao Kambhampati
https://openreview.net/forum?id=zDbsSscmuj
Keywords: LLMs, Planning, Domain Model, LLMs for Planning, LLMs for Heuristic Guidance
Compressor summary: The authors propose a method that uses large language models to construct PDDL domain models and then plan with sound planners, reducing human involvement and improving planning efficiency.
Carlo Alfano, Rui Yuan, Patrick Rebeschini
https://openreview.net/forum?id=zD6lXmTPPh
Keywords: Theory for Reinforcement Learning, Policy Optimization, Policy Gradient, Mirror Descent.
Compressor summary: The paper introduces a policy optimization framework based on mirror descent that allows general parameterizations and shows linear convergence, sample complexity improvements, and experimental validation.
Isaac Reid, Adrian Weller, Krzysztof Marcin Choromanski
https://openreview.net/forum?id=zCFfv49MjE
Keywords: Graph, discrete mathematics, quasi-Monte Carlo, kernel, scalability, Laplacian, clustering, random walks
Compressor summary: The paper introduces a novel method to improve graph random features by inducing negative correlations through antithetic termination, leading to lower-variance estimators and better performance on various tasks.
Ba-Hien Tran, Giulio Franzese, Pietro Michiardi, Maurizio Filippone
https://openreview.net/forum?id=zAXg8dW8ZO
Keywords: Generative Models, Normalizing Flows, Variational Autoencoders
Compressor summary: The paper introduces a simple technique to improve the quality of likelihood-based generative models by using data mollification, which enhances density estimation and low-density region handling without additional computational costs.
Siyan Zhao, Aditya Grover
https://openreview.net/forum?id=zAQK5r1enm
Keywords: reinforcement learning, generative models, offline RL, sequential decision making
Compressor summary: Decision Stacks is a framework that splits goal-conditioned policy agents into three independent modules for efficient learning and flexible generative decision making in sequential decision making problems.
Shashank Subramanian, Peter Harrington, Kurt Keutzer, Wahid Bhimji, Dmitriy Morozov, Michael W. Mahoney, Amir Gholami
https://openreview.net/forum?id=zANxvzflMl
Keywords: scientific machine learning, scaling, transfer learning, neural operators, foundation models
Compressor summary: The study shows that pre-training machine learning models can significantly improve performance in scientific applications by adapting them to different tasks with fewer examples and across various physics problems.
Sihan Xu, Ziqiao Ma, Yidong Huang, Honglak Lee, Joyce Chai
https://openreview.net/forum?id=z9d9DsjAPH
Keywords: Image to image translation, latent diffusion models, conditional diffusion models
Compressor summary: Cyclenet is a method that uses cycle consistency to improve image synthesis with diffusion models, enabling consistent unpaired image-to-image translation and robust generation of high-quality images.
Matthew C Bendel, Rizwan Ahmad, Philip Schniter
https://openreview.net/forum?id=z4vKRmq7UO
Keywords: Generative adversarial network, inverse problems, posterior sampling, cGAN, GAN
Compressor summary: The paper proposes a fast method to generate multiple high-quality images from noisy measurements using a conditional Wasserstein GAN with regularization and shows its effectiveness in MRI and inpainting tasks.
Shurui Gui, Meng Liu, Xiner Li, Youzhi Luo, Shuiwang Ji
https://openreview.net/forum?id=z3HACY5CMa
Keywords: deep learning, graph neural network, out-of-distribution generalization, distribution shift
Compressor summary: The paper introduces LECI, a method to improve graph out-of-distribution generalization by using label and environment information and developing an adversarial training strategy.
Samuel McCauley, Benjamin Moseley, Aidin Niaparast, Shikha Singh
https://openreview.net/forum?id=z37ki6nqAY
Keywords: Algorithms with Predictions, Data Structures, Learned Indices, Online List Labeling, Resource Allocation, Beyond Worst Case Analysis
Compressor summary: The paper proposes a new list labeling data structure that leverages predictions to improve performance and provides theoretical guarantees for both worst-case and stochastic error models.
Daniel Augusto de Souza, Alexander V Nikitin, S. T. John, Magnus Ross, Mauricio A Álvarez, Marc Peter Deisenroth, João Paulo Pordeus Gomes, Diego Mesquita, César Lincoln Mattos
https://openreview.net/forum?id=z2BHMLA8pM
Keywords: Gaussian Processes, Deep Gaussian Processes, non-stationary kernels
Compressor summary: TDGP combines the benefits of deep GPs and shallow GPs by learning interpretable low-dimensional embeddings while maintaining the flexibility to adjust kernel hyperparameters, making it better suited for uncertainty quantification than previous models.
Zixuan Jiang, Jiaqi Gu, Hanqing Zhu, David Z. Pan
https://openreview.net/forum?id=z06npyCwDq
Keywords: Transformer, Normalization, Layer Normalization, RMSNorm, Efficient Machine Learning
Compressor summary: The paragraph discusses different normalization techniques for Transformers and proposes a solution to unify and improve their efficiency without compromising performance.
Rainer Engelken
https://openreview.net/forum?id=yzZbwQPkmP
Keywords: spiking networks, event-based simulation, sparse networks, backpropagation, algorithm, neuroscience
Compressor summary: SparseProp is a novel event-based algorithm that simulates and trains sparse SNNs with reduced computational cost, enabling efficient and exact simulations of large spiking networks.
Fnu Suya, Xiao Zhang, Yuan Tian, David Evans
https://openreview.net/forum?id=yyLFUPNEiT
Keywords: poisoning attacks; adversarial machine learning; machine learning security
Compressor summary: The study explores how linear learners can resist indiscriminate poisoning attacks on some datasets with well-separated and low-variance data distributions.
Xinli Yue, Ningping Mou, Qian Wang, Lingchen Zhao
https://openreview.net/forum?id=ywrPcBEXdC
Keywords: Deep Learning, Knowledge Distillation, Adversarial Training, Fairness
Compressor summary: This paper introduces Fair-ARD, a novel framework to enhance robust fairness in student models by adjusting the weights of difficult classes during adversarial robustness distillation, and shows its effectiveness in improving both overall and class-wise robustness.
Brian Hu Zhang, Gabriele Farina, Ioannis Anagnostides, Federico Cacciamani, Stephen Marcus McAleer, Andreas Alexander Haupt, Andrea Celli, Nicola Gatti, Vincent Conitzer, Tuomas Sandholm
https://openreview.net/forum?id=yw1v4RqvPk
Keywords: extensive-form games, deep reinforcement learning, mechanism design, correlated equilibria
Compressor summary: A new method for finding optimal equilibria in extensive-form games is presented, which applies existing zero-sum game learning techniques and achieves state-of-the-art results in tabular games and auction design.
Alexander Meulemans, Simon Schug, Seijin Kobayashi, Nathaniel Daw, Greg Wayne
https://openreview.net/forum?id=yvqqkOn9Pi
Keywords: Reinforcement learning, Long-term credit assignment, contribution analysis, hindsight credit assignment, policy gradient methods
Compressor summary: COCOA is a new model-based credit assignment algorithm that measures the contribution of actions to future rewards by answering counterfactual questions, improving sample efficiency in reinforcement learning.
Kyriakos Flouris, Ender Konukoglu
https://openreview.net/forum?id=yubwSWol6K
Keywords: manifold learning flows, normalizing flows, optimization, orthogonalization, sparsity, sparse learning, generative modeling, Riemannian manifold, geometry, metric tensor, orthogonal basis
Compressor summary: The proposed canonical manifold learning flow method generates a sparse and orthogonal low-dimensional representation of the data that improves the efficiency and accuracy of generative modeling techniques compared to other methods.
Martijn De Vos, Sadegh Farhadkhani, Rachid Guerraoui, Anne-marie Kermarrec, Rafael Pires, Rishi Sharma
https://openreview.net/forum?id=ytrhsvGP0r
Keywords: Epidemic, Decentralized Learning, Randomized Communication, Peer sampling
Compressor summary: Epidemic Learning (EL) is a decentralized learning algorithm that uses changing communication topologies to converge faster than conventional approaches, achieving better performance with less communication volume.
Fabian Jogl, Maximilian Thiessen, Thomas Gärtner
https://openreview.net/forum?id=ytTfonl9Wd
Keywords: Graph Neural Networks, GNNs, Graphs, Message Passing, Expressiveness, Graph Transformations, Message Passing Graph Neural Networks
Compressor summary: The paragraph discusses how graph transformations enable message passing operations for simulating graph neural networks (GNNs) without losing expressivity and improving code optimization, while distinguishing between weak and strong simulation methods.
Nika Haghtalab, Michael Jordan, Eric Zhao
https://openreview.net/forum?id=ysqlhW0v26
Keywords: multicalibration, multi-objective learning, learning theory, calibration, fairness, games
Compressor summary: The authors present a unified framework for designing and analyzing multi-calibrated predictors by using game dynamics and achieving better guarantees than existing methods.
Luming Tang, Menglin Jia, Qianqian Wang, Cheng Perng Phoo, Bharath Hariharan
https://openreview.net/forum?id=ypOiXjdfnU
Keywords: Correspondence, Diffusion Model
Compressor summary: The paper introduces DIFT, a method to establish correspondences between images using diffusion models without explicit supervision, and shows that it can outperform weakly-supervised methods and off-the-shelf features on various tasks.
Xiaohan Wang, Yuehu Liu, Xinhang Song, Beibei Wang, Shuqiang Jiang
https://openreview.net/forum?id=yoZTVn0T50
Keywords: Embodied AI, Interactive Navigation, Causal Reinforcement Learning, Hierarchical Reinforcement Learning
Compressor summary: The paper introduces a causal diagram for visual navigation in complex scenes, proposes a multi-policy model to explore counterfactual interactions and reduce exploration, and presents a large-scale dataset for evaluation.
Xiaoying Xing, Mingfu Liang, Ying Wu
https://openreview.net/forum?id=yoAmURKDJi
Keywords: knowledge-based visual question answering, task-oriented, active image understanding, large language model, visual reasoning, multi-round dialogue
Compressor summary: The authors propose a new method for knowledge-based visual question answering using large language models that actively collects relevant visual evidence to verify their hypotheses, improving performance and interpretability on open-ended datasets.
El Mehdi Saad, Gilles Blanchard, Nicolas Verzelen
https://openreview.net/forum?id=ymHM1qRUeb
Keywords: Multi-armed bandits, Best-arm identification, Adaptive identification
Compressor summary: The paper proposes new algorithms for identifying the best arm in a multi-armed bandit model with dependent arms and covariance, improving on standard methods in applications like clinical trials.
Ziang Liu, Genggeng Zhou, Jeff He, Tobia Marcucci, Li Fei-Fei, Jiajun Wu, Yunzhu Li
https://openreview.net/forum?id=ymBG2xs9Zf
Keywords: model learning, model-based control, neural network sparsification, mixed-integer programming, trajectory optimization
Compressor summary: The paper proposes a new framework for learning and controlling with deep neural networks by sparsifying them and using efficient optimization algorithms, achieving better performance in tasks involving contact dynamics.
Yan Sun, Li Shen, Dacheng Tao
https://openreview.net/forum?id=ylPX5D7It7
Keywords: federated learning, local consistency, personalized initialization, excess risk
Compressor summary: The paper proposes FedInit, an efficient federated learning algorithm that uses personalized relaxed initialization to reduce the "client drift" problem and improve performance in distributed training of global models.
Tianrong Chen, Guan-Horng Liu, Molei Tao, Evangelos Theodorou
https://openreview.net/forum?id=ykvvv0gc4R
Keywords: Schrödinger Bridge, Trajectory Inference, Optimal Transport
Compressor summary: The paper presents DMSB, a novel computational framework for inferring high-dimensional multi-marginal trajectories from unlabeled samples at coarse time intervals, which outperforms baselines and can reconstruct velocity distributions from position snapshots.
Yuanshao Zhu, Yongchao Ye, Shiyao Zhang, Xiangyu Zhao, James Yu
https://openreview.net/forum?id=ykMdzevPkJ
Keywords: Trajectory Generation, Diffusion Model, Urban Computing, Spatial-temporal Data Mining
Compressor summary: The authors propose a model called DiffTraj that generates realistic and privacy-preserving GPS trajectories by denoising white noise using spatial-temporal features and a deep neural network.
Moritz Haas, David Holzmüller, Ulrike von Luxburg, Ingo Steinwart
https://openreview.net/forum?id=yjYwbZBJyl
Keywords: benign overfitting, kernels, neural tangent kernel, consistency, learning theory
Compressor summary: The key to benign overfitting is not high dimension but large derivatives in smooth estimators; this can be achieved by adding small fluctuations to the activation function of wide neural networks.
Oscar Michel, Anand Bhattad, Eli VanderBilt, Ranjay Krishna, Aniruddha Kembhavi, Tanmay Gupta
https://openreview.net/forum?id=yjWVd8Fhqt
Keywords: computer vision, image editing, generative modeling, diffusion models, 3D
Compressor summary: The authors propose a new task called language-guided 3D-aware editing, where edits are done according to language instructions while preserving the 3D scene, and release a large dataset for this task along with models that can perform well on it.
Xiuhong Lin, Changjie Qiu, zhipeng cai, Siqi Shen, Yu Zang, Weiquan Liu, Xuesheng Bian, Matthias Müller, Cheng Wang
https://openreview.net/forum?id=yiehppUCO2
Keywords: event camera, 2D-3D registration, representation learning
Compressor summary: E2PNet is a novel method that registers 2D RGB images to 3D point clouds using event cameras, which improves robustness and enables other vision tasks.
GUOJUN XIONG, Jian Li
https://openreview.net/forum?id=yhNHpLWJDl
Keywords: Restless bandits, Whittle index policy, Q-learning, Two-timescale stochastic approximation, Neural network function approximation
Compressor summary: Neural-Q-Whittle is a Q-learning algorithm that uses neural networks and Whittle indices to solve intractable restless multi-armed bandits problems with an empirical convergence rate of $\mathcal{O}(1/k^{2/3})$.
Jaemin Cho, Abhay Zala, Mohit Bansal
https://openreview.net/forum?id=yhBFG9Y85R
Keywords: text-to-image generation; visual programming; text-to-image evaluation; step-by-step generation; interpretability; explainability
Compressor summary: The paragraph introduces two novel visual programming frameworks, VPGen and VPEval, for text-to-image tasks that improve spatial control, handle predefined object classes, and provide interpretable and explainable evaluations with experts in different skills.
Chanakya Ekbote, Ajinkya Deshpande, Arun Iyer, SUNDARARAJAN SELLAMANICKAM, Ramakrishna B Bairi
https://openreview.net/forum?id=yh0OkiUk5h
Keywords: Graph Neural Networks, Unsupervised Representation Learning, Graph Filters
Compressor summary: The paper introduces FiGURe, a method to improve node representations using filter-based augmentations and random Fourier feature projections, achieving up to 4.4% better performance on downstream tasks than existing unsupervised models.
Kexin Huang, Ying Jin, Emmanuel Candes, Jure Leskovec
https://openreview.net/forum?id=ygjQCOyNfh
Keywords: Graph Neural Networks, Conformal Prediction, Uncertainty Quantification
Compressor summary: CF-GNN is a graph-based model that provides guaranteed uncertainty estimates by extending conformal prediction to graph data and using a topology-aware output correction model to reduce the prediction set size.
Tom George, Kim Stachenfeld, Caswell Barry, Claudia Clopath, Tomoki Fukai
https://openreview.net/forum?id=yft4JlxsRf
Keywords: hippocampus, path integration, local learning, generative models, oscillations, inference, Helmholtz machine, wake-sleep
Compressor summary: The authors propose a biologically plausible model of the hippocampus that uses theta-band oscillations and a wake-sleep algorithm to learn and generate realistic sensory predictions and path integration.
Jiancong Xiao, Ruoyu Sun, Zhi-Quan Luo
https://openreview.net/forum?id=ydKWoqWZ3t
Keywords: Pac-Bayes, Adversarial Robustness, Generalization
Compressor summary: The paper proposes a new robust generalization bound for deep neural networks that does not require additional assumptions and is tighter than existing bounds, offering insights into the causes of poor robust generalization.
Polina Kirichenko, Mark Ibrahim, Randall Balestriero, Diane Bouchacourt, Shanmukha Ramakrishna Vedantam, Hamed Firooz, Andrew Gordon Wilson
https://openreview.net/forum?id=yageaKlk7S
Keywords: data augmentation, class-dependent bias
Compressor summary: The paper studies how data augmentation affects image classification, especially for classes with ambiguous or fine-grained distinctions, and proposes class-conditional augmentation strategies to improve performance.
Xingang Guo, Darioush Keivan, Geir Dullerud, Peter Seiler, Bin Hu
https://openreview.net/forum?id=yaJ4vZPnHX
Keywords: Structured $\mathcal{H}_\infty$ Control, Nonsmooth Optimization, Complexity Analysis
Compressor summary: This paper presents new theoretical results on the difficulty of optimizing policies without derivatives for a class of robust control tasks, and shows how to find approximate solutions using zeroth-order oracles.
Andrey Okhotin, Dmitry Molchanov, Arkhipkin Sergeevich Vladimir, Grigory Bartosh, Viktor Ohanesian, Aibek Alanov, Dmitry P. Vetrov
https://openreview.net/forum?id=yYUdgbmhh9
Keywords: Generative models, Diffusion, Exponential Family
Compressor summary: The paper introduces Star-Shaped Diffusion Probabilistic Models, which enable generative modeling with non-Gaussian distributions using a simpler and more efficient approach than existing methods.
Wesley Khademi, Li Fuxin
https://openreview.net/forum?id=yVMlYSL1Bp
Keywords: multimodal shape completion, point cloud completion, 3d shape generation, generative modeling, generative adversarial networks
Compressor summary: The paper proposes a new network that can generate diverse, plausible 3D shape completions from partial observations using style modulation and multiple discriminators to prevent mode collapse.
Peiqing Yang, Shangchen Zhou, Qingyi Tao, Chen Change Loy
https://openreview.net/forum?id=yThjbzhIUP
Keywords: Face Restoration, Diffusion
Compressor summary: The paper introduces a new restoration approach called partial guidance that adapts to real-world degradations by modeling desired properties of high-quality images and applying them during the reverse diffusion process, achieving good results across various tasks.
Xiang Cheng, Bohan Wang, Jingzhao Zhang, Yusong Zhu
https://openreview.net/forum?id=yT0f93CeTw
Keywords: Sampling, MCMC, Conditional Mixing, Non-log-concave Distributions
Compressor summary: The paragraph explains that MCMC algorithms can work efficiently when a subset of the state space has Poincaré-style inequality, even if global mixing is slow, and this can help in various applications like sampling from mixtures of Gaussians or parameter estimation.
Lisha Chen, Heshan Devaka Fernando, Yiming Ying, Tianyi Chen
https://openreview.net/forum?id=yPkbdJxQ0o
Keywords: Generalization, algorithm stability, multi-objective optimization, gradient conflict
Compressor summary: The MoDo algorithm studies the stability of dynamic weighting methods for multi-objective learning and reveals how updating along conflict-avoidant directions may hinder optimal generalization performance.
Weitian Huang, Sirui Yang, Hongmin Cai
https://openreview.net/forum?id=yN6NHZOXkg
Keywords: information bottleneck, multi-view clustering, variational autoencoders
Compressor summary: The paper proposes an information-based multi-view clustering model that uses deep neural networks and Stochastic Gradient Variational Bayes for representation learning and clustering, achieving better results than existing methods.
Yiding Chen, Jerry Zhu, Kirthevasan Kandasamy
https://openreview.net/forum?id=yKCLfOOIL7
Keywords: Mechanism design, statistical minimax estimation, federated learning
Compressor summary: The paper proposes a novel mechanism for estimating the mean of a normal distribution collaboratively among strategic agents, using corrupted data sharing and minimax optimal estimators to incentivize truthful reporting and reduce estimation errors.
Xingdong Feng, Xin HE, Caixing Wang, Chao Wang, Jingnan Zhang
https://openreview.net/forum?id=yIcCkMUCtL
Keywords: kernel methods, covariate shift, reproducing kernel Hilbert space (RKHS)
Compressor summary: The paper proposes a general analysis of nonparametric methods under covariate shift in RKHS, with applications to various learning tasks and theoretical and numerical results.
Yongliang Shen, Kaitao Song, Xu Tan, Dongsheng Li, Weiming Lu, Yueting Zhuang
https://openreview.net/forum?id=yHdTscY6Ci
Keywords: LLM, ChatGPT, Hugging Face, Autonomous LLM
Compressor summary: HuggingGPT is an LLM-powered agent that uses ChatGPT to connect various AI models in Hugging Face to solve complicated AI tasks across different domains and modalities, moving toward artificial general intelligence.
Pedro Sandoval-Segura, Vasu Singla, Jonas Geiping, Micah Goldblum, Tom Goldstein
https://openreview.net/forum?id=yGs9vTRjaE
Keywords: data poisoning, poisons, unlearnable dataset, data protection, imperceptible perturbations, adversarial machine learning
Compressor summary: Unlearnable dataset methods may not protect data privacy as they can enable neural networks to learn useful features, and linear separability of perturbations is not a necessary condition for this.
Suresh kumar Amalapuram, Sumohana S. Channappayya, Bheemarjuna Tamma
https://openreview.net/forum?id=yGLokEhdh9
Keywords: Continual learning, Class imbalance, scalability, Network intrusion detection, and Cybersecurity
Compressor summary: The paper proposes two methods to improve continual learning-based intrusion detection: one for class imbalance and another for reducing computations, which perform better than baselines on various benchmarks.
Owen Queen, Thomas Hartvigsen, Teddy Koker, Huan He, Theodoros Tsiligkaridis, Marinka Zitnik
https://openreview.net/forum?id=yEfmhgwslQ
Keywords: Explainability, Interpretability, Time Series, Explanations, Temporal patterns, Model Understanding, Latent space, Self-supervised learning
Compressor summary: TimeX is a model that trains an interpretable surrogate to mimic a pretrained time series model and provides discrete attribution maps and latent space explanations for interpreting time series predictions.
Haochuan Li, Alexander Rakhlin, Ali Jadbabaie
https://openreview.net/forum?id=yEewbkBNzi
Keywords: Non-convex optimization, Adam, Convergence, Variance reduction
Compressor summary: The paper proves that the Adam algorithm for optimization converges to stationary points under realistic conditions and introduces a variance-reduced version with faster convergence.
Seunghun Lee, Jaewon Chu, Sihyeon Kim, Juyeon Ko, Hyunwoo J. Kim
https://openreview.net/forum?id=yE62KM4qsO
Keywords: Bayesian optimization, smoothness regularization, variational autoencoder
Compressor summary: Correlated latent space Bayesian Optimization (CoBO) is a method that learns correlated latent spaces to reduce the gap between optimization in the latent space and the input space, achieving high performance on discrete data tasks with limited function evaluations.
Bohan Wang, Jingwen Fu, Huishuai Zhang, Nanning Zheng, Wei Chen
https://openreview.net/forum?id=yDvb3mlogA
Keywords: Adam, Convergence, Upper Bound, Lower Bound
Compressor summary: This paper derives a new convergence guarantee for Adam optimization with smooth conditions and bounded noise variance, filling a gap in existing literature and providing a tight upper bound.
Radu Marinescu, Debarun Bhattacharjya, Junkyu Lee, Fabio Cozman, Alexander G. Gray
https://openreview.net/forum?id=yCBqKTvYe9
Keywords: graphical models, credal networks, probabilistic inference
Compressor summary: Credal Marginal MAP inference is explored in this paper, with new exact and approximation methods proposed and evaluated on various problems.
Agrim Gupta, Jiajun Wu, Jia Deng, Li Fei-Fei
https://openreview.net/forum?id=yC3q7vInux
Keywords: Representation Learning, Visual Correspondence, Self-supervised learning, Videos
Compressor summary: Siamese Masked Autoencoders (SiamMAE) is a method for learning visual correspondence from videos by predicting missing patches in one frame based on another frame with masked patches, focusing on object motion and achieving competitive results on various tasks.
Yifei Wang, Liangchen Li, Jiansheng Yang, Zhouchen Lin, Yisen Wang
https://openreview.net/forum?id=yBoVwpGa5E
Keywords: Adversarial Training
Compressor summary: Adversarial Training (AT) has become arguably the state-of-the-art algorithm for extracting robust features. However, researchers recently notice that AT suffers from severe robust overfitting problems, particularly after learning rate (LR) decay. In this paper, we explain this phenomenon by viewing adversarial training as a dynamic minimax game between the model trainer and the attacker. Specifically, we analyze how LR decay breaks the balance between the minimax game by empowering the trainer with a stronger memorization ability, and show such imbalance induces robust overfitting as a result of memorizing non-robust features. We validate this understanding with extensive experiments, and provide a holistic view of robust overfitting from the dynamics of both the two game players. This understanding further inspires us to alleviate robust overfitting by rebalancing the two players by either regularizing the trainer's capacity or improving the attack strength. Experiments show that the proposed ReBalanced Adversarial Training (ReBAT) can attain good robustness and does not suffer from robust overfitting even after very long training. Code is available at https://github.com/PKU-ML/ReBAT.
Fulton Wang, Julius Adebayo, Sarah Tan, Diego Garcia-Olano, Narine Kokhlikyan
https://openreview.net/forum?id=yBVLXvJ1sb
Keywords: Debugging, interpretability, influence functions
Compressor summary: The authors introduce InfEmden, a slice discovery method that finds coherent groups of test examples where a model under-performs by using influence embeddings and K-Means clustering.
Shen Jiang, Zipeng Ji, Guanghui Zhu, Chunfeng Yuan, Yihua Huang
https://openreview.net/forum?id=yAOwkf4FyL
Keywords: Differentiable neural architecture search; Image classification; Failure of DARTS
Compressor summary: The paper proposes a method called operation-level early stopping (OLES) to improve Differentiable NAS (DARTS) by addressing the overfitting issue caused by skip connections in neural architecture search.
Hugues Van Assel, Titouan Vayer, Rémi Flamary, Nicolas Courty
https://openreview.net/forum?id=y9U0IJ2uFr
Keywords: Dimension Reduction, Optimal Transport, Affinities
Compressor summary: The paragraph discusses a novel way of symmetrizing entropic affinities using optimal transport, which improves clustering performance and robustness in dimensionality reduction algorithms like t-SNE.
Moni Naor, Kobbi Nissim, Uri Stemmer, Chao Yan
https://openreview.net/forum?id=y8UAQQHVTX
Keywords: Differential privacy, private learning, private prediction
Compressor summary: The paper proposes private everlasting prediction, a method to protect the privacy of both training data and queries in answering a stream of classification questions, with lower sample complexity than existing methods.
Yiding Jiang, J Zico Kolter, Roberta Raileanu
https://openreview.net/forum?id=y5duN2j9s6
Keywords: reinforcement learning, generalization, procgen, crafter
Compressor summary: The paper proposes EDE, a method that encourages exploration in deep RL by using an ensemble of Q-value distributions to handle epistemic uncertainty and improve generalization to new environments.
Wanxing Chang, Ye Shi, Jingya Wang
https://openreview.net/forum?id=y50AnAbKp1
Keywords: Learning with Noisy Labels, Optimal Transport, Curriculum Learning
Compressor summary: The paper proposes a novel optimal transport method, CSOT, to address challenges in learning with noisy labels by considering sample distribution structure and assigning reliable labels incrementally.
Sarah Mameche, David Kaltenpoth, Jilles Vreeken
https://openreview.net/forum?id=y0OlQSZsyp
Keywords: independent mechanisms, causal discovery, information theory, gaussian processes
Compressor summary: The paper proposes a method to discover causal relationships in systems with changing components using Gaussian Process models and independence assumptions.
Yuxin Jia, Youfang Lin, Xinyan Hao, Yan Lin, Shengnan Guo, Huaiyu Wan
https://openreview.net/forum?id=y08bkEtNBK
Keywords: long-range time series forecasting, information transmission, long- and short-term repetitive patterns, global and local correlations
Compressor summary: The WITRAN framework captures semantic information for long-range time series forecasting using a bi-granular information transmission and horizontal vertical gated selective unit, while reducing time and memory complexity compared to previous methods.
weitao Du, Jiujiu Chen, Xuecang Zhang, Zhi-Ming Ma, Shengchao Liu
https://openreview.net/forum?id=xzmaFfw6oh
Keywords: Molecule Joint Auto-encoding, Molecule Joint Self-supervised Learning, Markov processes, contrastive learning, molecule representation learning
Compressor summary: The proposed method, MoleculeJAE, learns both 2D and 3D molecular information using a pretraining technique, achieving state-of-the-art results on various tasks.
Attila Lengyel, Ombretta Strafforello, Robert-Jan Bruintjes, Alexander Gielisse, Jan van Gemert
https://openreview.net/forum?id=xz8j3r3oUA
Keywords: color equivariance, equivariance, color robustness, equivariant convolutions
Compressor summary: The paper introduces CEConvs, a novel deep learning building block that allows shape feature sharing across different colors while preserving important color information, improving robustness to color changes and performance on various tasks.
Fuyuan Lyu, Xing Tang, Dugang Liu, Chen Ma, Weihong Luo, Liang Chen, xiuqiang He, Xue Liu
https://openreview.net/forum?id=xxfHMqNcum
Keywords: Feature Interaction Search, Deep Sparse Network
Compressor summary: The paper introduces a hybrid-grained feature interaction selection approach for deep sparse networks that targets both feature field and feature value, and shows its effectiveness on three benchmark datasets.
Zelin Ni, Hang Yu, Shizhan Liu, Jianguo Li, Weiyao Lin
https://openreview.net/forum?id=xx3qRKvG0T
Keywords: time series forecasting, basis learning, self-supervised learning
Compressor summary: BasisFormer is a time series forecasting architecture that uses adaptive self-supervised learning, cross-attention, and similarity coefficients to create tailored bases for accurate predictions.
Qing Wu, Lixuan Chen, Ce Wang, Hongjiang Wei, S Kevin Zhou, Jingyi Yu, Yuyao Zhang
https://openreview.net/forum?id=xx3QgKyghS
Keywords: Medical Image, Computed Tomography, Metal Arftiacts, Implicit Neural Representation, Unsupervised Learning
Compressor summary: The paragraph describes a novel polychromatic neural representation (Polyner) method to improve computed tomography (CT) imaging with metallic implants by modeling the metal artifact reduction problem from a nonlinear inverse perspective and achieving better performance than supervised methods.
Geyu Liang, Naichen Shi, Raed Al Kontar, Salar Fattahi
https://openreview.net/forum?id=xw6Szwu4xz
Keywords: Dictionary Learning, Data Heterogeneity, Personalization
Compressor summary: The paper introduces PerDL, a problem where sparse linear representations are learned from heterogeneous datasets with shared and unique features, and proposes PerMA, an efficient meta-algorithm that can recover the global and local dictionaries for various learning tasks.
Prateek Yadav, Derek Tam, Leshem Choshen, Colin Raffel, Mohit Bansal
https://openreview.net/forum?id=xtaX3WyCj1
Keywords: Model Merging, Fusing, Collaborative Training, Robust Fine-tuning, Federated Learning
Compressor summary: TIES-Merging is a novel method to combine task-specific models into a multitask model by resolving parameter interference and exploiting sign information, which leads to better performance in various settings.
Taihei Oki, Shinsaku Sakaue
https://openreview.net/forum?id=xtQ9IGRzIW
Keywords: algorithms with predictions, beyond the worst-case analysis of algorithms, time complexity, combinatorial optimization, discrete convex analysis, submodular functions
Compressor summary: The paper presents a framework to accelerate discrete optimization algorithms, especially laminar convex minimization, by using machine-learned predictions.
Kai Zhao, Qiyu Kang, Yang Song, Rui She, Sijie Wang, Wee Peng Tay
https://openreview.net/forum?id=xtADRDRsM2
Keywords: adversarial robustness, graph neural networks
Compressor summary: This paper explores how using conservative Hamiltonian neural flows can improve the adversarial robustness of graph neural networks, which are otherwise vulnerable to perturbations affecting both node features and graph topology.
Jerry Chee, Yaohui Cai, Volodymyr Kuleshov, Christopher De Sa
https://openreview.net/forum?id=xrk9g5vcXR
Keywords: Quantization, Large Language Models, Adaptive Rounding, Theoretical Guarantees
Compressor summary: The paper introduces QuIP, a new method for post-training parameter quantization in large language models that uses incoherence processing to improve results and reduce bits per weight needed.
Jiayu Wang, Kang Zhao, Yifeng Ma, Shiwei Zhang, Yingya Zhang, Yujun Shen, Deli Zhao, Jingren Zhou
https://openreview.net/forum?id=xrK3QA9mLo
Keywords: diffusion model; talking face generation; face generation
Compressor summary: FaceComposer is a versatile generative model for creating facial content with text control and animation, based on the latent diffusion framework, and it includes a large-scale multi-modal face database and an easy-to-use interface.
Yuxuan Lu, Yuqing Kong
https://openreview.net/forum?id=xr3KAzboHY
Keywords: Peer prediction, Peer review, Calibration
Compressor summary: The paper proposes a one-shot noise calibration process for ranking papers fairly by predicting others' scores, which reduces error probability and outperforms average ratings.
Myong Chol Jung, He Zhao, Joanna Dipnall, Lan Du
https://openreview.net/forum?id=xq1QvViDdW
Keywords: Uncertainty estimation, multimodality, neural processes
Compressor summary: MNPs are a novel approach for estimating uncertainty in multimodal data that adapts Neural Processes, providing robustness and reliability with faster computation.
Yonglong Tian, Lijie Fan, Phillip Isola, Huiwen Chang, Dilip Krishnan
https://openreview.net/forum?id=xpjsOQtKqx
Keywords: representation learning, synthetic images, text-to-image models
Compressor summary: The paper explores using text-to-image models to generate synthetic images for learning visual representations, and shows that their method outperforms existing methods on large scale datasets.
Joon-Hyeok Yim, Anna Gilbert
https://openreview.net/forum?id=xo2lbfQE8I
Keywords: tree metric fitting, ultrametric fitting, $\ell_1$-hyperbolicity
Compressor summary: The paper proposes an algorithm to fit trees based on hyperbolicity values, which improves upon previous algorithms and reveals differences between real and synthetic data sets.
Zibo Zhao, Wen Liu, Xin Chen, Xianfang Zeng, Rui Wang, Pei Cheng, BIN FU, Tao Chen, Gang YU, Shenghua Gao
https://openreview.net/forum?id=xmxgMij3LY
Keywords: Conditional 3D Shape Generation, Neural 3D Representation, 3D Reconstruction
Compressor summary: The paper proposes a new method for generating 3D shapes from 2D images or texts by aligning the representations of 3D shapes in different modalities and using two models to encode and decode them.
Yuheng Zha, Yichi Yang, Ruichen Li, Zhiting Hu
https://openreview.net/forum?id=xkkBFePoFn
Keywords: Text Alignment, Efficient Unified Model, NLU Tasks, Factual Consistency Evaluation, QA with Unanswerable Question
Compressor summary: The paper proposes a small and efficient text alignment model that performs well on various NLP tasks, including text entailment, similarity, question answering, and factual consistency, outperforming larger LLMs in some cases.
Youngsoo Baek, Samuel Berchuck, Sayan Mukherjee
https://openreview.net/forum?id=xgzkuTGBTx
Keywords: asymptotics, random features model, Bayesian inference
Compressor summary: This paper studies how the uncertainty of Bayesian models relates to the performance of their maximum likelihood estimates in high-dimensional settings, finding agreement in certain regimes and similarity in others.
Dmitry Chistikov, Matthias Englert, Ranko Lazic
https://openreview.net/forum?id=xgY4QcOiEZ
Keywords: implicit bias, implicit regularization, training dynamics, ReLU networks, gradient flow, theoretical analysis
Compressor summary: The paper investigates how one-hidden layer ReLU networks learn a single neuron task by gradient flow and shows that they implicitly minimize the rank of network parameters while converging to zero loss.
Steven Adriaensen, Herilalaina Rakotoarison, Samuel Müller, Frank Hutter
https://openreview.net/forum?id=xgTV6rmH6n
Keywords: learning curve extrapolation, prior-data fitted networks, transformers, Bayesian inference, uncertainty estimation, model selection
Compressor summary: The authors propose LC-PFN, a fast and accurate Bayesian method for learning curve extrapolation using prior-data fitted neural networks, which can improve model selection efficiency by enabling predictive early stopping.
Kim Andrea Nicoli, Christopher J. Anders, Lena Funcke, Tobias Hartung, Karl Jansen, Stefan Kuhn, Klaus Robert Muller, Paolo Stornati, Pan Kessel, Shinichi Nakajima
https://openreview.net/forum?id=xfBeVGJwyL
Keywords: Bayesian optimization, Expected improvement, Quantum computing, Variational Quantum Eigensolvers
Compressor summary: The paper presents a new Bayesian optimization method for VQEs that uses a novel kernel and acquisition function to reduce uncertainty and improve performance.
Jiaming Guo, Rui Zhang, Shaohui Peng, Qi Yi, Xing Hu, Ruizhi Chen, Zidong Du, Xishan Zhang, Ling Li, Qi Guo, Yunji Chen
https://openreview.net/forum?id=xdQpmUPNHC
Keywords: reinforcement learning, context variables, symbolic policy
Compressor summary: The paper proposes ESPL, an efficient gradient-based method to learn simple and interpretable symbolic policies from scratch for sequential decision-making tasks using a differentiable symbolic expression and a path selector.
Hoang Pham, The-Anh Ta, Shiwei Liu, Lichuan Xiang, Dung D. Le, Hongkai Wen, Long Tran-Thanh
https://openreview.net/forum?id=xdOoCWCYaY
Keywords: Pruning Neural Network, Sparsity, Neural Architecture Search
Compressor summary: The paper proposes a new framework and method for pruning neural networks at initialization that balances effective nodes and paths, improving accuracy and computational efficiency compared to existing methods.
Surbhi Goel, Steve Hanneke, Shay Moran, Abhishek Shetty
https://openreview.net/forum?id=xcGhx9FdxM
Keywords: Sequential prediction, adversarial examples, abstention, out-of-distribution, VC Classes
Compressor summary: The paper proposes a new model for sequential prediction that allows learners to abstain from making predictions on adversarial examples and designs learners with better error guarantees than existing methods.
Yunqing Zhao, Tianyu Pang, Chao Du, Xiao Yang, Chongxuan Li, Ngai-man Cheung, Min Lin
https://openreview.net/forum?id=xbbknN9QFs
Keywords: Large Vision-Language Models, Adversarial Robustness
Compressor summary: The paragraph discusses evaluating the robustness of vision-language models (VLMs) to targeted adversarial examples and their potential security flaws before deployment.
Daolang Huang, Manuel Haussmann, Ulpu Remes, S. T. John, Grégoire Clarté, Kevin Sebastian Luck, Samuel Kaski, Luigi Acerbi
https://openreview.net/forum?id=xax5eWeObb
Keywords: neural processes, equivariance, Gaussian processes
Compressor summary: Relational Conditional Neural Processes (RCNPs) are a new model that improves the performance of equivariant neural processes in higher dimensions for various machine learning tasks.
Scott Fujimoto, Wei-Di Chang, Edward J. Smith, Shixiang Shane Gu, Doina Precup, David Meger
https://openreview.net/forum?id=xZvGrzRq17
Keywords: Deep reinforcement learning, representation learning
Compressor summary: The paper introduces SALE, a method for learning embeddings from low-level states that improve representation learning in reinforcement learning, and combines it with TD3 to create TD7, which performs better than existing continuous control algorithms on OpenAI gym benchmark tasks.
Benjamin Samuel Ruben, Cengiz Pehlevan
https://openreview.net/forum?id=xXfDB8kJUs
Keywords: ridge regression, ensembling methods
Compressor summary: Feature bagging reduces prediction variance by combining predictions of many estimators trained on subsets or projections of features, and can mitigate double-descent in noisy least-squares ridge ensembles.
Dongmin Park, Seola Choi, Doyoung Kim, Hwanjun Song, Jae-Gil Lee
https://openreview.net/forum?id=xWCp0uLcpG
Keywords: Data Pruning, Data Subset Selection, Noisy Labels, Relabeling, Self-training
Compressor summary: Prune4Rel is a novel data pruning algorithm that maximizes re-labeling accuracy and generalization performance by finding a subset of training examples with high prediction confidence of their neighbors in the subset.
Qizhang Feng, Zhimeng Jiang, Ruiquan Li, Yicheng Wang, Na Zou, Jiang Bian, Xia Hu
https://openreview.net/forum?id=xW0ayZxPWs
Keywords: Graph Distillation, Algorithmic Fairness
Compressor summary: The paper proposes a method to create smaller graphs that are both accurate and fair for graph neural networks, by developing a new metric called coherence and using a bi-level optimization algorithm.
Fan Yao, Chuanhao Li, Karthik Abinav Sankararaman, Yiming Liao, Yan Zhu, Qifan Wang, Hongning Wang, Haifeng Xu
https://openreview.net/forum?id=xUyBP16Q5J
Keywords: Recommender system, Mechanism design, Potential function, Optimization
Compressor summary: The text discusses how online content recommendation platforms affect media creators' choices and welfare, and proposes a new reward mechanism called Backward Rewarding Mechanisms to improve the platform's design.
Haoyu Guo, Sida Peng, Yunzhi Yan, Linzhan Mou, Yujun Shen, Hujun Bao, Xiaowei Zhou
https://openreview.net/forum?id=xTgM7XLN9P
Keywords: Computer Vision, 3D Vision, Volumetric Video
Compressor summary: The paper introduces a novel neural representation called dynamic codebook that compresses volumetric videos by merging similar features and compensating with dynamic codes, achieving better quality and storage efficiency.
Peng Jin, Yang Wu, Yanbo Fan, Zhongqian Sun, Yang Wei, Li Yuan
https://openreview.net/forum?id=xSEhb2j3TK
Keywords: Text-driven Motion Synthesis, Diffusion Models, Graph networks
Compressor summary: The paper proposes using hierarchical semantic graphs to generate fine-grained human motion from text descriptions, improving over existing methods that may focus too much on action names and lack detail.
Xingyu Jiang, Jiayi Ma
https://openreview.net/forum?id=xRfTcZdQxq
Keywords: Model reasoning; Model fitting; Outliers; Sparse subspace learning; Feature matching
Compressor summary: The paper proposes a unified optimization model for outlier rejection, true model reasoning and parameter estimation, which uses sparse subspace recovery to search for independent bases in an over-embedded data space.
Wojciech Masarczyk, Mateusz Ostaszewski, Ehsan Imani, Razvan Pascanu, Piotr Miłoś, Tomasz Trzcinski
https://openreview.net/forum?id=xQOHOpe1Fv
Keywords: representation learning, continual learning, training dynamics
Compressor summary: The paper investigates how deep neural networks split into two parts during image classification training, where the initial layers create separable representations and the latter layers, called "the tunnel," compress them without significantly affecting performance.
Yan Wang, Huaiqing Wu, Dan Nettleton
https://openreview.net/forum?id=xPqINp0Eu1
Keywords: Stability, Prediction Intervals, Random Forests
Compressor summary: The paper shows how random forests are stable and can provide good interval predictions using out-of-bag error and weak assumptions.
Panagiotis Misiakos, Chris Wendler, Markus Püschel
https://openreview.net/forum?id=xPLaXSuSvQ
Keywords: directed acyclic graph, few root causes, structural equation models, linear SEMs, additive noise
Compressor summary: The paper proposes a new algorithm for learning directed acyclic graphs (DAGs) from linear structural equation models (SEMs) with fewer root causes and noise, which improves upon previous methods.
Qihe Huang, Lei Shen, Ruixin Zhang, Shouhong Ding, Binwu Wang, Zhengyang Zhou, Yang Wang
https://openreview.net/forum?id=xOzlW2vUYc
Keywords: Time Series Forecasting;
Compressor summary: CrossGNN is a linear complexity GNN model that improves multivariate time series forecasting by refining cross-scale and cross-variable interactions using an adaptive multi-scale identifier and two specialized GNN components.
Miao Xiong, Ailin Deng, Pang Wei Koh, Jiaying Wu, Shen Li, Jianqing Xu, Bryan Hooi
https://openreview.net/forum?id=xOJUmwwlJc
Keywords: Calibration, Uncertainty Estimation, Trustworthiness, Fairness, Multicalibration
Compressor summary: The text discusses a common issue in calibration algorithms called proximity bias, which causes models to be more overconfident in low proximity data, and proposes a new algorithm (ProCal) to address it.
Antoine Scardigli, Lukas Cavigelli, Lorenz K Muller
https://openreview.net/forum?id=xNyR7DXUzJ
Keywords: computer graphics, rendering, ray tracing, GPU acceleration, RL, spatiotemporal latent space
Compressor summary: The paper proposes a framework that uses reinforcement learning and neural networks to improve Monte-Carlo path tracing, reducing noise and rendering times for realistic image synthesis in real-time applications.
Alexander Pak, Justin Ko, Florent Krzakala
https://openreview.net/forum?id=xNUmTRYtV1
Keywords: Spectral Method, Community detection, Wigner Spike model, Random Matrix, BBP transition, Approximate Message Passing, Spin glasses, Statistical Physics
Compressor summary: The paper proposes an AMP algorithm for a spiked Wigner problem with inhomogeneous noise and compares it with a spectral method, showing good performance and matching information-theoretic results.
Mingyu Yang, Yaodong Yang, Zhenbo Lu, Wengang Zhou, Houqiang Li
https://openreview.net/forum?id=xMgO04HDOS
Keywords: Multi-Agent Reinforcement Learning, Hierarchical Skill Discovery, Probabilistic Graphical Model
Compressor summary: The text describes a hierarchical algorithm called Hierarchical Multi-Agent Skill Discovery (HMASD) that learns team and individual skills for multi-agent tasks without extrinsic rewards, outperforming existing methods.
Minki Kang, Seanie Lee, Jinheon Baek, Kenji Kawaguchi, Sung Ju Hwang
https://openreview.net/forum?id=xJLEQQrFia
Keywords: language model, distillation, reasoning, knowledge augmentation
Compressor summary: KARD is a novel method that enhances small LMs with rationales from large and external sources, achieving better results on knowledge-intensive reasoning tasks.
Artun Saday, Y. Cahit Yıldırım, Cem Tekin
https://openreview.net/forum?id=xINPCvgULc
Keywords: robust satisficing, regret minimization, Gaussian processes
Compressor summary: The paper introduces RoBOS, a novel robust Bayesian satisficing algorithm for noisy black-box optimization that performs well under distributional shifts.
Le Yu, Leilei Sun, Bowen Du, Weifeng Lv
https://openreview.net/forum?id=xHNzWHbklj
Keywords: dynamic graph learning, Transformer-based architecture, dynamic graph library
Compressor summary: DyGFormer is a new Transformer-based architecture for dynamic graph learning that leverages historical interactions and uses DyGLib, a unified library for rigorous evaluations.
Devleena Das, Sonia Chernova, Been Kim
https://openreview.net/forum?id=xGz0wAIJrS
Keywords: Concept-Based Explanations, Reinforcement Learning, Human-AI Interaction
Compressor summary: The paper proposes State2Explanation (S2E), a framework that learns to embed state-action pairs and concept-based explanations, which can improve both an RL agent's learning and users' understanding of the agent's decisions in sequential decision making tasks.
Fangchen Yu, Runze Zhao, Zhan Shi, Yiwen Lu, Jicong Fan, Yicheng Zeng, Jianfeng Mao, Wenye Li
https://openreview.net/forum?id=xFtuNq23D5
Keywords: Spectral Clustering, Incomplete Data, Kernel Correction, Self-expressive Affinity Learning
Compressor summary: The paper proposes two new methods to improve spectral clustering on incomplete data by enhancing the kernel matrix and learning adaptive affinity matrices.
Hao Qin, Kwang-Sung Jun, Chicheng Zhang
https://openreview.net/forum?id=xF89MjFbWp
Keywords: multi-armed bandits, bounded rewards
Compressor summary: The paper analyzes KL-Maillard Sampling, a variant of Maillard Sampling that achieves finite-time gap-dependent regret bounds and asymptotic optimality in bandit problems with Bernoulli rewards.
Yingcong Li, Kartik Sreenivasan, Angeliki Giannou, Dimitris Papailiopoulos, Samet Oymak
https://openreview.net/forum?id=xEhKwsqxMa
Keywords: chain-of-thought, in-context learning, attention, compositional learning, approximation, length generalization
Compressor summary: The study examines how chain-of-thought (CoT) helps transformers learn compositional functions more efficiently by breaking down the learning process into two phases and reducing sample complexity.
Duy Minh Ho Nguyen, Hoang Nguyen, Nghiem Tuong Diep, Tan Ngoc Pham, Tri Cao, Binh T. Nguyen, Paul Swoboda, Nhat Ho, Shadi Albarqouni, Pengtao Xie, Daniel Sonntag, Mathias Niepert
https://openreview.net/forum?id=xE7oH5iVGK
Keywords: medical imaging; self-supervised learning; graph matching; large-vision model
Compressor summary: LVM-Med is a family of deep networks trained on large-scale medical datasets that uses a novel self-supervised contrastive learning algorithm to improve performance on various medical tasks.
Dominik Straub, Matthias Schultheis, Heinz Koeppl, Constantin A. Rothkopf
https://openreview.net/forum?id=xDHzQQ4lnC
Keywords: inverse optimal control, probabilistic modeling, motor control, cognitive science
Compressor summary: The paper presents a probabilistic approach to inverse optimal control for partially observable stochastic non-linear systems with unobserved action signals, which can characterize behavior in sequential decision-making tasks under uncertainty and disentangle perceptual factors and behavioral costs.
Zige Wang, Yonggang Zhang, Zhen Fang, Long Lan, Wenjing Yang, Bo Han
https://openreview.net/forum?id=xBqjoG0NxM
Keywords: test-time data adaptation, zeroth-order optimization, out-of-distribution generalization
Compressor summary: The paper proposes a method called SODA, which uses reliable labels and preserves data information, to improve test-time data adaptation and adapt deployed models to distribution shifts while maintaining privacy.
Moritz Willig, Matej Zečević, Devendra Singh Dhami, Kristian Kersting
https://openreview.net/forum?id=xBhvMu4J03
Keywords: Structural Causal Models, Marginalization, Consolidation, Compression
Compressor summary: Consolidating causal mechanisms helps simplify large-scale structural causal models while maintaining their causality and interventional behavior.
Xinran Zhao, Hongming Zhang, Xiaoman Pan, Wenlin Yao, Dong Yu, Jianshu Chen
https://openreview.net/forum?id=x9FOu3W6iy
Keywords: knowledge-intensive natural language processing, pre-trained language models, instance-level adaptive knowledge usage
Compressor summary: The paper proposes IAPEK, a technique that adaptively retrieves external knowledge for PTLMs using Thrust, a metric that measures models' instance-level knowledgeability, achieving higher cost-efficiency and performance improvement on 26% of the tasks.
Soochan Lee, Jaehyeon Son, Gunhee Kim
https://openreview.net/forum?id=x816mCbWpR
Keywords: meta-continual learning, sequence modeling, Transformers, efficient Transformers
Compressor summary: The authors propose a novel approach to continual learning by framing it as a sequence modeling problem and demonstrate its effectiveness using Transformers and their efficient variants on various benchmarks.
Jiarui Jin, Xianyu Chen, Fanghua Ye, Mengyue Yang, Yue Feng, Weinan Zhang, Yong Yu, Jun Wang
https://openreview.net/forum?id=x7q7w07r6Y
Keywords: Conversational Agent, Recommender System, Conversational Recommendation
Compressor summary: CORE is a framework for chatbots to improve online recommendations by minimizing uncertainty through querying attributes or items.
Anudhyan Boral, Zhong Yi Wan, Leonardo Zepeda-Nunez, James Lottes, Qing Wang, Yi-Fan Chen, John Roberts Anderson, Fei Sha
https://openreview.net/forum?id=x6cOcxRnxG
Keywords: partial differential equations, physics, turbulence, stochastic differential equations, physical simulation, neural differential equations
Compressor summary: The authors propose a neural network approach that combines ideal large eddy simulation and stochastic differential equations to model turbulent flows, achieving better accuracy and stability than traditional methods.
Junbo Li, Ang Li, Chong Tian, Qirong Ho, Eric Xing, Hongyi Wang
https://openreview.net/forum?id=x5fs7TXKDc
Keywords: Federated learning, weight decay, adaptive hyperparameters
Compressor summary: The paper proposes FedNAR, a method to improve federated learning by adjusting weight decay and gradient co-clipping, which enhances convergence speed and model accuracy.
Oussama Boussif, Ghait Boukachab, Dan Assouline, Stefano Massaroli, Tianle Yuan, Loubna Benabbou, Yoshua Bengio
https://openreview.net/forum?id=x5ZruOa4ax
Keywords: Time series forecasting, multi-modal learning, solar irradiance, context-enriched learning
Compressor summary: The paper proposes a deep learning architecture using satellite data to improve solar irradiance forecasting accuracy and uncertainty estimation, as well as introducing a new multi-modal dataset for training and testing.
Dachao Lin, Yuze Han, Haishan Ye, Zhihua Zhang
https://openreview.net/forum?id=x5JCDCvR4b
Keywords: distributed optimization, convex optimization, second-order similarity, client sampling
Compressor summary: The paper proposes two new algorithms for distributed optimization problems with improved communication complexity and shows their effectiveness in smoothness-free and ill-conditioned scenarios.
Siyuan Xu, Minghui Zhu
https://openreview.net/forum?id=x2xQEszznV
Keywords: meta-learning; generalization
Compressor summary: The paper introduces an online meta-learning framework that learns from constrained tasks and analyzes its optimality gaps and constraint violations, with a practical algorithm and experiments on imitation learning and few-shot classification.
Andrea Cini, Ivan Marisca, Daniele Zambon, Cesare Alippi
https://openreview.net/forum?id=x2PH6q32LR
Keywords: time series forecasting, spatiotemporal forecasting, graph-based spatiotemporal forecasting, graph neural networks
Compressor summary: The paper explores the trade-off between global and local models in spatiotemporal forecasting using graph neural networks, and proposes a method to incorporate trainable node embeddings for better adaptation to different time series dynamics.
Jia Gu, Caizhi Tang, Han Yan, Qing Cui, Longfei Li, JUN ZHOU
https://openreview.net/forum?id=wzg0BsV8rQ
Keywords: Data fusion, heterogeneous treatment effects estimation, shrinkage estimation, tree-based method
Compressor summary: The paper introduces FAST, a method for estimating treatment effects that uses both trial and observational data to improve accuracy and robustness by combining optimal weighting and tree-based techniques.
Frederik Kunstner, Victor S. Portella, Mark Schmidt, Nick Harvey
https://openreview.net/forum?id=wzPcffMZ3b
Keywords: line-search, gradient descent, hypergradient, adaptive methods, smooth, convex, optimization, preconditioning
Compressor summary: The authors introduce multidimensional backtracking, a technique that finds optimal per-coordinate step-sizes for smooth convex problems by using hyper-gradients and cutting-plane methods.
Ruida Zhou, Tao Liu, Min Cheng, Dileep Kalathil, Panganamala Kumar, Chao Tian
https://openreview.net/forum?id=wxkBdtDbmH
Keywords: robust reinforcement learning, policy-based approach, function approximation, actor-critic
Compressor summary: The authors propose two new uncertainty sets for large-scale robust reinforcement learning and a robust natural actor-critic algorithm that converges to the optimal policy in finite time.
Ziqing Fan, Ruipeng Zhang, Jiangchao Yao, Bo Han, Ya Zhang, Yanfeng Wang
https://openreview.net/forum?id=wwmKVO8bsR
Keywords: federated learning, data heterogeneity, partially class-disjoint data
Compressor summary: FedGELA is a novel approach for federated learning with partially class-disjoint data that uses a simplex ETF classifier to achieve fair and equal discrimination for all classes, improving performance and providing convergence guarantees.
Raymond Feng, Flavio Calmon, Hao Wang
https://openreview.net/forum?id=wwkQUiaKbo
Keywords: algorithmic fairness, discrimination, missing values, machine learning
Compressor summary: The paper analyzes how missing values in data affect algorithmic fairness, showing that the standard "impute-then-classify" approach can worsen fairness and accuracy, and presents new algorithms that handle missing patterns while preserving their information for better fairness and accuracy.
Wenhu Chen, Hexiang Hu, YANDONG LI, Nataniel Ruiz, Xuhui Jia, Ming-Wei Chang, William W. Cohen
https://openreview.net/forum?id=wv3bHyQbX7
Keywords: Diffusion Model, Image Generation, Image Editing, In-Context Learning
Compressor summary: SuTI is a fast and high-quality text-to-image generator that learns from millions of online image clusters to produce customized images without fine-tuning for each subject.
Jacek Dmochowski
https://openreview.net/forum?id=wqIm0Qsgy0
Keywords: components analysis, unsupervised learning, Granger Causality
Compressor summary: The paper introduces a new unsupervised learning technique for time series data that uses Granger causality to identify driving and driven signals in multivariate data sets.
WEI W. XING, Yuxin Wang, Zheng Xing
https://openreview.net/forum?id=wpfsnu5syT
Keywords: Gaussian process, autoregression, multi fidelity, nonparametric Bayesian
Compressor summary: The authors propose ContinuAR, a novel method for multi-fidelity fusion that improves efficiency and performance compared to existing techniques.
Yashas Annadani, Nick Pawlowski, Joel Jennings, Stefan Bauer, Cheng Zhang, Wenbo Gong
https://openreview.net/forum?id=woptnU6fh1
Keywords: Causal Discovery, Structure Learning, Bayesian Inference, Variational Inference, MCMC, Generative Model
Compressor summary: The text introduces a new Bayesian causal discovery method that samples DAGs from the posterior without regularization, using a combination of SG-MCMC and VI, and applies gradient-based MCMC sampling for the first time in this field.
Gang Li, Wei Tong, Tianbao Yang
https://openreview.net/forum?id=wm5Ane9VRO
Keywords: Adversarial Average Precision Maximization, Robust Average Precision, Adversarial Ranking Robustness, Adversarial Training
Compressor summary: The paper proposes a new method to optimize both average precision and adversarial robustness, outperforming current methods by a significant margin.
Sheng-Yen Chou, Pin-Yu Chen, Tsung-Yi Ho
https://openreview.net/forum?id=wkIBfnGPTA
Keywords: backdoor, diffusion model, trustworthy
Compressor summary: The paper introduces a backdoor attack framework (VillanDiffusion) for diffusion models, which are generative AI models that can be vulnerable to output manipulation attacks when exposed to malicious patterns at the input.
Lin Li, Jun Xiao, Guikun Chen, Jian Shao, Yueting Zhuang, Long Chen
https://openreview.net/forum?id=wiv21EJ0Vd
Keywords: Visual relation detection, Zero-short learning, Scene graph generation
Compressor summary: RECODE is a novel zero-shot VRD method that uses composite description prompts to improve relation detection by leveraging LLMs and spatial information.
Congye Wang, Wilson Ye Chen, Heishiro Kanagawa, Chris J. Oates
https://openreview.net/forum?id=wiidCRA3at
Keywords: Bayesian, discrepancy, kernel, sampling, Stein's method
Compressor summary: The paper proposes a new method for improving Monte Carlo output using Stein importance sampling and provides a novel variational construction and convergence conditions for it.
Sloan Nietert, Ziv Goldfeld, Soroosh Shafiee
https://openreview.net/forum?id=wg3d2FKAm8
Keywords: distributionally robust optimization, robust statistics, optimal transport, Wasserstein distance
Compressor summary: The paper proposes a new robust optimization approach that handles both geometric and non-geometric data perturbations and shows improved performance on regression and classification tasks.
Silviu Pitis
https://openreview.net/forum?id=wcdF6jR0Sp
Keywords: Normative Agency Design, Reward Design, Sequential Decision Making, Reinforcement Learning, Intertemporal Fairness, Multi-Objective Decision Making
Compressor summary: The paper shows that optimal AI agents must use non-Markovian rewards when serving multiple objectives with different discount factors, and proposes a practical aggregation scheme for this purpose.
Tiancheng Jin, Junyan Liu, Haipeng Luo
https://openreview.net/forum?id=wbg4JEM5Jp
Keywords: multi-armed bandit, best of both worlds, Follow-the-Regularized-Leader, Tsallis entropy, Shannon entropy, Log-barrier
Compressor summary: The paper studies adaptive multi-armed bandit algorithms and improves the performance of FTRL with various regularizers and a new learning rate schedule, removing the uniqueness assumption and achieving better regret bounds.
Hugo Cui, Lenka Zdeborova
https://openreview.net/forum?id=wbbTqsiKzl
Keywords: statistical physics, replica method, autoencoder, exact asymptotics
Compressor summary: The paper proposes a two-layer non-linear autoencoder with tied weights and a skip connection for denoising high-dimensional data from a Gaussian mixture, and compares it to other methods.
Chirag Pabbaraju, Dhruv Rohatgi, Anish Sevekari, Holden Lee, Ankur Moitra, Andrej Risteski
https://openreview.net/forum?id=waXoG35kbb
Keywords: theory, score matching, exponential families, sample complexity, computational hardness
Compressor summary: The paper presents a new exponential family of distributions where score matching is as easy to optimize as maximum likelihood, unlike other cases where score matching is harder or intractable.
Chen Cheng, Gary Cheng, John Duchi
https://openreview.net/forum?id=waDF0oACu2
Keywords: Collaborative Learning, Missing Data, Sensors, Linear Regression
Compressor summary: The paper proposes Collab, a distributed, semi-supervised algorithm for learning least squares estimates for m agents with different features, without communicating the labeled data, achieving near-optimal performance and generalizability.
Zixing Song, Yifei Zhang, Irwin King
https://openreview.net/forum?id=wYkfog48Bq
Keywords: Graph-based Semi-supervised Learning, Affinity Graph Construction
Compressor summary: This paper proposes an optimal asymmetric graph structure for label inference in graph-based semi-supervised learning that differentiates the roles of labeled and unlabeled nodes, leading to improved performance on synthetic and real datasets.
An Dinh Vuong, Minh Nhat VU, Toan Tien Nguyen, Baoru Huang, Dzung Nguyen, Thieu Vo, Anh Nguyen
https://openreview.net/forum?id=wYKU1C77sa
Keywords: scene synthesis, language-driven, diffusion models, multi-conditional generation, 3D point cloud
Compressor summary: The paper proposes a new language-driven scene synthesis task that combines text, human motion, and objects, and presents a multi-conditional diffusion model to solve it effectively.
Hengyu Fu, Tianyu Guo, Yu Bai, Song Mei
https://openreview.net/forum?id=wX8GuzDSJR
Keywords: transformers, attention, deep learning theory, random features
Compressor summary: The paper analyzes how attention layers in Transformers learn and generalize, showing they can efficiently learn certain functions with random features and have advantages over standard neural networks.
Abhra Chaudhuri, Massimiliano Mancini, Zeynep Akata, Anjan Dutta
https://openreview.net/forum?id=wUNPmdE273
Keywords: Interpretability, Robustness, Fine-Grained Representation Learning, Graph Theory, Information Theory
Compressor summary: The paper presents a method called TRD that converts abstract relational representations in image analysis into interpretable graphs and shows its effectiveness and robustness to noisy views.
Ling Yang, Jingwei Liu, Shenda Hong, Zhilong Zhang, Zhilin Huang, Zheming Cai, Wentao Zhang, Bin CUI
https://openreview.net/forum?id=wRhLd65bDt
Keywords: Diffusion Model, Image Generation
Compressor summary: ConPreDiff is a novel diffusion model that predicts neighborhood context for better image synthesis and achieves state-of-the-art results in text-to-image generation.
Darshil Doshi, Tianyu He, Andrey Gromov
https://openreview.net/forum?id=wRJqZRxDEX
Keywords: Criticality, Gaussian Process, Jacobian, LayerNorm, Residual connections, ResNet
Compressor summary: The authors propose a method to diagnose criticality in deep neural networks using partial Jacobians, which helps select optimal initialization and show that LayerNorm and residual connections lead to a critical architecture for any initialization.
Johan Samir Obando Ceron, Marc G Bellemare, Pablo Samuel Castro
https://openreview.net/forum?id=wPqEvmwFEh
Keywords: Reinforcement Learning, Deep Reinforcement Learning, Value based, Batch Size
Compressor summary: Reducing the batch size in value-based deep reinforcement learning can improve performance, contrary to the common belief that larger batches are better.
Don Dennis, Abhishek Shetty, Anish Sevekari, Kazuhito Koishida, Virginia Smith
https://openreview.net/forum?id=wNxyDofh74
Keywords: Edge computing, compression, efficient inference, distillation and inference, run-time tradeoff, inference-time tradeoff, on-device, user-side, client-side
Compressor summary: B-DISTIL is a method to create ensembles of smaller student models from a large pretrained teacher model using function composition based aggregation rules, achieving similar performance while allowing flexibility in accuracy vs. inference cost trade-offs.
Yulhwa Kim, Dongwon Jo, Hyesung Jeon, Taesu Kim, Daehyun Ahn, Hyungjun Kim, jae-joon kim
https://openreview.net/forum?id=wNpsGwixjG
Keywords: diffusion models, post-training quantization
Compressor summary: The paper proposes a faster noise estimation method for diffusion models by using low-bit activations in the early reverse diffusion process, while keeping high-bit activations for later stages to maintain image quality.
Weijie Tu, Weijian Deng, Tom Gedeon
https://openreview.net/forum?id=wMNpMe0vp3
Keywords: CLIP
Compressor summary: This paper investigates the safety measures of CLIP models, focusing on their resilience to visual factor variations, uncertainty estimations, and anomaly detection, by testing them on various datasets and conditions.
Maciej Falkiewicz, Naoya Takeishi, Imahn Shekhzadeh, Antoine Wehenkel, Arnaud Delaunoy, Gilles Louppe, Alexandros Kalousis
https://openreview.net/forum?id=wLiMhVJ7fx
Keywords: simulation-based inference, inverse problem, bayesian inference, uncertainty quantification, generative modeling
Compressor summary: Bayesian inference uses probability to express uncertainty, but existing methods can be overconfident; a new method adds calibration error to neural models for more accurate uncertainty quantification.
Mengzhao Wang, Lingwei Lv, Xiaoliang Xu, Yuxiang Wang, Qiang Yue, Jiongkang Ni
https://openreview.net/forum?id=wLFXTAWa5V
Keywords: approximate nearest neighbor search, attribute filtering, high-dimensional vector, proximity graph
Compressor summary: The paper presents a fast and accurate framework for hybrid query processing using composite indexes based on proximity graphs, improving both approximate nearest neighbor search and attribute filtering.
Ajil Jalal, Justin Kang, Ananya Uppal, Kannan Ramchandran, Eric Price
https://openreview.net/forum?id=wImYhdu4VF
Keywords: Generative models, distribution learning, maximum likelihood estimation
Compressor summary: A conditional generative model is a method to sample from a conditional distribution without assumptions on the input distribution, and it can learn deep models using few samples.
Yuanyuan Liu, Fanhua Shang, Weixin An, Junhao Liu, Hongying Liu, Zhouchen Lin
https://openreview.net/forum?id=wIlmx4bHrO
Keywords: Constrained Minimax Optimization; nonconvex- nonconcave
Compressor summary: The paper presents a new algorithm for solving constrained nonconvex-nonconcave problems that improves convergence rates and complexity bounds compared to existing methods.
Zhiyong Wang, Jize Xie, Tong Yu, Shuai Li, John C.S. Lui
https://openreview.net/forum?id=wHhPIv5G8Q
Keywords: online learning, online corrupted user detection, clustering of bandits
Compressor summary: The paper proposes an online learning problem and two novel algorithms to learn from user relations and detect corrupted users in web systems.
Yilun Xu, Mingyang Deng, Xiang Cheng, Yonglong Tian, Ziming Liu, Tommi S. Jaakkola
https://openreview.net/forum?id=wFuemocyHZ
Keywords: Generative models, diffusion models, PFGM, sampling
Compressor summary: Restart is a novel sampling algorithm that combines ODE and SDE methods to achieve faster and higher quality results in generative models.
Alexandre Capone, Sandra Hirche, Geoff Pleiss
https://openreview.net/forum?id=wFH5hZAwYz
Keywords: Gaussian Processes, Frequentist Statistics, Kernel Methods, Model Selection and Structure Learning, Regression
Compressor summary: The paragraph describes a new method for improving the uncertainty estimates of Gaussian processes by using a different set of hyperparameters that better satisfy calibration constraints and yield tighter predictive quantiles.
Yuankai Luo, Lei Shi, Veronika Thost
https://openreview.net/forum?id=wEiUGpcr0M
Keywords: Graph Neural Networks, Molecular Representation Learning, Persistent Homology, Contrastive Learning, Self-supervised Learning
Compressor summary: The paper proposes a self-supervised learning method for molecular representation learning using persistent homology, which offers unique features and improves predictive power, especially in small datasets.
Giannis Daras, Kulin Shah, Yuval Dagan, Aravind Gollakota, Alex Dimakis, Adam Klivans
https://openreview.net/forum?id=wBJBLy9kBY
Keywords: corrupted data, generative models, ambient gan, inverse problems, learning from measurements
Compressor summary: The paper proposes a diffusion-based method to learn an unknown distribution using highly-corrupted samples, which can train generative models that are less likely to memorize individual samples and work on standard benchmarks with significant pixel loss or corrupted datasets.
Yudi Zhang, Yali Du, Biwei Huang, Ziyan Wang, Jun Wang, Meng Fang, Mykola Pechenizkiy
https://openreview.net/forum?id=w7TyuWhGZP
Keywords: Reinforcement learning, sparse reward, return decomposition, causal modeling
Compressor summary: This paper proposes GRD, a framework that uses causal generative models to redistribute rewards in reinforcement learning for delayed reward scenarios and improves policy optimization by training over the most favorable state subspace.
Fangzhou Lin, Yun Yue, Ziming Zhang, Songlin Hou, Kazunori Yamada, Vijaya B Kolachalama, Venkatesh Saligrama
https://openreview.net/forum?id=w7LxAZfDfv
Keywords: Contrastive learning; Point cloud completion
Compressor summary: The paper introduces InfoCD, a contrastive Chamfer distance loss that improves point cloud completion and surface similarity estimation by maximizing mutual information between underlying surfaces, achieving state-of-the-art results on benchmark datasets.
Timur Garipov, Sebastiaan De Peuter, Ge Yang, Vikas Garg, Samuel Kaski, Tommi S. Jaakkola
https://openreview.net/forum?id=w79RtqIyoM
Keywords: generative model composition, GFlowNets, diffusion models, classifier guidance, probabilistic methods
Compressor summary: The paper proposes Compositional Sculpting, a method for combining and sampling from iterative generative models like GFlowNets and diffusion models using classifier guidance.
Caroline Lee, Jane Han, Ma Feilong, Guo Jiahui, James Haxby, Christopher Baldassano
https://openreview.net/forum?id=w6krZiUa7t
Keywords: Brain Imaging, Other Cognitive Science, Other Neuroscience
Compressor summary: The Hyper-HMM is a hybrid model that aligns both temporal and spatial features across brains when processing naturalistic stimuli, allowing for mapping of individual cognitive dynamics and semantic content.
Dat Do, Huy Nguyen, Khai Nguyen, Nhat Ho
https://openreview.net/forum?id=w3ghbKBJg4
Keywords: Mixture Model, Minimax Rate, Maximum Likelihood Estimation
Compressor summary: The paper studies how to estimate parameters in a multivariate model with a mixed density function, using a new condition called \emph{distinguishability} to deal with challenges in convergence rates.
Yite Wang, Jing Wu, Naira Hovakimyan, Ruoyu Sun
https://openreview.net/forum?id=w2F8Fm6Sg3
Keywords: Dynamics sparse training; pruning; neural network pruning; empirical deep learning
Compressor summary: The paper introduces a new metric and method for balancing sparse training in generative adversarial networks (GANs) to improve efficiency without sacrificing performance.
Idan Attias, Steve Hanneke, Alkis Kalavasis, Amin Karbasi, Grigoris Velegkas
https://openreview.net/forum?id=w116w62fxH
Keywords: Learning Theory, Regression, PAC Learning, Online Learning
Compressor summary: The authors study the complexity of learning realizable regression in different settings, introduce new dimensions to characterize learnability, and design optimal learners for these settings.
Haotian Liu, Chunyuan Li, Qingyang Wu, Yong Jae Lee
https://openreview.net/forum?id=w0H2xGHlkw
Keywords: visual instruction tuning, instruction tuning, multimodal, LLM, GPT
Compressor summary: The paper introduces LLaVA, a multimodal model that uses language to generate image instructions for GPT-4, improving its performance on various tasks, including Science QA.
Ryan-Rhys Griffiths, Leo Klarner, Henry Moss, Aditya Ravuri, Sang T. Truong, Yuanqi Du, Samuel Don Stanton, Gary Tom, Bojana Ranković, Arian Rokkum Jamasb, Aryan Deshwal, Julius Schwartz, Austin Tripp, Gregory Kell, Simon Frieder, Anthony Bourached, Alex James Chan, Jacob Moss, Chengzhi Guo, Johannes P. Dürholt, Saudamini Chaurasia, Ji Won Park, Felix Strieth-Kalthoff, Alpha Lee, Bingqing Cheng, Alan Aspuru-Guzik, Philippe Schwaller, Jian Tang
https://openreview.net/forum?id=vzrA6uqOis
Keywords: Gaussian processes, Bayesian optimization, Chemistry, Molecular Machine Learning, Applications, Software
Compressor summary: GAUCHE is an open-source library that uses Gaussian processes to optimize molecular representations for applications like molecular discovery, chemical reaction optimization, and protein design.
Lyndon Duong, Eero P Simoncelli, Dmitri Chklovskii, David Lipshutz
https://openreview.net/forum?id=vz7SdRqWGM
Keywords: neuroscience, adaptation, whitening, efficient coding, recurrent neural network, gain modulation, synaptic plasticity, local learning rules
Compressor summary: The paragraph describes a new model that combines synaptic plasticity and gain modulation to rapidly adapt sensory neurons to changing input statistics.
Lingchen Meng, Xiyang Dai, Jianwei Yang, Dongdong Chen, Yinpeng Chen, Mengchen Liu, Yi-Ling Chen, Zuxuan Wu, Lu Yuan, Yu-Gang Jiang
https://openreview.net/forum?id=vybQs1Gbuk
Keywords: Long-tail object detection, visual semantics, soft supervision
Compressor summary: RichSem is a simple and effective method for long-tailed object detection that uses semantic branch to learn rich semantics from coarse locations as additional soft supervision without needing accurate bounding boxes.
Ziyi Huang, Henry Lam, Amirhossein Meisami, Haofeng Zhang
https://openreview.net/forum?id=vwr4bHHsRT
Keywords: Bayesian bandits, approximate Bayesian inference, Bayesian Upper Confidence Bound, optimal regret order, bounded inference error
Compressor summary: The EBUCB framework improves on existing Bayesian bandit algorithms by achieving optimal regret bounds under constant approximate inference error using two different $\alpha$-divergences, while previous methods only had worst-case linear regret with one $\alpha$-divergence.
Wenliang Dai, Junnan Li, Dongxu Li, Anthony Tiong, Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, Steven Hoi
https://openreview.net/forum?id=vvoWPYqZJA
Keywords: Vision-Language Models, Instruction Tuning, Zero-shot
Compressor summary: The paper explores vision-language instruction tuning using pretrained BLIP-2 models and introduces a Query Transformer, achieving state-of-the-art results on various tasks.
Shenzhi Wang, Qisen Yang, Jiawei Gao, Matthieu Gaetan Lin, HAO CHEN, Liwei Wu, Ning Jia, Shiji Song, Gao Huang
https://openreview.net/forum?id=vtoY8qJjTR
Keywords: reinforcement learning, offline-to-online reinforcement learning, offline reinforcement learning, policy improvement, policy constraint
Compressor summary: FamO2O is a framework that adapts improvement and constraint balances for offline-to-online reinforcement learning based on each state, improving performance over existing methods.
Agustinus Kristiadi, Felix Dangel, Philipp Hennig
https://openreview.net/forum?id=vtLNwa6uX0
Keywords: neural network, invariance, equivariance, reparametrization, riemannian geometry, parameter space
Compressor summary: The paper studies how to represent the metric of neural nets under reparametrization using Riemannian geometry, which has implications for flatness analysis and optimization.
Denis Tarasov, Vladislav Kurenkov, Alexander Nikulin, Sergey Kolesnikov
https://openreview.net/forum?id=vqGWslLeEw
Keywords: Offline Reinforcement Learning
Compressor summary: The authors analyze recent offline reinforcement learning algorithms, propose ReBRAC that integrates several design elements, and show its state-of-the-art performance in various settings and datasets.
Xihan Li, Xiang Chen, Rasul Tutunov, Haitham Bou Ammar, Lei Wang, Jun Wang
https://openreview.net/forum?id=vq11gurmUY
Keywords: Self-consistent Field Equation, Computational Science, Online PCA
Compressor summary: The authors propose a novel approach to solve the SCF equation using online PCA techniques, which improves convergence and applies to electronic structure problems.
Wenhao Wang, Yifan Sun, Wei Li, Yi Yang
https://openreview.net/forum?id=vpQuCsZXz2
Keywords: hierarchical image classification, hierarchical prompting, vision transformer
Compressor summary: The paper proposes a novel hierarchical image classification method that uses tokenized ancestor-class hints to improve accuracy, efficiency, and explainability of image classification models.
Felix Chalumeau, Shikha Surana, Clément Bonnet, Nathan Grinsztajn, Arnu Pretorius, Alexandre Laterre, Thomas D Barrett
https://openreview.net/forum?id=vpMBqdt9Hl
Keywords: Reinforcement Learning, Combinatorial Optimization, TSP, CVRP, JSSP
Compressor summary: COMPASS is a novel RL approach that pre-trains diverse policies in a latent space to improve search performance for combinatorial optimization problems.
Lorenzo Baldassari, Ali Siahkoohi, Josselin Garnier, Knut Solna, Maarten V. de Hoop
https://openreview.net/forum?id=voG6nEW9BV
Keywords: score-based generative models, diffusion models, inverse problems, bayesian inference, infinite dimensions
Compressor summary: The paper proposes a new method for solving infinite-dimensional Bayesian linear inverse problems using amortized conditional score-based diffusion models, which improves upon previous heuristic approaches by being more theoretically grounded and computationally efficient.
Sarah Asad Toonsi, Jeff S Shamma
https://openreview.net/forum?id=vnTUuecp2v
Keywords: Learning in games, Nash equilibrium, Uncoupled Dynamics
Compressor summary: The paper introduces higher-order gradient play dynamics for multi-agent learning that can lead to Nash Equilibrium or fail depending on the game and dynamics.
Zeke Xie, zhiqiang xu, Jingzhao Zhang, Issei Sato, Masashi Sugiyama
https://openreview.net/forum?id=vnGcubtzR1
Keywords: Weight Decay, Regularization, Optimization, Deep Learning
Compressor summary: The paper proposes a new method called Scheduled Weight Decay (SWD) to reduce large gradient norms caused by weight decay, which can improve convergence and generalization in deep neural networks.
Chenhang Cui, Yazhou Ren, Jingyu Pu, Jiawei Li, Xiaorong Pu, Tianyi Wu, Yutao Shi, Lifang He
https://openreview.net/forum?id=vlDbqzwczj
Keywords: multi-view learning, clustering
Compressor summary: SUMVC is a new information-theoretic approach for multi-view clustering that improves performance by reducing redundancy and enhancing consistent information across views.
Afra Amini, Li Du, Ryan Cotterell
https://openreview.net/forum?id=vf77fTbgG3
Keywords: Natural Language Processing, Text Generation, Controlled Generation, MCMC, HMC, Langevin Dynamics
Compressor summary: The paper introduces Structured Voronoi Sampling (SVS), a principled gradient-based technique for sampling from language models that generates fluent and diverse texts while adhering to control targets.
Milena Gazdieva, Alexander Korotin, Daniil Selikhanovych, Evgeny Burnaev
https://openreview.net/forum?id=vZRiMjo826
Keywords: optimal transport, partial optimal transport, neural networks, domain translation
Compressor summary: The paper introduces extremal transport, a method for unpaired image domain translation based on neural optimal transport and inspired by theoretical limits of possible translations.
Junguang Jiang, Baixu Chen, Junwei Pan, Ximei Wang, Dapeng Liu, jie jiang, Mingsheng Long
https://openreview.net/forum?id=vZHk1QlBQW
Keywords: Auxiliary-Task Learning, Negative Transfer
Compressor summary: ForkMerge is a novel approach that helps mitigate negative transfer in auxiliary-task learning by periodically forking the model into multiple branches and dynamically merging them based on task weights.
Junren Chen, Jonathan Scarlett, Michael Ng, Zhaoqiang Liu
https://openreview.net/forum?id=vUXNNLatFv
Keywords: Compressed sensing, generative models, nonlinearity, uniform recovery
Compressor summary: The paper presents a unified framework for uniform recovery guarantees in nonlinear generative compressed sensing, handling discontinuous or unknown observation models, and achieves similar performance to existing non-uniform results with little additional cost.
Guanghui Wang, Zihao Hu, Vidya Muthukumar, Jacob Abernethy
https://openreview.net/forum?id=vTug54Uunq
Keywords: Implicit bias, margin maximization, zero-sum game, online learning
Compressor summary: The paper proposes new techniques to improve the implicit bias rates of generic optimization methods like mirror descent and steepest descent in binary classification problems, using a unified framework based on regularized bilinear games.
Zhanpeng Zhou, Yongyi Yang, Xiaojiang Yang, Junchi Yan, Wei Hu
https://openreview.net/forum?id=vORUHrVEnH
Keywords: Linear Mode Connectivity, Permutation Invariance, Optimization Landscape, Science of Deep Learning
Compressor summary: This paper introduces a concept called Layerwise Linear Feature Connectivity (LLFC), which shows how different neural networks can be connected by linear paths in their feature maps, based on the Linear Mode Connectivity phenomenon.
Haoru Tan, Sitong Wu, Fei Du, Yukang Chen, Zhibin Wang, Fan Wang, XIAOJUAN QI
https://openreview.net/forum?id=vO6ZdPWaHc
Keywords: Data Valuation, Deep Learning, Data Pruning, Coreset Selection.
Compressor summary: The paper presents MoSo, a data-pruning method that uses gradient information to identify and remove uninformative samples from the training set, improving performance at high pruning ratios.
Ziniu Li, Tian Xu, Zeyu Qin, Yang Yu, Zhi-Quan Luo
https://openreview.net/forum?id=vO04AzsB49
Keywords: imitation learning, distribution shift, policy optimization, data selection
Compressor summary: The paper presents a framework for imitation learning with supplementary data that addresses distribution shift issues using importance sampling to improve performance in various tasks.
Fuqi Jia, Yuhang Dong, Minghao Liu, Pei Huang, Feifei Ma, Jian Zhang
https://openreview.net/forum?id=vNsdFwjPtL
Keywords: Reinforcement Learning, Graph Neural Network, Cylindrical Algebraic Decomposition.
Compressor summary: The paper introduces two reinforcement learning methods with graph neural networks for suggesting variable order in cylindrical algebraic decomposition, improving efficiency and generalizing well to different datasets.
Arun Verma, Zhongxiang Dai, Yao Shu, Bryan Kian Hsiang Low
https://openreview.net/forum?id=vM5VnNQ4n7
Keywords: Parameterized Bandits, Auxiliary Feedback, Control Variate, Regret Minimization
Compressor summary: The paper proposes a method that uses auxiliary feedback to improve reward estimation, leading to reduced regret in parameterized bandits problems.
Dong-Kyum Kim, Jea Kwon, Meeyoung Cha, C. Justin Lee
https://openreview.net/forum?id=vKpVJxplmB
Keywords: Transformer, NMDA, long-term memory, reference memory, memory consolidation
Compressor summary: The paper proposes a new activation function for deep learning models that mimics the hippocampus's NMDA receptor dynamics, improving working memory and spatial representation in transformers.
Ruichen Jiang, Aryan Mokhtari
https://openreview.net/forum?id=vIGNYQ4Alv
Keywords: convex optimization, quasi-Newton methods, Monteiro-Svaiter acceleration, Nesterov's accelerated gradient, online learning
Compressor summary: The paper presents a new optimization method that beats Nesterov's accelerated gradient in certain regimes and achieves optimal or faster convergence rates based on an online learning approach.
Nikita Kornilov, Ohad Shamir, Aleksandr Lobanov, Darina Dvinskikh, Alexander Gasnikov, Innokentiy Andreevich Shibaev, Eduard Gorbunov, Samuel Horváth
https://openreview.net/forum?id=vHSQTEIFkp
Keywords: stochastic optimization, gradient-free optimization, zero-order oracle, gradient clipping, infinite variance
Compressor summary: The paper presents an optimal algorithm for non-smooth stochastic convex optimization with infinite noise variance, adapted from a clipped accelerated gradient method.
Dengwei Zhao, Shikui Tu, Lei Xu
https://openreview.net/forum?id=vHRLS8HhK1
Keywords: Monte Carlo Tree Search, Reinforcement learning, Path consistency.
Compressor summary: The paper proposes GW-PCZero, a generalized version of PCZero that uses a weighting mechanism to reduce variance and improve learning efficiency in reinforcement learning with neural-guided MCTS for real applications with non-zero immediate reward.
Nikhil Parthasarathy, S. M. Ali Eslami, Joao Carreira, Olivier J Henaff
https://openreview.net/forum?id=vF8ukt5l1R
Keywords: self-supervised learning, contrastive, video pretraining, representation learning, visual representation, human alignment, robustness, shape-bias, saliency
Compressor summary: Video pretraining using VITO can learn visually general, robust, and human-like representations from complex transformations in videos.
Federico Errica
https://openreview.net/forum?id=vEzcRdiTkP
Keywords: Deep Graph Networks, Graph Neural Networks, Graph Representation Learning, Nearest Neighbors, Node Classification, Tabular Data
Compressor summary: The paper questions the benefits of using nearest neighbor graphs as a way to represent missing graph structures in machine learning problems and suggests exploring other techniques instead.
Gleb Rodionov, Liudmila Prokhorenkova
https://openreview.net/forum?id=vBwSACOB3x
Keywords: neural algorithmic reasoning, graph neural networks, self-supervised regularization
Compressor summary: This paper proposes an unsupervised method to learn neural algorithms that can generalize to larger input sizes and achieve state-of-the-art results on classic algorithmic problems like sorting.
Shinji Ito, Kei Takemura
https://openreview.net/forum?id=vBHKSTgcYQ
Keywords: bandit, linear bandit, best of both worlds, exploration by optimization
Compressor summary: The paper presents a linear bandit algorithm that performs near-optimally in both stochastic and adversarial settings, using exploration by optimization with specific regret bounds depending on dimension, time horizon, and sub-optimality gap.
Noah Shinn, Federico Cassano, Ashwin Gopinath, Karthik R Narasimhan, Shunyu Yao
https://openreview.net/forum?id=vAElhFcKW6
Keywords: language model, reasoning, decision making, programming
Compressor summary: Reflexion is a framework that helps language agents learn from feedback by reflecting on it verbally and storing it in memory for better decision-making in future trials, improving performance across various tasks.
Shitao Tang, Fuyang Zhang, Jiacheng Chen, Peng Wang, Yasutaka Furukawa
https://openreview.net/forum?id=vA0vj1mY77
Keywords: multiview; image generation; generative model; diffusion models
Compressor summary: MVDiffusion is a method for generating multi-view images with high resolution and rich content by concurrently interacting with all views using a correspondence-aware attention mechanism, which includes modules for generation, interpolation, and super-resolution.
Peter Súkeník, Marco Mondelli, Christoph H Lampert
https://openreview.net/forum?id=v9yC7sSXf3
Keywords: neural collapse, unconstrained features model, deep learning
Compressor summary: The paper studies deep neural collapse, a phenomenon where neural networks lose structure in earlier layers, and proposes a generalized analytical framework to explain it.
Taehyun Cho, Seungyub Han, Heesoo Lee, Kyungjae Lee, Jungwoo Lee
https://openreview.net/forum?id=v8u3EFAyW9
Keywords: distributional reinforcement learning, risk
Compressor summary: The paper proposes a new distributional reinforcement learning method that uses randomized risk criterion for exploration and proves its convergence, optimality, and effectiveness in various environments.
Alexander Shmakov, Kevin Greif, Michael James Fenton, Aishik Ghosh, Pierre Baldi, Daniel Whiteson
https://openreview.net/forum?id=v7WWesSiOu
Keywords: Diffusion, Variational, VAE, LDM, Physics, Unfolding
Compressor summary: The paragraph discusses a new deep learning method for correcting detector effects in particle physics measurements at the Large Hadron Collider, which outperforms existing methods in accuracy and efficiency.
Cyrille Kone, Emilie Kaufmann, Laura Richert
https://openreview.net/forum?id=v6jIxRRDyD
Keywords: bandit, pure-exploration, pareto front, pareto set
Compressor summary: The paper proposes a new sampling strategy called Adaptive Pareto Exploration that can identify a relevant subset of the Pareto optimal set in multi-objective bandit problems, and demonstrates its effectiveness on vaccination strategies against Covid-19.
Tanya Marwah, Ashwini Pokle, J Zico Kolter, Zachary Chase Lipton, Jianfeng Lu, Andrej Risteski
https://openreview.net/forum?id=v6YzxwJlQn
Keywords: Deep Equilibrium Models, Partial Differential Equations, Neural Operators
Compressor summary: The authors propose FNO-DEQ, a deep learning architecture that solves steady-state PDEs more efficiently and robustly than existing methods by using weight-tied neural networks and exploiting the fixed point nature of their solutions.
Nathan Grinsztajn, Daniel Furelos-Blanco, Shikha Surana, Clément Bonnet, Thomas D Barrett
https://openreview.net/forum?id=v6VpqGcGAR
Keywords: Combinatorial Optimization, Reinforcement Learning, TSP, CVRP, JSSP
Compressor summary: The paper introduces Poppy, a training method for populations of reinforcement learning policies that induce unsupervised specialization to solve hard combinatorial optimization problems effectively.
Huafeng Kuang, Hong Liu, YONGJIAN WU, Shin'ichi Satoh, Rongrong Ji
https://openreview.net/forum?id=v5Aaxk4sSy
Keywords: Information Bottleneck, Adversarial training, Adversarial robustness, Knowledge distillation
Compressor summary: The study proposes an Information Bottleneck Distillation approach that uses prior knowledge from a pre-trained model to improve the information bottlenecks and robustness of deep neural networks against adversarial attacks.
Can Qin, Shu Zhang, Ning Yu, Yihao Feng, Xinyi Yang, Yingbo Zhou, Huan Wang, Juan Carlos Niebles, Caiming Xiong, Silvio Savarese, Stefano Ermon, Yun Fu, Ran Xu
https://openreview.net/forum?id=v54eUIayFh
Keywords: Image Generation, Multi-modal, HyperNet
Compressor summary: UniControl is a new generative foundation model that can handle various visual conditions and generate images with precise controls, while still allowing for arbitrary language prompts.
Jinrang Jia, Zhenjia Li, Yifeng Shi
https://openreview.net/forum?id=v2oGdhbKxi
Keywords: 3D detection, deep learning, autonomous driving
Compressor summary: The paper proposes a unified optimization target called normalized depth for monocular 3D detection of vehicle and infrastructure sides in autonomous driving, and introduces 3D normalized cube depth to improve accuracy.
Yi Yu, Xue Yang, Qingyun Li, Yue Zhou, Feipeng Da, Junchi Yan
https://openreview.net/forum?id=v1VVKaMYbk
Keywords: Oriented object detection, self-supervised learning
Compressor summary: H2RBox-v2 is a new method for oriented object detection that uses reflection symmetry and self-supervision to improve performance and reduce the need for high-quality annotations.
Xiaotian Liu, Hector Palacios, Christian Muise
https://openreview.net/forum?id=v0lkbp66Uw
Keywords: Embodied AI, High-Level Actions, Symbolic Reasoning, Replanning, ALFRED Challenge, Flexible Task Achievement, User-Goal Understanding, Object Types and Actions, Perception Grounding
Compressor summary: Egocentric Planning combines symbolic planning and Object-oriented POMDPs to solve complex domestic tasks with visual perception and natural language processing, achieving a high success rate in ALFRED benchmark and winning the CVPR Embodied AI workshop challenge.
Jianwei Tang, Jiangxin Sun, Xiaotong Lin, lifang zhang, Wei-Shi Zheng, Jian-Fang Hu
https://openreview.net/forum?id=v0GzRLvVp3
Keywords: Human Motion Prediction; Temporal Continual Learning; Prior Compensation Factor
Compressor summary: The paper proposes a new framework called Temporal Continual Learning (TCL) for human motion prediction that better preserves prior information and adapts to different models, datasets, and applications.
Abishek Sankararaman, Murali Balakrishnan
https://openreview.net/forum?id=uzOBDerK1j
Keywords: Estimation, heavy-tails, distribution shifts, regret
Compressor summary: The text describes a new algorithm for estimating time-varying parameters from noisy and corrupted high-dimensional data-streams that is adaptive to drift, robust to heavy-tails and corruptions, requires no distributional knowledge, and can be implemented online.
Sotiris Anagnostidis, Dario Pavllo, Luca Biggio, Lorenzo Noci, Aurelien Lucchi, Thomas Hofmann
https://openreview.net/forum?id=uvdJgFFzby
Keywords: Transformers, Context-pruning, Efficient Transformer
Compressor summary: The study proposes a method that prunes uninformative tokens from context during generation, reducing computational and memory requirements while maintaining model expressiveness and interpretability.
Man Zhou, Naishan Zheng, Yuan Xu, Chun-Le Guo, Chongyi Li
https://openreview.net/forum?id=uv3ge0goPa
Keywords: Image restoration, low-light image enhancement, image de-noising
Compressor summary: The study proposes using random weight networks as constraints for training better image restoration networks, and explores four prototypes with different strategies and variants.
Alexandru Tifrea, Gizem Yüce, Amartya Sanyal, Fanny Yang
https://openreview.net/forum?id=utreNaM1VY
Keywords: semi-supervised learning, statistical lower bound
Compressor summary: The paragraph discusses the limitations of existing theory on semi-supervised learning (SSL) algorithms, which may not always improve upon unsupervised and supervised learning algorithms, but can still perform well in real-world scenarios.
Liyao Tang, Zhe Chen, Shanshan Zhao, Chaoyue Wang, Dacheng Tao
https://openreview.net/forum?id=utQms7PPx5
Keywords: point cloud segmentation, weak supervision
Compressor summary: The paper proposes ERDA, a learning strategy that uses entropy regularization and distribution alignment to improve weakly supervised 3D segmentation tasks with pseudo-labels, achieving state-of-the-art performance with minimal annotations.
Xinyu Ma, Xu Chu, Yasha Wang, Yang Lin, Junfeng Zhao, Liantao Ma, Wenwu Zhu
https://openreview.net/forum?id=uqkUguNu40
Keywords: Graph Data Augmentation, Graph Mixup, Fused Gromov Wasserstein
Compressor summary: The paper introduces FGWMixup, a graph data augmentation method that uses optimal transport to find an optimal inter-graph node matching strategy, improving the generalizability and robustness of GNNs.
Paul Geuchen, Felix Voigtlaender
https://openreview.net/forum?id=uotGmrcooz
Keywords: complex-valued neural networks, approximation rates
Compressor summary: The paper analyzes the expressivity and approximation properties of complex-valued neural networks (CVNNs) with various activation functions, showing that their error scales as $m^{-k/(2n)}$ for smooth target functions and is optimal under a continuity assumption.
Neeratyoy Mallik, Eddie Bergman, Carl Hvarfner, Danny Stoll, Maciej Janowski, Marius Lindauer, Luigi Nardi, Frank Hutter
https://openreview.net/forum?id=uoiwugtpCH
Keywords: Hyperparameter Optimization, Deep Learning
Compressor summary: PriorBand is an HPO algorithm for DL that combines expert knowledge and cheap proxy tasks to efficiently optimize hyperparameters.
Veronica Alvarez, Santiago Mazuelas, Jose A. Lozano
https://openreview.net/forum?id=uoRiO855Sj
Keywords: Concept drift, Continual learning, Minimax classification, Performance guarantees
Compressor summary: IMRCs use forward and backward learning to improve classification accuracy on evolving tasks with few samples per task.
Yiheng Zhu, Jialu Wu, Chaowen Hu, Jiahuan Yan, Chang-Yu Hsieh, Tingjun Hou, Jian Wu
https://openreview.net/forum?id=uoG1fLIK2s
Keywords: drug discovery, multi-objective molecular optimization, Bayesian optimization, generative flow networks
Compressor summary: The paper introduces HN-GFN, a novel algorithm that uses a single hypernetwork to optimize multiple objectives and diversity in molecule design, and a hindsight-like strategy to improve learning and sharing of high-quality molecules.
Khai Nguyen, Nhat Ho
https://openreview.net/forum?id=umvV3yvo4N
Keywords: Sliced Wasserstein, Monte Carlo Methods, Point-Cloud, Optimal Transport
Compressor summary: The paper introduces a new way to choose the slicing distribution for the Wasserstein distance, called energy-based sliced Wasserstein (EBSW), which improves on previous approaches in terms of discrimination, cost, and stability.
Qiang Gao, Xiaojun Shan, Yuchen Zhang, Fan Zhou
https://openreview.net/forum?id=uj9PxVTVqq
Keywords: data-free subnetwork, task-incremental learning, knowledge transfer, mask
Compressor summary: DSN is a novel method for neuron-wise incremental learning that transfers knowledge across tasks using masks and data-free replay, avoiding forgetting and privacy issues.
Binhui Xie, Shuang Li, qingju guo, Chi Harold Liu, Xinjing Cheng
https://openreview.net/forum?id=uiiVSVADDc
Keywords: Active Learning, LiDAR Semantic Segmentation, Domain Adaptation
Compressor summary: Annotator is an active learning method that uses voxel-centric online selection to efficiently annotate LiDAR point clouds for semantic segmentation under various distribution shifts.
Ashok Cutkosky, Aaron Defazio, Harsh Mehta
https://openreview.net/forum?id=uhKtQMn21D
Keywords: optimization, deep learning, online convex optimization
Compressor summary: Mechanic is a technique for automatically adjusting the learning rate in any optimization algorithm, which performs well in various deep learning tasks and can match or surpass manual tuning.
Eric Nguyen, Michael Poli, Marjan Faizi, Armin W Thomas, Michael Wornow, Callum Birch-Sykes, Stefano Massaroli, Aman Patel, Clayton M. Rabideau, Yoshua Bengio, Stefano Ermon, Christopher Re, Stephen Baccus
https://openreview.net/forum?id=ubzNoJjOKj
Keywords: genomics, hyena, foundation models, large language models, transformers
Compressor summary: HyenaDNA is a large language model that can process up to 1 million tokens of human genome data, enabling long-range interactions and single nucleotide resolution for genomic tasks, and outperforms previous models on various benchmarks.
Yibo Jiang, Bryon Aragam, Victor Veitch
https://openreview.net/forum?id=ubp5s2tgXq
Keywords: embedding, representation, graphical models, partial orthogonality, Markov boundary
Compressor summary: The paper explores how semantic structure is encoded in vector embeddings using partial orthogonality and introduces embeddings that preserve conditional independence structures.
Yanyu Li, Huan Wang, Qing Jin, Ju Hu, Pavlo Chemerys, Yun Fu, Yanzhi Wang, Sergey Tulyakov, Jian Ren
https://openreview.net/forum?id=ubgdInLSF9
Keywords: Text-to-Image, Diffusion model, mobile devices, distillation
Compressor summary: The authors introduce a generic approach to make text-to-image diffusion models faster and more accessible on mobile devices, improving efficiency by modifying the network architecture and step distillation, and achieving better results than Stable Diffusion v$1.5$.
Ning Liu, Siavash Jafarzadeh, Yue Yu
https://openreview.net/forum?id=ubap5FKbJs
Keywords: Operator-Regression Neural Networks, Neural Operators, Data-Driven Physics Modeling, Geometrical and Topological Shape Changes
Compressor summary: The authors propose a novel neural operator architecture, DAFNO, that can learn nonlinear mappings between function spaces on irregular geometries and evolving domains using FFT, achieving state-of-the-art accuracy in material modeling and airfoil simulation, and demonstrating generalizability to unseen crack patterns.
Evgenii E Chzhen, Christophe Giraud, Zhen LI, Gilles Stoltz
https://openreview.net/forum?id=uZvG0HLkOB
Keywords: mutli-armed bandits, bandits with knapsacks, primal-dual approaches
Compressor summary: The paper proposes a new algorithm for contextual bandit problems with knapsacks that can handle cost constraints of the order of $\sqrt{T}$, which is more efficient and simpler than existing approaches.
Xiuzhe Wu, Peng Dai, Weipeng DENG, Handi Chen, Yang Wu, Yan-Pei Cao, Ying Shan, XIAOJUAN QI
https://openreview.net/forum?id=uZjpSBTPik
Keywords: Neural Radiance Field; Continual Learning; Scene Representation
Compressor summary: The paper introduces CL-NeRF, a method for efficiently adapting Neural Radiance Fields to scene changes using continual learning and a new benchmark for evaluating it.
Pengchong Hu, Zhizhong Han
https://openreview.net/forum?id=uWNqy09dFW
Keywords: 3D Reconstruction, SDF, Neural Rendering, Implicit Representations, SLAM
Compressor summary: The paragraph introduces a novel method for 3D reconstruction from multi-view images using neural implicit representations, volume rendering with an attentive depth fusion prior, and a Truncated Signed Distance Function (TSDF).
Saurabh Garg, Amrith Setlur, Zachary Chase Lipton, Sivaraman Balakrishnan, Virginia Smith, Aditi Raghunathan
https://openreview.net/forum?id=uWGH6jDTVv
Keywords: contrastive learning; self training; distribution shift; semi supervised learning; unsupervised domain adaptation
Compressor summary: The paper investigates how combining self-training and contrastive learning techniques affects unsupervised domain adaptation and semi-supervised learning, showing improved accuracy in the former but not the latter.
Hassan Akbari, Dan Kondratyuk, Yin Cui, Rachel Hornung, Huisheng Wang, Hartwig Adam
https://openreview.net/forum?id=uTlKUAm68H
Keywords: Alternating Gradient Descent, Multimodal, Mixture of Experts, AGD, MoE, Deep Learning, Optimization
Compressor summary: IMP is a multimodal multi-task training method that uses alternating gradient descent and mixture-of-experts to efficiently improve model performance on various tasks, achieving state-of-the-art results in zero-shot video classification.
Boyuan Chen, Chuning Zhu, Pulkit Agrawal, Kaiqing Zhang, Abhishek Gupta
https://openreview.net/forum?id=uRewSnLJAa
Keywords: deep reinforcement learning, self-supervised learning
Compressor summary: The paper proposes a self-supervised RL method that combines model-free and model-based approaches for transferring behaviors across tasks with different rewards in complex environments.
Erik Lien Bolager, Iryna Burak, Chinmay Datar, Qing Sun, Felix Dietrich
https://openreview.net/forum?id=uRHpgo6TMR
Keywords: random sampling, neural network parameters, iterative optimization
Compressor summary: The authors propose a method for quickly training fully-connected neural networks without gradient computations by using a data-dependent probability distribution and an efficient sampling algorithm.
William Merrill, Ashish Sabharwal
https://openreview.net/forum?id=uR8TtWCIsr
Keywords: transformers, logic, reasoning, circuit complexity, mechanistic interpretability
Compressor summary: The paper investigates whether log-precision transformers, which can attend universally, can be represented by a generalized first-order logic with majority-vote quantifiers.
Sang Michael Xie, Shibani Santurkar, Tengyu Ma, Percy Liang
https://openreview.net/forum?id=uPSQv0leAu
Keywords: Language models, pretraining, data selection, fine-tuning
Compressor summary: DSIR is a data selection method that uses importance resampling in a reduced feature space to match a target distribution, achieving comparable or better results than expert curation or simple heuristics.
Marcel Torne Villasevil, Max Balsells I Pamies, Zihan Wang, Samedh Desai, Tao Chen, Pulkit Agrawal, Abhishek Gupta
https://openreview.net/forum?id=uOEeui0rL7
Keywords: Learning from human preferences, self-supervised learning, exploration in reinforcement learning
Compressor summary: HUGE is a technique that uses low-quality human feedback to guide exploration for reinforcement learning without requiring reward specification or exploration bonuses.
Zhenyu Zhu, Francesco Locatello, Volkan Cevher
https://openreview.net/forum?id=uNnPWR66b8
Keywords: Causal discovery, Score matching, Score-based generative modeling
Compressor summary: The paper studies how well deep neural networks can estimate a score function for causal discovery using score-matching methods.
Mufang Ying, Koulik Khamaru, Cun-Hui Zhang
https://openreview.net/forum?id=uNmKBZrRZC
Keywords: bandit algorithm, statistical inference, adaptively collected data, asymptotic normality
Compressor summary: The paper proposes a debiased estimator for sequential data collection that improves statistical inference by achieving both non-asymptotic performance and asymptotic normality.
Lee M. Gunderson, Gecia Bravo-Hermsdorff, Peter Orbanz
https://openreview.net/forum?id=uN71BdBEG8
Keywords: Stochastic block model, SBM, graphons, matrix pencil method, method of moments
Compressor summary: The paper presents a method to find the parameters of a stochastic block model from a finite set of subgraph densities using a unique translation map that is efficient and easy to compute.
Bo Li, Fangxiao Wang, Yu Zhou
https://openreview.net/forum?id=uJmsYZiu3E
Keywords: fair allocation of chores, beyond additive cost functions, bin packing, job scheduling
Compressor summary: The paper explores the limitations of achieving exact or good approximations of maximin share fairness in allocating indivisible tasks with subadditive costs and studies two specific settings, bin packing and job scheduling, where constant approximate allocations exist.
Sriram Balasubramanian, Gaurang Sriramanan, Vinu Sankar Sadasivan, Soheil Feizi
https://openreview.net/forum?id=uJ3qNIsDGF
Keywords: Neural networks, Vision models, blind spots, undersensitivity, invariance, level set geometry, input connectivity
Compressor summary: The authors study how vision models like CNNs and Transformers are under-sensitive to input perturbations and propose a method to explore the geometry of "equi-confidence" level sets in these networks, revealing a star-like structure.
Zheng Zhang, Qi Liu, Hao Jiang, Fei Wang, Yan Zhuang, Le Wu, Weibo Gao, Enhong Chen
https://openreview.net/forum?id=uFpjPJMkv6
Keywords: fairness, user modeling
Compressor summary: The paper proposes a new framework called FairLISA to create user models that are fair and accurate even when some sensitive data is missing or unknown.
Haoqing Wang, Shibo Jie, Zhi-Hong Deng
https://openreview.net/forum?id=uFlE0qgtRO
Keywords: few-shot image classification, fine-tuning, vision transformers
Compressor summary: The paper proposes a method to improve few-shot image classification with pre-trained vision transformers by using attention and gradient information to locate key entities in support images, which helps the model focus on class-related features and generalize better.
CHENXU ZHAO, Wei Qian, Zhitao Ying, Mengdi Huai
https://openreview.net/forum?id=uEJfW3OtUm
Keywords: Selective forgetting, static setting, sequential setting, security and robustness
Compressor summary: The paper explores malicious data update attacks on machine unlearning systems and proposes new attack methods and algorithms to exploit them.
Lin Yang, Junlong Lyu, Wenlong Lyu, Zhitang Chen
https://openreview.net/forum?id=uDV4lA0gZ6
Keywords: bayesian optimization, robust optimization
Compressor summary: The paper proposes AIRBO, a robust Bayesian Optimization algorithm that models uncertain inputs using Maximum Mean Discrepancy and Nystrom approximation, achieving state-of-the-art performance under various input uncertainties.
Mathieu Even, Scott Pesme, Suriya Gunasekar, Nicolas Flammarion
https://openreview.net/forum?id=uAyElhYKxg
Keywords: SGD, GD, implicit bias, large stepsizes, edge of stability, diagonal linear networks
Compressor summary: The paper studies how randomness and big steps affect gradient descent and stochastic gradient descent in linear networks with two layers, proving their convergence and analysing how these factors influence solution quality in regression problems.
Anikait Singh, Aviral Kumar, Quan Vuong, Yevgen Chebotar, Sergey Levine
https://openreview.net/forum?id=u8srPlinoj
Keywords: offline RL, support constraints, heteroskedastic data
Compressor summary: Offline reinforcement learning methods struggle to learn from data with non-uniform variability due to distribution constraints; conservative Q-learning (CQL) with reweighting improves performance by allowing the learned policy to choose how closely to follow the behavior policy per state.
Haonan Duan, Adam Dziedzic, Nicolas Papernot, Franziska Boenisch
https://openreview.net/forum?id=u6Xv3FuF8N
Keywords: differential privacy, in-context learning, trustworthy ML
Compressor summary: The authors propose a method to privately learn how to prompt large language models without exposing the data used for prompting, which can maintain high accuracy while protecting user privacy.
Yuechen ZHANG, Jinbo Xing, Eric Lo, Jiaya Jia
https://openreview.net/forum?id=u6Ibs4hTJH
Keywords: image variation, diffusion model, image generation, text-driven image editing
Compressor summary: RIVAL is a novel inference pipeline that generates high-quality image variations by aligning the diffusion process to the source image's inversion chain using cross-image self-attention and step-wise distribution normalization.
Aveen Dayal, Vimal K B, Linga Reddy Cenkeramaddi, C Krishna Mohan, Abhinav Kumar, Vineeth N. Balasubramanian
https://openreview.net/forum?id=u6BYyPuD29
Keywords: Domain Generalization, Margin Loss, Adversarial Learning, Domain Adaptation
Compressor summary: The paper proposes a new adversarial learning method for domain generalization called MADG, which uses a margin loss-based discrepancy metric to learn domain-invariant features and achieves better generalization to unseen domains.
Kai Yi, Bingxin Zhou, Yiqing Shen, Pietro Lio, Yu Guang Wang
https://openreview.net/forum?id=u4YXKKG5dX
Keywords: inverse folding, graph neural networks, roto-translation equivariance, diffusion model
Compressor summary: The proposed graph denoising diffusion model generates diverse amino acid sequences for a given protein backbone using physiochemical properties, local environment, and prior knowledge of amino acid replacements, outperforming existing methods in inverse protein folding.
Evren Gokcen, Anna Ivic Jasper, Alison Xu, Adam Kohn, Christian K. Machens, Byron M. Yu
https://openreview.net/forum?id=u39QQh5L8Q
Keywords: neuroscience, multi-population neural recordings, dimensionality reduction, latent variable models, Gaussian processes
Compressor summary: The paragraph describes a new framework for analyzing signals among multiple neuronal populations in different brain networks, which reveals how they communicate and evolve over time.
Weida Li, Yaoliang Yu
https://openreview.net/forum?id=u359tNBpxF
Keywords: data valuation, robustness, weighted Banzhaf values
Compressor summary: The paper proposes a new robust data valuation method using weighted Banzhaf values and introduces Kronecker noise to analyze stochasticity.
Radoslav Dimitrov, Zeyang Zhao, Ralph Abboud, Ismail Ilkan Ceylan
https://openreview.net/forum?id=u2RJ0I3o3j
Keywords: Graph Representation Learning; Planar Graphs; Graph Property Prediction
Compressor summary: The paragraph introduces PlanE, a framework for learning complete invariants of planar graphs, which are special graph classes with efficient isomorphism testing algorithms, and reports its successful performance on benchmarks.
Mher Safaryan, Alexandra Peste, Dan Alistarh
https://openreview.net/forum?id=tzxP9Rx0LV
Keywords: knowledge distillation, stochastic optimization, variance reduction
Compressor summary: The paper explores how knowledge distillation works as a stochastic variance reduction technique for enhancing student models with teacher models, and highlights the importance of careful parametrization for optimal performance.
SungYub Kim, Kyungsu Kim, Eunho Yang
https://openreview.net/forum?id=tz4ECtAu8e
Keywords: Influence Function, Geometric Ensemble, Loss Landscape
Compressor summary: The paper proposes a new method to improve Influence Function (IF) approximations for neural networks by addressing limitations and enhancing performance in various tasks.
Muyang Li, Runze Wu, Haoyu Liu, Jun Yu, Xun Yang, Bo Han, Tongliang Liu
https://openreview.net/forum?id=txv7TnPvOi
Keywords: semi-supervised learning, pseudo-labeling
Compressor summary: The paper proposes a new SSL method that uses instance-dependent thresholds to select confident unlabeled instances and assign pseudo-labels with probabilistic guarantees.
Zhiyuan Zhang, Deli Chen, Hao Zhou, Fandong Meng, Jie Zhou, Xu Sun
https://openreview.net/forum?id=txPdKZrrZF
Keywords: federated learning, backdoor learning, robust federated aggregation, data divergence
Compressor summary: The paper proposes a method called Federated F-Divergence-Based Aggregation (Fed-FA) to detect and exclude suspicious clients in federated NLP systems by modeling data divergence using f-divergence indicators and synthetic datasets.
Alexander H. Liu, Heng-Jui Chang, Michael Auli, Wei-Ning Hsu, James R. Glass
https://openreview.net/forum?id=twmHKU3Ds4
Keywords: speech representation learning, self-supervised learning, self-distillation, discrete representation learning
Compressor summary: The paper presents a self-supervised speech representation learning model called DinoSR that combines masked language modeling, self-distillation, and online clustering to achieve state-of-the-art performance in various downstream tasks.
Shay Moran, Hilla Schefler, Jonathan Shafer
https://openreview.net/forum?id=tw4QaiiJex
Keywords: Algorithmic stability, Replicability, Differential Privacy, KL Stability, Mutual Information Stability, Global Stability, Perfect Generalization, PAC Learning, Littlestone Dimension, Clique Dimension, PAC Bayes
Compressor summary: The paper explores the connections between different definitions of stability in learning theory, including various types of privacy, generalization, and divergence-based stability notions.
Shengzhuang Chen, Long-Kai Huang, Jonathan Richard Schwarz, Yilun Du, Ying Wei
https://openreview.net/forum?id=tt7bQnTdRm
Keywords: meta-generalization, out-of-distribution tasks
Compressor summary: The EBML framework uses two neural networks to characterize any meta-training task distribution, enabling detection and adaptation of out-of-distribution tasks for safe and effective meta-learning.
Changsheng Lv, Shuai Zhang, Yapeng Tian, Mengshi Qi, Huadong Ma
https://openreview.net/forum?id=trHfuGQyyr
Keywords: Physical Audiovisual;Commonsense Reasoning
Compressor summary: The paper proposes a Disentangled Counterfactual Learning (DCL) method for inferring physical commonsense from audiovisual data, which decouples video factors and models causal relationships using a variational autoencoder and a counterfactual learning module.
Ulyana Piterbarg, Lerrel Pinto, Rob Fergus
https://openreview.net/forum?id=tp2nEZ5zfP
Keywords: imitation learning, NetHack
Compressor summary: The paper studies why neural policy learning struggles in NetHack and proposes improvements based on a winning symbolic agent's strategy, leading to a better neural agent that still falls behind symbolic and human players.
Meshi Bashari, Amir Epstein, Yaniv Romano, Matteo Sesia
https://openreview.net/forum?id=toYvRJ7Zmy
Keywords: Conformal inference, Derandomization, E-values, False discovery rate, Out-of-distribution testing, Testing for outliers, Uncertainty
Compressor summary: The authors propose a method to make conformal inference more stable for novelty detection by using conformal e-values instead of p-values and weighting them with additional data information.
Youngsoo Jang, Geon-Hyeong Kim, Jongmin Lee, Sungryull Sohn, Byoungjip Kim, Honglak Lee, Moontae Lee
https://openreview.net/forum?id=toEGuA9Qfn
Keywords: Imitation learning, Preference-based learning, Safe imitation learning
Compressor summary: The paper introduces SafeDICE, an offline imitation learning algorithm that learns a safe policy from non-preferred demonstrations and unlabeled data by estimating stationary distribution corrections.
Xuefeng Du, Yiyou Sun, Jerry Zhu, Yixuan Li
https://openreview.net/forum?id=tnRboxQIec
Keywords: Outlier imagination, machine learning
Compressor summary: The paper introduces Dream-OOD, a framework that generates realistic outlier images for machine learning models using in-distribution data and classes, without requiring manual data collection or cleaning.
David Lüdke, Marin Biloš, Oleksandr Shchur, Marten Lienen, Stephan Günnemann
https://openreview.net/forum?id=tn9Dldam9L
Keywords: Point Processes, Diffusion, Temporal Data, Generative Model, Forecasting, Density Estimation, Denoising
Compressor summary: The paper introduces ADD-THIN, a probabilistic denoising diffusion model that improves long-term forecasting for event sequences with discrete and continuous components by using temporal point process frameworks.
Yi-Chung Chen, Hsi-Wen Chen, Shun-Guei Wang, Ming-Syan Chen
https://openreview.net/forum?id=tmxjuIFSEc
Keywords: Federated Learning, Contribution Evaluation, Shapley Value, Knowledge Amalgamation
Compressor summary: The paper introduces SPACE, an efficient method for evaluating participant contributions in federated learning using Federated Knowledge Amalgamation and Prototype-based Model Evaluation.
Ioannis Panageas, Nikolas Patris, Stratis Skoulakis, Volkan Cevher
https://openreview.net/forum?id=tkenkPYkxj
Keywords: fictitious play, convergence rate, potential games
Compressor summary: This paper studies the convergence rate of Fictitious Play in potential games and shows that it can take exponential time even for two-agent games.
Taiji Suzuki, Denny Wu, Kazusato Oko, Atsushi Nitanda
https://openreview.net/forum?id=tj86aGVNb3
Keywords: mean-field regime, feature learning, Neural network optimization, sparse parity function, classification, sample complexity
Compressor summary: The paragraph discusses how mean-field Langevin dynamics (MFLD) can optimize and improve generalization performance of neural networks for binary classification problems by learning features, and provides a new analysis of its sample complexity and convergence rate.
Jean Kaddour, Oscar Key, Piotr Nawrot, Pasquale Minervini, Matt Kusner
https://openreview.net/forum?id=thbXgJ8gNK
Keywords: language models, transformers, efficient training
Compressor summary: The authors investigate various efficient training algorithms for Transformer-based language models, but find that a fully-decayed learning rate outperforms them all.
Yukang Yang, Dongnan Gui, Yuhui Yuan, Weicong Liang, Haisong Ding, Han Hu, Kai Chen
https://openreview.net/forum?id=thPI8hrA4V
Keywords: Generative Models, Visual Text Generation, Diffusion Models
Compressor summary: The paper introduces a new method called GlyphControl to improve visual text generation by using glyph instructions with an off-the-shelf Stable-Diffusion model and creates a dataset called LAION-Glyph for further research.
Yizi Zhang, Tianxiao He, Julien Boussard, Charlie Windolf, Olivier Winter, Eric M. Trautmann, Noam Roth, Hailey Barrel, Mark M Churchland, Nick Steinmetz, Erdem Varol, Cole Lincoln Hurwitz, Liam Paninski
https://openreview.net/forum?id=tgQRMrsxht
Keywords: neural decoding, brain-computer interfaces, spike sorting, variational inference, generative models
Compressor summary: The authors propose a new method for decoding neural activity without spike sorting, using mixture of Gaussians to model uncertainty and improve performance in brain-computer interfaces.
Hongxin Li, Jingran Su, Yuntao Chen, Qing Li, Zhaoxiang Zhang
https://openreview.net/forum?id=tfyr2zRVoK
Keywords: Large Language Model; Task Planning; Embodied AI; Robotics; Software Automation
Compressor summary: The authors propose SheetCopilot, an agent that uses natural language to control spreadsheets and complete various tasks, and present a dataset and evaluation pipeline to benchmark its performance.
Chengbin Du, Yanxi Li, Zhongwei Qiu, Chang Xu
https://openreview.net/forum?id=tesBViWnbx
Keywords: Adversarial Attack, Generative Model, Diffusion Model, Latent Diffusion Model, Conditional Latent Diffusion Model
Compressor summary: Auto-attack on Text-to-image Models (ATM) is a method to make text-to-image models more robust by generating perturbations that can prevent them from blending or losing primary subjects in generated images.
Zuhao Yang, Yingfang Yuan, Yang Xu, SHUO ZHAN, Huajun Bai, Kefan Chen
https://openreview.net/forum?id=tdyLryDebq
Keywords: natural language generation; evaluation metrics; cross-entropy; language model
Compressor summary: The proposed FACE metrics use Fourier Analysis of Cross-Entropy to measure similarity between model-generated and human language, effectively identifying the gap between them and reflecting various factors such as model size and sampling methods.
Seungjae Lee, Daesol Cho, Jonghae Park, H. Jin Kim
https://openreview.net/forum?id=tcotyjon2a
Keywords: Reinforcement Learning, Curriculum Learning, Goal-conditioned RL
Compressor summary: The authors propose a new method for curriculum RL that automatically defines a semantic goal space from continuous observations and suggests uncertainty and temporal distance-aware goals, improving exploration and performance on goal-reaching tasks.
David Lindner, Janos Kramar, Sebastian Farquhar, Matthew Rahtz, Thomas McGrath, Vladimir Mikulik
https://openreview.net/forum?id=tbbId8u7nP
Keywords: interpretability, transformers, language models, RASP, Tracr, mechanistic interpretability
Compressor summary: The authors introduce a compiler called Tracr that generates structured transformer models from human-readable programs, enabling experiments and interpretability evaluations.
Manan Tomar, Riashat Islam, Matthew E. Taylor, Sergey Levine, Philip Bachman
https://openreview.net/forum?id=tW2KSph9o8
Keywords: representation learning, mutual information
Compressor summary: Information gating is a method for learning efficient and robust representations by selectively revealing or hiding information depending on the task at hand.
Wenxuan Ma, Shuang Li, Lincan Cai, Jingxuan Kang
https://openreview.net/forum?id=tUyW68cRqr
Keywords: Data-Efficient Learning, Language Semantic Graph
Compressor summary: The paragraph discusses a new approach called Language Semantic Graph (LSG) that uses semantic information from labels to improve data efficiency in machine learning tasks, and demonstrates its effectiveness across different modalities and scenarios.
Samuel Dooley, Gurnoor Singh Khurana, Chirag Mohapatra, Siddartha Venkat Naidu, Colin White
https://openreview.net/forum?id=tScBQRNgjk
Keywords: Forecasting, Zero-shot, Synthetic Data
Compressor summary: ForecastPFN is a new zero-shot forecasting model that uses synthetic data and Bayesian inference to make more accurate and faster predictions than existing methods, even with limited real-life observations.
Andi Peng, Mycal Tucker, Eoin M. Kenny, Noga Zaslavsky, Pulkit Agrawal, Julie Shah
https://openreview.net/forum?id=tSEeRl7ACo
Keywords: human-in-the-loop, representation learning, interpretability
Compressor summary: The authors propose training neural networks to generate a spectrum of discrete representations and control their complexity based on the task, which leads to better generalization and can be informed by human intuition.
Zitang Sun, Yen-Ju Chen, Yung-Hao Yang, Shin'ya Nishida
https://openreview.net/forum?id=tRKimbAk5D
Keywords: motion perception, optical flow estimation, attention mechanism, psychophysics, In silico neurophysiology, human vision
Compressor summary: The paper proposes a new image-computable model that combines biological and computer vision approaches to simulate human motion perception in complex scenes.
Fenggen Yu, Qimin Chen, Maham Tanveer, Ali Mahdavi Amiri, Hao Zhang
https://openreview.net/forum?id=tQYGjnxPOm
Keywords: 3D reconstruction, constructive solid geometry, unsupervised learning, compact shape assembly
Compressor summary: D$^2$CSG is a neural model that learns to reconstruct 3D CAD shapes using quadric primitives and a dedicated residual branch for complex shape complements.
Qinyi Chen, Negin Golrezaei, Djallel Bouneffouf
https://openreview.net/forum?id=tP50lLiZIo
Keywords: non-stationary bandits; autoregressive model; low-regret policy; online learning algorithms
Compressor summary: The paper introduces a new non-stationary MAB framework that considers temporal dynamics in real-world applications like recommendation systems, and proposes an algorithm with two mechanisms for exploration-exploitation and discarding outdated information.
Bhavya Sukhija, Lenart Treven, Cansu Sancaktar, Sebastian Blaes, Stelian Coros, Andreas Krause
https://openreview.net/forum?id=tLrkjK128n
Keywords: Active Exploration, Reinforcement Learning, Dynamical Systems
Compressor summary: OPAX is an algorithm for active exploration in reinforcement learning that uses probabilistic models to quantify uncertainty and maximizes information gain, enabling zero-shot solutions for multiple tasks.
Emanuele Marconato, Stefano Teso, Antonio Vergari, Andrea Passerini
https://openreview.net/forum?id=tLTtqySDFb
Keywords: Neuro-Symbolic Integration, Trustworthy AI, Concept Learning, Learning Shortcuts, Mitigation Strategies
Compressor summary: The paragraph discusses how neuro-symbolic predictive models can suffer from reasoning shortcuts that compromise their advantages, and proposes mitigation strategies to address this issue.
Zixing Lei, Yiming Zhang, Yuxin Xiong, Siheng Chen
https://openreview.net/forum?id=tLEDsaKuDh
Keywords: Emergent communication, Interactive, Question Answering
Compressor summary: The authors propose a new task (ISQA) and system for vision-based emergent communication, which uses sketches to answer questions about images in multiple rounds, improving accuracy, complexity, and interpretability.
Can Chen, Christopher Beckham, Zixuan Liu, Xue Liu, Christopher Pal
https://openreview.net/forum?id=tJwyg9Zg9G
Keywords: offline model-based optimization, bi-level optimization
Compressor summary: The authors propose parallel-mentoring, a novel method that uses majority voting and adaptive soft-labeling to improve model-based optimization with multiple proxies, addressing out-of-distribution issues in black-box objective function maximization.
Pengjie Gu, Xinyu Cai, Dong Xing, Xinrun Wang, Mengchen Zhao, Bo An
https://openreview.net/forum?id=tJN664ZNVG
Keywords: Offline RL, POMDP
Compressor summary: ORDER is a probabilistic framework for offline RL that uses discrete state representations and proxy representations to improve robustness against diverse masked observabilities.
Boris van Breugel, Nabeel Seedat, Fergus Imrie, Mihaela van der Schaar
https://openreview.net/forum?id=tJ88RBqupo
Keywords: model evaluation, tabular, synthetic data
Compressor summary: The paper proposes 3S Testing, a method that uses generative models to create synthetic test data for small subgroups and simulate distributional shifts, improving the evaluation of machine learning model performance on diverse and underrepresented groups.
Zuobai Zhang, Minghao Xu, Aurelie Lozano, Vijil Chenthamarakshan, Payel Das, Jian Tang
https://openreview.net/forum?id=tIzbNQko3c
Keywords: Protein representation learning, diffusion models, self-supervised learning
Compressor summary: The paper proposes a pre-training method for protein encoders using joint sequence-structure diffusion models and enhances it with a technique to capture conformational variations among protein conformers.
Jianglin Lu, Yi Xu, Huan Wang, Yue Bai, Yun Fu
https://openreview.net/forum?id=tGuMwFnRZX
Keywords: Latent Graph Inference, CUR Matrix Decomposition, Graph Neural Networks
Compressor summary: The paper proposes a method to improve latent graph inference by restoring corrupted affinities and replenishing missed supervision for better node representation learning.
Shang Liu, Zhongze Cai, Xiaocheng Li
https://openreview.net/forum?id=tGPx7HdBr4
Keywords: Regression calibration, model recalibration, conditional quantile, nonparametric method
Compressor summary: The paper proposes nonparametric methods for individual calibration of regression models that are efficient, consistent, and provide theoretical guarantees, addressing limitations of existing heuristic methods.
Arnav Kumar Jain, Lucas Lehnert, Irina Rish, Glen Berseth
https://openreview.net/forum?id=tFsxtqGmkn
Keywords: Reinforcement Learning, Maximum state entropy exploration, Non-Markovian exploration, Successor Representation
Compressor summary: The authors propose a new exploration algorithm, $\eta\psi$-Learning, that learns efficient policies by using past experience to predict state visitation entropy and maximize coverage.
Sayantan Choudhury, Eduard Gorbunov, Nicolas Loizou
https://openreview.net/forum?id=tFeaLw9AWn
Keywords: Optimization, Machine Learning, Extragradient Methods, Min-Max Optimization
Compressor summary: The paper studies single-call stochastic extragradient methods for solving large-scale min-max optimization and variational inequalities problems, improves their convergence guarantees, and applies them to two classes of structured non-monotone VIPs using the expected residual condition.
Xinwen Zhang, Yihan Zhang, Tianbao Yang, Richard Souvenir, Hongchang Gao
https://openreview.net/forum?id=tF7W8ai8J3
Keywords: federated learning, compositional optimization, minimax optimization, AUC maximization
Compressor summary: The paragraph discusses a novel federated learning method for imbalanced data that optimizes the AUC score and provides theoretical and empirical evidence of its effectiveness.
Zhizhang Yuan, Daoze Zhang, Yang Yang, Junru Chen, Yafeng Li
https://openreview.net/forum?id=tEmFyqjaJh
Keywords: Brain signal, Seizure detection, Pretraining, Domain generalization
Compressor summary: The study proposes a patient-independent seizure detection model using stereoelectroencephalography (SEEG) data that overcomes challenges like domain shift and pattern evolution with novel self-supervised tasks and domain generalization techniques.
Christian Schilling, Anna Lukina, Emir Demirović, Kim Guldstrand Larsen
https://openreview.net/forum?id=tEKBU5XOTw
Keywords: safety verification, decision tree, reinforcement learning, controller, continuous time
Compressor summary: The paper introduces an algorithm to verify the safety of decision tree controlled systems in continuous time, overcoming the challenges of existing methods.
John P Dickerson, Seyed A. Esmaeili, Jamie Heather Morgenstern, Claire Jie Zhang
https://openreview.net/forum?id=tECyQO1QOp
Keywords: Fairness, Clustering, Approximation Algorithms
Compressor summary: The paper investigates how two common fairness criteria for clustering, group fairness and diversity in center selection, can be simultaneously satisfied using a constant approximation algorithm, and shows that they are compatible with some but not all distance-based fairness notions.
Ruozi Huang, Xipeng Wu, Hongsheng Yu, Zhong Fan, Haobo Fu, QIANG FU, Yang Wei
https://openreview.net/forum?id=tDAu3FPJn9
Keywords: StarCraft II, league training, AlphaStar, opponent-modeling, reinforcement learning
Compressor summary: The paper presents improvements to AlphaStar's league training for StarCraft II, enabling better and more resource-efficient AI agents with advanced exploitation and opponent modeling abilities.
Xinyu Mao, Jiapeng Zhang
https://openreview.net/forum?id=tC0r8duG9z
Keywords: Clustering Algorithms, Stochastic Block Model, Spectral Algorithms
Compressor summary: The paper investigates how spectral methods like SVD can improve clustering algorithms and shows that it works well for a specific type of network model called the stochastic block model.
ERIC EATON, Marcel Hussing, Michael Kearns, Jessica Sorrell
https://openreview.net/forum?id=tBwRbgsol1
Keywords: Reinforcement Learning, Learning Theory, Replicability, Reproducibility
Compressor summary: The text discusses the development of algorithm frameworks for replicability in various fields and introduces provably replicable algorithms for parallel value iteration and R-Max in episodic reinforcement learning.
Haobo Zhang, Junyuan Hong, Yuyang Deng, Mehrdad Mahdavi, Jiayu Zhou
https://openreview.net/forum?id=tBib2fWr3r
Keywords: Deep Learning, Privacy, Federated Learning, Influence Function
Compressor summary: The paper proposes a method called Inversion Influence Function (I$^2$F) to analyze and understand the privacy leakage caused by Deep Gradient Leakage (DGL) attack in distributed learning settings, and show its effectiveness on various scenarios.
Yan Zhuang, Qi Liu, GuanHao Zhao, Zhenya Huang, Weizhe Huang, Zachary Pardos, Enhong Chen, Jinze Wu, Xin Li
https://openreview.net/forum?id=tAwjG5bM7H
Keywords: adaptive learning, computerized adaptive testing, educational measurement, cognitive diagnosis
Compressor summary: The paper proposes a new method (BECAT) for computerized adaptive testing that selects questions to estimate student's ability more accurately and efficiently by matching the gradient of full responses.
Hongzheng Yang, Cheng Chen, Yueyao Chen, Markus Scheppach, Hon Chi Yip, Qi Dou
https://openreview.net/forum?id=t9Swbo82dB
Keywords: uncertainty estimation, semantic segmentation, medical application
Compressor summary: The paper proposes a novel framework that uses reinforcement learning and reward optimization to improve uncertainty estimation in deep segmentation models for safety-critical applications like medical imaging.
Gen Luo, Yiyi Zhou, Tianhe Ren, Shengxin Chen, Xiaoshuai Sun, Rongrong Ji
https://openreview.net/forum?id=t877958UGZ
Keywords: vision-language instruction tuning, multimodal LLM, efficient training
Compressor summary: The paper proposes a cost-effective method called Mixture-of-Modality Adaptation (MMA) to improve the multimodal capabilities of large language models for vision-language tasks, achieving competitive performance and efficiency with low expenditure.
Zhicheng Sun, Yadong MU
https://openreview.net/forum?id=t7ozN4AXd0
Keywords: continual learning, reinforcement learning, brain-inspired learning
Compressor summary: The paragraph describes a novel rewiring approach for the human brain in continual reinforcement learning that promotes adaptivity and exploration, with a focus on stability and plasticity-stability tradeoffs.
Tristan Deleu, Mizu Nishikawa-Toomey, Jithendaraa Subramanian, Nikolay Malkin, Laurent Charlin, Yoshua Bengio
https://openreview.net/forum?id=t7lnhhi7De
Keywords: bayesian network, bayesian, structure learning, causal discovery, gflownet
Compressor summary: The paper proposes JSP-GFN, a method that uses GFlowNets to approximate the joint posterior over Bayesian Network structure and parameters, achieving good results on various datasets.
Yan Xia, Hai Huang, Jieming Zhu, Zhou Zhao
https://openreview.net/forum?id=t7ZowrDWVw
Keywords: multi-modal, discrete representation, mutual information estimation
Compressor summary: The paper proposes a novel task called Cross Modal Generalization (CMG) that learns a unified discrete representation from paired multimodal data and introduces Uni-Code, which contains two key contributions for fine-grained unified representation of multimodal sequences.
Andrew Campbell, William Harvey, Christian Dietrich Weilbach, Valentin De Bortoli, Tom Rainforth, Arnaud Doucet
https://openreview.net/forum?id=t6nA7x3GAC
Keywords: diffusion, score-based, score, markov chain, jump diffusion, poisson
Compressor summary: The proposed generative model can handle data of different dimensions by switching between them during the generation process, leading to better performance on tasks like imputation and interpolation.
Thomas Paniagua, Ryan Grainger, Tianfu Wu
https://openreview.net/forum?id=t3vPEjgNtj
Keywords: Ordered Top-K Clear-Box Targeted Adversarial Attack, Deep Neural Networks, Quadratic Programming, Robustness
Compressor summary: The paper introduces QuadAttac$K$, a method to learn more aggressive targeted attacks on DNNs by solving quadratic programming in the feature embedding space.
Martin Andres Bertran, Shuai Tang, Aaron Roth, Michael Kearns, Jamie Heather Morgenstern, Steven Wu
https://openreview.net/forum?id=t3WCiGjHqd
Keywords: machine learning, privacy, membership inference
Compressor summary: Our proposed attack estimates the distribution of confidence scores on non-training points using quantile regression, and it is more efficient and black-box than shadow model attacks.
Siyuan Huang, Yunchong Song, Jiayue Zhou, Zhouhan Lin
https://openreview.net/forum?id=t2hEZadBBk
Keywords: Graph Based Learning
Compressor summary: The paper introduces Subtree Attention, a novel graph attention mechanism that addresses the limitations of local and global attention by allowing computation of attention weights among multi-hop neighbors and has linear time complexity.
Andrii Zadaianchuk, Maximilian Seitzer, Georg Martius
https://openreview.net/forum?id=t1jLRFvBqm
Keywords: object-centric learning, video, representation learning, self-supervised learning, unsupervised learning
Compressor summary: The paper proposes a new method to learn object-centric representations from videos using pre-trained features and temporal feature similarity loss, achieving state-of-the-art results on synthetic and unconstrained video datasets.
Yifeng Chu, Maxim Raginsky
https://openreview.net/forum?id=t0fkjO4aZj
Keywords: generalization bounds, information theory, chaining, PAC-Bayes, couplings
Compressor summary: The paper proposes a method to derive generalization bounds for learning algorithms using a probabilistic decorrelation lemma and other techniques.
Zihao Hu, Guanghui Wang, Jacob Abernethy
https://openreview.net/forum?id=szFqlNRxeS
Keywords: Online learning, Riemannian optimization, projection-free optimization
Compressor summary: The paper proposes projection-free optimization methods for online geodesically convex problems on curved spaces with separation or linear optimization oracles, achieving sub-linear regret guarantees in different settings.
Vignesh Kothapalli, Tom Tirer, Joan Bruna
https://openreview.net/forum?id=sxao2udWXi
Keywords: Neural collapse, Graph neural networks, Community detection
Compressor summary: The paper investigates how graph topology affects feature evolution in node-wise classification using graph neural networks (GNNs) and compares it to the Neural Collapse phenomenon in instance-wise deep classifiers.
Yuanhao Wang, Qinghua Liu, Chi Jin
https://openreview.net/forum?id=sxZLrBqg50
Keywords: reinforcement learning theory, reinforcement learning from human feedback, preference-based reinforcement learning
Compressor summary: The paper shows that existing reward-based RL algorithms can be used to solve preference-based RL with minimal modifications, and applies the theory to various models with preferences.
Julian Coda-Forno, Marcel Binz, Zeynep Akata, Matthew Botvinick, Jane X Wang, Eric Schulz
https://openreview.net/forum?id=sx0xpaO0za
Keywords: Large language models, in-context learning, meta-learning, GPT-3
Compressor summary: The paper introduces meta-in-context learning, a method to improve large language models' in-context learning abilities by recursively applying them to different tasks, and shows its effectiveness on various domains and benchmarks.
Ingvar Ziemann, Stephen Tu, George J. Pappas, Nikolai Matni
https://openreview.net/forum?id=swNtr6vGqg
Keywords: Learning Theory, Learning with dependent data, Time-Series
Compressor summary: We find upper limits for random linear regression with dependent data without assuming realizability, and our analysis shows how the error improves as we reduce misspecification.
Mehdi Azabou, Vinam Arora, Venkataramana Ganesh, Ximeng Mao, Santosh B Nachimuthu, Michael Jacob Mendelson, Blake Aaron Richards, Matthew G Perich, Guillaume Lajoie, Eva L Dyer
https://openreview.net/forum?id=sw2Y0sirtM
Keywords: neural population, brain decoder, transformer, tokenization, sequence-to-sequence, electrophysiology, brain-computer interfaces
Compressor summary: The paper introduces a method to model neural activity across diverse recordings using tokenization, cross-attention, and PerceiverIO, enabling few-shot performance with minimal labels in new sessions.
Jinxin Liu, Li He, Yachen Kang, Zifeng Zhuang, Donglin Wang, Huazhe Xu
https://openreview.net/forum?id=suzMI2P1rT
Keywords: imitation learning, reinforcement learning, offline imitation learning
Compressor summary: The paper introduces ContExtual Imitation Learning (CEIL), a versatile algorithm for imitation learning that learns from expert behaviors using hindsight embeddings, and demonstrates its effectiveness on various settings and benchmarks.
Jihun Yun, Eunho Yang
https://openreview.net/forum?id=strvrjSi3C
Keywords: optimization, riemannian, manifolds, sharpness-aware
Compressor summary: The paper introduces Riemannian SAM, a novel optimization algorithm for training geometric deep learning models, and shows its benefits on knowledge graph completion and machine translation tasks.
Elia Turner, Omri Barak
https://openreview.net/forum?id=stDm3S0CV7
Keywords: Computational Neural Models; Recurrent Neural Networks; Multiple Tasks; Geometry;Dynamical Systems;Attractors;Neuroscience
Compressor summary: The authors study how Recurrent Neural Networks (RNNs) handle multiple tasks and find that they tend to reuse existing dynamics and opt for simple solutions, which they call the "simplicity bias".
Mitchell Wortsman, Tim Dettmers, Luke Zettlemoyer, Ari S. Morcos, Ali Farhadi, Ludwig Schmidt
https://openreview.net/forum?id=sqqASmpA2R
Keywords: CLIP, int8, stability
Compressor summary: The authors propose SwitchBack, a fast linear layer for int8 training, and recommend a hybrid of AdamW and Adafactor to prevent loss spikes in large language-vision models.
Shengcao Cao, Dhiraj Joshi, Liangyan Gui, Yu-Xiong Wang
https://openreview.net/forum?id=sqkGJjIRfG
Keywords: self-supervised learning, object detection
Compressor summary: HASSOD is a novel self-supervised object detection method that learns to detect objects and understand their compositions without human supervision, achieving improved performance and interpretability over existing methods.
Badih Ghazi, Yangsibo Huang, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Amer Sinha, Chiyuan Zhang
https://openreview.net/forum?id=sqTcCXkG4P
Keywords: Differential Privacy, Recommendation Systems, Embedding Models, Efficient Machine Learning
Compressor summary: The paper introduces DP-FEST and DP-AdaFEST, two new privacy-preserving algorithms for training large embedding models that reduce gradient size and maintain accuracy.
Joy Hsu, Jiayuan Mao, Joshua B. Tenenbaum, Jiajun Wu
https://openreview.net/forum?id=sq4o3tjWaj
Keywords: concept learning, visual reasoning, large language models, neuro-symbolic learning
Compressor summary: The paper proposes LEFT, a framework that learns to ground and reason with concepts across domains using a logic-based program executor and an LLM interpreter, outperforming previous methods in reasoning tasks.
Jangwon Kim, Hangyeol Kim, Jiwook Kang, Jongchan Baek, Soohee Han
https://openreview.net/forum?id=sq0m11cUMV
Keywords: time-delay system, reinforcement learning
Compressor summary: The paragraph describes a new actor-critic algorithm called BPQL that solves delayed feedback problems in reinforcement learning by reducing the state space size and outperforming traditional methods in continuous control tasks.
Eric Hedlin, Gopal Sharma, Shweta Mahajan, Hossam Isack, Abhishek Kar, Andrea Tagliasacchi, Kwang Moo Yi
https://openreview.net/forum?id=sovxUzPzLN
Keywords: Semantic Correspondence, Stable Diffusion, Optimization-based Inference
Compressor summary: The authors propose a method to find semantic correspondences between images using diffusion models' semantic knowledge without any training, achieving results comparable to strongly supervised methods.
Wentian Zhang, Haozhe Liu, Bing Li, Jinheng Xie, Yawen Huang, Yuexiang Li, Yefeng Zheng, Bernard Ghanem
https://openreview.net/forum?id=sodl2c3aTM
Keywords: Generative model, Generative Adversarial Network
Compressor summary: The paper proposes a new GAN method that helps the discriminator adapt faster to changes in generated data, improving the quality of generated results.
Susan Liang, Chao Huang, Yapeng Tian, Anurag Kumar, Chenliang Xu
https://openreview.net/forum?id=snY3FOnlQi
Keywords: Scene synthesis, audio-visual, NeRF
Compressor summary: The authors propose a novel approach to synthesize realistic audio-visual scenes using NeRF, integrating knowledge of sound propagation and source-centric acoustic fields, and present a new dataset and demo videos.
Chenze Shao, Zhengrui Ma, Min Zhang, Yang Feng
https://openreview.net/forum?id=sla7V80uWA
Keywords: Maximum likelihood estimation, Convex function, Text generation
Compressor summary: The paper proposes convex functions as an alternative to maximum likelihood estimation (MLE) for training text generation models, which can improve performance and efficiency on various tasks.
Anqi Li, Dipendra Misra, Andrey Kolobov, Ching-An Cheng
https://openreview.net/forum?id=shePL2nbwl
Keywords: Offline RL, safe RL
Compressor summary: Offline reinforcement learning can produce good and safe policies even with wrong reward labels due to a survival instinct from pessimism and implicit biases in data collection.
Kun Song, Huimin Ma, Bochao Zou, Huishuai Zhang, Weiran Huang
https://openreview.net/forum?id=shXnfALjuH
Keywords: few-shot learning, CLIP, fine-tuning
Compressor summary: The paper introduces FD-Align, a method that improves few-shot learning by preserving spurious feature consistency and generalizing better to ID and OOD tasks.
Sai Srivatsa Ravindranath, Yanchen Jiang, David C. Parkes
https://openreview.net/forum?id=sgCrNMOuXp
Keywords: Data Markets, Information Design, Differentiable Economics, Economics, Deep Learning, Mechanism Design, Algorithmic Game Theory
Compressor summary: The paper proposes a deep learning approach to design revenue-optimal data markets by learning signaling schemes and handling obedience and incentive constraints.
Thanh Nguyen-Tang, Raman Arora
https://openreview.net/forum?id=sdlh4gVOj8
Keywords: reinforcement learning, offline reinforcement learning
Compressor summary: The paper investigates what makes offline reinforcement learning sample-efficient and unifies three classes of algorithms using a new concept of data diversity.
Yufei CUI, Ziquan Liu, Yixin CHEN, Yuchen Lu, Xinyue Yu, Xue Liu, Tei-Wei Kuo, Miguel R. D. Rodrigues, Chun Jason Xue, Antoni B. Chan
https://openreview.net/forum?id=scaKiAtbI3
Keywords: Multiple Instance Learning, Whole Slide Imaging, Nearest Neighbor Retrieval
Compressor summary: The paper introduces RAM-MIL, a framework that uses Optimal Transport to improve MIL performance on out-of-domain data in medical diagnosis tasks based on whole slide images.
Sihan Chen, Handong Li, Qunbo Wang, Zijia Zhao, Mingzhen Sun, Xinxin Zhu, Jing Liu
https://openreview.net/forum?id=scYa9DYUAy
Keywords: Cross-Modality Foundation Model, Cross-Modality Pretraining Dataset
Compressor summary: This paper introduces a large dataset (VAST-27M) for multi-modal video captioning and a model (VAST) that can process vision, audio, and subtitle modalities for various tasks.
Yu Liang, Shiliang Zhang, Kenli Li, Xiaoyu Wang
https://openreview.net/forum?id=scG0cwftEe
Keywords: Deep Hash, Image Retrieval, Product Quantization
Compressor summary: The paragraph introduces a novel deep hashing framework using product quantization that addresses the limitations of current methods on large-scale datasets and achieves better performance.
Yelysei Bondarenko, Markus Nagel, Tijmen Blankevoort
https://openreview.net/forum?id=sbusw6LD41
Keywords: transformers, LLM, softmax, attention, outliers, quantization, post-training quantization
Compressor summary: The paper proposes two modifications to the attention mechanism in transformer models to reduce outliers and enable efficient quantization without compromising performance.
Paul Yoo, Jiaxian Guo, Yutaka Matsuo, Shixiang Shane Gu
https://openreview.net/forum?id=sZNBYvunEr
Keywords: Novel View Synthesis, Diffusion Model
Compressor summary: DreamSparse is a framework that uses a pre-trained diffusion model to generate high-quality novel view images from sparse views by incorporating 3D geometry information as a prior.
Yangtian Zhang, Zuobai Zhang, Bozitao Zhong, Sanchit Misra, Jian Tang
https://openreview.net/forum?id=sXMQPKbLXf
Keywords: protein side-chain packing, diffusion models, autoregressive models, geometric deep learning
Compressor summary: The authors propose DiffPack, a torsional diffusion model that predicts protein side-chain packing by learning joint distributions of side-chain angles and improves accuracy while reducing model size.
Konstantinos P. Panousis, Sotirios Chatzis
https://openreview.net/forum?id=sWNOvNXGLP
Keywords: Interpretability, Explainability, Network Dissection, Competitive Networks, Sparsity, Multimodal Models
Compressor summary: The authors present a method to generate textual descriptions of individual neuron functions in deep vision networks, enabling easier interpretation and exploration of their decision processes.
Jincheng Mei, Bo Dai, Alekh Agarwal, Mohammad Ghavamzadeh, Csaba Szepesvari, Dale Schuurmans
https://openreview.net/forum?id=sW8yGZ4uVJ
Keywords: reinforcement learning, policy gradient, policy optimization, function approximation, global convergence
Compressor summary: The paper shows that policy gradient methods can converge globally for finite-arm bandits with linear function approximation, depending on properties between the policy update and the representation, without relying heavily on approximation error as a key quantity.
Shentong Mo, Bhiksha Raj
https://openreview.net/forum?id=sUqG96QqZM
Keywords: audio-visual learning, visual sound localization, audio-visual segmentation
Compressor summary: The paragraph describes a novel framework called WS-AVS that simplifies the supervision for audio-visual segmentation by using instance-level annotations and multi-scale contrastive learning, and shows its effectiveness on various scenarios.
Wei Liu, Jun Wang, Haozhao Wang, Ruixuan Li, Zhiying Deng, YuanKai Zhang, Yang Qiu
https://openreview.net/forum?id=sUFGPYS25Q
Keywords: interpretability, causal inference, rationalization, self-explaining
Compressor summary: The Minimum Conditional Dependence (MCD) criterion is a novel way to uncover the causal rationale of NLP models by minimizing the dependence between non-selected input parts and the target label, which improves interpretability and performance compared to previous MMI-based methods.
Dinghuai Zhang, Hanjun Dai, Nikolay Malkin, Aaron Courville, Yoshua Bengio, Ling Pan
https://openreview.net/forum?id=sTjW3JHs2V
Keywords: graph; combinatorial optimization; sampling; gflownets
Compressor summary: The paper presents a method to train conditional GFlowNets on MDPs for solving combinatorial optimization problems by sampling from the solution space and demonstrates their effectiveness in finding high-quality solutions.
Hongyu Zang, Xin Li, Leiji Zhang, Yang Liu, Baigui Sun, Riashat Islam, Remi Tachet des Combes, Romain Laroche
https://openreview.net/forum?id=sQyRQjun46
Keywords: Bisimulation metrics, Reinforcement Learning, Representation Learning, Offline RL
Compressor summary: The paper analyzes why bisimulation methods work well in online RL but not in offline RL, and proposes using the expectile operator and reward scaling to improve performance on two benchmarks.
Minghui Hu, Jianbin Zheng, Daqing Liu, Chuanxia Zheng, Chaoyue Wang, Dacheng Tao, Tat-Jen Cham
https://openreview.net/forum?id=sQBGVw5qH9
Keywords: Multi-modality, Image Generation, Diffusion
Compressor summary: Cocktail is a pipeline that mixes different modalities into one embedding and uses a hyper-network called gControlNet to align and infuse control signals from various sources into a pre-trained diffusion model, resulting in high-quality image synthesis with diverse contents.
Junhyung Park, Simon Buchholz, Bernhard Schölkopf, Krikamol Muandet
https://openreview.net/forum?id=sPLTQSf6GI
Keywords: Causality, probability theory, causal models
Compressor summary: The text introduces a new concept of causal space that combines probability theory with causality to overcome some limitations of current approaches.
Yihe Wang, Yu Han, Haishuai Wang, Xiang Zhang
https://openreview.net/forum?id=sOQBHlCmzp
Keywords: Deep Learning, Contrastive Learning, Self-supervised Learning, Time Series, Healthcare
Compressor summary: COMET is a hierarchical framework that leverages data consistencies at multiple levels in medical time series to improve contrastive representation learning and outperforms existing methods on diverse datasets.
Khashayar Gatmiry, Zakaria Mhammedi
https://openreview.net/forum?id=sOOg1xJADA
Keywords: Online Learning, Online convex optimization, projection-free, Newton method
Compressor summary: The paper proposes a new online convex optimization algorithm that uses self-concordant barriers to avoid projections and achieve low regret without computing matrix inverses often.
Samuel Lanthaler, Nicholas H. Nelsen
https://openreview.net/forum?id=sLr1sohnmo
Keywords: random features, random feature model, operator learning, vector-valued
Compressor summary: The paper analyzes vector-valued random features ridge regression in an infinite-dimensional setting, providing strong consistency and minimax optimal convergence rates without relying on random matrix theory.
Lachlan Ewen MacDonald, Jack Valmadre, Hemanth Saratchandran, Simon Lucey
https://openreview.net/forum?id=sLhXMkI0kx
Keywords: optimisation, optimization, skip, connection, normalisation, normalization, deep, learning, polyak, lojasiewicz, lipschitz
Compressor summary: The paper presents a theory for gradient optimisation of deep neural networks that explains the roles of normalisation layers and skip connections, and shows how they enable training to global optima and acceleration of training.
Zizhao Wang, Jiaheng Hu, Peter Stone, Roberto Martín-Martín
https://openreview.net/forum?id=sL4pJBXkxu
Keywords: reinforcement learning; intrinsic motivation; exploration
Compressor summary: ELDEN is a novel intrinsic reward function that encourages reinforcement learning agents to explore environments with complex dependencies by exploiting uncertainty in the learned dynamics.
Zihan Luo, Hong Huang, Jianxun Lian, Xiran Song, Xing Xie, Hai Jin
https://openreview.net/forum?id=sJDkwMVqb9
Keywords: Cross-links, Debias, Graph Neural Networks, Link Prediction
Compressor summary: The paper proposes a twin-structure framework to reduce data bias between internal-links and cross-links in GNN-based link prediction, enhancing both bias mitigation and utility.
Jiayi Guan, Guang Chen, Jiaming Ji, Long Yang, Ao Zhou, Zhijun Li, changjun jiang
https://openreview.net/forum?id=sIU3WujeSl
Keywords: Offline safe reinforcement learning, Pessimistic conservative estimation, Variational optimization, Reinforcement Learning
Compressor summary: VOCE is a new offline safe RL algorithm that uses probabilistic inference and pessimistic estimation to learn policies that satisfy safety constraints while optimizing rewards.
Yanjing Li, Sheng Xu, Xianbin Cao, Xiao Sun, Baochang Zhang
https://openreview.net/forum?id=sFGkL5BsPi
Keywords: network quantization, diffusion model, image synthesize
Compressor summary: The paper proposes TaQ and NeM methods to improve low-bit quantized diffusion models, reducing computation and memory costs while maintaining high-quality data generation.
Shane Bergsma, Tim Zeyl, Lei Guo
https://openreview.net/forum?id=sC4RbbVKbu
Keywords: time series, probabilistic forecasting, autoregressive generative models, neural networks
Compressor summary: SutraNets are a new method for predicting long time series by factoring likelihood into conditional probabilities and generating multivariate outputs to reduce error accumulation.
Yuankun Jiang, Nuowen Kan, Chenglin Li, Wenrui Dai, Junni Zou, Hongkai Xiong
https://openreview.net/forum?id=sABYNWKcwK
Keywords: Meta-reinforcement learning, doubly robust (DR), sample transfer
Compressor summary: The paper proposes a new Meta-RL method called DRaT that addresses dynamics and reward function variations across tasks by using a doubly robust estimator to transfer informative samples from other tasks.
Hongting Ye, Yalu Zheng, Yueying Li, Ke Zhang, Youyong Kong, Yonggui Yuan
https://openreview.net/forum?id=s97ezbqoDZ
Keywords: Multimodal, Neuroscience, Subgraph, Transformer
Compressor summary: The authors present RH-BrainFS, a novel multimodal brain network fusion strategy that addresses regional heterogeneity between structural and functional connectivity, and shows its effectiveness in various neuroscience tasks.
Fangzhou Luo, Xiaolin Wu, Yanhui Guo
https://openreview.net/forum?id=s8QsYV1VZ2
Keywords: Blind Image Super-Resolution
Compressor summary: The paper proposes an adversarial neural degradation (AND) model that improves image super-resolution by generating a wide range of complex degradation effects without explicit supervision, resulting in better generalization and performance on real-world images.
Jaron Maene, Luc De Raedt
https://openreview.net/forum?id=s86M8naPSv
Keywords: neuro-symbolic AI, probabilistic logic, embeddings
Compressor summary: The paragraph discusses soft-unification, a technique for combining logic and neural concepts in AI, and proposes a new framework called DeepSoftLog that improves upon previous systems by introducing probabilistic semantics and achieving better performance on neuro-symbolic benchmarks.
Zhengxiang Shi, Aldo Lipani
https://openreview.net/forum?id=s7xWeJQACI
Keywords: Continued Pre-training, Prompt-based Fine-tuning, Language Models
Compressor summary: The study proposes Prompt-based Continued Pre-training (PCP), which improves prompt-based fine-tuning in natural language processing tasks by combining instruction tuning with conventional continued pre-training using unsupervised objectives.
Bastian Boll, Christoph Schnoerr
https://openreview.net/forum?id=s1jQ91yFAb
Keywords: Structured Prediction, PAC-Bayes, Concentration Inequalities, Statistical Learning Theory, Knothe-Rosenblatt Rearrangement
Compressor summary: The paper proposes a new way to measure how well structured prediction models generalize using generative models and Wasserstein dependency matrices.
Szymon Tworkowski, Konrad Staniszewski, Mikołaj Pacek, Yuhuai Wu, Henryk Michalewski, Piotr Miłoś
https://openreview.net/forum?id=s1FjXzJ0jy
Keywords: Transformers, Language Models, Natural Language Processing
Compressor summary: The Focused Transformer (FoT) technique addresses the distraction issue in large language models by enhancing the structure of the (key, value) space and extending the effective context length through contrastive learning.
Yejiang Wang, Yuhai Zhao, Daniel Zhengkui Wang, Ling Li
https://openreview.net/forum?id=rzlqOVExUA
Keywords: graph neural network; self-supervised learning; optimal transport;
Compressor summary: The paper proposes a self-supervised graph learning method that preserves both structural and matching information between graphs using optimal transport plans, and shows its advantages over existing methods in different scenarios.
Jingfeng Wu, Wennan Zhu, Peter Kairouz, Vladimir Braverman
https://openreview.net/forum?id=rzDBoh1tBh
Keywords: sketch, federated analytics, privacy
Compressor summary: The paper proposes an improved sketch algorithm for federated frequency estimation with multiple communication rounds, and a two-phase approach that adapts the sketch size based on problem complexity and ensures differential privacy.
Tushar Nagarajan, Santhosh Kumar Ramakrishnan, Ruta Desai, James Hillis, Kristen Grauman
https://openreview.net/forum?id=rybsHQ4DXy
Keywords: egocentric video, 3D environment, sim2real, sim-to-real, episodic memory
Compressor summary: The paper proposes an approach to understand human activities in their environment using egocentric video and 3D simulation, and shows its effectiveness on real-world videos.
Rajarshi Saha, Varun Srivastava, Mert Pilanci
https://openreview.net/forum?id=rxsCTtkqA9
Keywords: Matrix compression, Randomized low rank factorization, Randomized SVD, Sketching, Quantized embeddings, Random matrices
Compressor summary: The paper proposes an algorithm to decompose large matrices into low rank factors and quantize them for efficient storage and processing.
Paul Steven Scotti, Atmadeep Banerjee, Jimmie Goode, Stepan Shabalin, Alex Nguyen, Cohen Ethan, Aidan James Dempster, Nathalie Verlinde, Elad Yundler, David Weisberg, Kenneth Norman, Tanishq Mathew Abraham
https://openreview.net/forum?id=rwrblCYb2A
Keywords: fMRI, computational neuroscience, mind reading, diffusion models
Compressor summary: MindEye is a novel method to reconstruct and retrieve images from brain activity using fMRI and specialized submodules for retrieval and reconstruction.
Michael A. Lepori, Thomas Serre, Ellie Pavlick
https://openreview.net/forum?id=rwbzMiuFQl
Keywords: Deep Learning, Compositionality, Cognitive Science
Compressor summary: The authors investigate whether neural networks use modular subnetworks or template matching to solve complex tasks, and find evidence that they can learn structural compositionality without relying on symbolic mechanisms.
Jiuhn Song, Seonghoon Park, Honggyu An, Seokju Cho, Min-Seop Kwak, Sungjin Cho, Seungryong Kim
https://openreview.net/forum?id=rsrfEIdawr
Keywords: Neural Radiance Fields, 3D Reconstruction, Few-shot NeRF, Monocular Priors
Compressor summary: The DäRF framework combines NeRF and monocular depth estimation to improve novel view synthesis and 3D geometry reconstruction with a small number of real-world images.
Guangyan Chen, Meiling Wang, Yi Yang, Kai Yu, Li Yuan, Yufeng Yue
https://openreview.net/forum?id=rqE0fEQDqs
Keywords: Generative Pre-training Transformer; GPT; Auto-regressively Generative Pre-training; Self-supervised Learning; Point clouds
Compressor summary: PointGPT extends the GPT language model to point clouds, using a novel auto-regressive generation task and a dual masking strategy, achieving state-of-the-art results on various downstream tasks.
Alireza Fathollah Pour, Hassan Ashtiani
https://openreview.net/forum?id=rpuEARqB54
Keywords: PAC Learning, Recurrent Neural Networks, Noise, Sample Complexity
Compressor summary: The paper studies how adding noise to multi-layered sigmoid networks affects their sample complexity for classifying sequences, and shows an exponential gap between noisy and non-noisy networks.
Senthil Purushwalkam, Nikhil Naik
https://openreview.net/forum?id=roGYQvarnC
Keywords: 3D, generation, diffusion, viewpoint
Compressor summary: The paper introduces ConRad, a novel method that uses pretrained image generation models to reconstruct 3D objects from a single RGB image, capturing the appearance and preserving details while producing realistic 3D reconstructions.
Haotian Sun, Yuchen Zhuang, Lingkai Kong, Bo Dai, Chao Zhang
https://openreview.net/forum?id=rnKgbKmelt
Keywords: Large language models, decision making, closed-loop planning
Compressor summary: AdaPlanner is a closed-loop approach that allows large language models to adaptively refine their plans in response to environmental feedback and outperforms state-of-the-art methods with fewer samples.
Zichen Zhang, Johannes Kirschner, Junxi Zhang, Francesco Zanini, Alex Ayoub, Masood Dehghan, Dale Schuurmans
https://openreview.net/forum?id=rmQgQCZWiP
Keywords: Reinforcement Learning, Policy Evaluation, Temporal Discretization, Continuous Time, LQR
Compressor summary: The paper analyzes how time discretization affects Monte-Carlo policy evaluation for LQR systems and finds a trade-off between approximation and statistical error, leading to an optimal choice of temporal resolution for data efficiency.
Alexandre Blain, Bertrand Thirion, Olivier Grisel, Pierre Neuvial
https://openreview.net/forum?id=rlPUJ60bwM
Keywords: Knockoffs, Derandomization of Knockoffs, False Discoveries Proportion control, Controlled variable selection, Statistical inference, High-dimensional inference
Compressor summary: KOPI is a new method that improves Knockoff-based inference by controlling the actual proportion of false discoveries and using a novel aggregation technique.
Andong Wang, Chao Li, Mingyuan Bai, Zhong Jin, Guoxu Zhou, Qibin Zhao
https://openreview.net/forum?id=rih3hsSWx8
Keywords: Tensor SVD; Tensor Neural Networks; Transformed Low-rankness; Adversarial Generalization; Implicit Bias.
Compressor summary: The paper analyzes the theoretical generalization error of neural networks with t-product layers (t-NNs) and shows that transformed low-rank parameterization can improve their adversarial robustness.
Nikos Kolotouros, Thiemo Alldieck, Andrei Zanfir, Eduard Gabriel Bazavan, Mihai Fieraru, Cristian Sminchisescu
https://openreview.net/forum?id=rheCTpRrxI
Keywords: text to 3d; 3d avatars
Compressor summary: DreamHuman is a method that creates realistic 3D human avatars from text descriptions by combining different models and optimization techniques, resulting in diverse and high-quality animatable models.
Martin Saveski, Steven Jecmen, Nihar B Shah, Johan Ugander
https://openreview.net/forum?id=rhIfzCZoXG
Keywords: peer review, causal inference, counterfactual policy evaluation
Compressor summary: This paper proposes and applies novel methods to evaluate how changes in peer-review assignment algorithms affect review quality using randomness and data from two computer science venues.
Hanyang Zhao, Wenpin Tang, David Yao
https://openreview.net/forum?id=rfcak9EV99
Keywords: exploratory stochastic control, occupation time, performance difference, policy optimization
Compressor summary: The authors develop a new concept called occupation time for continuous reinforcement learning and show how it can improve performance and optimization methods in this setting.
Wei Fang, Zhaofei Yu, Zhaokun Zhou, Ding Chen, Yanqi Chen, Zhengyu Ma, Timothée Masquelier, Yonghong Tian
https://openreview.net/forum?id=rfTFJvTkr2
Keywords: Spiking Neural Network, SNN, deep learning, spiking neuron, neuromorphic computing
Compressor summary: The authors propose Parallel Spiking Neurons, which improve simulation speed and accuracy in spiking neural networks by reformulating neuronal dynamics without reset and enabling parallel processing of inputs.
Ben Prystawski, Michael Y. Li, Noah Goodman
https://openreview.net/forum?id=rcXXNFVlEn
Keywords: chain-of-thought; language models; reasoning
Compressor summary: The text discusses how chain-of-thought reasoning helps language models make better inferences when trained on locally structured data with strong variable dependencies, reducing bias and improving efficiency.
Wenxuan Bao, Tianxin Wei, Haohan Wang, Jingrui He
https://openreview.net/forum?id=rbw9xCU6Ci
Keywords: federated learning, personalized federated learning, test-time adaptation
Compressor summary: The paper introduces ATP, a novel test-time personalized federated learning algorithm that adapts global models in an unsupervised way without labeled data from multiple clients with different distributions, achieving strong generalization and outperforming existing methods.
Niklas Freymuth, Philipp Dahlinger, Tobias Daniel Würth, Simon Reisch, Luise Kärger, Gerhard Neumann
https://openreview.net/forum?id=rZqRu8e4uc
Keywords: Adaptive Mesh Refinement, Finite Element Method, Swarm Reinforcement Learning, Graph Neural Networks
Compressor summary: The paper proposes ASMR, a novel Adaptive Swarm Markov Decision Process for adaptive mesh refinement in engineering simulations, which achieves significant speedup and quality improvements compared to existing methods.
Amir Joudaki, Hadi Daneshmand, Francis Bach
https://openreview.net/forum?id=rY4sA9qYKy
Keywords: dynamical isometry, Lyapunov analysis, random neural networks
Compressor summary: In this paper, we explore the structure of the penultimate Gram matrix in deep neural networks, which contains the pairwise inner products of outputs corresponding to a batch of inputs. In several architectures it has been observed that this Gram matrix becomes degenerate with depth at initialization, which dramatically slows training. Normalization layers, such as batch or layer normalization, play a pivotal role in preventing the rank collapse issue. Despite promising advances, the existing theoretical results do not extend to layer normalization, which is widely used in transformers, and can not quantitatively characterize the role of non-linear activations. To bridge this gap, we prove that layer normalization, in conjunction with activation layers, biases the Gram matrix of a multilayer perceptron towards the identity matrix at an exponential rate with depth at initialization. We quantify this rate using the Hermite expansion of the activation function.
Hanna Ziesche, Leonel Rozo
https://openreview.net/forum?id=rW4mNcDxpS
Keywords: Policy optimization, robot learning, reinforcement learning, Gaussian mixture models, optimal transport, robotics
Compressor summary: The paper presents a novel policy optimization method for robots based on optimal transport and Riemannian optimization of GMMs, which enhances performance in different robot tasks.
Hao Yang, Haiyang Wang, Di Dai, Liwei Wang
https://openreview.net/forum?id=rUldfB4SPT
Keywords: Pre-Train, Autonomous Driving, LiDAR, 3D Object Detection
Compressor summary: The paper proposes a new method, PRED, that uses images and neural rendering to help train point cloud encoders for 3D perception tasks in autonomous driving, addressing the challenges of incompleteness and occlusions.
Indradyumna Roy, Rishi Agarwal, Soumen Chakrabarti, Anirban Dasgupta, Abir De
https://openreview.net/forum?id=rUf0GV5CuU
Keywords: Locality sensitive hashing, Fourier transform, Order embeddings
Compressor summary: The paper proposes FourierHashNet, a data-sensitive and trainable index for fast retrieval of relevant documents using hinge distance transformed into dominance similarity in the frequency domain.
Joon-Hyuk Ko, Hankyul Koh, Nojun Park, Wonho Jhe
https://openreview.net/forum?id=rUFckPrzXR
Keywords: neural ordinary differential equations, synchronization, homotopy optimization, loss landscape, dynamical systems
Compressor summary: This paper introduces a new training method for NeuralODEs that improves their performance and speed by using synchronization and homotopy optimization without changing the model architecture.
Seiyun Shin, Ilan Shomorony, Han Zhao
https://openreview.net/forum?id=rQI3FOzo1f
Keywords: Graph neural networks, Random sampling, Regression
Compressor summary: Graph Neural Networks (GNNs) aim to learn from large graphs, but their computational cost can be high; this paper proposes an efficient training method for a specific type of GNN by subsampling nodes and using leverage score sampling.
Jaewook J. Suh, Jisun Park, Ernest K. Ryu
https://openreview.net/forum?id=rN99gLCBe4
Keywords: acceleration, convex optimization, continuous-time analysis, monotone operator, monotone inclusion, minimax optimization, fixed-point problem, anchor acceleration
Compressor summary: This paper analyzes the continuous-time models of anchor acceleration, a minimization technique with an unclear mechanism, and proposes an adaptive method based on the findings.
Xixi Jia, Hailin Wang, Jiangjun Peng, Xiangchu Feng, Deyu Meng
https://openreview.net/forum?id=rLpLjCBW4J
Keywords: Non-convex optimization, matrix factorization, low rank, scaled gradient descent
Compressor summary: The paper proposes two preconditioned gradient descent methods for low-rank matrix factorization that converge faster and do not require small learning rates or initializations.
Chun-Han Yao, Amit Raj, Wei-Chih Hung, Michael Rubinstein, Yuanzhen Li, Ming-Hsuan Yang, Varun Jampani
https://openreview.net/forum?id=rJc5Lsn5QU
Keywords: 3D articulated shape, animal body estimation, diffusion for 3D
Compressor summary: ARTIC3D is a self-supervised framework that reconstructs 3D shapes from sparse images using skeleton-based surface representation and 2D diffusion priors, achieving high-quality results even with noisy and occluded inputs.
Xiong-Hui Chen, Yang Yu, Zhengmao Zhu, ZhiHua Yu, Chen Zhenjun, Chenghe Wang, Yinan Wu, Rong-Jun Qin, Hongqiu Wu, Ruijin Ding, Huang Fangsheng
https://openreview.net/forum?id=rHAX0LRwk8
Keywords: environment model learning, offline reinforcement learning, off-policy evaluation, individual treatment effects estimation, causal inference, adversarial learning
Compressor summary: The paper proposes a new model-learning objective called AWRM to address the selection bias of behavior policies in environment dynamics models and shows that it improves counterfactual prediction and downstream tasks using the GALILEO algorithm.
Colin Bredenberg, Ezekiel Williams, Cristina Savin, Blake Aaron Richards, Guillaume Lajoie
https://openreview.net/forum?id=rGN3X9jnEg
Keywords: synaptic plasticity, computational neuroscience
Compressor summary: The authors propose formal definitions of locality for machine learning algorithms that aim to be biologically plausible, in order to make clear what quantities cannot be included in a learning rule and test their predictions against them.
Irene Wang, Prashant J. Nair, Divya Mahajan
https://openreview.net/forum?id=rG1M3kOVba
Keywords: Federated Learning
Compressor summary: FLuID is a method to balance the training load in Federated Learning by creating sub-models for low-performance devices without sacrificing accuracy.
Pingsheng Li, Jonathan Cornford, Arna Ghosh, Blake Aaron Richards
https://openreview.net/forum?id=rDiMgZulwi
Keywords: Dale's Law, RNNs, brain-inspired neural networks, DANNs, computational neuroscience, spectral properties, inhibition
Compressor summary: The authors extend Dale's ANNs to RNNs, showing that good performance is possible while respecting Dale's Law, which states that neurons must be excitatory or inhibitory. They also find that the spectral properties of the recurrent weight matrix at initialisation are more impactful on network performance than sign constraints.
Xiang Zhuang, Qiang Zhang, Keyan Ding, Yatao Bian, Xiao Wang, Jingsong Lv, Hongyang Chen, Huajun Chen
https://openreview.net/forum?id=r9fzp8eyhZ
Keywords: molecular representation learning, out-of-distribution
Compressor summary: The authors propose a new molecular representation learning framework that improves generalization and robustness against distribution shifts, using a novel strategy of first encoding and then separating features in the latent space.
Tinglin Huang, Ziniu Hu, Zhitao Ying
https://openreview.net/forum?id=r9eZH6WNm2
Keywords: molecule, routing mechanism, meta gradient
Compressor summary: MolGroup is a method that uses graph structure similarity and task similarity to select the best auxiliary datasets for small molecule machine learning models, improving their performance.
Zifan Wang, Saranya Vijayakumar, Kaiji Lu, Vijay Ganesh, Somesh Jha, Matt Fredrikson
https://openreview.net/forum?id=r8snfquzs3
Keywords: Satisfiability Modulo Theories, Solver Layer, Combinatorial Problem, MAXSAT, SAT
Compressor summary: The paper proposes SMTLayer, a technique that integrates SMT solvers into DNNs to encode domain knowledge as mathematical formulas, enabling more efficient, robust, and interpretable learning.
Xinrui Chen, Yizhi Wang, Renao Yan, Yiqing Liu, Tian Guan, Yonghong He
https://openreview.net/forum?id=r8LYNleLf9
Keywords: Zero-shot quantization, Texture feature calibration, Post-training quantization, low bit width, Neural network compression
Compressor summary: The paper proposes TexQ, a novel zero-shot quantization method for neural networks that synthesizes calibration images and uses mixup knowledge distillation to generate diverse samples, improving performance in ultra-low bit width quantization.
YUANHAO WANG, Ramzi Idoughi, Wolfgang Heidrich
https://openreview.net/forum?id=r7g9nFsulw
Keywords: Neural density fields, Coordinate-based representations, Quadtree structure, Cryo-electron microscope
Compressor summary: The authors propose a new learning-based method for improving 3D structure reconstruction from cryo-ET data, which addresses challenges such as missing data and high noise levels by using an adaptive tensor representation and a novel loss function.
Mikhail Khodak, Ilya Osadchiy, Keegan Harris, Nina Balcan, Kfir Yehuda Levy, Ron Meir, Steven Wu
https://openreview.net/forum?id=r6xGZ0XL2g
Keywords: online learning, multi-armed bandits, meta-learning, multi-task learning, bandit linear optimization
Compressor summary: The authors propose meta-algorithms that improve performance across multiple similar tasks using bandit feedback, by tuning hyperparameters of inner learners in online-within-online settings for MAB and BLO.
Sai Aparna Aketi, Abolfazl Hashemi, Kaushik Roy
https://openreview.net/forum?id=qyixBZl8Ph
Keywords: Federated Learning, Decentralized Learning, Non-IID Data, Heterogeneous data distribution, Peer-to-peer connectivity
Compressor summary: The paper proposes Global Update Tracking (GUT), a decentralized learning algorithm that reduces the impact of heterogeneous data distribution and achieves state-of-the-art performance on various Computer Vision datasets.
Arman Zharmagambetov, Brandon Amos, Aaron M Ferber, Taoan Huang, Bistra Dilkina, Yuandong Tian
https://openreview.net/forum?id=qyEm4tF2p1
Keywords: learning surrogates, predict+optimize framework, combinatorial nonlinear optimization, argmin differentiation
Compressor summary: The paper proposes a learnable landscape surrogate to improve learning-integrated optimization by addressing challenges such as problem uncertainty and sparse gradients.
Yuyang Shi, Valentin De Bortoli, Andrew Campbell, Arnaud Doucet
https://openreview.net/forum?id=qy07OHsJT5
Keywords: diffusion Schrödinger bridge, bridge matching, optimal transport
Compressor summary: The paragraph discusses a new method, Iterative Markovian Fitting (IMF), for solving Schrödinger bridges (SBs) problems and a novel algorithm, Diffusion Schrödinger Bridge Matching (DSBM), which significantly improves SB numerics and recovers various recent transport methods.
Tankred Saanum, Noemi Elteto, Peter Dayan, Marcel Binz, Eric Schulz
https://openreview.net/forum?id=qxF8Pge6vM
Keywords: Deep Reinforcement Learning, Compression, Sequence learning, Information bottleneck, Mutual information
Compressor summary: The paper proposes an RL algorithm that learns simple action sequences, either learned by autoregressive models or compressed by data algorithms, and shows it outperforms model-free approaches in continuous control tasks.
Aiwen Xu, Yuchen Hou, Cris M. Niell, Michael Beyeler
https://openreview.net/forum?id=qv5UZJTNda
Keywords: neuroscience, cognitive science, multimodal learning, representation learning, network architecture, computational biology, visual perception
Compressor summary: The authors introduce a multimodal recurrent neural network that can predict visual cortex activity in free-moving mice by integrating visual input, behavior, and temporal dynamics, revealing new insights into cortical function.
Anuran Makur, Marios Mertzanidis, Alexandros Psomas, Athina Terzoglou
https://openreview.net/forum?id=qumBHr77ht
Keywords: mechanism design, revenue maximization, correlated distributions, total variation distance
Compressor summary: The paper studies how to design truthful mechanisms for agents with unknown and correlated valuation functions, showing that some existing mechanisms are robust to variations in the distribution, and providing various applications and extensions.
Feiyang Kang, Hoang Anh Just, Anit Kumar Sahu, Ruoxi Jia
https://openreview.net/forum?id=quMBEd27x9
Keywords: data-centric AI, data acquisition, data valuation, performance prediction, data markets, optimal transport, scaling laws
Compressor summary: The paper introduces
Manbir S Gulati, Paul F Roysdon
https://openreview.net/forum?id=qs4swxtIAQ
Keywords: Tabular Data, Deep Learning, Generative Modeling, Transformers, Masked Transformers, Synthetic data
Compressor summary: TabMT is a new Masked Transformer model that generates synthetic tabular data with strong performance, handling missing data and various data types, and providing good privacy tradeoffs.
Jing Gu, Yilin Wang, Nanxuan Zhao, Tsu-Jui Fu, Wei Xiong, Qing Liu, Zhifei Zhang, HE Zhang, Jianming Zhang, HyunJoon Jung, Xin Eric Wang
https://openreview.net/forum?id=qqcIM8NiiB
Keywords: image editing, diffusion model, text to image generation
Compressor summary: The text introduces Photoswap, a novel approach for personalized subject swapping in images without training, which preserves the pose and coherence of the image while swapping subjects.
Praneeth Kacham, David Woodruff
https://openreview.net/forum?id=qptO6YDZEP
Keywords: Compressed Sensing, Matrix Recovery, Low rank approximation
Compressor summary: The authors study the lower bounds on adaptive algorithms for recovering low rank matrices using linear measurements, showing that any such algorithm must perform a certain number of rounds to achieve a satisfactory approximation.
Shenghuan Sun, Gregory Goldgof, Atul Butte, Ahmed Alaa
https://openreview.net/forum?id=qlnlamFQEa
Keywords: Synthetic clinical data, Machine learning for healthcare
Compressor summary: The paper introduces a pathologist-in-the-loop framework that uses human feedback to improve the quality and plausibility of synthetic medical images generated by conditional diffusion models, addressing challenges in assessing clinical sensibility and incorporating domain knowledge.
David Liu, Máté Lengyel
https://openreview.net/forum?id=qlJoo2y3gY
Keywords: Gaussian processes, renewal processes, point processes, neural data analysis, Bayesian machine learning, non-stationary time series
Compressor summary: The authors propose a Bayesian approach to model neural spiking activity that captures instantaneous variability in response to covariates, and apply it to two datasets of animal navigation, showing improved predictive power and richer patterns of modulation.
Miaoxi Zhu, Li Shen, Bo Du, Dacheng Tao
https://openreview.net/forum?id=ql6LVyi2Dg
Keywords: decentralized algorithm, minimax problem, algorithmic stability, generalization analysis
Compressor summary: The paper studies how decentralized algorithms like D-SGDA can generalize well despite their structure, by analyzing their primal-dual generalization bound using algorithmic stability.
Daesung Kim, Hye Won Chung
https://openreview.net/forum?id=qjqJL2lfkH
Keywords: Matrix completion, gradient descent, random initialization
Compressor summary: The paper analyzes gradient descent algorithm's convergence behavior on rank-1 symmetric matrix completion problem with small random initialization, and shows that it converges to the ground truth when more samples are available.
Rithesh Kumar, Prem Seetharaman, Alejandro Luebs, Ishaan Kumar, Kundan Kumar
https://openreview.net/forum?id=qjnl1QUnFA
Keywords: audio generation, audio compression, GAN, audio, speech
Compressor summary: The authors introduce a universal neural audio compression algorithm that achieves 90x compression of various audio domains with high fidelity, outperforming existing methods and providing open-source code and weights.
Emmanuel Abbe, Samy Bengio, Enric Boix-Adserà, Etai Littwin, Joshua M. Susskind
https://openreview.net/forum?id=qieeNlO3C7
Keywords: transformers, low-rank bias, incremental learning
Compressor summary: The text describes how transformers learn incrementally, leading to a growing difference between trained and initial weights, and provides theoretical and empirical evidence for this behavior.
Dave Epstein, Allan Jabri, Ben Poole, Alexei A Efros, Aleksander Holynski
https://openreview.net/forum?id=qgv56R2YJ7
Keywords: generative models, image editing, diffusion, guidance
Compressor summary: Self-guidance is a technique that uses internal representations of diffusion models to control various aspects of image generation and manipulation without additional training or models.
Yun-Yun Tsai, Chengzhi Mao, Junfeng Yang
https://openreview.net/forum?id=qgmrC8jhCo
Keywords: self-supervised learning, representation learning, visual prompts, domain generalization, input adaptation
Compressor summary: Convolutional visual prompts (CVP) are a lightweight method for adapting vision models in test-time settings without labels, improving their robustness against out-of-distribution samples.
Ameya Velingker, Ali Kemal Sinop, Ira Ktena, Petar Veličković, Sreenivas Gollapudi
https://openreview.net/forum?id=qgiG7WZohZ
Keywords: graph neural networks, message passing, effective resistance, hitting time
Compressor summary: The paper explores using random walk measures as features in graph neural networks to improve performance on various node and graph property prediction tasks with low computational complexity.
Zhuo Chen, Laker Newhouse, Eddie Chen, Di Luo, Marin Soljacic
https://openreview.net/forum?id=qdsDy0zbn4
Keywords: Autoregressive neural network, tensor network, quantum many-body physics, variational Monte Carlo
Compressor summary: The Autoregressive Neural TensorNet (ANTN) is a novel architecture that combines tensor networks and autoregressive neural networks to improve quantum many-body physics simulation, outperforming existing methods.
Kang Xu, Chenjia Bai, Xiaoteng Ma, Dong Wang, Bin Zhao, Zhen Wang, Xuelong Li, Wei Li
https://openreview.net/forum?id=qdM260dXsa
Keywords: Reinforcement Learning; Domain Adaptation; Online Dynamics Adaptation
Compressor summary: The paper proposes a new reinforcement learning method that adapts policies across domains with different dynamics by using value consistency to selectively share data, and shows its superior performance.
Eric Zelikman, Qian Huang, Gabriel Poesia, Noah Goodman, Nick Haber
https://openreview.net/forum?id=qd9qcbVAwQ
Keywords: reasoning, language models, code synthesis, decomposition
Compressor summary: Parsel is a framework that helps large language models generate complex algorithms with code by breaking down tasks into hierarchical natural language function descriptions and searching over combinations of possible function implementations using tests.
Lucas Caccia, Edoardo Ponti, Zhan Su, Matheus Pereira, Nicolas Le Roux, Alessandro Sordoni
https://openreview.net/forum?id=qcQhBli5Ho
Keywords: Parameter Efficient Finetuning, Multitask Learning, Transfer Learning, Natural Language Processing
Compressor summary: The paper explores how adapter routing affects PEFT and proposes $\texttt{MHR}$, a new routing method that improves performance by combining subsets of adapter parameters, and $\texttt{MHR}$-$\mu$, a variant that discards routing for single-adapter fine-tuning and zero-shot transfer.
Marc Rigter, Bruno Lacerda, Nick Hawes
https://openreview.net/forum?id=qZjl2TKvUY
Keywords: offline reinforcement learning, model-based reinforcement learning, risk, uncertainty
Compressor summary: The paper proposes a model-based offline reinforcement learning algorithm that combines risk-aversion to aleatoric and epistemic uncertainty to address distributional shift and safety in safety-critical domains.
Zhendong Chu, Nan Wang, Hongning Wang
https://openreview.net/forum?id=qYAp31KwU2
Keywords: Conversational Recommendation, Reinforcement Learning, Meta Learning
Compressor summary: The paper proposes a new method to improve Conversational Recommender Systems by learning user preferences and designing task-specific rewards from user interactions, leading to better recommendations with fewer turns.
Nimrah Mustafa, Aleksandar Bojchevski, Rebekka Burkholz
https://openreview.net/forum?id=qY7UqLoora
Keywords: graph attention networks, gradient flow, conservation law
Compressor summary: The paper explores the optimization and learning dynamics of Graph Attention Networks (GATs), derives a conservation law for their gradient flow dynamics, proposes an initialization scheme to improve trainability and speedup, and serves as a foundation for future studies on positive homogeneous models with attention.
Wenxuan Zeng, Meng Li, Haichuan Yang, Wen-jie Lu, Runsheng Wang, Ru Huang
https://openreview.net/forum?id=qVeDwgYsho
Keywords: Private Inference, Network/Protocol Co-Optimization, Winograd Convolution, Structural Re-parameterization
Compressor summary: CoPriv is a framework that optimizes both the DNN architecture and the 2PC inference protocol to significantly reduce communication overhead while maintaining accuracy for secure privacy protection in DNN inference.
Kensen Shi, Hanjun Dai, Wen-Ding Li, Kevin Ellis, Charles Sutton
https://openreview.net/forum?id=qVMPXrX4FR
Keywords: Program Synthesis, Programming By Example, Lambdas, Functional Programming
Compressor summary: LambdaBeam is a search algorithm that uses neural networks to synthesize lambda functions for longer and more general programs, overcoming the limitations of prior approaches in handling iterative loops and higher-order functions.
Hyuna Cho, Minjae Jeong, Sooyeon Jeon, Sungsoo Ahn, Won Hwa Kim
https://openreview.net/forum?id=qUlpDjYnsp
Keywords: graph wavelet transform, multi-scale wavelet filtering, graph generation, diffusion model
Compressor summary: The paper introduces a new method called Wavelet Graph Diffusion Model (Wave-GD) that generates realistic graphs with accurate node and edge frequency characteristics by modeling their joint distribution in the spectral space.
Tao Fang, Qian Zheng, Gang Pan
https://openreview.net/forum?id=qSS9izTOpo
Keywords: fMRI, image reconstruction, brain decoding
Compressor summary: The paper proposes a method (GESS) to improve fMRI-to-image reconstruction by using CLIP features to bridge the semantic gap between training and testing data, as well as leveraging structural information and uncertainty estimation.
Spencer Frei, Gal Vardi, Peter Bartlett, Nathan Srebro
https://openreview.net/forum?id=qSCziWQBPD
Keywords: adversarial robustness, neural networks, implicit bias, generalization
Compressor summary: The paper investigates how the biased gradient flow in two-layer ReLU networks affects generalization and adversarial robustness, showing that it leads to solutions that are good at fitting data but vulnerable to small perturbations.
Hanlin Zhu, Amy Zhang
https://openreview.net/forum?id=qS9aHF8bXz
Keywords: offline goal-conditioned RL, provably efficient algorithm, single-policy concentrability, general function approximation
Compressor summary: The paper provides a rigorous theoretical analysis of an existing offline goal-conditioned reinforcement learning algorithm, showing its efficiency, stability, and effectiveness in various real-world environments.
Jiaxin Bai, Xin Liu, Weiqi Wang, Chen Luo, Yangqiu Song
https://openreview.net/forum?id=qQnO1HLQHe
Keywords: Knowledge Graph, Complex Query Answering, Eventuality Graph
Compressor summary: The paper proposes a new framework, CEQA, that uses implicit logical constraints and theorem provers to answer complex queries on eventuality-centric knowledge graphs, improving neural query encoders' performance.
Tao Zhang, Yaowu Zhang, Tingyou Zhou
https://openreview.net/forum?id=qPyvuFT0U9
Keywords: High dimensionality; Independence test; Kernel method; Nonlinear dependency.
Compressor summary: The paper studies how a popular measure of dependence, HSIC, behaves when the dimensions of the variables it measures grow at different rates and provides conditions for its performance in high-dimensional settings.
Lakshya Agrawal, Aditya Kanade, Navin Goyal, Shuvendu K Lahiri, Sriram Rajamani
https://openreview.net/forum?id=qPUbKxKvXq
Keywords: Language models, code generation, correctness, program analysis
Compressor summary: Monitor-guided decoding (MGD) helps language models of code achieve better compilation rates and agreement with ground truth by using static analysis to provide context, especially when dealing with global information.
Yue Wu, Yewen Fan, Paul Pu Liang, Amos Azaria, Yuanzhi Li, Tom Mitchell
https://openreview.net/forum?id=qP0Drg2HuH
Keywords: Games, Instruction Manual, Atari Games, Large Language Models, Language Models, Zero-shot, In-context prompting
Compressor summary: The Read and Reward framework uses instruction manuals to enhance the performance and efficiency of reinforcement learning agents on Atari games.
Preetha Vijayan, Prashant Shivaram Bhat, Bahram Zonooz, Elahe Arani
https://openreview.net/forum?id=qL3zPoWJda
Keywords: Continual Learning, Catastrophic Forgetting, Experience Replay, Lifelong Learning, Bio-Inspired, Active Forgetting, Scalable Neurogenesis
Compressor summary: The paragraph introduces TriRE, a novel continual learning paradigm that combines multiple neurophysiological processes to reduce catastrophic forgetting and promote knowledge transfer in deep neural networks.
Sebastian Salazar
https://openreview.net/forum?id=qJRlz3SucN
Keywords: Probabilistic Machine Learning, Variational Inference, Bayesian Inference, Bayesian Nonparametrics
Compressor summary: The paper presents a new Bayesian model for decision trees that uses variational inference and shows its effectiveness in regression and causal inference problems, with a PyTorch implementation available.
Yue Tan, Chen Chen, Weiming Zhuang, Xin Dong, Lingjuan Lyu, Guodong Long
https://openreview.net/forum?id=qJJmu4qsLO
Keywords: Federated Learning, Test-Time Shift, Contrastive Learning
Compressor summary: FedICON is a framework that uses contrastive learning to capture and adapt to heterogeneous data in federated learning, improving performance under test-time shifts.
Gregory Dexter, Petros Drineas, David Woodruff, Taisuke Yasuda
https://openreview.net/forum?id=qHzEFxtheD
Keywords: dictionary learning, k means clustering, sketching, ptas, streaming
Compressor summary: The paper develops new techniques to extend sketching-based approaches to sparse dictionary learning and Euclidean k-means clustering problems, and obtains new upper and lower bounds for these problems in various settings.
Leonardo Galli, Holger Rauhut, Mark Schmidt
https://openreview.net/forum?id=qHrZszJSXj
Keywords: line search, nonmonotone, stochastic gradient descent, over-parametrized models, Polyak step size, optimization
Compressor summary: Nonmonotone line search methods can improve the speed and generalization of SGD/Adam by allowing larger step sizes without compromising convergence rates, as demonstrated by the PoNoS method and a new resetting technique.
Guhao Feng, Bohang Zhang, Yuntian Gu, Haotian Ye, Di He, Liwei Wang
https://openreview.net/forum?id=qHrADgAdYu
Keywords: Chain-of-Thought Prompting, Large Language Models, Theory, Circuit Complexity, Dynamic Programming
Compressor summary: The paper investigates how Chain-of-Thought prompting enhances Large Language Models' performance in solving mathematical and decision-making problems, using circuit complexity theory and experiments.
Anastasia Koloskova, Ryan McKenna, Zachary Charles, J Keith Rush, Hugh Brendan McMahan
https://openreview.net/forum?id=qCglMj6A4z
Keywords: optimization, machine learning, differential privacy
Compressor summary: The paper analyzes how gradient descent works with linearly correlated noise, which arises in some privacy-preserving optimization methods, and proposes improved matrix factorizations for better performance.
Ziyi Yin, Muchao Ye, Tianrong Zhang, Tianyu Du, Jinguo Zhu, Han Liu, Jinghui Chen, Ting Wang, Fenglong Ma
https://openreview.net/forum?id=qBAED3u1XZ
Keywords: vision-language, adversarial attacks, pre-trained model, fine-tuned model
Compressor summary: This paper proposes VLATTACK, a method to craft image and text perturbations that can attack black-box fine-tuned Vision-Language models on different tasks by using block-wise similarity attacks for images and existing methods for texts.
Jin-Hui Wu, Shao-Qun Zhang, Yuan Jiang, Zhi-Hua Zhou
https://openreview.net/forum?id=qA0uHmaVKk
Keywords: Complex-valued Neural Networks; Learning Neurons; Real-valued Neural Networks; Convergence Rate
Compressor summary: Complex-valued neural networks can efficiently learn functions of real-valued neurons and complex-valued neurons faster than real-valued neurons can learn each other or be learned by complex-valued neurons.
Dohyeong Kim, Kyungjae Lee, Songhwai Oh
https://openreview.net/forum?id=q9WMXjUxxT
Keywords: Reinforcement learning, Safety, Multiple Constraints, Distributional Critic
Compressor summary: The paper proposes a trust region-based safe reinforcement learning algorithm for robotic tasks with multiple constraints, and shows its effectiveness in experiments.
Tonghan Wang, Paul Duetting, Dmitry Ivanov, Inbal Talgam-Cohen, David C. Parkes
https://openreview.net/forum?id=q8mH2d6uw2
Keywords: Automated contract design, discontinuous neural networks
Compressor summary: The paper proposes using deep learning and a new type of network (DeLU) to automate the design of optimal contracts, considering both agent's incentives and principal's utility.
Zhenghao Peng, Wenjie Mo, Chenda Duan, Quanyi Li, Bolei Zhou
https://openreview.net/forum?id=q8SukwaEBy
Keywords: Machine Learning, Human-in-the-loop Reinforcement Learning, Safety, Sample Efficiency, Reward-free
Compressor summary: The proposed method uses a proxy value function to express human intents and label state-action pairs for policy optimization, allowing the AI agent to learn from active human involvement in various control tasks.
Kai Han, You Wu, He Huang, Shuang Cui
https://openreview.net/forum?id=q6bVqOgGxP
Keywords: mechanism design, budget-feasible, truthful
Compressor summary: The paper introduces TripleEagle, a new algorithmic framework for designing Budget-Feasible Mechanisms with better approximation ratios, linear complexities, and strategyproofness for submodular valuation functions.
Marcel Kollovieh, Abdul Fatir Ansari, Michael Bohlke-Schneider, Jasper Zschiegner, Hao Wang, Bernie Wang
https://openreview.net/forum?id=q6X038vKgU
Keywords: diffusion models, time series forecasting, generative modeling, deep learning
Compressor summary: The authors propose a task-agnostic diffusion model for time series applications and show its effectiveness in forecasting, refinement, and synthetic data generation.
Simian Luo, Chuanhao Yan, Chenxu Hu, Hang Zhao
https://openreview.net/forum?id=q5FAZAIooz
Keywords: Video-to-Audio Generation; Contrastive Audio-Visual Pretraining; Latent Diffusion Model;
Compressor summary: Diff-Foley is a new method for generating high-quality audio from silent videos with improved synchronization and relevance, using a latent diffusion model and contrastive pretraining.
Usha Bhalla, Suraj Srinivas, Himabindu Lakkaraju
https://openreview.net/forum?id=q4HlFS7B7Y
Keywords: Machine Learning Explainability, Machine Learning Interpretability
Compressor summary: The paragraph discusses a method called Distractor Erasure Tuning (DiET), which improves the faithfulness of post hoc explanations for machine learning models by making them robust to distractor erasure in inputs.
Eeshaan Jain, Tushar Nandy, Gaurav Aggarwal, Ashish V. Tendulkar, Rishabh K Iyer, Abir De
https://openreview.net/forum?id=q3fCWoC9l0
Keywords: Data Subset Selection, Efficient Learning
Compressor summary: The paragraph describes $\texttt{SubSelNet}$, a non-adaptive framework for subset selection in efficient learning that uses an attention-based neural gadget and can be either transductive or inductive, achieving better results than other methods on various real datasets.
Gabriel Herbert Sarch, Michael J. Tarr, Katerina Fragkiadaki, Leila Wehbe
https://openreview.net/forum?id=q3fA5tTod3
Keywords: Computational Neuroscience, Deep Neural Networks, Visual Neuroscience, Visual Streams, Scene Perception, Brain Imaging
Compressor summary: The paper trains neural networks to predict brain responses to images and uses "network dissection" to explore how different regions of the visual cortex interpret various features of natural scenes.
Yining Ma, Zhiguang Cao, Yeow Meng Chee
https://openreview.net/forum?id=q1JukwH2yP
Keywords: learning to optimize, vehicle routing problem, combinatorial optimization
Compressor summary: The paper introduces NeuOpt, a learning-to-search solver for routing problems that learns flexible k-opt exchanges and uses GIRE, D2A, and reward shaping to explore both feasible and infeasible regions effectively, achieving superior performance on TSP and CVRP compared to existing methods.
Simon Buchholz, Goutham Rajendran, Elan Rosenfeld, Bryon Aragam, Bernhard Schölkopf, Pradeep Kumar Ravikumar
https://openreview.net/forum?id=q131tA7HCT
Keywords: Causal Representation Learning, Interventional data, Gaussian Structural Causal models
Compressor summary: The authors study how to learn causal representations from unknown interventions using Gaussian latent distributions and a general mixing function, and they prove identifiability results without access to intervention targets.
Dohyeok Lee, Seungyub Han, Taehyun Cho, Jungwoo Lee
https://openreview.net/forum?id=q0sdoFIfNg
Keywords: Deep Reinforcement Learning, Ensemble Q-learning
Compressor summary: The paper proposes a new method, spiked Wishart Q-ensemble independence regularization (SPQR), to improve deep reinforcement learning by ensuring diversity of multiple Q-functions using random matrix theory.
Arghya Datta, Sayak Chakrabarty
https://openreview.net/forum?id=q0RfX96un8
Keywords: maximum likelihood estimate, non-identifiability, Redner approach, quotient topological spaces, consistency
Compressor summary: The text describes PPCA, a widely used statistical tool with applications in various fields, and proposes a novel approach to resolve the identifiability issue of its maximum likelihood solution using quotient topological spaces.
Zelei Cheng, Xian Wu, Jiahao Yu, Wenhai Sun, Wenbo Guo, Xinyu Xing
https://openreview.net/forum?id=pzc6LnUxYN
Keywords: deep reinforcement learning, interpretation, explanation
Compressor summary: StateMask is a novel method to identify critical states for an agent's final reward by learning a mask net that temporarily blinds the agent without affecting its performance.
Chendi Wang, Buxin Su, Jiayuan Ye, Reza Shokri, Weijie J Su
https://openreview.net/forum?id=pw5hEuEroL
Keywords: Differential privacy, $f$-DP, mixture mechanisms, shuffling, differentially private gradient descent
Compressor summary: The paper proposes an improved analysis of differential privacy for shuffling models and one-iteration DP-GD using $f$-DP, with applications to random initialization and mixture mechanisms.
Yuxin Cao, Yian Li, Yumeng Zhu, Derui Wang, Minhui Xue
https://openreview.net/forum?id=pvSKVt3EsM
Keywords: 3D mask detection, spatio-temporal aggregation, optical flow, deep learning
Compressor summary: FASTEN is a novel 3D mask detection framework that uses flow attention and spatio-temporal aggregation to quickly and accurately detect spoofing attacks in face recognition systems with low computational overhead and only five frames of input.
Amit Daniely, Nathan Srebro, Gal Vardi
https://openreview.net/forum?id=pvPujuvjQd
Keywords: learning neural networks, computational complexity, random networks
Compressor summary: The paper proposes a polynomial-time approximation scheme (PTAS) for learning random constant-depth neural networks with an additive error of epsilon under certain conditions on the network size and activation functions.
Mingxuan Ju, Tong Zhao, Wenhao Yu, Neil Shah, Yanfang Ye
https://openreview.net/forum?id=puupdGOWUp
Keywords: Graph neural network, Test-time Augmentation
Compressor summary: The paragraph discusses a test-time augmentation framework, GraphPatcher, that improves the performance of graph neural networks on low-degree nodes without compromising their ability to handle high-degree nodes.
Elliot Catt, Jordi Grau-Moya, Marcus Hutter, Matthew Aitchison, Tim Genewein, Gregoire Deletang, Li Kevin Wenliang, Joel Veness
https://openreview.net/forum?id=psXVkKO9No
Keywords: General Reinforcement Learning, Reinforcement Learning, Self-Modeling, Bayes-optimality, Policy Distillation, Uncertainty, Universal AI
Compressor summary: Self-AIXI is a universal agent that uses learning to obtain good policies by self-predicting its own action data, converging to AIXI and having maximal intelligence and self-optimization properties.
Rishi Dev Jha, Jonathan Hayase, Sewoong Oh
https://openreview.net/forum?id=prftZp6mDH
Keywords: security, backdoor attack
Compressor summary: The paper introduces FLIP, a novel label-only backdoor attack method that can successfully manipulate image predictions with corrupted training labels, while maintaining high clean test accuracy.
Xin-Qiang Cai, Pushi Zhang, Li Zhao, Jiang Bian, Masashi Sugiyama, Ashley Juan Llorens
https://openreview.net/forum?id=prIwYTU9PV
Keywords: Multi-Objective Reinforcement Learning
Compressor summary: DPMORL is a novel approach that learns policies balancing multiple objectives while considering return uncertainty and satisfying diverse distributional preferences.
Zhengyi Wang, Cheng Lu, Yikai Wang, Fan Bao, Chongxuan Li, Hang Su, Jun Zhu
https://openreview.net/forum?id=ppJuFSOAnM
Keywords: diffusion models, text to 3D
Compressor summary: VSD is a particle-based variational framework that improves text-to-3D generation by modeling the 3D parameter as a random variable and addressing over-saturation, over-smoothing, and low-diversity issues.
Kirill Neklyudov, Jannes Nys, Luca Thiede, Juan Felipe Carrasquilla Alvarez, qiang liu, Max Welling, Alireza Makhzani
https://openreview.net/forum?id=pjSzKhSrfs
Keywords: Quantum Monte Carlo, Schrödinger equation, Wasserstein Fisher-Rao gradient flow
Compressor summary: The paper introduces Wasserstein Quantum Monte Carlo, a new method for solving quantum many-body problems that uses transportation-based gradients and converges faster than traditional methods.
Sattar Vakili, Julia Olkhovskaya
https://openreview.net/forum?id=pirH9ycaNg
Keywords: Reinforcement Learning, Kernel ridge regression, Gaussian processes, LSVI
Compressor summary: The paper proposes $\pi$-KRVI, an optimistic modification of least-squares value iteration for reinforcement learning with nonlinear function approximation using kernel ridge regression, and proves its sublinear regret bound under a general setting.
Jinwoo Kim, Dat Tien Nguyen, Ayhan Suleymanzade, Hyeokjun An, Seunghoon Hong
https://openreview.net/forum?id=phnN1eu5AX
Keywords: equivariant machine learning, transformers, graphs, general-purpose architectures
Compressor summary: The authors propose a framework that uses a non-equivariant backbone and an equivariant network to learn functions with group symmetries, achieving competitive results against tailored equivariant architectures.
Zhiqin Yang, Yonggang Zhang, Yu Zheng, Xinmei Tian, Hao Peng, Tongliang Liu, Bo Han
https://openreview.net/forum?id=phnGilhPH8
Keywords: Federated Learning, Data Heterogeneity
Compressor summary: The paper proposes a method called FedFed that shares some features of data to improve federated learning while preserving privacy.
Jacob Beck, Risto Vuorio, Zheng Xiong, Shimon Whiteson
https://openreview.net/forum?id=pefAAzu8an
Keywords: meta-RL, RL, reinforcement learning, memory, rnn, recurrent, hypernetwork, few-shot
Compressor summary: The paper investigates how using hypernetworks can improve the performance of simple recurrent networks in meta-RL tasks.
Julien Grand-Clément, Marek Petrik
https://openreview.net/forum?id=pcuC65JWAa
Keywords: Markov Decision Process, Blackwell optimality, average optimality, robust optimization
Compressor summary: The paper introduces a discount factor for Markov Decision Processes called the Blackwell discount factor and shows how it can be used to find optimal policies without strict assumptions on the MDP structure.
Roey Magen, Ohad Shamir
https://openreview.net/forum?id=pcpjtYNJCH
Keywords: sample complexity; learning theory; neural networks; linear predictors
Compressor summary: The text discusses new findings on how many examples are needed to learn vector-valued linear predictors and neural networks, showing different sample complexity behavior compared to scalar-valued linear predictors.
Yuanhao Cai, Yuxin Zheng, Jing Lin, Xin Yuan, Yulun Zhang, Haoqian Wang
https://openreview.net/forum?id=pcKwgdVAlq
Keywords: Applications, Computer Vision, Low-level Vision, Image Restoration, Snapshot Compressive Imaging, Hyperspectral Image Reconstruction
Compressor summary: The paper proposes a novel method, BiSRNet, for efficient HSI restoration from compressed measurements in SCI systems using binarized convolutional modules that outperform existing binarization algorithms.
Sili Huang, Yanchao Sun, Jifeng Hu, Siyuan Guo, Hechang Chen, Yi Chang, Lichao Sun, Bo Yang
https://openreview.net/forum?id=pb1OwZNgr2
Keywords: reinforcement learning, generalization
Compressor summary: SGFD is a method to improve generalization in visual RL by decorrelating features using saliency maps and random Fourier functions.
Hanzhuo Huang, Yufan Feng, Cheng Shi, Lan Xu, Jingyi Yu, Sibei Yang
https://openreview.net/forum?id=paa2OU5jN8
Keywords: Text-to-Video, Zero-Shot Generation, Large Language Model, Latent Diffusion Models
Compressor summary: Text-to-video generation using large language models and latent diffusion models to create vivid and high-quality videos without video data or training.
Patrik Robert Gerber, Tianze Jiang, Yury Polyanskiy, Rui Sun
https://openreview.net/forum?id=paTESG8iSE
Keywords: Kernel methods, Maximum mean discrepancy, Likelihood-free inference, Hypothesis testing, Minimax statistics
Compressor summary: The paper studies the trade-off between labeling more inputs and using less training data in two applications: detecting the Higgs boson and identifying planted images in CIFAR-10 dataset.
Tianyi Chen, Qidi Wang, Zhen Dong, Liwei Shen, Xin Peng
https://openreview.net/forum?id=pZ2Ww45GkL
Keywords: program synthesis, partial envrionment, robotic programming, domain-specific language
Compressor summary: The paper presents a framework for synthesizing programs for robots in partially observed environments by learning an embedding space and using graph structure to rectify errors.
Shuyue Hu, Harold Soh, Georgios Piliouras
https://openreview.net/forum?id=pXtVyj4R33
Keywords: Multi-Agent Learning, Consensus Formation, Smooth Fictitious Play, Network Game, Population Game
Compressor summary: The paper explores how consensus and equilibrium are related in multi-agent systems with multiple interacting sub-populations and shows that smooth fictitious play can achieve both goals.
Pieter-Jan Hoedt, Günter Klambauer
https://openreview.net/forum?id=pWZ97hUQtQ
Keywords: initialization, signal propagation, input-convex networks
Compressor summary: ICNNs are networks with guaranteed convexity that require a new weight initialization method for better learning and generalization, which can also be trained without skip-connections and applied to drug discovery tasks.
Elie Bursztein, Marina Zhang, Owen Skipper Vallis, Xinyu Jia, Alexey Kurakin
https://openreview.net/forum?id=pVlC0reMKq
Keywords: language modeling, text embedding, adversarial text attack, text vectorization
Compressor summary: RETVec is a robust, multilingual text vectorizer with a character encoding and an optional embedding model that can handle typos and adversarial attacks.
Yixing Lao, Xiaogang Xu, zhipeng cai, Xihui Liu, Hengshuang Zhao
https://openreview.net/forum?id=pTCZWSDltG
Keywords: Neural Radiance Fields, 3D Reconstruction, Few View
Compressor summary: CorresNeRF uses image correspondence priors to supervise NeRF training, improving performance under sparse-view settings for novel view synthesis and surface reconstruction tasks.
Haixin Wang, Xinlong Yang, Jianlong Chang, Dian Jin, Jinan Sun, Shikun Zhang, Xiao Luo, Qi Tian
https://openreview.net/forum?id=pT8DIhsJCw
Keywords: parameter-efficient transfer learning; multi-modal learning; prompt learning
Compressor summary: The paper introduces AURORA, a framework for multimodal transfer learning that uses mode approximation and Informative Context Enhancement to achieve high performance with few parameters.
Yichao Cao, Qingfei Tang, Xiu Su, Song Chen, Shan You, Xiaobo Lu, Chang Xu
https://openreview.net/forum?id=pQvAL40Cdj
Keywords: Human-object interaction, Commonsense Knowledge, Foundation Models
Compressor summary: UniHOI is a method for recognizing complex human-object interactions in open world settings using vision-language foundation models and large language models to guide decoding, interpret interactions, and handle various input types.
Linyan Huang, Zhiqi Li, Chonghao Sima, Wenhai Wang, Jingdong Wang, Yu Qiao, Hongyang Li
https://openreview.net/forum?id=pQF9kbM8Ea
Keywords: camera-only detection, multi-modal distillation, multi-view object detection
Compressor summary: The paper proposes VCD, a framework that improves camera-only 3D object detection by using a multi-modal expert with LiDAR input and trajectory-based distillation supervision, achieving state-of-the-art results on nuScenes.
Qiyao Huang, Yingyue Zhang, Zhihong Zhang, Edwin Hancock
https://openreview.net/forum?id=pO7d6iFdnc
Keywords: Temporal Network, Graph Neural Network, Von Neumann Entropy
Compressor summary: ESSEN is a novel framework that uses von Neumann entropy and thermodynamic temperature to measure temporal network evolution and improve link prediction performance.
Qingxiu Dong, Jingjing Xu, Lingpeng Kong, Zhifang Sui, Lei Li
https://openreview.net/forum?id=pNtG6NAmx0
Keywords: Large Language Models, Knowledge Assessment, Evaluation
Compressor summary: The paper introduces KaRR, a statistical approach that quantifies how well large language models (LLMs) can generate factually correct answers for given facts, based on the ratio of LLM text to random chances. The paper tests 20 LLMs and finds strong agreement with human judgments.
Qian Huang, Hongyu Ren, Peng Chen, Gregor Kržmanc, Daniel Zeng, Percy Liang, Jure Leskovec
https://openreview.net/forum?id=pLwYhNNnoR
Keywords: Graph Neural Network, in-context learning, pretraining
Compressor summary: PRODIGY is a novel framework that enables in-context learning over graphs using a prompt graph representation and a graph neural network architecture, achieving strong performance on citation networks and knowledge graphs tasks.
Patric Bonnier, Harald Oberhauser, Zoltán Szabó
https://openreview.net/forum?id=pLsPFxqn7J
Keywords: kernel, cumulant, mean embedding, Hilbert-Schmidt independence criterion, maximum mean discrepancy
Compressor summary: The paper introduces kernelized cumulants for reproducing kernel Hilbert spaces, which are computationally efficient and provide new all-purpose statistics for data analysis.
Abdellah Aznag, Rachel Cummings, Adam N. Elmachtoub
https://openreview.net/forum?id=pLcSrn8NpJ
Keywords: Active learning, mean estimation, bandit feedback, data acquisition
Compressor summary: The paper proposes an active learning algorithm called Variance-UCB to learn the means of multiple unknown groups by minimizing the $p$-norm of variances, with optimal regret bounds for finite and infinite $p$.
Quanqi Hu, Dixian Zhu, Tianbao Yang
https://openreview.net/forum?id=pLOWV1UGF6
Keywords: non-smooth optimization, weakly-convex optimization, compositional optimization, AUC maximization
Compressor summary: The paper studies non-smooth weakly-convex finite-sum coupled compositional optimization problems and proposes algorithms for solving them in machine learning applications.
Ian Char, Jeff Schneider
https://openreview.net/forum?id=pKnhUWqZTJ
Keywords: Reinforcement Learning, Control, POMDP
Compressor summary: The paragraph discusses using PID features for history encoding in deep reinforcement learning to improve robustness and performance on various control tasks.
Odelia Melamed, Gilad Yehudai, Gal Vardi
https://openreview.net/forum?id=pJbEXBBN88
Keywords: Adversarial Examples, Robustness, Neural Networks, Classification
Compressor summary: This paper studies how gradient methods affect the robustness of two-layer neural networks trained on low dimensional linear subspaces and suggests ways to improve their resistance to adversarial examples in certain directions.
Andrew Wagenmaker, Guanya Shi, Kevin Jamieson
https://openreview.net/forum?id=pJQu0zpKCS
Keywords: reinforcement learning, control theory, system identification, experiment design, active learning
Compressor summary: The paper proposes an algorithm for exploring nonlinear dynamical systems to learn relevant parameters for controller optimization more efficiently.
Yuanhan Zhang, Kaiyang Zhou, Ziwei Liu
https://openreview.net/forum?id=pIXTMrBe7f
Keywords: computer vision, visual in-context learning, prompt learning
Compressor summary: This paper proposes prompt retrieval frameworks for large vision models to improve their performance in in-context learning by selecting appropriate input-output pairs without updating any internal model parameters.
Sina Akbari, Fateme Jamshidi, Ehsan Mokhtarian, Matthew James Vowels, Jalal Etesami, Negar Kiyavash
https://openreview.net/forum?id=pH4Fv7C3yC
Keywords: Causal effect, identifiability, causal DAGs, probabilistic graphs
Compressor summary: The paper studies how to find the most plausible probabilistic causal graph that can identify a specific causal effect, using algorithms that solve an NP-hard optimization problem called edge ID.
Zhuo Huang, Li Shen, Jun Yu, Bo Han, Tongliang Liu
https://openreview.net/forum?id=pE3yaP0Eqg
Keywords: Semi-Supervised Learning
Compressor summary: FlatMatch is a new SSL method that balances learning on labeled and unlabeled data by penalizing the prediction difference between worst-case and original models, improving generalization performance.
Valentino Maiorca, Luca Moschella, Antonio Norelli, Marco Fumero, Francesco Locatello, Emanuele Rodolà
https://openreview.net/forum?id=pBa70rGHlr
Keywords: latent space translation, relative representation, Procrustes analysis, zero-shot, stitching, latent communication, representation learning, manifold alignment, multimodal
Compressor summary: The paper proposes a method to translate and stitch representations learned from different neural networks, enabling better understanding of their intrinsic similarity and improving performance in various tasks and domains.
Minoh Jeong, Martina Cardone, Alex Dytso
https://openreview.net/forum?id=p9k5MS0JAL
Keywords: Bayes error, estimation, classification, minimum error probability
Compressor summary: The authors propose a new method to estimate the Bayes error rate for multi-class classification problems and analyze its properties and performance on synthetic and real data.
Yuchen Yan, Yuzhong Chen, Huiyuan Chen, Minghua Xu, Mahashweta Das, Hao Yang, Hanghang Tong
https://openreview.net/forum?id=p8lowHbuv8
Keywords: graph convolutional network
Compressor summary: The paper proposes a new GCN model that can adjust its depth continuously to handle both homophilic and heterophilic graph topologies, and shows improved performance on node classification tasks.
Yung-Hsuan Lai, Yen-Chun Chen, Yu-Chiang Frank Wang
https://openreview.net/forum?id=p8gTWkFIvx
Keywords: Audio-Visual Video Parsing, Audio-Visual Learning
Compressor summary: The paper introduces VALOR, a method that uses pre-trained models as teachers to help learn audio and visual labels from weak labels in an unaligned setting, improving performance on the Look, Listen, and Parse dataset and Audio-Visual Event Localization tasks.
Mohammadreza Pourreza, Davood Rafiei
https://openreview.net/forum?id=p53QDxSIc5
Keywords: In-Context Learning, Text-to-SQL, Task Decomposition, Spider Challenge, Natural Language Interfaces to Databases
Compressor summary: The paper proposes decomposing the text-to-SQL task into smaller sub-tasks and using LLMs to solve them, leading to improved performance and setting new benchmarks.
Ya-Ping Hsieh, Mohammad Reza Karimi Jaghargh, Andreas Krause, Panayotis Mertikopoulos
https://openreview.net/forum?id=p4SjKPchJy
Keywords: Riemannian optimization, saddle points, stochastic approximation
Compressor summary: The paper investigates if retraction-based stochastic Riemannian optimization algorithms can avoid saddle points with certain assumptions, ensuring the convergence to a local minimizer.
Michael Hanna, Ollie Liu, Alexandre Variengien
https://openreview.net/forum?id=p4PckNQR8k
Keywords: interpretability, language models, NLP
Compressor summary: The paper examines how pre-trained language models like GPT-2 small use basic math skills and identifies the specific components and mechanisms involved in solving a simple mathematical task.
Zhiqing Sun, Yikang Shen, Qinhong Zhou, Hongxin Zhang, Zhenfang Chen, David Daniel Cox, Yiming Yang, Chuang Gan
https://openreview.net/forum?id=p40XRfBX96
Keywords: AI Alignment, Large Language Models, In Context Learning, Neural Symbolics
Compressor summary: The proposed SELF-ALIGN approach enables AI assistant agents to align themselves with human intentions using minimal human supervision by combining LLMs and principle-driven reasoning.
Janaka Chathuranga Brahmanage, Jiajing Ling, Akshat Kumar
https://openreview.net/forum?id=p1gzxzJ4Y5
Keywords: action-constrained reinforcement learning, decision making
Compressor summary: ACRL uses a normalizing flow model to learn an invertible mapping between feasible actions and latent variables, enabling faster and more accurate action sampling and constraint satisfaction in DDPG-based RL problems.
Yefan Zhou, Tianyu Pang, Keqin Liu, charles h martin, Michael W. Mahoney, Yaoqing Yang
https://openreview.net/forum?id=oyV9FslE3j
Keywords: Heavy-tail self-regularization, learning rate schedule
Compressor summary: The paper introduces TempBalance, a layer-wise learning rate method for neural network training, which uses Heavy-Tailed Self-Regularization Theory to balance the temperature across layers and improve performance on various datasets.
Kanchana Ranasinghe, Michael S Ryoo
https://openreview.net/forum?id=oyFyOPZUCs
Keywords: self-supervised learning for videos, zero-shot action recognition
Compressor summary: The paper proposes a method to adapt image CLIP models to the video domain using language tied self-supervised learning, which improves action recognition performance on benchmarks.
Nicolas Menet, Michael Hersche, Geethan Karunaratne, Luca Benini, Abu Sebastian, Abbas Rahimi
https://openreview.net/forum?id=ox7aynitoW
Keywords: Computation in superposition, Vector-symbolic architectures, Convolutional neural networks, Transformers
Compressor summary: The authors propose MIMONets, which can process multiple inputs at once using variable binding and unbinding. This enables computation in superposition, improving inference efficiency and performance-accuracy trade-offs for CNN and Transformer models. They also provide theoretical bounds on the interference between superposition channels.
Yufeng Zhang, Jialu Pan, Kenli Li, Wanwei Liu, Zhenbang Chen, Xinwang Liu, J Wang
https://openreview.net/forum?id=ouLe91yibj
Keywords: Kullback-Leibler divergence, statistical divergence, multivariate Gaussian distribution, mathematical optimization, Lambert $W$ function, machine learning, flow-based model, reinforcement learning
Compressor summary: The paper studies properties of KL divergence between multivariate Gaussian distributions, deriving bounds for different scenarios and showing that it follows a relaxed triangle inequality.
Junyu Huang, Qilong Feng, Ziyun Huang, Jinhui Xu, Jianxin Wang
https://openreview.net/forum?id=oss2jXD1Zs
Keywords: Approximation Algorithms, k-means Clustering, Local Search
Compressor summary: The paper proposes a multi-swap local search algorithm for the $k$-means problem with linear running time and improved approximation ratio, and also introduces a sampling-based method and a recombination mechanism to enhance the performance.
Cuong Pham, Cuong C. Nguyen, Trung Le, Dinh Phung, Gustavo Carneiro, Thanh-Toan Do
https://openreview.net/forum?id=os2BdbiGwX
Keywords: Bayesian Neural Networks, Deep Mutual Learning
Compressor summary: The paper proposes a deep mutual learning approach that increases diversity in parameter and feature distributions of Bayesian Neural Networks, leading to significant performance improvements in classification accuracy, negative log-likelihood, and expected calibration error.
Haolin Liu, Chen-Yu Wei, Julian Zimmert
https://openreview.net/forum?id=orh4e0AO9R
Keywords: adversarial linear contextual bandits, log-determinant barrier
Compressor summary: The paper proposes a new method for solving the adversarial linear contextual bandit problem that achieves lower regret and maintains computational efficiency in the presence of stochastic arm availability and additive misspecification errors.
Ziqiao Wang, Yongyi Mao
https://openreview.net/forum?id=oqDSDKLd3S
Keywords: generalization, information-theoretic bounds, stability
Compressor summary: The authors propose a new way to measure generalization in machine learning using a matrix and a stability concept that improve existing bounds, especially for stochastic convex optimization problems.
Zhongbin Fang, Xiangtai Li, Xia Li, Joachim M. Buhmann, Chen Change Loy, Mengyuan Liu
https://openreview.net/forum?id=ooXpTZYwXa
Keywords: In-context learning, Point cloud, Prompt tuning
Compressor summary: The paper introduces Point-In-Context, a novel framework for in-context learning in 3D point clouds, which addresses technical challenges and outperforms individual models.
Xuchuang Wang, Qingyun Wu, Wei Chen, John C.S. Lui
https://openreview.net/forum?id=oi45JlpSOT
Keywords: Multi-fidelity, multi-armed bandits
Compressor summary: The paper investigates the multi-fidelity multi-armed bandit problem, where each arm has different costs and observation accuracies, and proposes algorithms for best arm identification and regret minimization.
Changmin Yu, Neil Burgess, Maneesh Sahani, Samuel Gershman
https://openreview.net/forum?id=ohKbQp0jIY
Keywords: Exploration, reinforcement learning
Compressor summary: The paper introduces SPIE, an exploration algorithm for reinforcement learning that combines prospective and retrospective information to generate more efficient and ethologically plausible behavior in environments with sparse rewards and bottleneck states.
Xiangzhi Chen, Le Wu, Fei Liu, Lei Chen, Kun Zhang, Richang Hong, Meng Wang
https://openreview.net/forum?id=ogPBujRhiN
Keywords: Intelligent Education System, Cognitive Diagnosis, Disentangled Representation Learning, Interpretability
Compressor summary: The paper proposes a new method called Disentanglement based Cognitive Diagnosis (DCD) for measuring students' proficiency in specific knowledge concepts with limited exercise labels, using response records and two novel modules.
Thomas Edward Yerxa, Yilun Kuang, Eero P Simoncelli, SueYeon Chung
https://openreview.net/forum?id=og9V7NgOrQ
Keywords: computational neuroscience, theoretical neuroscience, efficient coding, representation geometry, neural manifolds, self-supervised learning, statistical physics of learning
Compressor summary: The efficient coding hypothesis suggests sensory systems adapt to their inputs for maximal information capture, but measuring this is difficult; researchers have developed Maximum Manifold Capacity Representations (MMCR) which use a novel efficiency metric and are competitive with self-supervised learning methods.
Yu-Jie Zhang, Masashi Sugiyama
https://openreview.net/forum?id=ofa1U5BJVJ
Keywords: Logistic Bandit, Generalized Linear Bandit, Regret Bound, Computation Cost
Compressor summary: The paper proposes a logistic bandit algorithm that is statistically efficient, computationally fast, and adaptable to complex decision-making problems, improving on existing binary and multinomial methods.
Haitao Mao, Zhikai Chen, Wei Jin, Haoyu Han, Yao Ma, Tong Zhao, Neil Shah, Jiliang Tang
https://openreview.net/forum?id=oef30oScVB
Keywords: Graph Neural Network
Compressor summary: The study shows that Graph Neural Networks perform well on nodes with similar structures in homophilic graphs and heterophilic nodes in heterophilic graphs, but struggle with nodes having different structures in the opposite graph type.
Yuhang Yao, Weizhao Jin, Srivatsan Ravi, Carlee Joe-Wong
https://openreview.net/forum?id=ody3RBUuJS
Keywords: Federated Graph Learning;
Compressor summary: FedGCN is a novel algorithm that trains GCN models on graphs distributed among clients using federated learning, achieving fast convergence, high accuracy, and low communication.
Ilias Diakonikolas, Daniel Kane, Yuxin Sun
https://openreview.net/forum?id=obCNIzeSrg
Keywords: mixtures models, linear classifier, Statistical Query model, spherical designs
Compressor summary: The paper studies how to learn mixtures of linear classifiers with Gaussian features and shows that current algorithms are nearly optimal using a new construction of spherical designs.
Xinyu Sun, Peihao Chen, Jugang Fan, Jian Chen, Thomas H. Li, Mingkui Tan
https://openreview.net/forum?id=oaJEB5Qcia
Keywords: Visual Navigation, Image-Goal Navigation, Embodied AI
Compressor summary: The paper proposes a Fine-grained Goal Prompting method to improve image-goal navigation for autonomous systems by using high-resolution feature maps in the goal image as prompts.
Xueyuan Lin, Haihong E, Chengjin Xu, Gengxian Zhou, Haoran Luo, Tianyi Hu, Fenglong Su, Ningyuan Li, Mingzhi Sun
https://openreview.net/forum?id=oaGdsgB18L
Keywords: Temporal Knowledge Graph Reasoning, Temporal Knowledge Graph Embedding, Temporal Knowledge Graph, Temporal Logic, Knowledge Graph Reasoning, Knowledge Graph Embedding, Knowledge Graph, Machine Learning
Compressor summary: TFLEX is a novel framework for multi-hop logical reasoning over temporal knowledge graphs that uses fuzzy logic to handle complex queries involving entities, timestamps, and set operations.
Chaoqi Wang, Ziyu Ye, Zhe Feng, Ashwinkumar Badanidiyuru, Haifeng Xu
https://openreview.net/forum?id=oaCDiKoJ2w
Keywords: linear stochastic bandits, online learning, partial information, contextual bandits
Compressor summary: The paragraph discusses a novel contextual bandit problem with post-serving contexts, where additional valuable information about user rewards can be observed after arm selection, and presents poLinUCB, a new algorithm that achieves tight regret under standard assumptions using a robustified version of the Elliptical Potential Lemma.
Daniel Thuerck, Boro Sofranac, Marc Pfetsch, Sebastian Pokutta
https://openreview.net/forum?id=oU4QHdcIWW
Keywords: Integer Programming, Cutting Planes, Optimization
Compressor summary: The paper presents a novel method for learning facets of polyhedra by embedding an enumeration oracle in a Frank-Wolfe algorithm to generate strong cutting planes.
Changhyeon Lee, Seulki Lee
https://openreview.net/forum?id=oScaeIibRx
Keywords: Memory efficient, Activation saving memory, NLP, Transformer
Compressor summary: The paper proposes a method to reduce memory usage in attention-based networks by approximating the softmax output during training and shows that it improves performance with less memory.
Marina Munkhoeva, Ivan Oseledets
https://openreview.net/forum?id=oSYjkJKHZx
Keywords: unsupervised learning, self-supervised learning, representation learning, matrix completion
Compressor summary: The paper explores how self-supervised learning methods use a Laplace operator and low-rank matrix completion to learn representations without labels, and analyzes their convergence and performance.
Tobit Klug, Dogukan Atik, Reinhard Heckel
https://openreview.net/forum?id=oRn953uhFq
Keywords: image reconstruction, denoising, accelerated MRI, self-supervised, sample complexity
Compressor summary: The paper compares the performance and sample complexity of supervised and self-supervised methods for image reconstruction tasks, finding that self-supervised methods require more training data but eventually achieve similar results.
Dayoung Gong, Joonseok Lee, Deunsol Jung, Suha Kwak, Minsu Cho
https://openreview.net/forum?id=oOXZ5JEjPb
Keywords: neuro-symbolic approach, Temporal action segmentation, grammar
Compressor summary: The paper introduces KARI, a grammar induction algorithm for temporal action segmentation, and BEP, a parser that transforms frame-level probabilities into actions according to the induced grammar.
Fangcheng Zhong, Kyle Thomas Fogarty, Param Hanji, Tianhao Walter Wu, Alejandro Sztrajman, Andrew Everett Spielberg, Andrea Tagliasacchi, Petra Bosilj, Cengiz Oztireli
https://openreview.net/forum?id=oO1IreC6Sd
Keywords: neural fields, constrained optimization
Compressor summary: The authors present Constrained Neural Fields (CNF), a method for imposing hard constraints on neural networks during optimization, inspired by meshless interpolation and spectral collocation techniques in scientific computing.
Ziheng Cheng, Shiyue Zhang, Longlin Yu, Cheng Zhang
https://openreview.net/forum?id=oNuam8eFz2
Keywords: Particle-based VI, generalized Wasserstein gradient flow
Compressor summary: The paper introduces GWG, a novel ParVI framework that improves on ParVIs by using a generalized Wasserstein gradient flow of the KL divergence, which allows for more flexible kernel design and stronger convergence guarantees.
Maxence Noble, Valentin De Bortoli, Alain Durmus
https://openreview.net/forum?id=oMm1dfo3tK
Keywords: Hamiltonian Monte Carlo, Riemannian manifold, self-concordant barrier, constrained sampling
Compressor summary: The paper introduces Barrier Hamiltonian Monte Carlo, a method for sampling from constrained Gibbs distributions using Hamiltonian dynamics and a new filter step, which overcomes bias issues in existing Riemannian HMC methods.
Siliang Zeng, Chenliang Li, Alfredo Garcia, Mingyi Hong
https://openreview.net/forum?id=oML3v2cFg2
Keywords: Inverse Reinforcement Learning, Model-based Offline Inverse Reinforcement Learning
Compressor summary: The paragraph introduces a bi-level optimization approach to improve offline inverse reinforcement learning by incorporating uncertainty in the expert's model of the world.
Piotr Indyk, Haike Xu
https://openreview.net/forum?id=oKqaWlEfjY
Keywords: Nearest neighbor search; graph-based algorithms; worst-case analysis
Compressor summary: The paper analyzes the theoretical guarantees of graph-based approximate nearest neighbor search algorithms, finding that some variants have slow preprocessing but good performance, while others have fast preprocessing but high query time.
Muchao Ye, Ziyi Yin, Tianrong Zhang, Tianyu Du, Jinghui Chen, Ting Wang, Fenglong Ma
https://openreview.net/forum?id=oGxE2Nvlda
Keywords: certified robust training, text adversarial defense
Compressor summary: The paragraph describes a new framework called UniT that unifies and improves existing certified robust training pipelines for text classification by working in the word embedding space and using decoupled regularization loss to enhance the base model's robustness.
Zhengyi Yang, Jiancan Wu, Zhicai Wang, Xiang Wang, Yancheng Yuan, Xiangnan He
https://openreview.net/forum?id=oFpBnt6bgC
Keywords: Sequential Recommendation, Recommendation System, Generative Model, Diffusion Model
Compressor summary: The text describes a new sequential recommendation method, DreamRec, that generates an oracle item based on the user's historical interactions, rather than relying on negative sampling and classification.
Sanghyun Son, Laura Yu Zheng, Ryan Sullivan, Yi-Ling Qiao, Ming Lin
https://openreview.net/forum?id=oFaLc6fHSt
Keywords: Reinforcement Learning, Analytic Gradient-Based Policy Learning, Proximal Policy Optimization, Differentiable Programming
Compressor summary: The paper introduces a new policy learning method that combines analytical gradients with PPO, using an α-policy to adjust the influence of these gradients and improve performance in different environments.
Run Yang, Yuling Yang, Fan Zhou, Qiang Sun
https://openreview.net/forum?id=oDtyJt5JLk
Keywords: diffusion models, graph representation learning, unsupervised learning
Compressor summary: The paper proposes directional diffusion models for unsupervised graph representation learning, which use data-dependent, anisotropic, and directional noises to better capture graph structures and outperform existing methods on 12 datasets.
Zeyu Zhang, Yi Su, Hui Yuan, Yiran Wu, Rishab Balasubramanian, Qingyun Wu, Huazheng Wang, Mengdi Wang
https://openreview.net/forum?id=oDcWnfZyZW
Keywords: learning to rank, off-policy learning, reinforcement learning, click model
Compressor summary: The paper proposes a method (CUOLR) for off-policy Learning to Rank that adapts to various click models using offline reinforcement learning without complex debiasing techniques or prior knowledge.
Riccardo Zamboni, Alberto Maria Metelli, Marcello Restelli
https://openreview.net/forum?id=o91in9tDEs
Keywords: reinforcement learning, distributional reinforcement learning, maximum entropy estimation, representation learning
Compressor summary: The paper proposes a new Max-Ent framework for evaluating policies in distributional RL, which can consider the state representation complexity and guide state space learning using progressive factorization.
Chandra Sekhar Mukherjee, Pan Peng, Jiapeng Zhang
https://openreview.net/forum?id=o7W0Zet6p3
Keywords: SBM, Unbalanced SBM, Spectral algorithms, Small cluster barrier
Compressor summary: The paper proposes an SVD-based algorithm to detect communities of varying sizes in the stochastic block model and improves upon previous results, also providing an efficient clustering algorithm with sublinear query complexity.
Xin Cheng, Yuzhou Cao, Haobo Wang, Hongxin Wei, Bo An, Lei Feng
https://openreview.net/forum?id=o7HckkxOZH
Keywords: regression, rejection costs, surrogate loss
Compressor summary: The paper proposes a regression problem with cost-based rejection, where models reject examples based on their variance and mean squared error, and shows that this approach can improve prediction performance.
Jian Chen, Ruiyi Zhang, Tong Yu, Rohan Sharma, zhiqiang xu, Tong Sun, Changyou Chen
https://openreview.net/forum?id=o778eWSr1S
Keywords: diffusion model, label noise, retrieval augmented learning
Compressor summary: The paper proposes a new diffusion model that uses generative uncertainty and neighbor consistency to learn from noisy labels and achieve state-of-the-art results on real-world datasets.
Jishnu Ray Chowdhury, Cornelia Caragea
https://openreview.net/forum?id=o6yTKfdnbA
Keywords: Recursive Neural Networks, Long Range Arena, RvNN, Long Range Sequence Modeling, Length Generalization, LRA, Structured Encoding, Inductive Bias, Hierarchical Model, Recursive Models
Compressor summary: RIR combines a balanced tree model with a Beam Tree RvNN inner recursion to achieve length-generalization on ListOps and scalability on long sequence tasks like LRA, outperforming Transformers.
Yulun Zhang, Matthew Christopher Fontaine, Varun Bhatt, Stefanos Nikolaidis, Jiaoyang Li
https://openreview.net/forum?id=o6Dnt1uEyZ
Keywords: Multi-robot systems, quality diversity, automatic environment generation, neural cellular automata
Compressor summary: The paper proposes optimizing Neural Cellular Automata (NCA) environment generators via Quality Diversity (QD) algorithms to create arbitrarily large environments for multi-robot systems, improving their scalability and throughput.
Minhua Lin, Teng Xiao, Enyan Dai, Xiang Zhang, Suhang Wang
https://openreview.net/forum?id=o50nH0sV9x
Keywords: Certifiable Robustness, Graph Contrastive Learning
Compressor summary: The paper proposes a unified criterion and a novel technique called RES to certify and enhance the robustness of Graph Contrastive Learning models against adversarial attacks, which can be proven to be effective in downstream tasks.
Siyuan Guo, Viktor Tóth, Bernhard Schölkopf, Ferenc Huszár
https://openreview.net/forum?id=o4RtDFMSNL
Keywords: Independent Causal Mechanism, Causal Discovery, Exchangeable, Bayesian Statistics
Compressor summary: The text discusses how exchangeable data can help constraint-based causal discovery methods identify more accurate causal structures than i.i.d. data by using independent causal mechanism generative processes.
Jiaqi Zhang, Kristjan Greenewald, Chandler Squires, Akash Srivastava, Karthikeyan Shanmugam, Caroline Uhler
https://openreview.net/forum?id=o16sYKHk3S
Keywords: Causality, Identifiability, Disentanglement
Compressor summary: The paper proposes a method for causal disentanglement using latent variables, which can identify the causal model from unpaired observational and interventional data even when some causal variables are unobserved, and demonstrates its application to predicting combinatorial perturbation effects in genomics.
Samuel Holt, Alihan Hüyük, Mihaela van der Schaar
https://openreview.net/forum?id=o0ggjFD24U
Keywords: Sensing, Model-Based Reinforcement Learning
Compressor summary: The paper introduces a new continuous-time control problem with costly observations, where irregular observation policies can improve expected utility, and presents a simple initial method to solve it.
Carol Xuan Long, Hsiang Hsu, Wael Alghamdi, Flavio Calmon
https://openreview.net/forum?id=nzkWhoXUpv
Keywords: predictive multiplicity, fairness in machine learning, Rashomon effect
Compressor summary: The paper discusses how group fairness and accuracy optimizations can increase predictive multiplicity, a phenomenon where similar models produce different outputs for individual samples, and proposes an ensemble algorithm to improve consistency in individual-level decision-making.
Xinran Zhu, Kaiwen Wu, Natalie Maus, Jacob R. Gardner, David Bindel
https://openreview.net/forum?id=nwK8UkK3uB
Keywords: Gaussian processes, variational inference, variational Gaussian processes, Bayesian optimization
Compressor summary: The paper proposes a new way to improve variational Gaussian processes by making the predictive mean and covariance more flexible, leading to better performance on regression tasks and Bayesian optimization.
Brahma S Pavse, Josiah P. Hanna
https://openreview.net/forum?id=nvX3MiQM0G
Keywords: reinforcement learning, off-policy evaluation, off-policy RL, representation learning, behavioral similarity metrics
Compressor summary: The paper proposes a new method to improve off-policy evaluation in reinforcement learning by using a learned encoder to transform the dataset before applying fitted q-evaluation, and introduces an OPE-specific state-action similarity metric to learn the encoder.
Alexander Gilbert Reisach, Myriam Tami, Christof Seiler, Antoine Chambaz, Sebastian Weichwald
https://openreview.net/forum?id=nrbR2F29vU
Keywords: Causal Discovery, Directed Acyclic Graph, Varsortability, Additive Noise Model, Structural Causal Model, Simulation, Benchmark
Compressor summary: The paper studies how different parameter choices in additive noise models affect causal discovery and introduces a new method called $R^2$-SortnRegress that leverages high $R^2$ values to find causal orders.
Yuhang Zhang, Yaqi Li, lixiong Qin, Xuannan Liu, Weihong Deng
https://openreview.net/forum?id=nrQif5tH7O
Keywords: Facial expression recognition, imbalanced learning
Compressor summary: The paper proposes a novel method for facial expression recognition that extracts extra knowledge from both major and minor class samples using re-balanced attention maps and smooth labels to handle the imbalance problem.
Yunyao Mao, Jiajun Deng, Wengang Zhou, Li Li, Yao Fang, Houqiang Li
https://openreview.net/forum?id=nqIIWnwe73
Keywords: human-object interaction detection, zero-shot learning, CLIP model adaptatiion
Compressor summary: CLIP4HOI is a novel framework for zero-shot human-object interaction detection that uses CLIP to independently identify humans and objects and adapts the model into a fine-grained classifier for proposal discrimination, improving transferability and performance on rare and unseen categories.
Hezhe Qiao, Guansong Pang
https://openreview.net/forum?id=nq4OhifyEe
Keywords: Anomaly Detection, Graph Neural Network, Graph Anomaly Detection, One-Class Homophily, Local Node Affinity
Compressor summary: The paper introduces a novel anomaly scoring method for graph anomaly detection, based on local node affinity, and a tailored representation learning approach called Truncated Affinity Maximization (TAM) that leverages the one-class homophily phenomenon.
Muhammad Faaiz Taufiq, Arnaud Doucet, Rob Cornish, Jean-Francois Ton
https://openreview.net/forum?id=noyleECBam
Keywords: contextual bandits, variance reduction, off-policy evaluation
Compressor summary: The Marginal Ratio (MR) estimator is a new off-policy evaluation method for contextual bandits that reduces variance compared to existing methods like Inverse Probability Weighting and Doubly Robust, and performs well in causal inference settings.
Lei Xu, Lei Chen, Rong Wang, Feiping Nie, Xuelong Li
https://openreview.net/forum?id=noMktb4ait
Keywords: Feature Selection, Differential k-NN Graph, Dirichlet Energy
Compressor summary: The paper proposes a deep feature selection method that uses differentiable k-NN graph learning based on Dirichlet Energy to identify important features and learn dynamic graphs, and applies Optimal Transport theory to address non-differentiability issues.
Dongkuk Si, Chulhee Yun
https://openreview.net/forum?id=nijJN0LHqM
Keywords: Sharpness-Aware Minimization, convex optimization
Compressor summary: The paper studies how varying the parameters in an optimizer called SAM affects its convergence properties on smooth and non-smooth functions.
Zunzhi You, Daochang Liu, Bohyung Han, Chang Xu
https://openreview.net/forum?id=niHkj9ixUZ
Keywords: self-supervised learning, adversarial robustness
Compressor summary: The paper proposes De^3, a method that uses noisy image modeling to improve adversarial robustness of deep neural networks by exploiting the pretrained decoder for denoising and sampling the noise scale hyperparameter from random distributions.
Hao Wang, Shivchander Sudalairaj, John Henning, Kristjan Greenewald, Akash Srivastava
https://openreview.net/forum?id=neu9JlNweE
Keywords: differential privacy, synthetic data
Compressor summary: The paper introduces a post-processing technique to improve the utility of synthetic data for specific end-user requirements while maintaining privacy and quality, by resampling and optimizing weights using efficient algorithms.
Xiangning Chen, Chen Liang, Da Huang, Esteban Real, Kaiyuan Wang, Hieu Pham, Xuanyi Dong, Thang Luong, Cho-Jui Hsieh, Yifeng Lu, Quoc V Le
https://openreview.net/forum?id=ne6zeqLFCZ
Keywords: AutoML
Compressor summary: The authors propose a method to discover optimization algorithms for deep neural network training and introduce Lion, an efficient and effective optimizer that performs well on various tasks.
Jacob A Zavatone-Veth, Cengiz Pehlevan
https://openreview.net/forum?id=nbG6zfJtIe
Keywords: random feature models, generalization, deep networks, ridge regression
Compressor summary: The study examines how weight anisotropy and structured Gaussian features impact generalization in deep learning models using the replica trick from statistical physics.
Shashank Hegde, Sumeet Batra, K.R. Zentner, Gaurav S. Sukhatme
https://openreview.net/forum?id=nafgeYknRT
Keywords: Latent Diffusion, Quality Diversity, Reinforcement Learning, Graph Neural Networks
Compressor summary: The paper proposes using diffusion models to compress a collection of diverse policies into one generative model that can select and sequence behaviors with language.
Elias Nehme, Omer Yair, Tomer Michaeli
https://openreview.net/forum?id=nZ0jnXizyR
Keywords: Uncertainty Quantification, Inverse Problems, Probabilistic Modelling, Principal Components Analysis, Deep Learning
Compressor summary: This paper proposes a fast method to predict the principal components of the posterior distribution for image restoration models, enabling better uncertainty quantification and visualization.
Gabriele Farina, Julien Grand-Clément, Christian Kroer, Chung-Wei Lee, Haipeng Luo
https://openreview.net/forum?id=nYgs0qZJ97
Keywords: Regret Matching, Predictive algorithms, Extensive-Form Games
Compressor summary: The paper investigates the instability of regret matching algorithms, proposes fixes, and demonstrates their effectiveness in various game scenarios using theory and experiments.
Changho Shin, Sonia Cromp, Dyah Adila, Frederic Sala
https://openreview.net/forum?id=nXPqMyWUnx
Keywords: Weak supervision, fairness
Compressor summary: Weak supervision can produce biased pseudolabels, and a fairness-based technique can improve accuracy and reduce bias in weak supervision without tradeoffs.
Kilian Pfeiffer, Ramin Khalili, Joerg Henkel
https://openreview.net/forum?id=nXNsqB4Yr1
Keywords: Federated Learning, Memory, Resource Constraints
Compressor summary: The paper proposes a new method for federated learning on edge devices that reduces resource requirements and improves model accuracy by enabling successive freezing and training of model parameters.
Jinyuan Jia, Zhuowen Yuan, Dinuka Sahabandu, Luyao Niu, Arezoo Rajabi, Bhaskar Ramasubramanian, Bo Li, Radha Poovendran
https://openreview.net/forum?id=nX0zYBGEka
Keywords: backdoor defense, federated learning, game theory
Compressor summary: Federated learning faces backdoor attacks from dynamic attackers, and the paper proposes FedGame, a minimax game-based interactive defense mechanism that enhances robustness against such attacks.
MORTEZA GHAHREMANI, Christian Wachinger
https://openreview.net/forum?id=nUbdkXqC8R
Keywords: Multimodal Data, Multimodality, Batch Normalization, Heterogeneous data, Regularization, Confounder, Confounding Effect Removal, Data Dependency
Compressor summary: The paper introduces RegBN, a novel approach for multimodal Batch Normalization with REGularization, which uses the Frobenius norm as a regularizer term to address confounders and dependencies among different data sources, simplifying training and inference, and improving performance across diverse modalities and architectures.
Xing Gao, Yu Cheng
https://openreview.net/forum?id=nSr2epejn2
Keywords: Matrix sensing, Optimization, Low-rank matrix recovery, Semi-random, Adversarial input, Robustness
Compressor summary: The paper proposes a new non-convex optimization algorithm for semi-random matrix sensing that avoids bad local optima by reweighting the input matrices based on the current solution and takes weighted gradient steps.
Souhaib Attaiki, Maks Ovsjanikov
https://openreview.net/forum?id=nSgMh5v5Ne
Keywords: shape matching
Compressor summary: Shape Non-rigid Kinematics (SNK) is a new method for matching shapes that requires no training or ground truth data, and uses an encoder-decoder architecture to deform the source shape to match the target shape.
Lawrence Stewart, Francis Bach, Felipe Llinares-López, Quentin Berthet
https://openreview.net/forum?id=nRfcVBsF9n
Keywords: Structured learning, Clustering, Differentiable, weakly supervised, semi-supervised, representation learning
Compressor summary: A new method for clustering data that can be trained end-to-end with efficient gradients and works well in noisy and complex settings is presented, along with a custom loss function for learning from partial clustering data.
Zhengxuan Wu, Atticus Geiger, Thomas Icard, Christopher Potts, Noah Goodman
https://openreview.net/forum?id=nRfClnMhVX
Keywords: Mechanistic Interpretability
Compressor summary: The paper introduces Boundless DAS, an improved method to find interpretable causal structures in large language models, and applies it to the Alpaca model, revealing its use of two boolean variables for numerical reasoning.
Nataly Brukhim, Amit Daniely, Yishay Mansour, Shay Moran
https://openreview.net/forum?id=nQ84YY9Iut
Keywords: Boosting, Multiclass classification, PAC Learning, List PAC Learning
Compressor summary: The paper presents a new boosting method for multiclass classification that generalizes weak learnability, has low sample complexity, and simplifies theoretical applications in list learning.
Jianqing Zhang, Yang Hua, Jian Cao, Hao Wang, Tao Song, Zhengui XUE, Ruhui Ma, Haibing Guan
https://openreview.net/forum?id=nO5i1XdUS0
Keywords: Federated Learning, Personalized Federated Learning, Representation, Knowledge Transfer
Compressor summary: The paper introduces a framework called Domain Bias Eliminator (DBE) to address representation degeneration in federated learning and improve generalization and personalization abilities.
Bingrui Li, Jianfei Chen, Jun Zhu
https://openreview.net/forum?id=nN8TnHB5nw
Keywords: memory efficiency, optimizer, Adam, quantization
Compressor summary: The paper proposes 4-bit quantization of optimizer states for neural networks, achieving memory efficiency without sacrificing accuracy on various tasks.
Zhonghang Li, Lianghao Xia, Yong Xu, Chao Huang
https://openreview.net/forum?id=nMH5cUaSj8
Keywords: Spatial Temporal Prediction, Deep Neural Networks, Pre-training Model
Compressor summary: The authors propose a spatio-temporal pre-training framework that enhances the performance of downstream models for traffic management and travel planning by learning customized representations and adaptive mask strategies.
Yuan Cheng, Jing Yang, Yingbin Liang
https://openreview.net/forum?id=nMB41QjLDY
Keywords: Reinforcement Learning, Nonstationary Environment, Representation Learning, Policy Optimization, Statistical Complexity
Compressor summary: The paper proposes two algorithms, PORTAL and Ada-PORTAL, for reinforcement learning in nonstationary low-rank Markov Decision Processes, with theoretical guarantees on their sample efficiency and suboptimality gap.
David Loiseaux, Mathieu Carrière, Andrew Blumberg
https://openreview.net/forum?id=nKCUDd9GYu
Keywords: Topological Data Analysis, Multiparameter Persistent Homology, Kernel Methods, Convergence Rate, Statistical Learning
Compressor summary: The article proposes a new representation framework for multiparameter persistent homology in topological data analysis, which is stable, efficient, and informative, and shows its effectiveness on geometric and point cloud data.
Gleb Bazhenov, Denis Kuznedelev, Andrey Malinin, Artem Babenko, Liudmila Prokhorenkova
https://openreview.net/forum?id=nJFJcgjnGo
Keywords: graph, distributional shift, structural shift, uncertainty, robustness, graph neural networks
Compressor summary: The authors propose a method to create diverse distributional shifts in graph learning based on node structure, evaluate its difficulty, and show that simple models often perform better than complex ones.
Cong Wang, Jinshan Pan, Wei Wang, Jiangxin Dong, Mengzhu Wang, Yakun Ju, Junyang Chen
https://openreview.net/forum?id=nIaNgaQvsV
Keywords: Degradation Vanishing, Prompting Learning, Image Restoration
Compressor summary: PromptRestorer is a novel image restoration model that uses degradation features to guide the restoration process and achieves state-of-the-art results on 4 tasks.
Anqi Mao, Mehryar Mohri, Yutao Zhong
https://openreview.net/forum?id=nI7EmXq2PL
Keywords: consistency, H-consistency, characterization, learning theory
Compressor summary: The paper introduces general characterizations and extensions of $H$-consistency bounds for surrogate losses in multi-class classification, covering constrained and comp-sum losses, and providing tighter bounds than previous studies.
Tran Phong, Haoran Wu, Cunjun Yu, Panpan Cai, Sifa Zheng, David Hsu
https://openreview.net/forum?id=nG35q8pNL9
Keywords: trajectory prediction; autonomous driving
Compressor summary: The paper studies the dynamics gap between trajectory prediction accuracy on fixed datasets and real-world driving performance, and emphasizes the trade-off between computational efficiency and prediction accuracy for autonomous driving systems.
Xiang Ji, Gen Li
https://openreview.net/forum?id=nFsbQHFmj2
Keywords: reinforcement learning theory, regret minimization, minimax optimality
Compressor summary: The paper presents a new model-free reinforcement learning algorithm that achieves optimal performance with low memory and computational cost, and fast sample efficiency.
Hui Guo, Boyu Wang, Grace Yi
https://openreview.net/forum?id=nFEQNYsjQO
Keywords: Noisy Label, Instance-Dependent Transition Matrix, Label Correction, Crowdsourcing
Compressor summary: The paper proposes a new label correction algorithm using a Bayesian network to model instance-dependent noise transitions in crowdsourced annotations, with theoretical guarantees and experimental validation.
Aleksandar Stanić, Anand Gopalakrishnan, Kazuki Irie, Jürgen Schmidhuber
https://openreview.net/forum?id=nF6X3u0FaA
Keywords: object-centric learning, complex-valued networks, unsupervised learning, temporal correlation hypothesis
Compressor summary: The authors propose a new method for synchrony-based object-centric models that improves their ability to learn from multi-object color datasets and store more than three objects.
Shai Ben-David, Alex Bie, Clement Louis Canonne, Gautam Kamath, Vikrant Singhal
https://openreview.net/forum?id=nDIrJmKPd5
Keywords: differential privacy, distribution learning, gaussians, mixture of gaussians, compression schemes, robust compression schemes, privacy
Compressor summary: The paper studies how to privately learn a distribution from public and private data, and shows that this depends on compressing the data and learning from lists of examples.
Qi Zhu, Man Zhou, Jie Huang, Naishan Zheng, Hongzhi Gao, Chongyi Li, Yuan Xu, Feng Zhao
https://openreview.net/forum?id=nCwStXFDQu
Keywords: Image restoration, Down-Sampling, Fourier transform
Compressor summary: The authors propose a new down-sampling method called FouriDown, which uses a learnable and context-adaptive Fourier function to improve image restoration tasks like de-blurring and low-light enhancement.
Victor Boone, Panayotis Mertikopoulos
https://openreview.net/forum?id=nCLdsEzZBV
Keywords: Regularized learning, dynamic stability, strategic stability, Nash equilibrium
Compressor summary: The paper studies how no-regret learning in multiplayer games leads to setwise rationality properties and closedness under better replies, and explores the convergence rate of different regularization methods.
Langzhang Liang, Xiangjing Hu, Zenglin Xu, Zixing Song, Irwin King
https://openreview.net/forum?id=nBFMCyEi0j
Keywords: graph neural networks, heterophily problem, global label relationship matrix
Compressor summary: LRGNN is a generic GNN that works on both homophilous and heterophilous graphs by predicting the low-rank label relationship matrix using robust low-rank approximation, which has two advantages for graph modeling.
Manuel Tran, Yashin Dicente Cid, Amal Lahiani, Fabian J Theis, Tingying Peng, Eldad Klaiman
https://openreview.net/forum?id=nArzDm353Y
Keywords: generative pre-training, causal modeling, masked modeling, commutative modeling, transitive modeling, multimodal learning
Compressor summary: LoReTTa is a self-supervised framework that links modalities with causal modeling and masked modeling to handle multimodal datasets with missing modalities.
Gon Buzaglo, Niv Haim, Gilad Yehudai, Gal Vardi, Yakir Oz, Yaniv Nikankin, michal Irani
https://openreview.net/forum?id=nA9Fh3HFHJ
Keywords: memorization, data reconstruction, implicit bias
Compressor summary: This paper extends a previous study on reconstructing training data from neural networks and explores factors affecting this process, such as weight decay and neuron count.
Haonan Yuan, Qingyun Sun, Xingcheng Fu, Ziwei Zhang, Cheng Ji, Hao Peng, Jianxin Li
https://openreview.net/forum?id=n8JWIzYPRz
Keywords: dynamic graph learning, out-of-distribution generalization, invariant learning, link prediction
Compressor summary: The EAGLE framework models and exploits spatio-temporal environments for out-of-distribution generalization in dynamic graph neural networks, achieving superior performance against existing methods under distribution shifts.
David Ruhe, Johannes Brandstetter, Patrick Forré
https://openreview.net/forum?id=n84bzMrGUD
Keywords: Clifford algebras, geometric deep dearning, Clifford group equivariance, E(n)-equivariant neural networks, O(n)-equivariant neural networks
Compressor summary: The paragraph introduces a new method for building equivariant neural networks using the Clifford group, which has desirable properties and shows state-of-the-art performance on various tasks.
Rattana Pukdee, Dylan Sam, J Zico Kolter, Nina Balcan, Pradeep Kumar Ravikumar
https://openreview.net/forum?id=n6ztJ3Lrdj
Keywords: Interpretable ML, Semi-supervised learning, Learning theory
Compressor summary: The paper proposes a learning framework that uses explanation constraints from prior knowledge to improve the performance and interpretability of deep learning models, especially when using gradient information as explanations.
Youssef Allouah, Rachid Guerraoui, Nirupam Gupta, Rafael Pinot, Geovani Rizk
https://openreview.net/forum?id=n3fPDW87is
Keywords: Optimization, Byzantine resilience, Distributed machine learning, federated learning
Compressor summary: The paper proposes a more realistic heterogeneity model for robust distributed learning algorithms, showing lower breakdown points and better matching with empirical observations.
Siran Dai, Qianqian Xu, Zhiyong Yang, Xiaochun Cao, Qingming Huang
https://openreview.net/forum?id=n3ZVdny7OH
Keywords: Robust Learning AUC
Compressor summary: The paper proposes a method to optimize AUC in long-tailed classification scenarios using distributionally robust optimization and a surrogate loss, while addressing label bias.
Guillaume Mahey, Laetitia Chapel, Gilles Gasso, Clément Bonet, Nicolas Courty
https://openreview.net/forum?id=n3XuYdvhNW
Keywords: Optimal Transport, Wasserstein distance, Generalized Geodesics, Sliced Wasserstein
Compressor summary: The paper introduces a new proxy for Wasserstein distance called min-SWGG, which is based on one-dimensional projections and has benefits in various applications such as gradient flows, shape matching, and image colorization.
Alexander Tyurin, Peter Richtárik
https://openreview.net/forum?id=n18MhTsSGb
Keywords: convex optimization, accelerated method, communication compression, bidirectional compression, distributed optimization
Compressor summary: The 2Direction method speeds up distributed convex optimization by using bidirectional compressed communication and a new error-feedback mechanism, outperforming previous methods in both communication complexity and acceleration.
Trung Dang, Jasper C.H. Lee, Maoyuan Song, Paul Valiant
https://openreview.net/forum?id=mvSDs51eqQ
Keywords: mean estimation, instance optimality
Compressor summary: The paper explores if mean estimation algorithms can benefit from knowing specific features of the input distribution and finds that it's possible in some cases, but generally not, while introducing a new framework for analyzing algorithm optimality.
Stephen Chung, Ivan Anokhin, David Krueger
https://openreview.net/forum?id=mumEBl0arj
Keywords: Reinforcement learning, model-based reinforcement learning, planning, Monte Carlo Tree Search, Markov Decision Process
Compressor summary: The Thinker algorithm is a novel reinforcement learning approach that allows agents to interact with and utilize a learned world model for autonomous planning, achieving state-of-the-art performance in Sokoban and competitive results in Atari 2600.
Jonas Schweisthal, Dennis Frauen, Valentyn Melnychuk, Stefan Feuerriegel
https://openreview.net/forum?id=muVKSb8gi5
Keywords: off-policy learning, causal inference, reliable machine learning, medicine, dosaging, normalizing flows
Compressor summary: The paper proposes a novel method for personalized medicine that estimates individualized dose-response functions and optimal dosage combinations using neural networks, conditional normalizing flows, and gradient-based learning.
Anant Raj, Umut Simsekli, Alessandro Rudi
https://openreview.net/forum?id=mookk2nLO9
Keywords: Kernel Methods, Sampling, Fokker-Planck Equation, Fractional Fokker-Planck Equation, Stochastic Differential Equations, Partial Differential Equations
Compressor summary: The paper proposes an efficient method to sample from a stochastic differential equation using a recent probabilistic model and shows that it works well for smooth solutions.
Anagh Malik, Parsa Mirdehghan, Sotiris Nousias, Kyros Kutulakos, David B. Lindell
https://openreview.net/forum?id=mmmd2vp0n0
Keywords: neural radiance fields, 3D reconstruction, single-photon lidar, computational imaging
Compressor summary: The paper proposes a novel method to render transient NeRFs using single-photon lidar data and captures light transport phenomena at picosecond timescales, showing improved geometry and appearance compared to point cloud-based supervision.
Constantine Caramanis, Dimitris Fotakis, Alkis Kalavasis, Vasilis Kontonis, Christos Tzamos
https://openreview.net/forum?id=mmTy1iyU5G
Keywords: Policy Gradient, Combinatorial Optimization, Gradient Descent
Compressor summary: The authors develop a theoretical framework to analyze effective deep neural network and reinforcement learning methods for solving combinatorial problems, showing they can find near-optimal solutions with a limited number of parameters and without getting stuck in sub-optimal states. They also propose a novel regularization process for gradient descent that improves performance.
Drago Plecko, Elias Bareinboim
https://openreview.net/forum?id=mm9svgvwvk
Keywords: Causal Inference, Confounding, Fair and Explainable AI
Compressor summary: The authors develop formal tools for decomposing spurious associations in causal mechanisms using Markovian and Semi-Markovian models, providing applications in various fields and demonstrating their approach on a real dataset.
Hao Liu, Wilson Yan, Pieter Abbeel
https://openreview.net/forum?id=mlxRLIy7kc
Keywords: Large Language Model, VQVAE, Vector Quantization, Multimodal
Compressor summary: The paper proposes LQAE, a method to align images and texts without supervision, using pretrained language models, which improves few-shot learning in vision tasks like image classification and VQA.
Krishna Pillutla, Galen Andrew, Peter Kairouz, Hugh Brendan McMahan, Alina Oprea, Sewoong Oh
https://openreview.net/forum?id=mlbes5TAAg
Keywords: Differential privacy auditing, multiple canaries, randomization, lifting, adaptive confidence intervals
Compressor summary: The authors propose a method to audit differentially private machine learning by adding multiple randomized examples (canaries) and use lifted differential privacy, statistical tests, and confidence intervals to improve sample complexity.
Gleb Novikov, David Steurer, Stefan Tiegel
https://openreview.net/forum?id=mkve1raJUc
Keywords: Robust Mean Estimation, Unbounded First Moment, Symmetric Distributions (Spherical, Elliptical, Product), Filtering Algorithm, Huber Loss
Compressor summary: The paper proposes efficient and robust algorithms for estimating the mean of symmetric distributions, including product Cauchy and elliptical distributions, without moment assumptions or strong distributional assumptions.
Moise Blanchard, Junhui Zhang, Patrick Jaillet
https://openreview.net/forum?id=mkKQr56xdB
Keywords: Convex optimization, feasibility problem, first-order methods, memory constraints, cutting planes, oracle complexity
Compressor summary: The paper introduces recursive cutting-plane algorithms for solving feasibility problems and convex optimization with constrained memory, achieving sub-polynomial complexity and optimal memory trade-offs in some regimes.
Alexandre Marthe, Aurélien Garivier, Claire Vernade
https://openreview.net/forum?id=mgNu8nDFwa
Keywords: Markov Decision Process, Dynamic Programming, statistical functionnals, Distributionnal Reinforcement Learning, Policy Evaluation, Planning
Compressor summary: The paragraph discusses how Dynamic Programming can handle certain operations efficiently in Markov Decision Processes, but only for specific statistics classes, and explores the use of Distributional Reinforcement Learning to evaluate functionals approximately.
Nate Gruver, Marc Anton Finzi, Shikai Qiu, Andrew Gordon Wilson
https://openreview.net/forum?id=md68e8iZK1
Keywords: large language models, time series, probabilistic forecasting
Compressor summary: The authors propose using large language models to predict future values in time series by encoding them as text and show their effectiveness in handling various aspects of time series data.
Ava Pun, Gary Sun, Jingkang Wang, Yun Chen, Ze Yang, Sivabalan Manivasagam, Wei-Chiu Ma, Raquel Urtasun
https://openreview.net/forum?id=mcx8IGneYw
Keywords: Scene Relighting, Lighting Estimation, Camera Simulation, Self-Driving, Lighting Simulation, Scene Editing
Compressor summary: LightSim is a neural lighting camera simulation system that creates realistic and diverse images under different illumination conditions for training image-based robot perception models.
Yuhang Li, Tamar Geller, Youngeun Kim, Priyadarshini Panda
https://openreview.net/forum?id=mbaN0Y0QTw
Keywords: Spiking Neural Networks, ANN-SNN Conversion, Conditional Computing
Compressor summary: This paper introduces Spiking Early-Exit Neural Networks (SEENNs), which adjust the number of timesteps in spiking neural networks based on input samples, achieving better accuracy and efficiency tradeoffs.
Liyuan Liu, Chengyu Dong, Xiaodong Liu, Bin Yu, Jianfeng Gao
https://openreview.net/forum?id=mayAyPrhJI
Keywords: discrete random variables, back-propagation, straight through
Compressor summary: ReinMax is a novel method that approximates gradients for discrete latent variables in deep learning using Heun's method, achieving second-order accuracy with negligible computation overheads and outperforming existing approaches.
Dongrui Liu, Huiqi Deng, Xu Cheng, Qihan Ren, Kangrui Wang, Quanshi Zhang
https://openreview.net/forum?id=mZ3hnyL9bS
Keywords: representation complexity, deep learning
Compressor summary: The paper explores why deep neural networks find it harder to learn complex concepts involving many input variables, and identifies the specific factor that increases learning difficulty.
Tiffany Ding, Anastasios Nikolas Angelopoulos, Stephen Bates, Michael Jordan, Ryan Tibshirani
https://openreview.net/forum?id=mYz6ApeU4J
Keywords: conformal prediction, uncertainty quantification, class imbalance
Compressor summary: Clustered conformal prediction is a method that groups similar classes together and improves the probability of correctly predicting their labels when there is limited labeled data per class.
Yi Ma, Hongyao Tang, Dong Li, Zhaopeng Meng
https://openreview.net/forum?id=mVywRIDNIl
Keywords: Offline Reinforcement Learning
Compressor summary: The paper proposes Representation Distinction (RD), a technique that improves offline reinforcement learning by differentiating between in-sample and out-of-distribution data representations to reduce overgeneralization.
Yizhou Zhang, Jingchao Ni, Wei Cheng, Zhengzhang Chen, Liang Tong, Haifeng Chen, Yan Liu
https://openreview.net/forum?id=mVTyeQIiE4
Keywords: Few-Shot Learning, Meta Learning, Task Representation
Compressor summary: The paper proposes a meta-learning framework that models the density of task instances using a Hierarchical Gaussian Mixture based Task Generative Model (HTGM), which helps adapt to new tasks from different or unseen distributions.
Junda Wu, Tong Yu, Rui Wang, Zhao Song, Ruiyi Zhang, Handong Zhao, Chaochao Lu, Shuai Li, Ricardo Henao
https://openreview.net/forum?id=mSNfjOcDUv
Keywords: soft prompt tuning
Compressor summary: The paragraph describes a new method called InfoPrompt that improves soft prompt tuning by maximizing the mutual information between prompts and other model parameters, leading to faster convergence and better performance.
Linhao Qu, xiaoyuan Luo, Kexue Fu, Manning Wang, Zhijian Song
https://openreview.net/forum?id=mSDfBXr8Py
Keywords: multiple instance learning, whole slide image classification, prompt learning, vision-language model, few-shot learning
Compressor summary: The paper proposes a novel few-shot weakly supervised learning method for pathology Whole Slide Image (WSI) classification using prompt learning and GPT-4 to handle challenges posed by the weak bag labels within the Multiple Instance Learning framework.
Matthias Minderer, Alexey A. Gritsenko, Neil Houlsby
https://openreview.net/forum?id=mQPNcBWjGc
Keywords: object detection, open-vocabulary object detection, vision transformers, vision-language models, scaling, self-training
Compressor summary: OWLv2 is a model that uses self-training with pseudo-annotations to scale up object detection data, achieving significant improvements in performance on rare classes.
Jian Meng, Li Yang, Kyungmin Lee, Jinwoo Shin, Deliang Fan, Jae-sun Seo
https://openreview.net/forum?id=mOVEJletyD
Keywords: Contrastive Learning, Self-supervised Learning, Energy-efficient contrastive learning
Compressor summary: SACL-XD is a new self-supervised contrastive learning scheme that improves the performance of lightweight models and reduces energy consumption in AI applications.
Jan Schuchardt, Yan Scholten, Stephan Günnemann
https://openreview.net/forum?id=mLe63bAYc7
Keywords: Adversarial robustness, Geometric machine learning, Equivariances, Robustness Certification, Graph neural networks
Compressor summary: The paragraph discusses a new approach to adversarial robustness that considers task equivariance in real-world tasks like molecular property prediction, and proposes methods to achieve provable robustness for various models and architectures.
Ilias Diakonikolas, Daniel Kane, Vasilis Kontonis, Sihan Liu, Nikos Zarifis
https://openreview.net/forum?id=mIm0hsUUt1
Keywords: Machine Learning
Compressor summary: The paper presents a fast algorithm for learning halfspaces under the Gaussian distribution with adversarial label noise, using iterative soft localization and testers to ensure data quality.
yatong sun, Bin Wang, Zhu Sun, Xiaochun Yang, Yan Wang
https://openreview.net/forum?id=mHsxsrLl0y
Keywords: recommender systems, sequential recommendation
Compressor summary: The paper proposes BirDRec, a framework that rectifies unreliable data in sequential recommender systems by adjusting both input and target items, and reduces its complexity with sampling and self-ensemble methods.
Youbang Sun, Tao Liu, Ruida Zhou, Panganamala Kumar, Shahin Shahrampour
https://openreview.net/forum?id=mA7nTGXjD3
Keywords: Multi Agent Reinforcement Learning, Markov Potential Games, Natural Policy Gradient, Nash Equilibrium
Compressor summary: The paper proposes an efficient algorithm for multi-agent reinforcement learning that finds approximate Nash Equilibria using policy gradients, and demonstrates its effectiveness on two examples.
Jiale Tao, Shuhang Gu, Wen Li, Lixin Duan
https://openreview.net/forum?id=m9uHv1Pxq7
Keywords: Face animation, Motion refinement, Structure correlation
Compressor summary: The paper presents a new unsupervised face animation approach that learns both coarse and fine facial motions using a local affine motion model and a novel motion refinement module based on dense correlation between source and driving images.
Oleg Platonov, Denis Kuznedelev, Artem Babenko, Liudmila Prokhorenkova
https://openreview.net/forum?id=m7PIJWOdlY
Keywords: graph characteristics, homophily, heterophily, label informativeness, constant baseline, GNN
Compressor summary: The paper discusses the drawbacks of common homophily measures in graph neural networks and proposes new measures called adjusted homophily and label informativeness.
Arnab Kumar Mondal, Siba Smarak Panigrahi, Sékou-Oumar Kaba, Sai Rajeswar, Siamak Ravanbakhsh
https://openreview.net/forum?id=m6dRQJw280
Keywords: deep learning, large pretrained models, symmetry, equivariance, group theory, computer vision, point clouds, foundation models
Compressor summary: Equivariant networks can be made from large pretrained models by using a canonicalization network that transforms input to a canonical form, improving robustness to deterministic data transformations.
Jiarong Xu, Renhong Huang, XIN JIANG, Yuxuan Cao, Carl Yang, Chunping Wang, Yang Yang
https://openreview.net/forum?id=m2WR1yJ8N9
Keywords: graph neural networks, pre-training
Compressor summary: The paper proposes a graph pre-training framework (APT) that uses fewer, carefully selected data points to improve downstream tasks, by choosing representative and instructive data based on graph properties and predictive uncertainty.
Ruicheng Xian, Honglei Zhuang, Zhen Qin, Hamed Zamani, Jing Lu, Ji Ma, Kai Hui, Han Zhao, Xuanhui Wang, Michael Bendersky
https://openreview.net/forum?id=m21rQusNgb
Keywords: learning to rank, domain adaptation, text ranking
Compressor summary: This paper introduces list-level alignment, a new invariant representation learning method for ranking problems that leverages the list structure of data and provides theoretical and empirical improvements over existing methods.
Zhenqian Shen, Hansi Yang, Yong Li, James Kwok, quanming yao
https://openreview.net/forum?id=m11TbsaQQI
Keywords: hyper-parameter optimization, cubic regularization
Compressor summary: The paper proposes a new hyper-parameter optimization method that uses cubic regularization and stochastic relaxation to avoid local optima and work without hyper-gradients.
Sarah Rastegar, Hazel Doughty, Cees G. M. Snoek
https://openreview.net/forum?id=m0vfXMrLwF
Keywords: Generalized category discovery, Open world learning, Open-set recognition
Compressor summary: The paper proposes a self-supervised method for discovering unknown categories at test time by viewing them as optimal solutions to well-defined problems and using minimal length codes to represent the hierarchy in real-world data.
Yinghao Aaron Li, Cong Han, Vinay S Raghavan, Gavin Mischler, Nima Mesgarani
https://openreview.net/forum?id=m0RbqrUM26
Keywords: Speech Processing, Text-to-Speech, Diffusion Model, Large Language Model, Self-Supervised Speech Model, WavLM
Compressor summary: StyleTTS 2 is a text-to-speech model that uses style diffusion, adversarial training, and large speech language models to generate high-quality, natural-sounding speech with different styles.
Ruihang Chu, Enze Xie, Shentong Mo, Zhenguo Li, Matthias Nießner, Chi-Wing Fu, Jiaya Jia
https://openreview.net/forum?id=lzqaQRsITh
Keywords: 3d shape completion, conditional generation, diffusion models
Compressor summary: The authors present a new diffusion-based method called DiffComplete that balances realism, multi-modality, and high fidelity in 3D shape completion by using hierarchical feature aggregation and occupancy-aware fusion.
Gabriel Raya, Luca Ambrogioni
https://openreview.net/forum?id=lxGFGMMSVl
Keywords: generative models;diffusion models;score-based generative models; symmetry-breaking
Compressor summary: The paper presents a new understanding of diffusion models' dynamics, leading to improved performance and diversity in generating high-dimensional data.
Xiao Luo, Haixin Wang, Zijie Huang, Huiyu Jiang, Abhijeet Sadashiv Gangan, Song Jiang, Yizhou Sun
https://openreview.net/forum?id=lwg3ohkFRv
Keywords: Dynamical System, Distribution Shift, Neural ODE, Graph Neural Network
Compressor summary: The paper proposes a probabilistic model called Context-attended Graph ODE (CARE) for time-varying interacting dynamical systems and shows its effectiveness on four datasets.
Yanjie Ze, Yuyao Liu, Ruizhe Shi, Jiaxin Qin, Zhecheng Yuan, Jiashun Wang, Huazhe Xu
https://openreview.net/forum?id=lvvaNwnP6M
Keywords: Visual Reinforcement Learning, Representation Learning, Dexterous Manipulation
Compressor summary: The paper proposes a framework called H-InDex that uses human hand poses to improve robotic dexterous manipulation tasks using reinforcement learning.
Shuai Li, Yingjie Zhang, Hongtu Zhu, Christina Dan Wang, Hai Shu, Ziqi Chen, Zhuoran Sun, Yanfeng Yang
https://openreview.net/forum?id=luyXPdkNSN
Keywords: Conditional Independence testing, causal inference, conditional mutual information, k-nearest neighbor, conditional randomization test, conditional permutation test
Compressor summary: The article introduces a novel conditional independence testing method that is efficient, powerful, and robust, without assumptions about distributions or dependencies.
Zhiqing Xiao, Haobo Wang, Ying Jin, Lei Feng, Gang Chen, Fei Huang, Junbo Zhao
https://openreview.net/forum?id=lpx9LZPVtZ
Keywords: Domain Adaptation, Self-training, Graph Spectra
Compressor summary: The paper proposes a new graph-based method for unsupervised domain adaptation, called SPA, which improves the performance of machine learning models in different target domains by aligning their structures and enhancing discriminability.
Yingjun Du, Zehao Xiao, Shengcai Liao, Cees G. M. Snoek
https://openreview.net/forum?id=lp9GR2t3hn
Keywords: Meta-learning, few-shot learning, diffusion model, prototype
Compressor summary: ProtoDiff is a novel framework that uses task-guided diffusion to generate overfitted prototypes for few-shot learning challenges, improving classification performance.
Alexander Tyurin, Peter Richtárik
https://openreview.net/forum?id=loxinzXlCx
Keywords: Nonconvex Optimization, Partial Participation, Variance Reduction, Compressed Communication, Distributed Optimization
Compressor summary: The new method optimizes distributed optimization and federated learning by combining variance reduction, partial participation, and compressed communication with optimal oracle and communication complexities.
Pengfei Li, Jianyi Yang, Adam Wierman, Shaolei Ren
https://openreview.net/forum?id=loixpHDZKj
Keywords: Online optimization, competitive algorithm, switching cost
Compressor summary: The paper proposes an online optimization method called Robustness-Constrained Learning that uses machine learning predictions to improve performance while accounting for multi-step costs and feedback delay.
Qian Yu, Yining Wang, Baihe Huang, Qi Lei, Jason D. Lee
https://openreview.net/forum?id=lnTpBUge5G
Keywords: optimization, quadratic bandits, sample complexity, optimality
Compressor summary: The paper studies how to efficiently optimize quadratic functions using stochastic zeroth-order methods and introduces the concept of energy allocation to analyze the information-theoretic limitations and proposes a Hessian-independent algorithm that works well for all cases.
Zih-Yun Chiu, Yi-Lin Tuan, William Yang Wang, Michael C. Yip
https://openreview.net/forum?id=lmXNcKhj4c
Keywords: Reinforcement Learning, Deep Reinforcement Learning, Sample Efficiency, Generalizability, Multi-Policy Decision Making, Multi-Policy Continuous Control
Compressor summary: The paper introduces Knowledge-Grounded RL (KGRL), a reinforcement learning paradigm that fuses multiple knowledge policies for efficient and flexible learning, and proposes Knowledge-Inclusive Attention Network (KIAN) as a new actor architecture.
Sophia Sanborn, Nina Miolane
https://openreview.net/forum?id=llP6lmMiXE
Keywords: equivariance, group-equivariant cnns, invariance, pooling, convolutional neural networks
Compressor summary: The $G$-TC layer is a robust group-invariance method for $G$-CNNs that uses triple correlation theory to preserve signal structure and improve classification accuracy, working for both commutative and non-commutative groups.
Konstantin Makarychev, Sayak Chakrabarty
https://openreview.net/forum?id=lkEiOZlmPm
Keywords: correlation clustering, Pivot algorithm, streaming
Compressor summary: The paper proposes a simplified and efficient algorithm for correlation clustering with improved memory usage compared to previous approaches.
Casey Meehan, Florian Bordes, Pascal Vincent, Kamalika Chaudhuri, Chuan Guo
https://openreview.net/forum?id=lkBygTc0SI
Keywords: self-supervised learning, privacy, data reconstruction, memorization
Compressor summary: SSL models can accidentally remember specific parts of images they've seen before, which could expose private information and be a security risk.
Tin Sum Cheng, Aurelien Lucchi, Anastasis Kratsios, Ivan Dokmanić, David Belius
https://openreview.net/forum?id=lk6KDG6qI7
Keywords: Kernel, regression, bias-variance, generalization
Compressor summary: Existing statistical learning guarantees for general kernel regressors often yield loose bounds when used with finite-rank kernels. Yet, finite-rank kernels naturally appear in a number of machine learning problems, e.g. when fine-tuning a pre-trained deep neural network's last layer to adapt it to a novel task when performing transfer learning. We address this gap for finite-rank kernel ridge regression (KRR) by deriving sharp non-asymptotic upper and lower bounds for the KRR test error of any finite-rank KRR. Our bounds are tighter than previously derived bounds on finite-rank KRR and, unlike comparable results, they also remain valid for any regularization parameters.
Chih-Yu Lai, Fan-Keng Sun, Zhengqi Gao, Jeffrey Lang, Duane S Boning
https://openreview.net/forum?id=ljgM3vNqfQ
Keywords: time series, anomaly detection, point anomalies, contextual anomalies, nominality score, induced anomaly score
Compressor summary: The paper presents a framework for unsupervised time series anomaly detection using point-based and sequence-based models, and shows it beats most existing methods.
Yu Bai, Fan Chen, Huan Wang, Caiming Xiong, Song Mei
https://openreview.net/forum?id=liMSqUuVg9
Keywords: in-context learning, transformers, deep learning theory, learning theory
Compressor summary: The paragraph discusses how neural sequence models based on the transformer architecture can perform in-context learning and adaptively select different algorithms or tasks without explicit prompting, using a comprehensive statistical theory and experimental evidence.
Bingliang Jiao, Lingqiao Liu, Liying Gao, Ruiqi Wu, Guosheng Lin, PENG WANG, Yanning Zhang
https://openreview.net/forum?id=leS8668NJm
Keywords: Re-identification, Category-generalizable
Compressor summary: The paper introduces a new task called Re-identify Any Animal in the Wild (ReID-AW) and proposes UniReID, a universal re-identification model that can handle any unseen wildlife category using dynamic prompts and semantic knowledge from GPT-4.
Junyi Zhang, Charles Herrmann, Junhwa Hur, Luisa Polania Cabrera, Varun Jampani, Deqing Sun, Ming-Hsuan Yang
https://openreview.net/forum?id=lds9D17HRd
Keywords: Semantic Correspondence, Diffusion Models, Vision Transformer, Representation
Compressor summary: The text describes how diffusion models can be used for image processing tasks and explores their properties, performance, and potential applications.
Qitao Zhao, Ce Zheng, Mengyuan Liu, Chen Chen
https://openreview.net/forum?id=lclQ2RvWYu
Keywords: Human Pose Estimation; 2D-to-3D Lifting; Context-Aware
Compressor summary: The proposed method improves 3D human pose estimation by using spatial context from 2D pose detectors without relying on temporal clues or additional computation.
Xin Cheng, Di Luo, Xiuying Chen, Lemao Liu, Dongyan Zhao, Rui Yan
https://openreview.net/forum?id=lYNSvp51a7
Keywords: natural language processing, retrieval-augmented text generation, self memory
Compressor summary: The paper proposes a novel framework called selfmem that uses its own output as memory to improve text generation tasks like translation, summarization, and dialogue generation.
Sang Michael Xie, Hieu Pham, Xuanyi Dong, Nan Du, Hanxiao Liu, Yifeng Lu, Percy Liang, Quoc V Le, Tengyu Ma, Adams Wei Yu
https://openreview.net/forum?id=lXuByUeHhd
Keywords: language models, pretraining, domain reweighting, data curation
Compressor summary: DoReMi is a method that uses a small proxy model to optimize mixture proportions of pretraining data domains and improve language model performance without knowing downstream tasks.
Weijian Luo, Boya Zhang, Zhihua Zhang
https://openreview.net/forum?id=lXOoR4KYcJ
Keywords: implicit sampler, learning to sample, generative models
Compressor summary: The paper introduces a neural implicit sampler that efficiently generates samples from un-normalized target distributions using two novel training methods, and shows its effectiveness, efficiency, and scalability on three sampling benchmarks.
Yang Yang, Yuxuan Zhang, XIN SONG, Yi Xu
https://openreview.net/forum?id=lV3LIGlc1w
Keywords: Out-of-Distribution, Active Learning
Compressor summary: Progressive Active Learning (PAL) is a sampling scheme that selects valuable out-of-distribution instances and balances pseudo-ID and pseudo-OOD instances to improve both ID classifier and OOD detector performance in open-set active learning.
Zhiyu Zhang, Ashok Cutkosky, Ioannis Paschalidis
https://openreview.net/forum?id=lT9n36RH1w
Keywords: Dynamic online learning, parameter-free online learning, time series forecasting, wavelet
Compressor summary: The paper proposes adaptive regret bounds for online convex optimization under nonstationary comparator sequences by using sparse coding and energy-sparse complexity measures.
Alexandre Rame, Guillaume Couairon, Corentin Dancette, Jean-Baptiste Gaya, Mustafa Shukor, Laure Soulier, Matthieu Cord
https://openreview.net/forum?id=lSbbC2VyCu
Keywords: Deep learning, Foundation models, Fine-tuning, Reward optimization, Linear mode connectivity, Weight averaging, Model soups, Robustness, Generalization, Alignment, Multi objective learning.
Compressor summary: The paper introduces rewarded soup, a multi-policy approach that aims for Pareto-optimal generalization across different rewards and preferences, by interpolating specialized networks' weights linearly.
Feng Wang, Zilong Chen, Guokang Wang, Yafei Song, Huaping Liu
https://openreview.net/forum?id=lSLYXuLqRQ
Keywords: NeRF, Dynamic Scenes
Compressor summary: MSTH is a novel method that efficiently reconstructs dynamic 3D scenes from videos by using a learnable mask to guide a weighted combination of 3D and 4D hash encodings, reducing redundancy and achieving better results than previous methods with less training time and memory storage.
Tao Ge, Jing Hu, Li Dong, Shaoguang Mao, Yan Xia, Xun Wang, Si-Qing Chen, Furu Wei
https://openreview.net/forum?id=lRxpVfDMzz
Keywords: large language model, prompt, imaginary words, OOD robustness, natural language, zero-shot
Compressor summary: X-Prompt uses imaginary words that can be understood by large language models to help them comprehend complex concepts and enable better interactions with humans.
Wentao Zhu, Jason Qin, Yuke Lou, Hang Ye, Xiaoxuan Ma, Hai Ci, Yizhou Wang
https://openreview.net/forum?id=lRu0dN7BY6
Keywords: multi-person motion prediction
Compressor summary: The study introduces a new benchmark, formulation, and framework for predicting 3D multi-person motions in team sports using behavioral cloning, generative adversarial imitation learning, and cognitive hierarchy.
Haitao Lin, Yufei Huang, Odin Zhang, Yunfan Liu, Lirong Wu, Siyuan Li, Zhiyuan Chen, Stan Z. Li
https://openreview.net/forum?id=lRG11M91dx
Keywords: sturcture-based drug design; molecule generation; diffusion model
Compressor summary: \textsc{D3FG} is a functional-group-based diffusion model that generates realistic and interacting molecules for protein pocket-specific drug design using graph neural networks.
Geon Yeong Park, Jeongsol Kim, Beomsu Kim, Sang Wan Lee, Jong Chul Ye
https://openreview.net/forum?id=lOCHMGO6ow
Keywords: Diffusion model, Energy-based model, Text-to-image generation
Compressor summary: The paper proposes a novel energy-based model framework that adapts context control for text-to-image diffusion models, improving semantic alignment and enabling zero-shot compositional generation for various image generation tasks.
Maryam Aliakbarpour, Mark Bun, Adam Smith
https://openreview.net/forum?id=lM1UnEssuX
Keywords: Hypothesis selection, memory constrained algorithms, density estimation, limited space
Compressor summary: The paper studies how to select the best distribution from a set of candidates using limited memory and queries, achieving nearly optimal tradeoffs between memory usage and sample complexity.
Ioannis Anagnostides, Tuomas Sandholm
https://openreview.net/forum?id=lM0xyViO90
Keywords: learning in games, optimistic gradient descent, Nash equilibrium, price of anarchy, smooth games, social welfare
Compressor summary: The paper connects efficiency, smoothness, and no-regret learning algorithms to study equilibrium computation in different game types and improve welfare bounds.
Minsik Cho, Saurabh Adya, Devang Naik
https://openreview.net/forum?id=lLztVBaBVU
Keywords: pruning, cnn, transformers
Compressor summary: PDP is an efficient and effective train-time pruning scheme for DNNs that works on various tasks, architectures, and pruning constraints, achieving state-of-the-art results in model size, accuracy, and training cost.
Amanda Bertsch, Uri Alon, Graham Neubig, Matthew R. Gormley
https://openreview.net/forum?id=lJWUJWLCJo
Keywords: retrieval augmentation, summarization, long-context, generation, long-input, encoder-decoder, transformers, language models, natural language generation, natural language processing, deep learning, neural networks
Compressor summary: The paper introduces Unlimiformer, a method that allows transformers to handle long inputs by using a kNN index for attention instead of attending to every token.
Chengcheng Wang, Wei He, Ying Nie, Jianyuan Guo, Chuanjian Liu, Yunhe Wang, Kai Han
https://openreview.net/forum?id=lJDoPAjkCV
Keywords: YOLO, object detection, computer vision
Compressor summary: The paper introduces Gold-YOLO, a new real-time object detection model that uses convolution and self-attention to improve feature fusion and achieves state-of-the-art performance while maintaining low latency.
Yuxuan Ding, Chunna Tian, Haoxuan Ding, Lingqiao Liu
https://openreview.net/forum?id=lHa7gFbmvS
Keywords: Diffusion Model, CLIP model, Image Variation, Customized Generation
Compressor summary: The paper shows how to use CLIP to convert images into text prompts for text-to-image generation, improving this ability with some training data or online steps, and enabling better image-text interactions for tasks like variation and editing.
Mohammad Salameh, Keith G Mills, Negar Hassanpour, Fred X. Han, Shuting Zhang, Wei Lu, SHANGLING JUI, CHUNHUA ZHOU, Fengyu Sun, Di Niu
https://openreview.net/forum?id=lDI3ZuyzM9
Keywords: Neural Architecture Search, Optimization Framework, Performance Prediction
Compressor summary: AutoGO is a framework that optimizes neural networks by evolving their low-level computation graphs, improving performance and hardware compatibility for various computer vision tasks.
Marco Celotto, Jan Bím, Alejandro Tlaie, Vito De Feo, Alessandro Toso, Stefan M Lemke, Daniel Chicharro, Hamed Nili, Malte Bieler, Ileana Livia Hanganu-Opatz, Tobias H. Donner, Andrea Brovelli, Stefano Panzeri
https://openreview.net/forum?id=lD8xaUWw24
Keywords: Information transmission; Brain data analysis; Sensory processing; Partial information decomposition
Compressor summary: The authors develop a new method called Feature-specific Information Transfer (FIT) that can measure how much information about specific features is transmitted between brain regions, beyond what traditional methods can reveal.
Stephen Marcus McAleer, Gabriele Farina, Gaoyue Zhou, Mingzhi Wang, Yaodong Yang, Tuomas Sandholm
https://openreview.net/forum?id=lCThtrJxoH
Keywords: PSRO, team games, TMECor, populations, equilibrium, game theory, RL
Compressor summary: The paper introduces two algorithms for multi-player team games that balance game-theoretic guarantees with scalability and outperform existing methods on some experiments.
Míriam Barrabés, Daniel Mas Montserrat, Margarita Geleta, Xavier Giró-i-Nieto, Alexander G Ioannidis
https://openreview.net/forum?id=lBhRTO2uWf
Keywords: feature shift detection, distribution shift, shift, data-centric AI
Compressor summary: The paper proposes a method to detect and fix feature shifts in data using adversarial learning and simple iterative heuristics, outperforming existing approaches.
Maximilian Mueller, Tiffany Joyce Vlaar, David Rolnick, Matthias Hein
https://openreview.net/forum?id=lArwl3y9x6
Keywords: sharpness-aware minimization, flatness, generalization, normalization layers
Compressor summary: The paper explores how perturbing only the normalization parameters in the adversarial step of Sharpness-aware minimization (SAM) can improve generalization performance compared to perturbing all parameters or using alternative sparse perturbation methods.
Bowen Gao, Bo Qiang, Haichuan Tan, Yinjun Jia, Minsi Ren, Minsi Lu, Jingjing Liu, Wei-Ying Ma, Yanyan Lan
https://openreview.net/forum?id=lAbCgNcxm7
Keywords: Application, Drug Discovery, Representation Learning, Dataset Augmentation
Compressor summary: The paper proposes a new contrastive learning method called DrugCLIP for faster and better virtual screening of potential drugs using protein pocket and molecule representations without explicit binding-affinity scores.
Yimeng Min, Yiwei Bai, Carla P Gomes
https://openreview.net/forum?id=lAEc7aIW20
Keywords: Combinatorial Optimization, Graph Neural Network, Travelling Salesman Problem
Compressor summary: UTSP is an unsupervised learning framework that uses a graph neural network and local search to solve the TSP, achieving better results than existing data-driven heuristics while being more efficient in terms of parameters and training samples.
Martin Ryner, Jan Kronqvist, Johan Karlsson
https://openreview.net/forum?id=l9MbuqzlZt
Keywords: Gromov-Wasserstein problem, QAP, Global optimization
Compressor summary: The paper proposes an efficient method for solving the Gromov-Wasserstein problem, which finds the best assignment between two sets of points preserving distances, using low-dimensional optimization.
Thao Nguyen, Yuheng Li, Utkarsh Ojha, Yong Jae Lee
https://openreview.net/forum?id=l9BsCh8ikK
Keywords: image editing, diffusion models, visual prompting
Compressor summary: The paper proposes a method for image editing using visual prompts and text-to-image diffusion models to generate editing directions from examples.
Jiyao Zhang, Mingdong Wu, Hao Dong
https://openreview.net/forum?id=l6ypbj6Nv5
Keywords: Category-Level Object Pose Estimation, Diffusion Model
Compressor summary: The authors propose a novel solution for category-level object pose estimation using conditional generative modeling with score-based diffusion models, achieving state-of-the-art results on the REAL275 dataset and generalizing well to new categories.
Mengxiao Zhang, Yuheng Zhang, Olga Vrousgou, Haipeng Luo, Paul Mineiro
https://openreview.net/forum?id=l6pYRbuHpO
Keywords: Online learning with feedback graphs, Contextual Bandits, Practical algorithms
Compressor summary: The paper proposes a method for contextual bandits with feedback graphs that simplifies learning by reducing it to regression and achieving minimal performance loss.
Lingfeng Yang, Yueze Wang, Xiang Li, Xinlong Wang, Jian Yang
https://openreview.net/forum?id=l6R4Go3noz
Keywords: visual prompting, zero-shot, visual language model, referring expression comprehension
Compressor summary: The authors propose a fine-grained visual prompting method that uses pixel-level masks to improve zero-shot instance-level recognition of objects in images, achieving significant improvements over existing methods on several benchmarks.
Liangliang Shi, Haoyu Zhen, Gu Zhang, Junchi Yan
https://openreview.net/forum?id=l61Kp1zBwC
Keywords: Optimal Transport; Unbalanced Classification
Compressor summary: The paper proposes a new optimal transport method for classification, which connects optimal transport and machine learning concepts like barycentric projection and transfer learning.
Shengzhong Liu, Tomoyoshi Kimura, Dongxin Liu, Ruijie Wang, Jinyang Li, Suhas Diggavi, Mani Srivastava, Tarek Abdelzaher
https://openreview.net/forum?id=l4CZCKXoSn
Keywords: Multimodal Time Series; Contrastive Learning; Factorized Latent Space
Compressor summary: FOCAL is a novel contrastive learning framework for multimodal time series that considers modality-specific features and temporal information locality, improving performance in downstream tasks.
Xuan Chen, Wenbo Guo, Guanhong Tao, Xiangyu Zhang, Dawn Song
https://openreview.net/forum?id=l3yxZS3QdT
Keywords: Backdoor Defense, Deep Reinforcement Learning
Compressor summary: BIRD is a technique that detects and removes backdoors from pretrained deep reinforcement learning policies without needing information about the attack or access to its training process.
Yuandong Tian, Yiping Wang, Beidi Chen, Simon Shaolei Du
https://openreview.net/forum?id=l3HUgVHqGQ
Keywords: transformer, training dynamics, theoretical analysis, self-attention, interpretability, neural network understanding
Compressor summary: The paper analyzes how self-attention in a 1-layer transformer works during training for next token prediction, showing that it acts as a discriminative scanning algorithm with a scan and snap dynamics.
Marco Jiralerspong, Joey Bose, Ian Gemp, Chongli Qin, Yoram Bachrach, Gauthier Gidel
https://openreview.net/forum?id=l2VKZkolT7
Keywords: Generative model, FID, Evaluation, Precision, Recall, Likelihood
Compressor summary: FLS is a new metric that assesses the novelty, fidelity, and diversity of generated data by deep generative models, addressing limitations of existing methods.
Peng Cheng, Xianyuan Zhan, Zhihao Wu, Wenjia Zhang, Youfang Lin, Shou cheng Song, Han Wang, Li Jiang
https://openreview.net/forum?id=kyXMU3H7RB
Keywords: sample efficiency; offline reinforcement learning; fundamental symmetry
Compressor summary: The paper proposes a new offline reinforcement learning algorithm (TSRL) that leverages time-reversal symmetry of system dynamics to improve performance on small datasets and enhance data efficiency and generalizability.
Zeyu Sun, Dogyoon Song, Alfred Hero
https://openreview.net/forum?id=kvXcHfBghm
Keywords: probability calibration, optimal number of bins, label shift adaptation
Compressor summary: The paper introduces minimum-risk recalibration using MSE decomposition, analyzes UMB recalibration, and proposes a two-stage approach for label shift adaptation.
Mingjia Shi, Yuhao Zhou, Kai Wang, Huaizheng Zhang, Shudong Huang, Qing Ye, Jiancheng Lv
https://openreview.net/forum?id=kuxu4lCRr5
Keywords: Federated Learning, Personalized Federated Learning, Expectation Maximization, Relaxed Mirror Descent
Compressor summary: The paper proposes a new method, pFedBreD, for personalized federated learning that injects prior knowledge into the global model to improve performance on various datasets.
Youzhi Zhang, Bo An, Venkatramanan Siva Subrahmanian
https://openreview.net/forum?id=kupNhxLc6k
Keywords: Algorithmic game theory, Optimal Nash equilibrium
Compressor summary: The paper proposes a new algorithm, CRM, to efficiently compute a Nash Equilibrium in multiplayer games by using correlation plans to reduce the solution space and improve the speed.
Ge Zheng, Bin Yang, Jiajin Tang, Hong-Yu Zhou, Sibei Yang
https://openreview.net/forum?id=ktYjrgOENR
Keywords: Chain-of-Thought Reasoning, Multimodal Science Question Answering, Vision and Langauge
Compressor summary: The text discusses challenges in applying chain of thought reasoning to multimodal contexts and proposes a new method called DDCoT that improves reasoning abilities, generalizability, and explainability of AI models.
Parker Knight, Rui Duan
https://openreview.net/forum?id=ktTSji9ZIs
Keywords: multi-task learning, genetic risk prediction, summary statistics
Compressor summary: The paper proposes a multi-task learning framework using summary statistics and adaptive parameter selection for applications where data-sharing is restricted, such as healthcare, and analyzes its performance theoretically and empirically.
Tyler LaBonte, Vidya Muthukumar, Abhishek Kumar
https://openreview.net/forum?id=kshC3NOP6h
Keywords: spurious correlations, group robustness, last-layer retraining, distribution shift
Compressor summary: The paper proposes a method called SELF that improves group robustness of neural networks by retuning the last layer using disagreements or misclassifications, without needing group annotations.
BANG AN, Xun Zhou, Yongjian Zhong, Tianbao Yang
https://openreview.net/forum?id=ks7Mf5lzSx
Keywords: urban event, NDCG optimization, ranking, traffic accident, crime, spatiotemporal data
Compressor summary: SpatialRank is a novel method for urban event ranking that uses adaptive graph convolution layers and a hybrid NDCG loss to predict the top-k most risky locations based on spatiotemporal dependencies and neighboring spatial proximity.
Robert Istvan Busa-Fekete, Heejin Choi, Travis Dick, Claudio Gentile, Andres Munoz medina
https://openreview.net/forum?id=kqBUgrkm1c
Keywords: learning with partial information, unbiased loss, classification, proportion matching
Compressor summary: EASYLLP is a debiasing technique for weakly supervised classification that uses aggregate labels and works with various loss functions, improving instance level performance on Learning from Label Proportions.
Mingze Wang, Chao Ma
https://openreview.net/forum?id=konBXvt2iS
Keywords: non-convex optimization, training dynamics, neural network
Compressor summary: The paper analyzes the training process of a two-layer ReLU network using Gradient Flow and identifies four phases and specific nonlinear behaviors that occur during learning.
Hao Sun, Alihan Hüyük, Daniel Jarrett, Mihaela van der Schaar
https://openreview.net/forum?id=kmbG9iBRIb
Keywords: Accountability, Reinforcement Learning, Batched Control, Accountable Decision-Making, Offline RL, Interpretability in RL
Compressor summary: The paper presents AOC, a controller that uses an offline dataset as the Decision Corpus and selects a subset of examples for accountable control in low-data scenarios, evaluating it in simulated and real healthcare settings.
Masatoshi Uehara, Nathan Kallus, Jason D. Lee, Wen Sun
https://openreview.net/forum?id=kjkLJ7NJJZ
Keywords: Reinforcement learning theory, PAC RL, Offline Reinforcement learning
Compressor summary: The paper proposes value-based algorithms for offline reinforcement learning that can achieve PAC guarantees even with partial coverage of data, using novel minimax loss functions derived from Lagrange functions in nonlinear convex optimization problems.
David Brandfonbrener, Ofir Nachum, Joan Bruna
https://openreview.net/forum?id=kjMGHTo8Cs
Keywords: representation learning, imitation learning
Compressor summary: This paper explores how large datasets with multitask demonstrations can be pretrained using inverse dynamics modeling to learn low dimensional representations for imitation learning in unknown environments.
Simone Rossi, Ankit Singh, Thomas Hannagan
https://openreview.net/forum?id=kj33zJ9Vue
Keywords: Bayesian deep learning, approximate inference, permutation symmetries
Compressor summary: The authors extend the formalism of marginalized loss barrier and solution interpolation to Bayesian neural networks, propose an algorithm to search for linearly connected solutions, and experiment with different architectures and datasets.
Alexandre Max Maraval, Matthieu Zimmer, Antoine Grosnit, Haitham Bou Ammar
https://openreview.net/forum?id=kfWzpZvEUh
Keywords: meta-learning, bayesian optimisation, neural process, transformer, end-to-end, reinforcement learning
Compressor summary: The paper presents a differentiable meta-BO framework that uses neural processes and transformers to learn acquisition functions with RL and an auxiliary task to handle sparse rewards.
Jingye Chen, Yupan Huang, Tengchao Lv, Lei Cui, Qifeng Chen, Furu Wei
https://openreview.net/forum?id=ke3RgcDmfO
Keywords: Diffusion Model; Text Rendering
Compressor summary: TextDiffuser is a new method for generating images with readable and coherent text that works by first generating keyword layouts with a Transformer model and then using diffusion models to create images based on the layout and text prompts. It also introduces a large-scale dataset and evaluation tool for text rendering quality.
Shuhuai Ren, Aston Zhang, Yi Zhu, Shuai Zhang, Shuai Zheng, Mu Li, Alex Smola, Xu Sun
https://openreview.net/forum?id=kdFR6IUEW6
Keywords: Prompt Pre-Training, CLIP, Open-Vocabulary Visual Recognition
Compressor summary: POMP is a prompt pre-training method for vision-language models that improves visual recognition tasks with zero-shot learning and achieves state-of-the-art performances on 21 datasets.
Joshua Engels, Benjamin Coleman, Vihan Lakshman, Anshumali Shrivastava
https://openreview.net/forum?id=kXfrlWXLwH
Keywords: Embedding Based Retrieval, Passage Ranking, Locality Sensitive Hashing, Randomized Algorithms
Compressor summary: The paper introduces DESSERT, a fast approximate search algorithm for vector set queries, which can be integrated into semantic search models like ColBERT to improve their speed and performance.
Xiuye Gu, Yin Cui, Jonathan Huang, Abdullah Rashwan, Xuan Yang, Xingyi Zhou, Golnaz Ghiasi, Weicheng Kuo, Huizhong Chen, Liang-Chieh Chen, David A Ross
https://openreview.net/forum?id=kXOXrVnwbb
Keywords: universal segmentation, multi-task segmentation, multi-dataset segmentation, panoptic segmentation, semantic segmentation, instance segmentation, weakly-supervised segmentation
Compressor summary: The paper proposes DaTaSeg, a universal multi-dataset multi-task segmentation model that leverages weak-supervision, text embeddings, and merge operations to improve performance across different tasks and datasets, including open-vocabulary segmentation.
Kun Huang, Xin Guo, Meng Wang
https://openreview.net/forum?id=kVfHQV668B
Keywords: Knowledge Distillation; Pre-Trained Language Model
Compressor summary: FCD is a new method for compressing PLMs that uses output feature relationships to transfer knowledge more effectively and achieves better results than previous KD approaches.
Zeyu Zhang, Chaozhuo Li, Xu Chen, Xing Xie
https://openreview.net/forum?id=kS8rIH43Zc
Keywords: Causal Discovery, Active Learning, Multi-fidelity
Compressor summary: The paper presents a probabilistic model for multi-fidelity active causal discovery, with a mutual-information based acquisition function and a cascading model to capture correlations between different fidelity oracles, and extends it to batch intervention scenarios.
Sohir Maskey, Raffaele Paolino, Aras Bacho, Gitta Kutyniok
https://openreview.net/forum?id=kS7ED7eE74
Keywords: Graph Neural Networks, Graph Neural ODE, Fractional Laplacian, Oversmoothing
Compressor summary: The paper introduces fractional graph Laplacian neural ODEs for directed graphs to capture long-range dependencies and mitigate oversmoothing in graph neural networks.
Hong-Xing Yu, Yang Zheng, Yuan Gao, Yitong Deng, Bo Zhu, Jiajun Wu
https://openreview.net/forum?id=kRdaTkaBwC
Keywords: neural scene representations, fluid dynamics, flow reconstruction, physics-based learning
Compressor summary: The paper presents HyFluid, a neural method to jointly estimate fluid density and velocity from sparse multiview videos, using physics-based losses and a hybrid neural velocity representation that handles visual ambiguities and turbulence in fluid flows.
Hyeong Kyu Choi, Seunghun Lee, Jaewon Chu, Hyunwoo J. Kim
https://openreview.net/forum?id=kR5ycmBclj
Keywords: Knowledge Graph Question Answering, Knowledge Graph, Graph Neural Networks
Compressor summary: NuTrea is a tree search-based GNN model that improves multi-hop KGQA by incorporating broader KG context and handling ambiguous nodes.
Yingbin Bai, Zhongyi Han, Erkun Yang, Jun Yu, Bo Han, Dadong Wang, Tongliang Liu
https://openreview.net/forum?id=kR21XsZeAr
Keywords: learning with noisy labels, weakly supervised learning
Compressor summary: The paper studies a type of label noise called subclass-dominant label noise (SDN) and proposes a method called NoiseCluster that uses long-trained representations to identify and correct it, achieving better results than existing methods on synthetic and real data.
Fabian Christian Spaeh, Alina Ene
https://openreview.net/forum?id=kPfd3pcwHV
Keywords: Learning Augmented Algorithms, Display Ads, Generalized Assignment Problem
Compressor summary: The paper presents a learning-augmented algorithm for online packing problems that uses machine-learned predictions to improve performance beyond worst-case algorithms, and shows its effectiveness on synthetic and real data.
Sibylle Marcotte, Rémi Gribonval, Gabriel Peyré
https://openreview.net/forum?id=kMueEV8Eyy
Keywords: Implicit bias, conservation laws, gradient flow, linear neural network, matrix factorization
Compressor summary: The article explores how to identify and count conservation laws, which describe quantities conserved during gradient flows of a given model, and how these laws relate to optimization initialization properties in over-parameterized models.
Yoni Kasten, Ohad Rahamim, Gal Chechik
https://openreview.net/forum?id=kMmAYbT0VL
Keywords: Point Cloud, Text, 3D
Compressor summary: SDS-Complete uses a text-to-image diffusion model to complete incomplete point clouds of various objects without needing extensive 3D data, improving performance on out-of-distribution objects.
Sadaf Salehkalaibar, Truong Buu Phan, Jun Chen, Wei Yu, Ashish J Khisti
https://openreview.net/forum?id=kLIieSS2P3
Keywords: Video Compression, Information Theory, Neural Compression
Compressor summary: The choice of perception loss function affects video compression quality, especially at low bit rates, but encoded representations can be near universal and work for either choice of function.
Lunhao Duan, Shanshan Zhao, Nan Xue, Mingming Gong, Gui-Song Xia, Dacheng Tao
https://openreview.net/forum?id=kKXJkiniOx
Keywords: Point Cloud, Transformer, 3D Segmentation, 3D object detection
Compressor summary: The paper proposes ConDaFormer, a new transformer block for 3D point cloud understanding that reduces costs and models local geometry prior by disassembling cubic windows into three orthogonal 2D planes and using depth-wise convolution to enhance local structure.
Hengrui Cai, Yixin Wang, Michael Jordan, Rui Song
https://openreview.net/forum?id=kKFDMtpeDW
Keywords: Causal structural learning, Necessity and sufficiency, Natural causal effects, Probabilities of causation, Variable selection
Compressor summary: The paper proposes a method to learn a subset of causally relevant variables from a complex graph using probabilities of causation, which can improve causal estimation in various fields.
Sitao Luan, Chenqing Hua, Minkai Xu, Qincheng Lu, Jiaqi Zhu, Xiao-Wen Chang, Jie Fu, Jure Leskovec, Doina Precup
https://openreview.net/forum?id=kJmYu3Ti2z
Keywords: Graph Neural Networks, Homophily, Heterophily, Low-pass filter, High-pass filter, Node Distinguishability, Metrics
Compressor summary: The paper studies how the similarity of node neighborhood patterns (intra-class Node Distinguishability) and differences between classes (inter-class ND) affect Graph Neural Networks' performance, proposing new metrics to measure them and a hypothesis-testing based performance metric that goes beyond homophily.
Ignavier Ng, Yujia Zheng, Xinshuai Dong, Kun Zhang
https://openreview.net/forum?id=kJIibP5bq2
Keywords: independent component analysis, second-order statistics, sparsity
Compressor summary: The paper presents a new method for identifying Gaussian sources using independent component analysis that does not require non-Gaussian assumptions or restrictive connective structures, and provides experimental validation.
Thien Le, Stefanie Jegelka
https://openreview.net/forum?id=kDQwossJuI
Keywords: graph neural networks, convolution, graph limits, size transferability
Compressor summary: The paper investigates how well graph neural networks (GNNs) can generalize to different graphs, especially sparse ones, using a theoretical approach based on graphops.
Jialu Gao, Kaizhe Hu, Guowei Xu, Huazhe Xu
https://openreview.net/forum?id=kChEBODIx9
Keywords: Visual Reinforcement Learning, Large Generative Models, Image Editing, Robotics
Compressor summary: LfVoid is a method that uses text-to-image models and image editing to guide robots based on natural language instructions, even without in-domain training or true goal observations.
Joe Watson, Sandy Huang, Nicolas Heess
https://openreview.net/forum?id=kCCD8d2aEu
Keywords: imitation learning, inverse reinforcement learning, behavioral cloning, learning from demonstration
Compressor summary: The authors propose a hybrid imitation learning method that combines behavioral cloning and reinforcement learning to learn from expert demonstrations and achieve complex tasks with minimal hyperparameter tuning.
Ziyuan Ye, Rihan Huang, Qilin Wu, Quanying Liu
https://openreview.net/forum?id=kBBsj9KRgh
Keywords: GNN explainability, Shapley value, Monte Carlo tree search, structure awareness, multi-grained explanation
Compressor summary: SAME is a new method for explaining graph neural networks that uses Monte Carlo tree search to explore structure-aware substructures and improves explanation fidelity on various benchmarks.
Matthew Thomas Jackson, Minqi Jiang, Jack Parker-Holder, Risto Vuorio, Chris Lu, Gregory Farquhar, Shimon Whiteson, Jakob Nicolaus Foerster
https://openreview.net/forum?id=kAU6Cdq1gV
Keywords: Reinforcement Learning, Meta-Learning, Meta-RL, Meta-Optimization, Policy Meta-Optimization, Environment Design, Unsupervised Environment Design, Auto-Curricula
Compressor summary: This paper explores how meta-learning update rules for deep reinforcement learning can improve generalization and proposes GROOVE, a novel method that generates curricula to maximize the regret of a meta-learned optimizer.
Tingliang Feng, Hao Shi, Xueyang Liu, Wei Feng, Liang Wan, Yanlin Zhou, Di Lin
https://openreview.net/forum?id=k9zSU3pdi4
Keywords: Object Style Compensation, Open Compound Domain Adaptation, Semantic Segmentation
Compressor summary: The paper introduces Object Style Compensation, a method that adapts object styles across domains for semantic image segmentation by using an Object-Level Discrepancy Memory.
Yuhan Ding, Fukun Yin, Jiayuan Fan, Hui Li, Xin Chen, Wen Liu, Chongshan Lu, Gang YU, Tao Chen
https://openreview.net/forum?id=k8U8ZijXHh
Keywords: implicit neural representation; diffusion; point cloud; volume rendering
Compressor summary: The authors propose a new method called Point Diffusion implicit Function (PDF) that uses point cloud super-resolution and background modeling to represent and synthesize large-scale outdoor scenes for novel view synthesis, achieving better results than existing methods.
Zhaoyang Hai, Liyuan Pan, Xiabi Liu, Zhengzheng Liu, Mirna Yunita
https://openreview.net/forum?id=k6yNi6DEqK
Keywords: Learning to teach, dynamic loss function, optimization
Compressor summary: The paper proposes a teacher model with memory units and a Dynamic Loss Network to adjust loss functions adaptively based on both the states of the student model and the loss function, leading to improved deep learning performance on various tasks.
Sepehr Assadi, Vihan Shah, Chen Wang
https://openreview.net/forum?id=k4ZCORSFEd
Keywords: Correlation Clustering, Graph Streaming Algorithms, Large-scale Clustering, Graph Learning
Compressor summary: This paper studies streaming correlation clustering with very limited memory, proposing two novel algorithms that estimate the optimal cost up to a constant factor and some additional error.
Julia Linhart, Alexandre Gramfort, Pedro L. C. Rodrigues
https://openreview.net/forum?id=k2UVKezeWn
Keywords: machine learning, calibration, simulation-based inference, neuroscience, normalizing flows, classifier two-sample tests
Compressor summary: The paper introduces a new method, $\ell$-C2ST, for evaluating deep generative models' approximations of complex posterior distributions at specific observations, improving interpretability and performance over existing approaches.
Jiangtao Zhang, Shunyu Liu, Jie Song, Tongtian Zhu, Zhengqi Xu, Mingli Song
https://openreview.net/forum?id=k1Xy5zCNOJ
Keywords: Deep Learning, Computer Vision, Mode Connectivity, Weight Average
Compressor summary: Lookaround is a new optimizer that improves diversity and generalization in deep networks by training multiple networks on transformed data and averaging them, leading to flatter minima.
Nic Fishman, Leo Klarner, Emile Mathieu, Michael John Hutchinson, Valentin De Bortoli
https://openreview.net/forum?id=jzseUq55eP
Keywords: diffusion model, generative modelling, manifold, constraints, proteins, robotics
Compressor summary: The paper proposes a simple noising scheme based on Metropolis sampling for diffusion models with domain-informed constraints, improving efficiency and performance over existing methods.
Erik Schultheis, Marek Wydmuch, Wojciech Kotlowski, Rohit Babbar, Krzysztof Dembczynski
https://openreview.net/forum?id=jze2r6RDFz
Keywords: extreme multi-label classification, long-tail labels performance, complex performance measures
Compressor summary: The paper proposes a new metric for extreme multi-label classification that captures the importance of rare labels and optimizes expected test utility using efficient prediction rules.
Huiqiao Fu, Kaiqiang Tang, Yuanyang Lu, Yiming Qi, Guizhou Deng, Flood Sung, Chunlin Chen
https://openreview.net/forum?id=jxhUNLoi4m
Keywords: Generative adversarial imitation learning, semi-supervised learning, multi-modal behaviors, imbalanced data
Compressor summary: Imitation learning can reproduce expert behaviors without explicit rewards, but real-world demos are challenging; a new semi-supervised method learns disentangled behavior representations from imbalanced data using generative adversarial networks and information maximization.
Ivaxi Sheth, Samira Ebrahimi Kahou
https://openreview.net/forum?id=jvYXln6Gzn
Keywords: Interpretability, concept bottleneck models, explainability
Compressor summary: The paper introduces coop-CBM, a transparent neural network model that preserves concept representations and improves performance under different data distributions.
Yihong Chen, Kelly Marchisio, Roberta Raileanu, David Ifeoluwa Adelani, Pontus Stenetorp, Sebastian Riedel, Mikel Artetxe
https://openreview.net/forum?id=jvEbQBxd8X
Keywords: plasticity, continual learning, meta-learning, embeddings, cross-lingual transfer, forgetting
Compressor summary: The paper proposes using active forgetting during pretraining to create PLMs that can adapt quickly and effectively to new languages, especially when there is little data available.
Evgenii Nikishin, Junhyuk Oh, Georg Ostrovski, Clare Lyle, Razvan Pascanu, Will Dabney, Andre Barreto
https://openreview.net/forum?id=jucDLW6G9l
Keywords: deep reinforcement learning, continual learning, loss of plasticity
Compressor summary: The paper proposes plasticity injection, a method to increase neural network plasticity in deep reinforcement learning without changing parameters or predictions, and demonstrates its usefulness as a diagnostic and improvement tool for RL agents.
Jade Copet, Felix Kreuk, Itai Gat, Tal Remez, David Kant, Gabriel Synnaeve, Yossi Adi, Alexandre Défossez
https://openreview.net/forum?id=jtiQ26sCJi
Keywords: Music generation, Generative AI, Transformer, Language Models
Compressor summary: MusicGen is a single transformer model that generates high-quality music based on text or melodic features, outperforming previous methods.
Jan Böker, Ron Levie, Ningyuan Teresa Huang, Soledad Villar, Christopher Morris
https://openreview.net/forum?id=jt10uWlEbc
Keywords: graphons, universal approximation, weisfeiler-leman, graph metric, tree homomorphisms, tree distance, optimal transport, GNNs
Compressor summary: The paper presents a continuous extension of the Weisfeiler--Leman test for analyzing the expressive power of message-passing graph neural networks on graphons, providing a theoretical and empirical framework to understand their limitations and performance.
Yi Wu, Ziqiang Li, Chaoyue Wang, Heliang Zheng, Shanshan Zhao, Bin Li, Dacheng Tao
https://openreview.net/forum?id=jown9RvYn7
Keywords: StyleGAN, Few-Shot Generative Domain Adaptation
Compressor summary: The study proposes a new generator structure called Domain Re-Modulation (DoRM) for few-shot Generative Domain Adaptation, which incorporates memory and domain association to achieve high quality, diversity, and consistency across domains.
Zhu Wang, Sourav Medya, Sathya N. Ravi
https://openreview.net/forum?id=jooPcatnVF
Keywords: Implicit layer, Out-of-distribution detection, multimodal learning
Compressor summary: The authors propose a method to combine deep neural networks and semantic knowledge for vision and language tasks, using an implicit out-of-distribution detector to filter irrelevant information.
Zhaoyu Li, Jinpei Guo, Yuhe Jiang, Xujie Si
https://openreview.net/forum?id=jnIBiP2di1
Keywords: Logical Reasoning, Rule Learning, Interpretation, SATNet
Compressor summary: The paper proposes a new framework that generates interpretable and verifiable logical rules for advanced AI systems using differentiable learning and MaxSAT solver.
Krzysztof Kacprzyk, Zhaozhi Qian, Mihaela van der Schaar
https://openreview.net/forum?id=jnCPN1vpSR
Keywords: differential equations, symbolic regression
Compressor summary: D-CIPHER is a new method for finding differential equations from data that works with noisy and infrequent observations and discovers more phenomena than existing approaches.
Benjamin Scellier, Maxence Ernoult, Jack Kendall, Suhas Kumar
https://openreview.net/forum?id=jl5a3t78Uh
Keywords: energy-based learning algorithm, contrastive learning, equilibrium propagation, coupled learning, convolutional Hopfield network
Compressor summary: The paragraph compares seven energy-based learning algorithms on deep convolutional Hopfield networks for five vision tasks and finds that negative perturbations perform better than positive ones.
Sokhna Diarra Mbacke, Florence Clerc, Pascal Germain
https://openreview.net/forum?id=jkPDRHff3s
Keywords: Variational Autoencoders, PAC-Bayes, Statistical Learning Theory
Compressor summary: The authors use PAC-Bayesian theory to provide statistical guarantees for Variational Autoencoders (VAEs) and bounds on their performance metrics.
Siyuan Zhou, Yilun Du, Shun Zhang, Mengdi Xu, Yikang Shen, Wei Xiao, Dit-Yan Yeung, Chuang Gan
https://openreview.net/forum?id=jhs8F63xI6
Keywords: Decision making, Robotics, Planning-based
Compressor summary: The paper proposes a method to improve diffusion planning by deciding when to replan and adjusting trajectories, leading to better performance in robotic control tasks.
Xuyang Chen, Lin Zhao
https://openreview.net/forum?id=jh3UNSQK0l
Keywords: Finite-time analysis, single-timescale actor-critic
Compressor summary: The paper analyzes the convergence of online single-timescale actor-critic with linear function approximation on continuous state space, achieving an $\epsilon$-approximate stationary point with improved sample complexity.
Susung Hong, Donghoon Ahn, Seungryong Kim
https://openreview.net/forum?id=jgIrJeHHlz
Keywords: Text-to-3D, Diffusion Models
Compressor summary: The paper proposes two methods to improve view consistency in text-to-3D generation by debiasing scores from 2D diffusion models and aligning user prompts with object views.
Roy Uziel, Or Dinari, Oren Freifeld
https://openreview.net/forum?id=jfsjKBDB1z
Keywords: Unsupervised, video segmentation, clustering
Compressor summary: The paper presents a training-free video object segmentation method that uses pre-trained deep features and classical clustering, achieving state-of-the-art results with low memory and fast inference.
Mohammad Mozaffari, Sikan Li, Zhao Zhang, Maryam Mehri Dehnavi
https://openreview.net/forum?id=jcnvDO96N5
Keywords: machine learning, deep learning, optimizers, distributed training;second-order optimization;
Compressor summary: MKOR is a new optimizer that improves training speed and convergence of DNNs by reducing complexity and increasing second-order updates frequency.
Zhaolu Liu, Robert Peach, Pedro A. M. Mediano, Mauricio Barahona
https://openreview.net/forum?id=jcRB6xHdJ2
Keywords: High-order interactions; Lattice theory; Kernel tests
Compressor summary: The text introduces a method to measure and test interactions between groups of more than two variables in complex systems using kernel-based tests and lattice theory connections.
Runpeng Yu, Xinchao Wang
https://openreview.net/forum?id=jcJVgIFY2r
Keywords: Generative Model
Compressor summary: The paper presents a new method to reconstruct an image generator using a pre-trained classifier, by training the generator to satisfy convergence conditions based on maximum-margin bias theory.
Jinxin Liu, Hongyin Zhang, Zifeng Zhuang, Yachen Kang, Donglin Wang, Bin Wang
https://openreview.net/forum?id=jZYf1GxH1V
Keywords: offline reinforcement learning, test-time adaptation
Compressor summary: The paper proposes a non-iterative bi-level offline RL method called DROP that learns an MBO score model and a behavior embedding for safe policy extraction during testing.
Sander Beckers
https://openreview.net/forum?id=jYIknUIgkd
Keywords: responsibility, causation, causal models
Compressor summary: The paper defines moral responsibility for AI systems using causal models, compares its approach to existing ones, and extends the concept to a degree of responsibility.
Florian Seligmann, Philipp Becker, Michael Volpp, Gerhard Neumann
https://openreview.net/forum?id=jX49iKr6vb
Keywords: bayesian deep learning, distribution shift, calibration
Compressor summary: The paper surveys recent BDL methods on realistic tasks, focusing on how they generalize and calibrate under distribution shift. It explores different aspects of BDL, such as signed error, fine-tuning, and ensembles, and compares various algorithms in terms of accuracy and calibration.
Zeyue Xue, Guanglu Song, Qiushan Guo, Boxiao Liu, Zhuofan Zong, Yu Liu, Ping Luo
https://openreview.net/forum?id=jUdZCcoOu3
Keywords: Diffusion Model, Text-to-Image Generation
Compressor summary: RAPHAEL is a text-conditional image diffusion model that generates highly artistic images by stacking tens of mixture-of-experts layers, outperforming other models in image quality and style switching, with a zero-shot FID score of 6.61 on the COCO dataset.
Yue Wu, So Yeon Min, Shrimai Prabhumoye, Yonatan Bisk, Ruslan Salakhutdinov, Amos Azaria, Tom Mitchell, Yuanzhi Li
https://openreview.net/forum?id=jU9qiRMDtR
Keywords: Games, Instruction Manual, Crafter, Open-world games, Large Language Models, Language Models, Zero-shot, In-context prompting
Compressor summary: SPRING is a novel approach that uses a large language model to reason and play open-world survival games like Crafter, outperforming state-of-the-art RL baselines without any training.
Mazda Moayeri, Wenxiao Wang, Sahil Singla, Soheil Feizi
https://openreview.net/forum?id=jSuhnO9QJv
Keywords: spurious correlations, interpretability, bias, distributional robustness
Compressor summary: The paper proposes a simple and effective way to reduce model biases due to spurious features by ranking images based on how much they rely on such cues, using deep neural features of an interpretable network. The method works on any data and model without major modifications, and shows that class-wise biases are strongly related across models.
Ryan Theisen, Hyunsuk Kim, Yaoqing Yang, Liam Hodgkinson, Michael W. Mahoney
https://openreview.net/forum?id=jS4DUGOtBD
Keywords: Ensembling, theory, deep learning
Compressor summary: The paper investigates when ensembling improves classification performance, finding that it depends on the disagreement rate and the average error rate of the models involved, with non-interpolating models benefiting more from ensembling than interpolating ones.
Guozheng Ma, Linrui Zhang, Haoyu Wang, Lu Li, Zilin Wang, Zhen Wang, Li Shen, Xueqian Wang, Dacheng Tao
https://openreview.net/forum?id=jRL6ErxMVB
Keywords: Data Augmentation, Visual Reinforcement Learning, Sample Efficiency
Compressor summary: This work investigates the attributes of data augmentation (DA) in visual reinforcement learning, introducing new DA operations and fusion schemes to improve sample efficiency.
William Brown, Jon Schneider, Kiran Vodrahalli
https://openreview.net/forum?id=jR2FkqW6GB
Keywords: learning in games, correlated equilibria, Stackelberg equilibria, swap regret, dynamic regret
Compressor summary: This text discusses tradeoffs between reward and regret in repeated games with two agents using generalized equilibrium, and shows that different algorithms can lead to different outcomes depending on the opponent's algorithm choice.
Shuwei Shao, Zhongcai Pei, Xingming Wu, Zhong Liu, Weihai Chen, Zhengguo Li
https://openreview.net/forum?id=jOuxQGRVoQ
Keywords: Monocular depth estimation, Iterative refinement, Deep learning
Compressor summary: The paper introduces a novel iterative elastic bins method for monocular depth estimation, which progressively optimizes the search range and utilizes temporal context modeling with GRU-based architecture to achieve state-of-the-art results.
Atsuki Sato, Yusuke Matsui
https://openreview.net/forum?id=jL2eJxPK88
Keywords: optimization, data structures, algorithms, theory, learned algorithms
Compressor summary: This paper presents two methods to construct a learned Bloom filter faster and with similar memory efficiency, while one method can achieve the same data structure as the original.
Yingjie Wang, Jiajun Deng, Yuenan Hou, Yao Li, Yu Zhang, Jianmin Ji, Wanli Ouyang, Yanyong Zhang
https://openreview.net/forum?id=jIhX7SpfCz
Keywords: 3D object detection ; Point clouds
Compressor summary: The paper proposes CluB, a 3D object detection framework that combines BEV-based and cluster-based detectors by enriching object representation at feature and query levels.
Rainer Engelken
https://openreview.net/forum?id=jEQRoJzDx8
Keywords: exploding/vanishing gradients, Lyapunov exponents, Lyapunov spectrum, chaos, RNN, condition number, Jacobian
Compressor summary: Gradient flossing is a new method to stabilize recurrent neural networks by regularizing Lyapunov exponents during learning, leading to improved convergence speed and success rate for tasks with long time horizons.
Saurabh Saxena, Charles Herrmann, Junhwa Hur, Abhishek Kar, Mohammad Norouzi, Deqing Sun, David J. Fleet
https://openreview.net/forum?id=jDIlzSU8wJ
Keywords: Monocular depth, optical flow, diffusion, depth, flow
Compressor summary: Denoising diffusion probabilistic models can estimate optical flow and monocular depth effectively without specialized architectures, providing uncertainty and multimodality information.
Jianwei Zhang, Suren Jayasuriya, Visar Berisha
https://openreview.net/forum?id=jCPRG3FuHV
Keywords: repeatability, embeddings, metric learning, intra-class correlation, intra-class variance
Compressor summary: The authors propose a method to measure and improve the repeatability of embeddings for machine learning tasks using intra-class correlation coefficients (ICC) and show its effectiveness on speech tasks.
Yangru Huang, Peixi Peng, Yifan Zhao, Haoran Xu, Mengyue Geng, Yonghong Tian
https://openreview.net/forum?id=jB4wsc1DQW
Keywords: vision-based reinforcement learning, multi-modal, event camera
Compressor summary: The paper proposes a local value estimation framework for vision-based reinforcement learning that adapts to different modalities and improves performance in autonomous driving.
Alexander Wei, Nika Haghtalab, Jacob Steinhardt
https://openreview.net/forum?id=jA235JGM09
Keywords: red teaming, safety, RLHF, large language models
Compressor summary: The text discusses how large language models trained for safety still face adversarial misuse, investigates why this happens, and evaluates new attacks on state-of-the-art models like GPT-4 and Claude v1.3.
Pablo Moreno-Muñoz, Pol G. Recasens, Søren Hauberg
https://openreview.net/forum?id=j9wGUcS30B
Keywords: Marginal likelihood, masked pre-training, Bayesian inference
Compressor summary: The paper explains how masked pre-training works by maximizing a model's generalization measure and explores its potential for Bayesian models.
Kunxun Qi, Jianfeng Du, Hai Wan
https://openreview.net/forum?id=j7x9wW3tCf
Keywords: Knowledge graph completion; Neural approximate rule learning; Neural rule-based system
Compressor summary: The paper proposes a two-stage framework for knowledge graph completion using both structural and textual knowledge, which improves performance over existing rule-based systems.
Kaichen Zhou, Jia-Xing Zhong, Sangyun Shin, Kai Lu, Yiyuan Yang, Andrew Markham, Niki Trigoni
https://openreview.net/forum?id=j7U4pFkCYB
Keywords: View Synthesis, Monocular Video
Compressor summary: DynPoint is an algorithm that accelerates view synthesis for unconstrained monocular videos by predicting 3D correspondence between frames and using hierarchical neural point clouds.
Niklas Muennighoff, Alexander M Rush, Boaz Barak, Teven Le Scao, Nouamane Tazi, Aleksandra Piktus, Sampo Pyysalo, Thomas Wolf, Colin Raffel
https://openreview.net/forum?id=j5BuTrEj35
Keywords: large language models, scaling laws, data engineering
Compressor summary: The paper explores how to scale language models when there is limited text data and proposes a scaling law for compute optimality that takes into account repeated tokens and excess parameters.
Michal Yarom, Yonatan Bitton, Soravit Changpinyo, Roee Aharoni, Jonathan Herzig, Oran Lang, Eran Ofek, Idan Szpektor
https://openreview.net/forum?id=j5AoleAIru
Keywords: Vision-and-language, Image-text alignment, Text-to-image generation, Image-to-text generation, Multi-modal models, Synthetic images, Meta-evaluation, Visual-question-answering
Compressor summary: The authors propose two methods for automatic evaluation of text-image alignment and introduce SeeTRUE, a dataset with human judgements for semantically aligned text-image pairs.
Junyi Li, Heng Huang
https://openreview.net/forum?id=j4QVhftpYM
Keywords: Federated Learning
Compressor summary: FedSep is a novel two-layer federated learning framework that separates communication and learning layers for better privacy and efficiency, and outperforms existing methods in communication-efficient and heterogeneous-model FL tasks.
Jinghuan Shang, Michael S Ryoo
https://openreview.net/forum?id=j2oYaFpbrB
Keywords: Reinforcement Learning, Active Reinforcement Learning, Visual Reinforcement Learning, Active Vision, Active Perception, Partial Observability, Sensorimotor
Compressor summary: The paper introduces SUGARL, a framework that learns motor and sensory policies together using an intrinsic sensorimotor reward, which enables agents to control their visual observations in partially observable environments.
Taoli Cheng, Aaron Courville
https://openreview.net/forum?id=j0U6XJubbP
Keywords: Generative modeling, Energy-based models, Out-of-distribution detection, Sciences, Application, Physics
Compressor summary: The authors propose a flexible energy-based probabilistic model for High Energy Physics events that can generate simulations, detect anomalies, and identify particles.
Wei-Ning Chen, Dan Song, Ayfer Ozgur, Peter Kairouz
https://openreview.net/forum?id=izNfcaHJk0
Keywords: Differential Privacy, Federated Learning, Communication
Compressor summary: The paper studies how to achieve the best balance of privacy, communication, and accuracy in federated learning and analytics, and proposes two methods that use compression and random data selection to improve this trade-off.
Ilias Diakonikolas, Daniel Kane, Ankit Pensia, Thanasis Pittas
https://openreview.net/forum?id=iyweRIXAeH
Keywords: robust statistics, high-dimensional inference, regression, nearly linear time algorithms
Compressor summary: The paper presents novel algorithms for Gaussian mean estimation and linear regression with Huber contamination, achieving near-optimal sample and time complexity with optimal error guarantees.
Amir Akbarnejad, Gilbert Bigras, Nilanjan Ray
https://openreview.net/forum?id=iy4Of0w8ML
Keywords: Gaussian processes, Explainable AI
Compressor summary: The authors propose a method that uses Gaussian processes to match and explain the outputs of deep neural networks without strict assumptions, and demonstrate its effectiveness on various datasets using computational techniques for scalability and GPU acceleration.
Zacharia Issa, Blanka Horvath, Maud Lemercier, Cristopher Salvi
https://openreview.net/forum?id=ixcsBZw5pl
Keywords: Neural SDEs, score-based generative models, signature kernels, time series
Compressor summary: This paper proposes a new way to train Neural SDEs non-adversarially using kernel scores, which improves stability, avoids mode collapse, and allows for generating spatiotemporal data.
Hejie Cui, Xinyu Fang, Zihan Zhang, Ran Xu, Xuan Kan, Xin Liu, Yue Yu, Manling Li, Yangqiu Song, Carl Yang
https://openreview.net/forum?id=ixVAXsdtJO
Keywords: Visual Knowledge Extraction, Multimodality, Large Model Prompting
Compressor summary: OpenVik is a new method for extracting open visual knowledge from images using an open relational region detector and a multimodal model, with enhanced data techniques to diversify the results.
Andy Zhou, Jindong Wang, Yu-Xiong Wang, Haohan Wang
https://openreview.net/forum?id=iwp3H8uSeK
Keywords: robustness, knowledge distillation, adversarial training, data augmentation, generalization
Compressor summary: The paper introduces a framework that improves vision models' robustness by combining knowledge distillation and data augmentation using a robust teacher and Discrete Adversarial Distillation (DAD).
Zhendong Wang, Yifan Jiang, Huangjie Zheng, Peihao Wang, Pengcheng He, Zhangyang Wang, Weizhu Chen, Mingyuan Zhou
https://openreview.net/forum?id=iv2sTQtbst
Keywords: diffusion models, training efficiency, data efficiency
Compressor summary: Patch Diffusion is a faster and more efficient way to train diffusion models that uses patches with random sizes and coordinates, improving the quality of generated images even on small datasets.
Scott Pesme, Nicolas Flammarion
https://openreview.net/forum?id=iuqCXg1Gng
Keywords: gradient flow, saddle-to-saddle, diagonal linear network, incremental learning
Compressor summary: The paper describes how gradient flow in two-layer linear networks for regression finds the minimum l1-norm solution by jumping between saddles using a recursive algorithm and arc-length time-reparametrization, with no assumptions on data or network size.
Yi Ren, Samuel Lavoie, Mikhail Galkin, Danica J. Sutherland, Aaron Courville
https://openreview.net/forum?id=irRHgjePdR
Keywords: compositional generalization, systematic generalization, iterated learning, representation learning, graph neural networks
Compressor summary: The authors propose iterated learning on models with simplicial embeddings to improve deep networks' compositional generalization, inspired by human language development and Kolmogorov complexity analysis.
Wenqiang Wang, Chongyang Du, Tao Wang, Kaihao Zhang, Wenhan Luo, Lin Ma, Wei Liu, Xiaochun Cao
https://openreview.net/forum?id=ir6WWkFR80
Keywords: Punctuation-level Attack, Textual Adversarial attack, Natural Language Processing
Compressor summary: The paper introduces a new way of textual attack using single punctuation changes and proposes a search method to find the best position for the attack without exhaustive search.
Vijay Veerabadran, Srinivas Ravishankar, Yuan Tang, Ritik Raina, Virginia R. de Sa
https://openreview.net/forum?id=iqezE0EyXq
Keywords: cognitive science, recurrent neural networks, adaptive computation time, visual reasoning
Compressor summary: The study investigates how adaptive recurrent neural networks can help vision models solve difficult visual reasoning problems by adjusting their computational resources based on the input's needs.
Junqi Wang, PEI WANG, Patrick Shafto
https://openreview.net/forum?id=iohoef1bfM
Keywords: Bayesian theory, Belief transport, Unbalanced optimal transport, parametrization, asymptotic behavior, environment drift detection
Compressor summary: The paragraph discusses a new mathematical framework called Generalized Belief Transport that unifies different learning models by considering them as points in a broader space of possibilities, and explores its properties and behavior.
Christopher Solinas, Doug Rebstock, Nathan R. Sturtevant, Michael Buro
https://openreview.net/forum?id=inIONNg8Sq
Keywords: search, game theory, multi-agent, learning, markov chain monte carlo, complexity
Compressor summary: The authors analyze the computational aspects of filtering histories for subgame decomposition, a key technique for AI in imperfect information games, and introduce a novel Markov Chain Monte Carlo-based generation algorithm for trick-taking card games.
Lu Qi, Jason Kuen, Weidong Guo, Jiuxiang Gu, Zhe Lin, Bo Du, Yu Xu, Ming-Hsuan Yang
https://openreview.net/forum?id=ikkdTD3hQJ
Keywords: Image Segmentation
Compressor summary: The paper introduces AIMS, a new task for image segmentation that divides visual regions into three levels, and proposes a unified model to tackle challenges like annotation inconsistency and task correlation.
Weihang Dai, Yao DU, Hanru Bai, Kwang-Ting Cheng, Xiaomeng Li
https://openreview.net/forum?id=ij3svnPLzG
Keywords: Semi-supervised learning, deep regression, contrastive learning
Compressor summary: The authors propose a new semi-supervised contrastive learning method for deep regression that uses unlabeled data and achieves better results than existing methods.
Kun Yi, Qi Zhang, Wei Fan, Shoujin Wang, Pengyang Wang, Hui He, Ning An, Defu Lian, Longbing Cao, Zhendong Niu
https://openreview.net/forum?id=iif9mGCTfy
Keywords: time series forecasting, multi-layer perceptrons, frequency domain
Compressor summary: The paper proposes FreTS, a time series forecasting method that uses frequency-domain MLPs to learn global dependencies and concentrate on key parts of frequency components, outperforming existing methods.
Xin Zheng, Miao Zhang, Chunyang Chen, Soheila Molaei, Chuan Zhou, Shirui Pan
https://openreview.net/forum?id=ihlT8yvQ2I
Keywords: graph neural networks, GNN model evaluation, node classification accuracy
Compressor summary: The paper proposes a two-stage framework to evaluate graph neural network (GNN) models by estimating their performance on unseen graphs without labels using a DiscGraph set and a GNNEvaluator.
Lorenzo Brusca, Lars C.P.M. Quaedvlieg, Stratis Skoulakis, Grigorios Chrysos, Volkan Cevher
https://openreview.net/forum?id=igE3Zbxvws
Keywords: Maximum Independent Set, Combinatorial Optimization, Graph Neural Networks, Dynamic Programming
Compressor summary: The paper introduces a GNN-based algorithm for solving the MIS problem using a DP-like recursive approach and shows its effectiveness on various datasets.
Angelica Chen, David Dohan, David So
https://openreview.net/forum?id=ifbF4WdT8f
Keywords: language models, evolution, prompting, neural architecture search, code generation
Compressor summary: The paper explores using language models as mutation and crossover operators for evolutionary neural architecture search (NAS), proposing EvoPrompting which combines prompt engineering with soft prompt-tuning, and shows its effectiveness in finding diverse and high performing models on various tasks.
Sangwoo Seo, Sungwon Kim, Chanyoung Park
https://openreview.net/forum?id=icWwBKyVMs
Keywords: Graph neural network, Explainable AI, Interpretability
Compressor summary: The paper introduces interpretable Prototype-based Graph Information Bottleneck (PGIB), a novel framework for explainable GNNs that learns prototypes from key subgraphs important for predictions, improving both performance and interpretability.
YoungJoong Kwon, Lingjie Liu, Henry Fuchs, Marc Habermann, Christian Theobalt
https://openreview.net/forum?id=iajxrSgOSX
Keywords: DELIFFAS: Avatar Modeling, Avatar Synthesis, Animatable Human, Light Fields, Human Performance Capture
Compressor summary: The DELIFFAS method creates realistic human avatars with fast inference speed by using a surface light field attached to a deformable mesh model, achieving state-of-the-art results in appearance synthesis and speed.
Caspar Oesterheld, Johannes Treutlein, Roger Baker Grosse, Vincent Conitzer, Jakob Nicolaus Foerster
https://openreview.net/forum?id=ia4AL3QnOv
Keywords: Program equilibrium, multi-agent learning, game theory, opponent shaping, superrationality, decision theory, cooperative AI, Newcomb's problem
Compressor summary: The paper proposes a realistic setting in which machine learning agents can achieve cooperative outcomes in social dilemmas by only observing a single number indicating their similarity, and shows that this can be learned using simple ML methods.
Hang Lou, Siran Li, Hao Ni
https://openreview.net/forum?id=iWWLgcUTZU
Keywords: Generative adversarial networks, time series generation, rough path theory, Lie group
Compressor summary: The paragraph introduces PCF-GAN, a novel GAN that uses path characteristic functions to represent time series distributions and improve their generation and reconstruction quality.
Omar Chehab, Aapo Hyvarinen, Andrej Risteski
https://openreview.net/forum?id=iWGC0Nsq9i
Keywords: noise-contrastive estimation, annealed importance sampling
Compressor summary: The paragraph discusses different design choices for Monte Carlo methods that estimate the normalization constant using annealing and their impact on estimation error, showing that some choices are more efficient than others depending on the context.
Jiawen Chen, Wancen Mu, Yun Li, Didong Li
https://openreview.net/forum?id=iVYInarGXg
Keywords: Gaussian process, Identifiability, Interpretability, Mixture kernel, Separable kernel
Compressor summary: The paper analyzes additive and multiplicative mixtures of Mat\'ern kernels in single- and multi-output Gaussian process models, showing that the latter are more suitable for multi-output tasks due to better identifiability properties.
Soichiro Kumano, Hiroshi Kera, Toshihiko Yamasaki
https://openreview.net/forum?id=iT9MOAZqsb
Keywords: adversarial training; mean field theory
Compressor summary: The study analyzes adversarial training in random deep neural networks using a new mean field theory framework, showing its effects on network capacity, width, and dimensions.
Steve Hanneke, Shay Moran, Jonathan Shafer
https://openreview.net/forum?id=iSd8g75QvP
Keywords: Online Learning, Transductive Online Learning, Offline Learning, Mistake Bound
Compressor summary: The paper presents new bounds on learner mistakes in transductive online learning, showing a trichotomy of possible mistake rates depending on VC and Littlestone dimensions.
Xiaoshuai Hao, Wanqian Zhang
https://openreview.net/forum?id=iQlK3VJxV7
Keywords: video-text retrieval; cross-domain;Unsupervised Domain Adaptation Video-text Retrieval;
Compressor summary: The paper presents UAN, a novel method to improve unsupervised domain adaptation video-text retrieval by addressing multimodal mutual information and uncertainty-aware alignment.
Max B. Paulus, Andreas Krause
https://openreview.net/forum?id=iPTF2hON1C
Keywords: Generative Modeling, Combinatorial Optimization, Mixed Integer Programming, Graph Neural Networks, Diving Heuristics
Compressor summary: L2Dive is a diving heuristic that uses graph neural networks to predict variable assignments and improve the performance of mixed integer linear programs on various combinatorial optimization problems.
Deeparnab Chakrabarty, Andrei Graur, Haotian Jiang, Aaron Sidford
https://openreview.net/forum?id=iMfPFPMsZo
Keywords: parallel computation, convex optimization, submodular function minimization
Compressor summary: The paragraph discusses two new methods for parallel submodular function minimization with different trade-offs between depth and query complexity, and also introduces a highly-parallel algorithm for minimizing an $\ell_\infty$-Lipschitz function over the hypercube.
Nataly Brukhim, Miroslav Dudík, Aldo Pacchiano, Robert E. Schapire
https://openreview.net/forum?id=iM0MWWBr4W
Keywords: Interactive learning, bandits, statistical queries
Compressor summary: The paper proposes a new interactive learning framework with a combinatorial measure called Dissimilarity dimension that captures learnability and unifies two classic models, with improved analyses.
Wenjing Chen, Victoria G. Crawford
https://openreview.net/forum?id=iKarSI2a73
Keywords: submodular, combinatorial optimization, approximation algorithms
Compressor summary: The paper presents scalable algorithms for solving submodular cover problems with different variants and applications in data summarization and graph cut.
Alessandro Sordoni, Xingdi Yuan, Marc-Alexandre Côté, Matheus Pereira, Adam Trischler, Ziang Xiao, Arian Hosseini, Friederike Niedtner, Nicolas Le Roux
https://openreview.net/forum?id=iImnbUVhok
Keywords: deep prompt optimization, llm, variational inference, graphical model, chaining
Compressor summary: The text describes how to optimize prompts for two-layer deep language networks (DLNs) using latent variable inference and shows their potential to match or surpass GPT-4's performance with smaller models.
Ravid Shwartz-Ziv, Micah Goldblum, Yucen Lily Li, C. Bayan Bruss, Andrew Gordon Wilson
https://openreview.net/forum?id=iGmDQn4CRj
Keywords: Class Imbalance, Hyperparameters, Long-Tailed Distributions
Compressor summary: The authors show that by adjusting some standard deep learning components, one can achieve excellent results on class-imbalanced datasets without using specialized techniques.
Feng Chen, Daniel Kunin, Atsushi Yamamura, Surya Ganguli
https://openreview.net/forum?id=iFxWrxDekd
Keywords: Implicit Bias, SGD Dynamics, Implicit regularization, Learning rate schedule, Stochastic Gradient Descent, Invariant set, Attractive saddle points, Stochastic collapse, Permutation invariance, Simplicity bias, Teacher-student
Compressor summary: The paper shows that stochastic gradient descent (SGD) often simplifies neural networks during training by converging to simpler subnetworks, improving generalization.
Vinitra Swamy, Malika Satayeva, Jibril Frej, Thierry Bossy, Thijs Vogels, Martin Jaggi, Tanja Käser, Mary-Anne Hartley
https://openreview.net/forum?id=iB3Ew6z4WL
Keywords: Deep Learning, Multimodal Learning, Multi-task learning, Missingness, Interpretability
Compressor summary: MultiModN is a multimodal network that fuses latent representations in a sequence of any modality type, providing interpretable feedback on real-world tasks while being resistant to missing not-at-random data.
Dennis Frauen, Valentyn Melnychuk, Stefan Feuerriegel
https://openreview.net/forum?id=iAcEmyhwk2
Keywords: Causal machine learning, treatment effect estimation, sensitivity analysis, unobserved confounding, uncertainty estimation
Compressor summary: The paper presents a general framework for causal sensitivity analysis under unobserved confounding, extending the marginal sensitivity model to various types of treatments and outcomes, and providing sharp bounds for different causal effects using observational data.
Dongho Lee, Jongseo Lee, Jinwoo Choi
https://openreview.net/forum?id=iATY9W5Xw7
Keywords: action recognition, video understanding, cross-attention, balanced spatio-temporal understanding
Compressor summary: Recognizing human actions in videos requires spatial and temporal understanding. Most existing action recognition models lack a balanced spatio-temporal understanding of videos. In this work, we propose a novel two-stream architecture, called Cross-Attention in Space and Time (CAST), that achieves a balanced spatio-temporal understanding of videos using only RGB input. Our proposed bottleneck cross-attention mechanism enables the spatial and temporal expert models to exchange information and make synergistic predictions, leading to improved performance. We validate the proposed method with extensive experiments on public benchmarks with different characteristics: EPIC-Kitchens-100, Something-Something-V2, and Kinetics-400. Our method consistently shows favorable performance across these datasets, while the performance of existing methods fluctuates depending on the dataset characteristics. The code is available at https://github.com/KHU-VLL/CAST.
Changdae Oh, Junhyuk So, Hoyoon Byun, YongTaek Lim, Minchul Shin, Jong-June Jeon, Kyungwoo Song
https://openreview.net/forum?id=iAAXq60Bw1
Keywords: multi-modal learning, robustness, fine-tuning, contrastive learning, CLIP, Mixup
Compressor summary: The authors study CLIP's multi-modal embeddings and propose a fine-tuning method to improve alignment and uniformity for better transferability and robustness.
Zijiao Chen, Jiaxin Qing, Juan Helen Zhou
https://openreview.net/forum?id=i913TUOvTK
Keywords: Video Reconstruction from Brain Activities, Diffusion Model, Contrastive Learning
Compressor summary: The paragraph describes a new method called Mind-Video that can reconstruct high-quality videos from brain activities using various techniques and outperforms previous methods in accuracy and interpretability.
Haoran You, Huihong Shi, Yipin Guo, Yingyan Celine Lin
https://openreview.net/forum?id=i6mMWNcTfu
Keywords: Efficient Vision Transformer; Multiplication-reduced networks; Hardware acceleration
Compressor summary: The authors propose a new multiplication-reduced model called ShiftAddViT, which uses bitwise shifts and additions instead of dense multiplications to speed up ViTs on GPUs without losing much accuracy.
Marie Maros, Gesualdo Scutari
https://openreview.net/forum?id=i5sSWKbF3b
Keywords: Distributed non-convex optimization, Low-rank matrix recovery
Compressor summary: The paper studies how agents in a network can estimate a low-rank matrix using a decentralized gradient algorithm and shows that it converges to the correct solution with good performance guarantees.
Youquan Liu, Lingdong Kong, Jun CEN, Runnan Chen, Wenwei Zhang, Liang Pan, Kai Chen, Ziwei Liu
https://openreview.net/forum?id=i39yXaUKuF
Keywords: autonomous driving, point cloud segmentation, self-supervised learning, 3D scene understanding
Compressor summary: Seal is a novel framework that leverages vision foundation models for efficient and versatile segmentation of diverse automotive point clouds, with properties such as scalability, consistency, and generalizability.
Weiwei Kong, Andres Munoz medina
https://openreview.net/forum?id=i2H2sEiq2T
Keywords: differential privacy, dp-sgd, gradient clipping, computational complexity
Compressor summary: The paper proposes a framework that enables fast norm computations for gradients in DP-SGD with arbitrary intermediate operations and shows improvements for specific operations using components of the framework.
Gaspard Beugnot, Julien Mairal, Alessandro Rudi
https://openreview.net/forum?id=i28zCSsQIc
Keywords: non-convex optimization, polynomial optimization, kernel sum-of-squares
Compressor summary: The authors propose a novel optimization method that leverages the regularity of smooth functions in the Fourier spectrum and neural network techniques, leading to better performance and scalability.
Shida Wang, Beichen Xue
https://openreview.net/forum?id=i0OmcF14Kf
Keywords: state space, approximation theory, sequence modelling
Compressor summary: The paper shows how adding nonlinear activation to state-space models improves their ability to learn complex sequence patterns, but does not solve the problem of exponential decaying memory.
Joey Hong, Sergey Levine, Anca Dragan
https://openreview.net/forum?id=hzND3ZEFg2
Keywords: offline reinforcement learning, human-aware reinforcement learning, multi-agent influence
Compressor summary: The text discusses how AI agents can use offline reinforcement learning to influence suboptimal humans towards better performance by learning from a dataset of human-human interactions.
Bong Gyun Kang, HyunGi Kim, Dahuin Jung, Sungroh Yoon
https://openreview.net/forum?id=hz33V7Tb2O
Keywords: Continual learning, Knowledge transfer, Algorithmic reasoning
Compressor summary: CLeAR is a novel algorithmic reasoning methodology that enables continual learning of abstract logical concepts with minimal forgetting and backward transfer.
Rakshith Sharma Srinivasa, Jaejin Cho, Chouchang Yang, Yashas Malur Saidutta, Ching-Hua Lee, Yilin Shen, Hongxia Jin
https://openreview.net/forum?id=hz10oiVMNE
Keywords: contrastive loss, multimodal representation learning, zero-shot learning, intent classification, pre-trained models, modality alignment, cross-modal transfer
Compressor summary: The paper introduces CWCL, a new contrastive loss function that improves cross-modal 0-shot transfer by using a continuous measure of similarity, achieving better results than existing methods on various tasks and datasets.
Pierre-Étienne H Fiquet, Eero P Simoncelli
https://openreview.net/forum?id=hyPUZX03Ks
Keywords: video prediction, neural coding, symmetry discovery, self-supervised representation-learning
Compressor summary: The paper proposes a self-supervised learning method that uses natural video patterns to predict future frames and discover simple transformations, achieving comparable performance to conventional deep networks while being interpretable and fast.
Yuan Wang, Naisong Luo, Tianzhu Zhang
https://openreview.net/forum?id=hxJu0386if
Keywords: Computer Vision, Few-shot Segmentation
Compressor summary: The paper introduces AMFormer, a new few-shot segmentation model that focuses on query-centric learning and achieves state-of-the-art results with weak support labels.
Martina G. Vilas, Timothy Schaumlöffel, Gemma Roig
https://openreview.net/forum?id=hwjmEZ8561
Keywords: transformers, computer vision, image classification, mechanistic interpretability, explainability
Compressor summary: This paper presents a method to analyze how Vision Transformers learn image representations for classification tasks and use attention and context to build categorical features.
Saurav Jha, Dong Gong, He Zhao, Lina Yao
https://openreview.net/forum?id=huh0XmSdBK
Keywords: Continual Learning, Neural Process, Uncertainty, Incremental Learning
Compressor summary: The paper proposes a new approach to continuous learning using neural processes, which can handle tasks with less interference, reduce forgetting, and provide reliable uncertainty estimates for deep neural networks.
Kai Klede, Thomas Altstidl, Dario Zanca, Bjoern Eskofier
https://openreview.net/forum?id=htkdwc6jDB
Keywords: Clustering, (Other) Machine Learning Topics
Compressor summary: The paper introduces a new clustering comparison method called p-value adjusted Rand Index (PMI_2) that is unbiased, monotonous, and efficient for selecting better clustering and community detection algorithms.
Jiarong Ding, Xuehu Zhu
https://openreview.net/forum?id=htM8yp2EwX
Keywords: Mediation analysis, Composite null hypothesis, Local false discovery rate, Optimal ranking rule, High-dimensional
Compressor summary: The paper introduces a new method for detecting significant mediators in high-dimensional mediation analysis that controls the false discovery rate by ranking and selecting hypotheses based on p-values and composite null probabilities.
Wenliang Zhao, Lujia Bai, Yongming Rao, Jie Zhou, Jiwen Lu
https://openreview.net/forum?id=hrkmlPhp1u
Keywords: diffusion models, fast sampling, predictor-corrector, training-free
Compressor summary: The paper introduces UniPC, a framework for fast and high-quality sampling of diffusion probabilistic models in image synthesis, improving on previous methods especially with few steps.
Jiaxu Tian, Dapeng Zhi, Si Liu, Peixin Wang, Cheng Chen, Min Zhang
https://openreview.net/forum?id=hpYb4eUinX
Keywords: Deep reinforcement learning, Reachability analysis, Hybrid system, State abstraction
Compressor summary: The paper proposes a new method to verify deep reinforcement learning systems using piece-wise linear decision neural networks, which can reduce overestimation and speed up verification by up to 438 times.
Yunqi Shi, Ke Xue, Lei Song, Chao Qian
https://openreview.net/forum?id=hoyL1Ypjoo
Keywords: Black Box Optimization, Macro Placement, Electronic Design Automation, Reinforcement Learning, Application
Compressor summary: The paper proposes a new optimization framework, WireMask-BBO, for improving chip floorplanning by minimizing wirelength and avoiding overlapping, achieving significant improvements in quality and efficiency over previous methods.
Swati Padmanabhan, David Woodruff, Qiuyi Zhang
https://openreview.net/forum?id=hn1oJO7lg6
Keywords: $\ell_p$ sensitivities, Lewis weights, leverage scores, approximation algorithms, total sensitivity
Compressor summary: The paper introduces efficient algorithms for computing $\ell_p$ sensitivities and other summary statistics of a matrix, which can help reduce dimensionality in regression tasks and improve data quality after removing low-sensitivity points.
Zechuan Zhang, Li Sun, Zongxin Yang, Ling Chen, Yi Yang
https://openreview.net/forum?id=hlkhPdhuAO
Keywords: 3D human avatar reconstruction, vision transformer, parametric body model, tri-plane representation
Compressor summary: The GTA model uses a transformer-based architecture to reconstruct 3D clothed human avatars from monocular images, overcoming limitations of previous methods by capturing global-correlated features and using a hybrid prior fusion strategy.
Aleksandra Nowak, Bram Grooten, Decebal Constantin Mocanu, Jacek Tabor
https://openreview.net/forum?id=hkPn7M9k1W
Keywords: Dynamic Sparse Training, Pruning, Deep Learning
Compressor summary: The paragraph discusses the importance of studying pruning criteria for Dynamic Sparse Training (DST) and how simplicity works best in the low-density regime.
Wei Fu, Weihua Du, Jingwei Li, Sunli Chen, Jingzhao Zhang, Yi Wu
https://openreview.net/forum?id=hiwF7aG1dt
Keywords: diverse behavior, multi-agent reinforcement learning, deep reinforcement learning
Compressor summary: The paper proposes a new diversity-driven reinforcement learning algorithm (SIPO) that incorporates state-space distance information and shows improved performance in discovering diverse and interpretable strategies across various domains.
Kanishk Jain, Shyamgopal Karthik, Vineet Gandhi
https://openreview.net/forum?id=hiQG8qGxso
Keywords: Hierarchical Classification, Fine-grained Classification, Ensembles, Mistake Severity
Compressor summary: The paper proposes HiE, a method that uses label hierarchy to correct coarse-grained predictions and improve fine-grained classification performance, achieving state-of-the-art results and practical applicability.
Ke Yi, Yansen Wang, Kan Ren, Dongsheng Li
https://openreview.net/forum?id=hiOUySN0ub
Keywords: Electroencephalogram, EEG Pre-training, EEG-based emotion recognition
Compressor summary: The authors propose a new EEG pre-training framework, MMM, which uses a unified channel topology and multi-dimensional position encoding to improve performance on emotional recognition tasks.
Yunzhang Zhu, Renxiong Liu
https://openreview.net/forum?id=hgLMht2Z3L
Keywords: regularization, optimization, tuning parameter selection
Compressor summary: The paper proposes a method for choosing grid points and an adaptive stopping criterion to efficiently compute approximation solution paths for regularized M-estimation problems in machine learning.
Michele Garibbo, Maxime Robeyns, Laurence Aitchison
https://openreview.net/forum?id=hcXDbbzgoh
Keywords: Reinforcement learning, TD-learning, model-based, variance reduction
Compressor summary: Taylor TD is a model-based RL framework that reduces variance in continuous state-action settings by using a first-order Taylor series expansion of TD updates, integrating over stochasticity, and providing stable learning guarantees.
Waïss Azizian, Franck Iutzeler, Jérôme Malick
https://openreview.net/forum?id=haniyY7zm1
Keywords: robust optimization, distributionally robust optimization, optimization under uncertainty, generalization, Wasserstein distance, optimal transport
Compressor summary: Wasserstein distributionally robust estimators are powerful models for prediction and decision-making under uncertainty with improved generalization guarantees on various classes of models without the curse of dimensionality or distribution shifts.
Yinshuang Xu, Jiahui Lei, Kostas Daniilidis
https://openreview.net/forum?id=haHIji0yFt
Keywords: equivariance, light field, equivariant convolution over homogeneous space
Compressor summary: The paper proposes an SE(3)-equivariant convolution and transformer for learning geometric priors from multiple views, which can improve 3D reconstruction and novel view rendering tasks.
Jiawei Ren, Mingyuan Zhang, Cunjun Yu, Xiao Ma, Liang Pan, Ziwei Liu
https://openreview.net/forum?id=hXevuspQnX
Keywords: Physics-based Animation; Human Motion Generation
Compressor summary: InsActor is a framework that uses diffusion policies and skill discovery to generate physics-based character animations from high-level human instructions.
weitao Du, Yuanqi Du, Limei Wang, Dieqiao Feng, Guifeng Wang, Shuiwang Ji, Carla P Gomes, Zhi-Ming Ma
https://openreview.net/forum?id=hWPNYWkYPN
Keywords: Graph Neural Networks, Geometric Deep Learning, Equivariance, Symmetry
Compressor summary: The paper introduces local hierarchy of 3D isomorphism to evaluate and improve the expressiveness of equivariant GNNs for 3D object modeling, and presents LEFTNet that achieves state-of-the-art results on molecular property prediction tasks.
Syed Talal Wasim, Kabila Haile Soboka, Abdulrahman Mahmoud, Salman Khan, David Brooks, Gu-Yeon Wei
https://openreview.net/forum?id=hVVp8TXIPs
Keywords: Hardware Resilience, Reliability, Image Classification, CLIP, Vision-Language, Multimodal
Compressor summary: The paper proposes a method to improve image classification model reliability using GPT-3 text embeddings as an initialization for the classification layer, resulting in significant hardware reliability improvements with minimal accuracy loss.
Kareem Ahmed, Kai-Wei Chang, Guy Van den Broeck
https://openreview.net/forum?id=hVAla2O73O
Keywords: Neuro-Symbolic, Tractable Models, Learning with Constraints
Compressor summary: The authors propose a neuro-symbolic method for enforcing constraints on auto-regressive distributions that approximates the likelihood with pseudolikelihood, improving performance on Sudoku, shortest-path prediction, and detoxifying language models.
Fang Wu, Stan Z. Li
https://openreview.net/forum?id=hV52oj0Sik
Keywords: Antibody Design
Compressor summary: The paper proposes HTP, a hierarchical training paradigm that uses GNNs and protein language models to design therapeutic antibodies with desired binding specificity by incorporating evolutionary information from various databases.
Ang Li, Yifei Wang, Yiwen Guo, Yisen Wang
https://openreview.net/forum?id=hSkEcIFi3o
Keywords: Adversarial Examples, Adversarial Training, Generalization, Generative Models
Compressor summary: The paper re-examines the theory of adversarial examples by showing that non-robust features are not as useful or transferable across different learning paradigms, suggesting they are more like shortcuts and robustness alone is insufficient for model reliability.
Shengqiong Wu, Hao Fei, Hanwang Zhang, Tat-Seng Chua
https://openreview.net/forum?id=hSTaTBIUCj
Keywords: Image Synthesis, Scene Graph, Diffusion Model
Compressor summary: The text describes a new method for generating detailed images from simple text prompts using scene graph hallucination and diffusion models.
Naishan Zheng, Man Zhou, Chong Zhou, Chen Change Loy
https://openreview.net/forum?id=hOOOvOMok5
Keywords: Image restoration, low-light image enhancement, image de-noising
Compressor summary: The paper introduces a new high-order channel-wise convolution operator called Rubik's cube convolution that improves image restoration performance efficiently and simply by expanding first-order interactions to arbitrary high orders.
Amir Zandieh, Insu Han, Haim Avron
https://openreview.net/forum?id=hNpedVWwoe
Keywords: spherical harmonic expansion, Gegenbauer polynomials, interpolation, leverage score sampling
Compressor summary: The paragraph describes an efficient algorithm to recover the spherical harmonic expansion of functions defined on the unit sphere using a near-optimal number of function evaluations and kernel regression.
Ti-Rong Wu, Hung Guei, Ting Han Wei, Chung-Chin Shih, Jui-Te Chin, I-Chen Wu
https://openreview.net/forum?id=hN4qpvGzWn
Keywords: Game Solving, Computer Games, AlphaZero, Online Fine-Tuning, Monte Carlo Tree Search, Deep Reinforcement Learning
Compressor summary: The paper proposes using online fine-tuning and tailor-designed heuristics to improve the AlphaZero algorithm's performance in solving games, especially those with poor lines of play from the losing side.
James Oldfield, Christos Tzelepis, Yannis Panagakis, Mihalis Nicolaou, Ioannis Patras
https://openreview.net/forum?id=hLoanbRrjM
Keywords: generative models, text-to-image, vision-language models, interpretability
Compressor summary: The paper proposes a method to separate different visual properties in CLIP's image representations using part-of-speech associations and apply it to text-to-image synthesis and style transfer tasks.
Zhaoying Pan, Daniel Geng, Andrew Owens
https://openreview.net/forum?id=hLPJ7xLbNF
Keywords: Video Processing, Motion Processing, Motion Magnification, Optical Flow
Compressor summary: The paper introduces a self-supervised technique to magnify subtle motions in video by adjusting the optical flow based on a given factor and training a model with a loss function that minimizes the deviation from the desired magnification.
Benjamin Coleman, Wang-Cheng Kang, Matthew Fahrbach, Ruoxi Wang, Lichan Hong, Ed H. Chi, Derek Zhiyuan Cheng
https://openreview.net/forum?id=hJzEoQHfCe
Keywords: embedding learning; recommendation systems; representation learning
Compressor summary: The text introduces Feature Multiplexing, a framework that uses one representation space for multiple features, which improves embedding efficiency and effectiveness for web-scale machine learning systems.
Yongduo Sui, Qitian Wu, Jiancan Wu, Qing Cui, Longfei Li, JUN ZHOU, Xiang Wang, Xiangnan He
https://openreview.net/forum?id=hIGZujtOQv
Keywords: Graph Neural Network, Graph Data Augmentation, Distribution Shift
Compressor summary: The text discusses the problem of distribution shifts in graph representation learning, proposing a new data augmentation strategy called Adversarial Invariant Augmentation (AIA) to handle covariate shift on graphs.
Hatef Otroshi Shahreza, Sébastien Marcel
https://openreview.net/forum?id=hI6EPhq70A
Keywords: Face Recognition (FR), Face reconstruction, Generative Adversarial Network (GAN), Privacy, Security, Template Inversion (TI) attack, Transferability
Compressor summary: The paper proposes a method to attack face recognition systems by reconstructing realistic face images from facial templates using a GAN-based framework, which can be applied in different scenarios and is transferable between systems.
Chengchang Liu, Cheng Chen, Luo Luo, John C.S. Lui
https://openreview.net/forum?id=hHv3UuffXV
Keywords: Broyden's method, nonlinear equations
Compressor summary: The paper introduces block variants of good and bad Broyden's methods for nonlinear equation solving, with improved convergence rates and reduced computational cost compared to existing methods.
Yuxuan Song, Jingjing Gong, Minkai Xu, Ziyao Cao, Yanyan Lan, Stefano Ermon, Hao Zhou, Wei-Ying Ma
https://openreview.net/forum?id=hHUZ5V9XFu
Keywords: Drug Design, Molecule Generation, Deep Learning, Computational Biology
Compressor summary: The paper introduces geometric flow matching, a hybrid probability path method that improves the stability and efficiency of diffusion models for generating 3D molecules.
James Cook, Milind Shyani, Nina Mishra
https://openreview.net/forum?id=hExFOGZTSt
Keywords: Differential privacy, machine learning, linear queries
Compressor summary: The paper proposes a non-interactive privacy-preserving method to publish datasets with sensitive attributes and enable joins with other datasets on those attributes, using private sketches and satisfying pure differential privacy.
Zaiwei Chen, Kaiqing Zhang, Eric Mazumdar, Asuman E. Ozdaglar, Adam Wierman
https://openreview.net/forum?id=hElNdYMs8Z
Keywords: Zero-sum stochastic games, payoff-based independent learning, best-response-type dynamics, finite-sample analysis
Compressor summary: The authors propose a novel learning dynamics for two-player stochastic games that combines matrix game and stochastic game methods, and provide theoretical guarantees on its convergence, rationality, and symmetry.
Qiufu Li, Xi Jia, Jiancan Zhou, Linlin Shen, Jinming Duan
https://openreview.net/forum?id=hE7PG1lUZx
Keywords: Face Recognition, Unified Threshold, USS Loss
Compressor summary: The paper proposes a new loss function (USS loss) for face recognition that uses a unified threshold to distinguish positive from negative facial pairs and improves performance compared to existing methods.
Bahar Taskesen, Dan Andrei Iancu, Çağıl Koçyiğit, Daniel Kuhn
https://openreview.net/forum?id=hE5RWzQyvf
Keywords: linear quadratic control, distributionally robust optimization, optimal transport, Wasserstein distance
Compressor summary: The paper proposes an optimal control policy for a generalization of the discrete-time, finite-horizon LQG problem with unknown noise distributions, using the Frank-Wolfe algorithm and Kalman filter estimation.
Yiheng Lin, James A Preiss, Emile Timothy Anand, Yingying Li, Yisong Yue, Adam Wierman
https://openreview.net/forum?id=hDajsofjRM
Keywords: Online policy selection, online control, online learning
Compressor summary: The GAPS algorithm optimizes online policy selection in systems with varying costs and dynamics by approximating an ideal gradient descent algorithm, achieving optimal or local regret depending on convexity, and outperforming current methods in adaptability.
Zepu Lu, Jin Chen, Defu Lian, ZAIXI ZHANG, Yong Ge, Enhong Chen
https://openreview.net/forum?id=hCg4w8L8Dt
Keywords: Approximate Nearest Neighbor Search, Knowledge Distillation, Product Quantization, Inverted Index
Compressor summary: The paper proposes a new learning algorithm, KDindex, that improves retrieval performance of compressed search indices by distilling knowledge from high-precision models and using additional supervision signals between queries and documents.
Liang Zhang, Junchi YANG, Amin Karbasi, Niao He
https://openreview.net/forum?id=hCdqDkA25J
Keywords: Reproducibility, Convex Optimization, Minimax Optimization, Saddle-Point Problem
Compressor summary: The paper shows how to achieve both optimal reproducibility and near-optimal convergence for machine learning algorithms under various error-prone oracles, challenging the previous perception of a trade-off between these factors.
Puheng Li, Zhong Li, Huishuai Zhang, Jiang Bian
https://openreview.net/forum?id=hCUG1MCFk5
Keywords: diffusion models, training dynamics, generalization gap, modes shift
Compressor summary: This paper explores the generalization capabilities of diffusion models and provides theoretical and empirical evidence for their performance on different scenarios, such as early stopping and mode shifts.
Ceyuan Yang, Qihang Zhang, Yinghao Xu, Jiapeng Zhu, Yujun Shen, Bo Dai
https://openreview.net/forum?id=h8vJVABiBP
Keywords: generative adversarial network, image synthesis, video synthesis
Compressor summary: The paper introduces a new module (MTM) for GANs that can handle geometry deformation by predicting spatial offsets based on latent codes, improving performance on various generative tasks.
Ankit Vishnubhotla, Charlotte Loh, Akash Srivastava, Liam Paninski, Cole Lincoln Hurwitz
https://openreview.net/forum?id=h6WUKM7PCI
Keywords: contrastive learning, representations, extracellular, high-density, spike sorting, cell-type classification, transformers, invariance
Compressor summary: The authors propose a new contrastive learning framework, CEED, for analyzing high-density extracellular recordings in neuroscience and show that it outperforms current methods.
Xiang Wang, Hangjie Yuan, Shiwei Zhang, Dayou Chen, Jiuniu Wang, Yingya Zhang, Yujun Shen, Deli Zhao, Jingren Zhou
https://openreview.net/forum?id=h4r00NGkjR
Keywords: Video Synthesis, Video Diffusion Model, Compositional Synthesis
Compressor summary: VideoComposer is a tool that enables users to create videos with various temporal conditions by using motion vectors from compressed videos and a Spatio-Temporal Condition encoder.
Eghbal A. Hosseini, Evelina Fedorenko
https://openreview.net/forum?id=h3lTrt4Ftb
Keywords: (Cognitive/Neuroscience) Language, Structured Prediction, (Application) Natural Language and Text Processing
Compressor summary: The authors investigate how predicting the next word in a sentence affects the way transformer models represent language, and find that these models tend to favor straighter trajectories for making predictions.
Abhin Shah, Karthikeyan Shanmugam, Murat Kocaoglu
https://openreview.net/forum?id=h3kuB4z2G9
Keywords: Causal Effect Estimation, Front-door Adjustment, Limited Graph Knowledge
Compressor summary: The text discusses a new method to estimate causal effects without knowing the graph structure, using testable conditional independence statements and front-door-like adjustment with limited side information.
Zhiheng Liu, Yifei Zhang, Yujun Shen, Kecheng Zheng, Kai Zhu, Ruili Feng, Yu Liu, Deli Zhao, Jingren Zhou, Yang Cao
https://openreview.net/forum?id=h3QNH3qeC3
Keywords: text2img, diffusion model, customization
Compressor summary: This paper proposes an efficient method to synthesize images with multiple user-specified subjects using text embeddings and layout guidance, achieving better results than existing approaches.
Badri Narayana Patro, Vijay Srinivas Agneeswaran
https://openreview.net/forum?id=h3MShWMxNt
Keywords: Vision Transformer, Scatter Network, Spectral Transformer, Token Mixing, Channel Mixing, and Einstein Blending Method
Compressor summary: The Scattering Vision Transformer (SVT) is a novel approach that captures intricate image details and reduces attention complexity in vision transformers using spectrally scattering networks and spectral gating.
Yifan Xu, Mengdan Zhang, Chaoyou Fu, Peixian Chen, Xiaoshan Yang, Ke Li, Changsheng Xu
https://openreview.net/forum?id=h3CGHf7457
Keywords: open-world object detection, multi-modal query, vision-language pre-training
Compressor summary: MQ-Det is a method to improve object detection using both text and visual information, which can be easily applied to existing detectors and boosts performance on various tasks.
Arun Ganesh, MAHDI HAGHIFAM, Thomas Steinke, Abhradeep Guha Thakurta
https://openreview.net/forum?id=h2lkx9SQCD
Keywords: differential privacy, logistic regression, optimization, Newton's method, second-order methods
Compressor summary: The paper proposes a private version of Newton's method that can accelerate differentially private optimization, and tests it on logistic regression with promising results.
Ta Duy Nguyen, Thien Hang Nguyen, Alina Ene, Huy Nguyen
https://openreview.net/forum?id=h1FhXVM0cB
Keywords: convex optimization, non-convex optimization, high probability convergence, heavy-tailed noise, clipped stochastic gradient descent, clipped stochastic mirror descent
Compressor summary: The authors propose a new analysis approach for clipped gradient methods with heavy-tailed noise that improves convergence guarantees, allows time-varying step sizes and clipping parameters, and does not need problem constants.
Ignacio Hounie, Alejandro Ribeiro, Luiz F. O. Chamon
https://openreview.net/forum?id=h0RVoZuUl6
Keywords: Constrained Learning, Relaxation, Lagrangian duality, Primal-Dual, Machine Learning, Federated Learning, Invariance
Compressor summary: The paper proposes a constrained learning approach that adapts requirements while solving the task, balancing performance gains against the cost of relaxing constraints, called resilient constrained learning.
Matthew Le, Apoorv Vyas, Bowen Shi, Brian Karrer, Leda Sari, Rashel Moritz, Mary Williamson, Vimal Manohar, Yossi Adi, Jay Mahadeokar, Wei-Ning Hsu
https://openreview.net/forum?id=gzCS252hCO
Keywords: speech generation, flow-matching, diffusion, in-context learning, text-to-speech
Compressor summary: Voicebox is a versatile text-guided generative model for speech that can perform various tasks, such as zero-shot TTS, noise removal, and style conversion, with better performance and faster speed than VALL-E.
Jinghua Hou, Zhe Liu, dingkang liang, Zhikang Zou, Xiaoqing Ye, Xiang Bai
https://openreview.net/forum?id=gySmwdmVDF
Keywords: 3D Object Detection, Temporal, LiDAR-only, Multi-modality, Autonomous Driving
Compressor summary: The paper introduces QTNet, a simple and effective method for 3D detection in autonomous driving that exploits object queries in previous frames to enhance current ones using Motion-guided Temporal Modeling (MTM).
Yuxuan Guo, Yifan Hao, Rui Zhang, Enshuai Zhou, Zidong Du, Xishan Zhang, Xinkai Song, Yuanbo Wen, Yongwei Zhao, Xuehai Zhou, Jiaming Guo, Qi Yi, Shaohui Peng, Di Huang, Ruizhi Chen, Qi Guo, Yunji Chen
https://openreview.net/forum?id=gx20B4ItIw
Keywords: Emergent communication, Multi-agent communication, Raven's Progressive Matrices, Representation learning
Compressor summary: The paper introduces a new game for training deep-learning-based agents to reason and communicate high-level rules, and presents a dataset and training method for this purpose.
Siddharth Gollapudi, Sepideh Mahabadi, Varun Sivashankar
https://openreview.net/forum?id=gwvwbsnTps
Keywords: Determinant Maximization, Composable Coresets, Greedy Algorithm, DPP
Compressor summary: The paper proposes and analyzes a Greedy algorithm for determinant maximization with an almost optimal approximation factor and shows its practicality and local optimality on real data sets.
Juyeon Heo, Vihari Piratla, Matthew Robert Wicker, Adrian Weller
https://openreview.net/forum?id=guyhQMSp2F
Keywords: Learning from explanation, Robustness, Interpretability, Shortcuts, Explanations
Compressor summary: The paper proposes a new approach to machine learning from explanations that improves performance by using robustness techniques instead of model smoothing.
Jinxi Li, Ziyang Song, Bo Yang
https://openreview.net/forum?id=gsi9lJ3994
Keywords: Physics Learning, Velocity Field, Dynamic Radiance Field, Future Frame Extrapolation
Compressor summary: The paper proposes a method to learn geometry, appearance, and velocity of 3D scenes from videos for various applications such as future frame prediction and 3D scene analysis.
Haibao Yu, Yingjuan Tang, Enze Xie, Jilei Mao, Ping Luo, Zaiqing Nie
https://openreview.net/forum?id=gsglrhvQxX
Keywords: vehicle-infrastructure cooperative autonomous driving, 3D object detection, feature flow, self-supervised learning
Compressor summary: The Feature Flow Net (FFNet) is a novel cooperative detection framework for autonomous driving that predicts future features and uses feature flow transmission to address temporal asynchrony and limited communication conditions in traffic environments, achieving better performance with low transmission cost and latency.
Ayush Tewari, Tianwei Yin, George Cazenavette, Semon Rezchikov, Joshua B. Tenenbaum, Fredo Durand, William T. Freeman, Vincent Sitzmann
https://openreview.net/forum?id=gq4xkwQZ1l
Keywords: 3D generative models, neural rendering, neural scene representations, NeRF, diffusion models, differentiable rendering, inverse graphics, inverse problems
Compressor summary: The text introduces a new type of generative models that can learn from indirect observations through a known forward model, enabling conditional generation of hidden signals and applications in computer vision tasks like inverse graphics.
Jang-Hyun Kim, Sangdoo Yun, Hyun Oh Song
https://openreview.net/forum?id=gpyeRyc858
Keywords: Dataset cleaning, Label error detection, Outlier detection, Neural Networks, Robustness
Compressor summary: The paper proposes a method to identify and clean data problems using the relational structure of data in the feature-embedded space, and demonstrates its effectiveness on various image, speech, and language tasks.
Zhichao Wang, Andrew William Engel, Anand Sarwate, Ioana Dumitriu, Tony Chiang
https://openreview.net/forum?id=gpqBGyKeKH
Keywords: Random matrix theory, Heavy tails, Feature learning, Linear-width neural networks, Spike phase transition
Compressor summary: The spectral properties of high-dimensional linear neural networks change under different training strategies, affecting their generalization and feature learning abilities.
Chen Sun, Wannan Yang, Thomas Jiralerspong, Dane Malenfant, Benjamin Alsbury-Nealy, Yoshua Bengio, Blake Aaron Richards
https://openreview.net/forum?id=gpJw8f4tIU
Keywords: Reinforcement learning, long term credit assignment, rapid credit assignment, contrastive learning, few-shot learning in RL
Compressor summary: The paper introduces Contrastive Retrospection (ConSpec), an RL algorithm that uses offline contrastive learning to identify critical steps in a task and provides intrinsic rewards for matching them, improving learning in various RL tasks.
Jianhao Zhang, Shihan Ma, Peihong Liu, Jinhui Yuan
https://openreview.net/forum?id=gmmXyAq8TI
Keywords: Tensor Rematerialization; Gradient Checkpointing; Activation Recomputing; Deep Learning; Deep Learning Frameworks; Memory Allocator
Compressor summary: Tensor rematerialization technique optimizes memory allocation and saves resources for deep neural networks by using a sliding window and cheap tensor partitioning.
Dmitry Yarotsky
https://openreview.net/forum?id=gmVoaAxB1R
Keywords: VC-dimension, neural networks, activation functions, approximation, polynomials, algebraic functions
Compressor summary: The paper studies expressive models like neural networks and their limitations in approximating continuous functions on different sets using a hierarchy of classes and a new family of functions with polynomial constraints.
Joshua Benjamin Evans, Özgür Şimşek
https://openreview.net/forum?id=gjBk6IQofa
Keywords: Reinforcement Learning, Hierarchical Reinforcement Learning, RL, HRL, Skill Discovery, Skill Hierarchies, Graph-Based, Graphs, Graph Clustering, Graph Partitioning
Compressor summary: The text proposes an automatic method to generate a skill hierarchy for autonomous agents based on modularity maximisation and interaction graphs, which improves their learning performance in various environments.
Jiaxiang Dong, Haixu Wu, Haoran Zhang, Li Zhang, Jianmin Wang, Mingsheng Long
https://openreview.net/forum?id=ginTcBUnL8
Keywords: Time-series analysis, pre-training, masked time-series modeling
Compressor summary: SimMTM is a simple framework that improves masked modeling for time series by aggregating complementary information from neighbors and uncovering local structure.
Arlind Kadra, Maciej Janowski, Martin Wistuba, Josif Grabocka
https://openreview.net/forum?id=ghzEUGfRMD
Keywords: hyperparameter optimization, multi-fidelity hyperparameter optimization, multi-fidelity hpo, power laws, deep neural networks, deep power laws, deep ensemble, deep learning, large language models, scaling laws, llm
Compressor summary: The authors propose Deep Power Laws (DPL), a neural network ensemble that optimizes hyperparameters using power law scaling and gray-box evaluations, outperforming seven other methods on various datasets.
Longlin Yu, Tianyu Xie, Yu Zhu, Tong Yang, Xiangyu Zhang, Cheng Zhang
https://openreview.net/forum?id=ghIBaprxsV
Keywords: Hierarchical semi-implicit variational inference, Score based training, Diffusion model
Compressor summary: HSIVI is a hierarchical extension of SIVI that improves expressiveness and allows acceleration of diffusion models using pre-trained score networks.
Qihang Fang, Yafei Song, Keqiang Li, Liefeng Bo
https://openreview.net/forum?id=gh9JNeqjzo
Keywords: Neural radiance field, Novel-view synthesis, Regularization
Compressor summary: The paper proposes a more adaptive method to reduce shape-radiance ambiguity in neural radiance fields by estimating the color field based on the density field and using it to regularize NeRF's rendering process.
Sirui Li, Wenbin Ouyang, Max B. Paulus, Cathy Wu
https://openreview.net/forum?id=gf5xJVQS5p
Keywords: Combinatorial Optimization, Branch-and-Cut, Learning Guided Optimization, Deep Learning
Compressor summary: This paper proposes a novel data-driven strategy for selecting combinatorial separators in mixed integer linear programs, which can improve the solve time by up to 72% and 37% on synthetic and real-world benchmarks.
Hua Yuan, Yu Shi, Ning Xu, Xu Yang, Xin Geng, Yong Rui
https://openreview.net/forum?id=gevmGxsTSI
Keywords: Soft label; knowledge distillation; weakly-supervised learning; Machine learning.
Compressor summary: The paper explores how biased soft labels can still be effective for knowledge distillation, proposes indicators to measure their effectiveness, and applies the framework to three weakly-supervised learning scenarios.
Zhihan Zhou, Jiangchao Yao, Feng Hong, Ya Zhang, Bo Han, Yanfeng Wang
https://openreview.net/forum?id=geLARFEK8O
Keywords: Long-tailed learning, self-supervised learning
Compressor summary: The text introduces Geometric Harmonization (GH), a method that improves self-supervised learning (SSL) under long-tailed distribution by ensuring category-level uniformity in representation learning, avoiding collapse of tail classes and dominance of head classes.
Chen Zeno, Greg Ongie, Yaniv Blumenfeld, Nir Weinberger, Daniel Soudry
https://openreview.net/forum?id=gdzxWGGxWE
Keywords: Denoiser, Denoising, Neural network, Function space
Compressor summary: The paper studies the theoretical properties of shallow ReLU neural network denoisers for image reconstruction tasks, deriving their functions and showing they generalize better than an alternative method in low noise settings.
Jing Gong, Minsheng Hao, Xingyi Cheng, Xin Zeng, Chiming Liu, Jianzhu Ma, Xuegong Zhang, Taifeng Wang, Le Song
https://openreview.net/forum?id=gdwcoBCMVi
Keywords: pre-train, encoder-decoder, scRNA-seq, scalable
Compressor summary: The xTrimoGene model is a novel transformer architecture that can efficiently process large single-cell RNA-seq data and achieve state-of-the-art performance on various downstream tasks, while reducing computational and memory costs significantly.
Lee Cohen, Yishay Mansour, Michal Moshkovitz
https://openreview.net/forum?id=gdVcFOvxT3
Keywords: Theoretical guarantees, algorithms, learning theory, MDP, computational complexity, Interpretability
Compressor summary: The paper studies safe zones in Markov decision processes, which are subsets of states that most trajectories stay within, and presents an algorithm to find approximate safe zones with low complexity.
Banghua Zhu, Ying Sheng, Lianmin Zheng, Clark Barrett, Michael Jordan, Jiantao Jiao
https://openreview.net/forum?id=gd20oaZqqF
Keywords: caching, model selection, large language models, foundation models, inference, bandit, regret
Compressor summary: The paper proposes two methods to reduce resource consumption and latency of large foundation models during inference: caching previous queries and selecting the best model for each query, and shows their theoretical and empirical benefits.
Abhineet Agarwal, Anish Agarwal, Suhas Vijaykumar
https://openreview.net/forum?id=gbhixjg2dX
Keywords: Causal Inference, Matrix Completion, Combinatorial Learning, Ranking
Compressor summary: The paper proposes Synthetic Combinations, a method to learn causal parameters from observational data by exploiting latent structure across units and interventions.
Jindong Jiang, Fei Deng, Gautam Singh, Sungjin Ahn
https://openreview.net/forum?id=gbOukzirpK
Keywords: Object-Centric Representation Learning, Diffusion Models, Unsupervised Representation Learning
Compressor summary: The paper introduces Latent Slot Diffusion, a novel model that combines diffusion models with object-centric learning for better image generation, especially in complex scenes, without supervised annotations.
DI QI, Tong Yang, Xiangyu Zhang
https://openreview.net/forum?id=ganlU27uvj
Keywords: 3D object-centric representation learning, NeRF, 3D-aware slot
Compressor summary: The paper introduces a new method called sVORF that learns 3D object-centric representations from images by decomposing scenes into objects using volumetric radiance fields and object slots.
yanwu xu, Mingming Gong, Shaoan Xie, Wei Wei, Matthias Grundmann, kayhan Batmanghelich, Tingbo Hou
https://openreview.net/forum?id=gaktiSjatl
Keywords: Diffusion Model, GAN, Semi-implicit Modeling
Compressor summary: The paragraph introduces a novel approach to improve sample quality and diversity in generative models by combining an implicit model with the forward diffusion process, enabling larger steps during inference without compromising performance.
Yulian Wu, Xingyu Zhou, Youming Tao, Di Wang
https://openreview.net/forum?id=gaXAjtHic2
Keywords: Bandits, privacy, robustness
Compressor summary: The paper investigates how to balance privacy, robustness, and accuracy in multi-armed bandits with heavy-tailed rewards under a limited budget.
Po-An Wang, Ruo-Chun Tzeng, Alexandre Proutiere
https://openreview.net/forum?id=gYetLsNO8x
Keywords: Best arm identification, Large deviation
Compressor summary: The paper studies how to find the best arm in stochastic Multi-Armed Bandits with a fixed sampling budget, and presents a new adaptive algorithm called Continuous Rejects that outperforms existing methods using Large Deviation techniques.
Andre Vauvelle, Benjamin Wild, Roland Eils, Spiros Denaxas
https://openreview.net/forum?id=gYWjI7wLhc
Keywords: Survival Analysis, Censored Data, Semi-supervised Learning, Time-to-event-data, Algorithmic Supervision, Sorting, Risk Prediction, Weakly-supervised Learning, Machine Learning, Cox's Partial Likelihood, Differentiable Sorting Networks, Transitive Inductive Bias, Ranking Losses, Listwise Ranking, Healthcare Applications, Deep Learning, Neural Networks, Top-k Risk Prediction
Compressor summary: Diffsurv is a new survival analysis method that uses differentiable sorting networks to handle censored data and improve risk prediction in healthcare applications.
Yajie Bao, Amarda Shehu, Mingrui Liu
https://openreview.net/forum?id=gVLKXT9JwG
Keywords: convolutional neural network, gaussian input, local SGD, global convergence, non-convex optimization
Compressor summary: This paper presents the first global convergence analysis of local SGD for two-layer neural networks without overparameterization or noise, using a self-correction mechanism and an exact recursive characterization of the direction of global parameters.
Anselm Krainovic, Mahdi Soltanolkotabi, Reinhard Heckel
https://openreview.net/forum?id=gUlcyeHzw1
Keywords: Adversarial Training, Jittering, Denoising, Deconvolution, Compressive Sensing, Inverse Problems, Robustness
Compressor summary: The paper explores if adding noise during training makes neural networks more resistant to worst-case perturbations in inverse problems like denoising and finds that jittering improves robustness but may not work for all inverse problems.
Mihir Prabhudesai, Tsung-Wei Ke, Alexander Cong Li, Deepak Pathak, Katerina Fragkiadaki
https://openreview.net/forum?id=gUTVpByfVX
Keywords: test-time adaptation, diffusion models, generative models, classification, segmentation, depth prediction
Compressor summary: Diffusion-TTA uses generative models to adapt pre-trained discriminative models for different test examples, improving their accuracy and performance in online settings.
Chen Fan, Gaspard Choné-Ducasse, Mark Schmidt, Christos Thrampoulidis
https://openreview.net/forum?id=gUEekxYr6D
Keywords: Adaptive step sizes, bi-level optimization, convergence rates, line searches
Compressor summary: The authors propose adaptive step-size methods for bi-level optimization in deep learning to improve convergence speed without requiring fine-tuning.
Anh-Dung Dinh, Daochang Liu, Chang Xu
https://openreview.net/forum?id=gThGBHhqcU
Keywords: Diffusion model, conditional generative model, guidance diffusion, generative models, classifier guidance
Compressor summary: The paper presents Progressive Guidance, a generalized method for diffusion generative models that addresses diversity and adversarial effects issues in classifier guidance by refining image details progressively.
Austin Watkins, Enayat Ullah, Thanh Nguyen-Tang, Raman Arora
https://openreview.net/forum?id=gQ4h6WvME0
Keywords: Learning Theory, Multi-task and Transfer Learning, Classification
Compressor summary: The paper investigates how multi-task representation learning can improve transfer learning by providing new statistical rates on the excess risk of the target task depending on the difficulty of the learning task.
Elad Hazan, Adam Tauman Kalai, Varun Kanade, Clara Mohri, Y. Jennifer Sun
https://openreview.net/forum?id=gPylY8sCbw
Keywords: matrix completion, online learning
Compressor summary: The paragraph introduces the Partial Matrix Completion problem, where the goal is to accurately complete a subset of matrix entries, and presents an algorithm that handles complex sampling distributions and performs well in online settings.
Zhaoyu Chen, Bo Li, Shuang Wu, Kaixun Jiang, Shouhong Ding, Wenqiang Zhang
https://openreview.net/forum?id=gO60SSGOMy
Keywords: unrestricted attack, adversarial example, diffusion model, black-box attack, adversarial transferability
Compressor summary: The proposed Content-based Unrestricted Adversarial Attack framework optimizes images along an adversarial direction to create photorealistic adversarial examples that deceive both human perception and deep neural networks with high attack performance.
Paribesh Regmi, Rui Li
https://openreview.net/forum?id=gMjIUZBKH8
Keywords: nonparametric Bayes, variational autoencoders
Compressor summary: The paragraph introduces a Bayesian inference framework for variational autoencoders (VAEs) that adapts network structures to data and prevents overfitting, leading to improved generative performance.
Tian Zhou, Peisong Niu, Xue Wang, Liang Sun, Rong Jin
https://openreview.net/forum?id=gMS6FVZvmF
Keywords: general time series analysis, time series forecasting, cross modality knowledge transfer; pretrained language model;
Compressor summary: The paper presents Frozen Pretrained Transformer (FPT), a model that uses pre-trained language or image models for time series analysis, and shows its comparable or state-of-the-art performance on various tasks.
Jiachen Zhao, Tao Yu, Liang An, Yipeng Huang, Fang Deng, Qionghai Dai
https://openreview.net/forum?id=gLwjBDsE3G
Keywords: 3D pose estimation, triangulation, animal pose estimation
Compressor summary: The paper proposes a self-supervised method called Triangulation Residual loss for 3D pose estimation that uses global multiview geometric consistency and achieves state-of-the-art performance with limited labeled data.
Kirill Bykov, Laura Kopf, Shinichi Nakajima, Marius Kloft, Marina MC Höhne
https://openreview.net/forum?id=gLfgyIWiWW
Keywords: Explainable AI, Mechanistic Interpretability, Machine Learning, Deep Neural Networks
Compressor summary: INVERT is a scalable method to link DNNs' complex data representations to human-interpretable concepts without relying on segmentation masks or high computational demands, enabling interpretable explanations and assessing their statistical significance.
Zikai Xiao, Zihan Chen, Songshang Liu, Hualiang Wang, YANG FENG, Jin Hao, Joey Tianyi Zhou, Jian Wu, Howard Hao Yang, Zuozhu Liu
https://openreview.net/forum?id=gJewjFjfN2
Keywords: Federated learning, Long-tailed learning, Data heterogeneity
Compressor summary: The paper introduces $\texttt{Fed-GraB}$, a method for federated long-tailed learning that re-weights client gradients based on global distribution feedback to improve performance on minority classes while preserving majority class performance.
Soham Deshmukh, Benjamin Elizalde, Rita Singh, Huaming Wang
https://openreview.net/forum?id=gJLAfO4KUq
Keywords: audio language model, audio representation learning, audio and speech processing, multi-task and transfer learning
Compressor summary: Pengi is an audio language model that uses transfer learning to enable open-ended and close-ended tasks without extra fine-tuning, achieving state-of-the-art performance on 21 downstream tasks.
Mikaela Angelina Uy, Kiyohiro Nakayama, Guandao Yang, Rahul Krishna Thomas, Leonidas Guibas, Ke Li
https://openreview.net/forum?id=gJHAT79cZU
Keywords: neural radiance fields, volumetric rendering, nerfs, numerical quadrature, importance sampling
Compressor summary: The paragraph discusses quadrature instability in Neural Radiance Fields (NeRF) and proposes a solution that improves texture, geometry, and depth supervision by reformulating the sample-based rendering equation to match piecewise linear volume density.
Kaiqi Jiang, Dhruv Malik, Yuanzhi Li
https://openreview.net/forum?id=gIG8LvTLuc
Keywords: optimization, adaptive algorithms, neural networks
Compressor summary: Adaptive optimization methods improve neural network training by biasing iterate trajectories towards regions with lower $R^{\text{OPT}}\_{\text{med}}$, a statistic similar to the condition number of the loss Hessian, compared to vanilla gradient methods like SGD.
Yujia Zheng, Kun Zhang
https://openreview.net/forum?id=gI1SOgW3kw
Keywords: Latent variable models, nonlinear independent component analysis
Compressor summary: The paragraph discusses new identifiability results for nonlinear ICA under different conditions, such as undercompleteness, partial sparsity, source dependence, and flexible grouping structures, addressing limitations of previous methods.
Andrea Schioppa, Katja Filippova, Ivan Titov, Polina Zablotskaia
https://openreview.net/forum?id=gGl0n7Onug
Keywords: Explainable AI, Influence Functions, Training Data Attribution
Compressor summary: Influence functions are useful for debugging and correcting deep neural networks but their predictive power is limited by parameter divergence, which causes influence to fade over training time.
Yuriy Biktairov, Jyotirmoy Deshmukh
https://openreview.net/forum?id=gAQCx61chN
Keywords: Neural network verification, Robustness, Linear bounding
Compressor summary: The paper proposes optimal methods for finding tight linear bounds for activation functions in neural networks, which are important for robustness certification tools, using a new sampling-based approach that is efficient and practical.
Joey Hejna, Dorsa Sadigh
https://openreview.net/forum?id=gAP52Z2dar
Keywords: preference learning, preference-based reinforcement learning, human-in-the-loop reinforcement learning
Compressor summary: The paper proposes Inverse Preference Learning (IPL), a new simple and efficient algorithm for learning from offline preference data in preference-based RL, which eliminates the need for a learned reward function by using the $Q$-function instead.
Chenyu You, Weicheng Dai, Yifei Min, Fenglin Liu, David A. Clifton, S Kevin Zhou, Lawrence Hamilton Staib, James s Duncan
https://openreview.net/forum?id=g9gjpFOiO4
Keywords: Long-tailed Medical Image Segmentation, Contrastive Learning, Variance Reduction, Imbalanced Learning, Semi-Supervised Learning
Compressor summary: The paper introduces $\texttt{ARCO}$, a semi-supervised contrastive learning framework that improves medical image segmentation by using stratified group theory, variance-reduced estimation, and universal sampling techniques.
Zhenting Wang, Chen Chen, Yi Zeng, Lingjuan Lyu, Shiqing Ma
https://openreview.net/forum?id=g8bjq0qxOl
Keywords: Origin attribution of generated images
Compressor summary: The paper proposes an alteration-free, model-agnostic method for determining if an image was generated by a specific image generation model using reverse-engineering and reconstruction loss.
Shiva Kanth Sujit, Somjit Nath, Pedro Braga, Samira Ebrahimi Kahou
https://openreview.net/forum?id=g78QqvhnDU
Keywords: reinforcement learning, sample efficiency, experience replay
Compressor summary: The paper proposes a method to prioritize reinforcement learning samples based on their learn-ability, which is defined as the decrease of training loss over time, to improve sample efficiency and robustness.
Dongxu Li, Junnan Li, Steven Hoi
https://openreview.net/forum?id=g6We1SwaY9
Keywords: Diffusion-based Models, Text-to-Image Generation, Image Editing, Vision-and-Language, Multimodal
Compressor summary: BLIP-Diffusion is a new model that generates images of a given subject based on both image and text inputs, using a pre-trained multimodal encoder and a diffusion model. It enables efficient fine-tuning and can be combined with other techniques for more advanced applications.
Yuankai Luo, Veronika Thost, Lei Shi
https://openreview.net/forum?id=g49s1N5nmO
Keywords: Graph Neural Networks, Transformers, Graph Classification, Node Classification, Scalability
Compressor summary: This paper proposes efficient and effective architecture adaptations for transformer models on directed acyclic graphs (DAGs), enabling them to outperform graph neural networks tailored for DAGs.
Florian E. Dorner, Nikola Konstantinov, Georgi Stoyanov Pashaliev, Martin Vechev
https://openreview.net/forum?id=g2ROKOASiv
Keywords: Game Theory, Federated Learning, Optimization, Strategic Behavior, Economics, Mechanisms
Compressor summary: This paper studies how to design mechanisms that encourage honest communication between competitors in collaborative machine learning, despite their incentive to manipulate their updates for personal gain, and demonstrates their effectiveness on a non-convex federated learning task.
Haoyu Chen, Hao Tang, Radu Timofte, Luc Van Gool, Guoying Zhao
https://openreview.net/forum?id=g27BggUT3L
Keywords: 3D motion transfer, 3D Transformer, geometric preservation, 3D generation, correspondence learning
Compressor summary: The paper introduces LART, a 3D Transformer framework for transferring realistic and high-fidelity motions between meshes without key point annotations or predefined correspondence, using latent metric regularization to ensure motion quality.
Toru Lin, Allan Jabri
https://openreview.net/forum?id=g1dMYenhe4
Keywords: reinforcement learning, exploration, intrinsic reward, intrinsic motivation, masked autoencoder
Compressor summary: The paper proposes MIMEx, a framework for deriving intrinsic rewards in high-dimensional environments using conditional prediction and masked input modeling, which can improve exploration in sparse-reward visuomotor tasks.
Yixiu Mao, Hongchang Zhang, Chen Chen, Yi Xu, Xiangyang Ji
https://openreview.net/forum?id=fze7P9oy6l
Keywords: offline reinforcement learning
Compressor summary: SVR is a value regularization method for offline RL that penalizes Q-values for out-of-distribution actions and maintains optimal convergence results while achieving state-of-the-art performance on continuous control tasks.
Lukas Uzolas, Elmar Eisemann, Petr Kellnhofer
https://openreview.net/forum?id=fyfmHi8ay3
Keywords: Radiance Fields, View Synthesis, Kinematics, Reposing, NeRF
Compressor summary: The paper introduces a new method for synthesizing realistic dynamic scenes using point-based representation, Linear Blend Skinning, and multi-view video, improving visual quality, learning time, and generalization.
Mariia Seleznova, Dana Weitzner, Raja Giryes, Gitta Kutyniok, Hung-Hsu Chou
https://openreview.net/forum?id=fyLvHzEssH
Keywords: Neural Collapse, Neural Tangent Kernel, NTK alignment, Local Elasticity, Gradient Flow
Compressor summary: This work connects two ideas in deep learning, shows how class labels affect neural network training dynamics, and proves that well-trained networks develop structure based on these assumptions.
Kevin Clark, Priyank Jaini
https://openreview.net/forum?id=fxNQJVMwK2
Keywords: diffusion models, zero-shot, text-to-image, generative models, foundation models, stable diffusion
Compressor summary: The authors propose a method to evaluate text-to-image diffusion models as zero-shot classifiers, showing they perform competitively with CLIP on image classification tasks and have advantages in attribute binding. They suggest generative pre-training could be useful for vision and vision-language problems.
Tianhao Wu, Mingdong Wu, Jiyao Zhang, Yunchong Gan, Hao Dong
https://openreview.net/forum?id=fwvfxDbUFw
Keywords: Human-asissting Dexterous Grasping, Score-matching, Reinforcement Learning
Compressor summary: The paper proposes a new task for training robotic hands to assist users in grasping objects, using a hand-object conditioned grasping primitive and a history-conditional policy.
Lalit Manam, Venu Madhav Govindu
https://openreview.net/forum?id=fvm9jVcpBn
Keywords: sensitivity, translation averaging, structure from motion, 3D computer vision
Compressor summary: The paper studies sensitivity in 3D computer vision's translation averaging under uncertainty, provides a criterion for well-conditioned problems, and offers an algorithm to remove ill-conditioned directions, improving 3D reconstructions.
Huaxiaoyue Wang, Gonzalo Gonzalez-Pumariega, Yash Sharma, Sanjiban Choudhury
https://openreview.net/forum?id=ftPoVcm821
Keywords: Large Language Model, Code Generation, Robotics
Compressor summary: The paper introduces Demo2Code, a framework that generates code for robotic tasks from demonstrations using a recursive summarization and synthesis process.
Jiaming Gu, Minchao Jiang, Hongsheng Li, Xiaoyuan Lu, Guangming Zhu, Syed Afaq Ali Shah, Liang Zhang, Mohammed Bennamoun
https://openreview.net/forum?id=fsCcGr8YFR
Keywords: NeRF, UE4, Large-scale scenes, Real-time rendering, Rasterization
Compressor summary: The paper introduces UE4-NeRF, a system for real-time rendering of large-scale scenes using NeRF and Unreal Engine 4, achieving high quality results at 4K resolution with fast frame rates.
Maya Okawa, Ekdeep Singh Lubana, Robert P. Dick, Hidenori Tanaka
https://openreview.net/forum?id=frVo9MzRuU
Keywords: Diffusion model; Science of deep learning; Mechanistic interpretability
Compressor summary: The paragraph discusses a study of how conditional diffusion models generate and reason over novel samples, focusing on the factors that affect their compositional generalization abilities.
Kirankumar Shiragur, Jiaqi Zhang, Caroline Uhler
https://openreview.net/forum?id=frSfSaRGXY
Keywords: Causality, Graphical Models
Compressor summary: The text discusses a new technique called Meek separator, which helps in learning part of the causal graph with fewer interventions and presents two randomized algorithms that achieve logarithmic approximation for subset search and causal matching.
Aravind Gollakota, Parikshit Gopalan, Adam Klivans, Konstantinos Stavropoulos
https://openreview.net/forum?id=frHPeRedHo
Keywords: generalized linear models, single-index models, agnostic learning, pac learning, logistic regression, omnipredictors, multiaccuracy, calibration
Compressor summary: The paper presents a new algorithm to learn Single-Index Models with arbitrary activations and weaker distributional assumptions, based on Omniprediction and Bregman divergences.
Xilie Xu, Jingfeng Zhang, Feng Liu, Masashi Sugiyama, Mohan Kankanhalli
https://openreview.net/forum?id=fpzA8uRA95
Keywords: robust pre-training, adversarial contrastive learning, coreset selection
Compressor summary: The paper proposes a fast method called robustness-aware coreset selection (RCS) for adversarial contrastive learning (ACL), which improves the efficiency of generating adversarial variants and achieves a good balance between speed and robustness.
Jiayi Huang, Han Zhong, Liwei Wang, Lin Yang
https://openreview.net/forum?id=fpHfRD3f4N
Keywords: machine learning, reinforcement learning, linear bandits, heavy-tailed rewards, instance-dependent regret
Compressor summary: The paper proposes efficient algorithms for reinforcement learning with heavy-tailed rewards, achieving near-optimal regret bounds in both linear bandits and linear function approximation settings.
Shuai Zhang, Wenqi Jiang
https://openreview.net/forum?id=fpElyckKkd
Keywords: Geometric representation learning
Compressor summary: The paper proposes a method to automatically choose and combine different geometric spaces for each input data point to improve machine learning tasks without human intervention.
Allan Zhou, Kaien Yang, Kaylee Burns, Adriano Cardace, Yiding Jiang, Samuel Sokota, J Zico Kolter, Chelsea Finn
https://openreview.net/forum?id=fmYmXNPmhv
Keywords: equivariance, permutation, implicit neural representation, generalization
Compressor summary: The text introduces a framework for building permutation equivariant neural functionals that process the weights of other neural networks using NF-Layers with an inductive bias based on symmetry.
Benno Krojer, Elinor Poole-Dayan, Vikram Voleti, Christopher Pal, Siva Reddy
https://openreview.net/forum?id=fmJv8Hj0yo
Keywords: diffusion model, automatic evaluation, vision-and-language, compositionality
Compressor summary: The authors present a method to adapt diffusion-based image generation models for vision-and-language tasks and introduce a new benchmark to evaluate their performance, finding that the proposed approach outperforms CLIP on compositional tasks and reduces bias in Stable Diffusion.
Weiliang Tang, Biqi YANG, Xianzhi Li, Yun-Hui Liu, Pheng-Ann Heng, Chi-Wing Fu
https://openreview.net/forum?id=fljrZsJ2I8
Keywords: 3D Point Cloud Object Detection, Few Shot Learning, Computer Vision, Geometric Prototype
Compressor summary: The paper proposes a VAE-based method (GP-VAE) for few-shot 3D object detection, which enhances feature diversity and distinctiveness using probabilistic latent space and improves performance over existing methods.
Ruiqi Zhang, Andrea Zanette
https://openreview.net/forum?id=fjXTcUUgaC
Keywords: offline RL, online RL, exploration, non-reactive, fine-tuning
Compressor summary: The paper proposes an algorithm that uses an offline dataset to create a single non-reactive exploration policy with guaranteed performance.
Minsu Kim, Federico Berto, Sungsoo Ahn, Jinkyoo Park
https://openreview.net/forum?id=fj0ZeRtUTU
Keywords: Biological sequence design, offline model based optimization, conditional generation, bootstrapping, ensemble
Compressor summary: The BootGen algorithm optimizes biological sequences by training a generator with rank-based weights and then bootstrapping it with self-generated data labeled by a proxy score function, resulting in diverse and high-scoring designs.
Sindy Löwe, Phillip Lippe, Francesco Locatello, Max Welling
https://openreview.net/forum?id=fg7iyNK81W
Keywords: Object Discovery, Object-Centric Representations, Structured Representation Learning
Compressor summary: The paper proposes Rotating Features, a method to represent objects in distributed representations, and demonstrates its applicability to real-world data.
Aran Nayebi, Rishi Rajalingham, Mehrdad Jazayeri, Guangyu Robert Yang
https://openreview.net/forum?id=ffOhY40Nrh
Keywords: neural coding, mental simulation, foundation models, primate frontal cortex
Compressor summary: The authors use various machine learning models to investigate the neural mechanisms behind human and animal understanding of the physical world and future prediction, finding that models trained on dynamic video scenes in latent space best match the data.
Zhuoqun Huang, Neil G Marchant, Keane Lucas, Lujo Bauer, Olga Ohrimenko, Benjamin I. P. Rubinstein
https://openreview.net/forum?id=ffFcRPpnWx
Keywords: certified robustness, randomized smoothing, malware detection, sequence classification, edit distance
Compressor summary: The paper proposes randomized deletion (RS-Del), a randomized smoothing technique for discrete sequence classifiers, which provides certified robustness against edit distance-bounded adversaries and is applied to malware detection.
Zhongli Jiang, Dabao Zhang
https://openreview.net/forum?id=fezV91IJIo
Keywords: causal inference, large graphs, multi-task learning, structural model, directed cyclic graph
Compressor summary: The authors propose a new method for constructing and comparing multiple directed cyclic graphs (DCGs) using a structural model, limited-information, and parallel computation, which can handle algorithmic difficulty, computational burden, and variational causalities.
Thalles Santos Silva, Adín Ramírez Rivera
https://openreview.net/forum?id=fem6BIJkdv
Keywords: representation learning, unsupervised learning, self-supervised learning, computer vision
Compressor summary: CARP is a self-supervised clustering method that learns visual feature representations using random partitions and improves downstream task performance.
Di Liu, Anastasis Stathopoulos, Qilong Zhangli, Yunhe Gao, Dimitris N. Metaxas
https://openreview.net/forum?id=fcYObrixSS
Keywords: 3D computer vision, deep learning
Compressor summary: LEPARD is a framework that uses learning to reconstruct 3D animal shapes from single images by breaking down the object into simpler, robust 3D parts.
Akifumi Imanishi, Zijian Xu, Masayuki Takagi, Sixue Wang, Emilio Castillo
https://openreview.net/forum?id=fbpTObq6TW
Keywords: Recomputation, Gradient checkpointing, Memory reduction, Computational graph optimization
Compressor summary: The paper introduces FastSA, a new simulated annealing-based recomputation algorithm that significantly reduces GPU memory usage and is faster than current methods for optimizing large neural networks.
Vaisakh Shaj, Saleh GHOLAM ZADEH, Ozan Demir, Luiz Ricardo Douat, Gerhard Neumann
https://openreview.net/forum?id=fY7dShbtmo
Keywords: Hierarchical Models; Multi Time Scale Learning; World Models
Compressor summary: The Multi Time Scale State Space (MTS3) model is a probabilistic framework for learning world models that can make accurate long-horizon predictions and uncertainty estimates over multiple time scales.
Che-Ping Tsai, Chih-Kuan Yeh, Pradeep Kumar Ravikumar
https://openreview.net/forum?id=fX64q0SNfL
Keywords: explainable machine learning, sample based explanation, representer point
Compressor summary: The paper introduces a new class of sample-based explanations for machine learning models that use both global and local importance measures and show how existing methods can be considered as special cases.
Zhiyuan Liu, Yaorui Shi, An Zhang, Enzhi Zhang, Kenji Kawaguchi, Xiang Wang, Tat-Seng Chua
https://openreview.net/forum?id=fWLf8DV0fI
Keywords: Molecular Representation Learning, Masked Graph Modeling, Graph Tokenizer
Compressor summary: The paragraph discusses a new approach to learning representations of molecular graphs, focusing on improving the tokenizer and decoder in the masked graph modeling process, and introduces a novel method called SimSGT that outperforms existing methods.
Haochen Li, Rui Zhang, Hantao Yao, Xinkai Song, Yifan Hao, Yongwei Zhao, Ling Li, Yunji Chen
https://openreview.net/forum?id=fW5ZUSVTkv
Keywords: domain adaptation, object detection, prompt tuning
Compressor summary: The paper proposes a method to improve domain adaptive object detection by using a domain-aware detection head with prompt tuning, which generates a dynamic detection head for each domain using learnable tokens and textual descriptions.
Junchi YANG, Xiang Li, Ilyas Fatkhullin, Niao He
https://openreview.net/forum?id=fUZUoSLXw3
Keywords: Nonconvex optimization, Stochastic Gradient Descent, Adaptive methods
Compressor summary: Untuned Stochastic Gradient Descent (SGD) has a fast convergence rate but an exponential dependency on smoothness, while adaptive methods prevent this issue and are better for practical use.
Yoav Kolumbus, Menahem Levy, Noam Nisan
https://openreview.net/forum?id=fU9U7OYxfE
Keywords: Asynchronous Dynamics, Fisher Markets, Proportional Response, Best Response, Game Dynamics, Competitive Equilibrium, Convergence
Compressor summary: In this study, researchers explore how Proportional Response Dynamics (PRD) lead to competitive equilibria in linear Fisher markets with asynchronous bidding and reveal new properties of these markets.
Zheng Zhang, Junxiang Wang, Liang Zhao
https://openreview.net/forum?id=fTyGT5fulj
Keywords: Graph neural networks, Curriculum learning, Graph structure learning
Compressor summary: The paper proposes a novel curriculum learning strategy for graph neural networks that gradually incorporates edges based on their difficulty, improving representation quality and generalization.
Eshaan Nichani, Alex Damian, Jason D. Lee
https://openreview.net/forum?id=fShubymWrc
Keywords: Deep Learning Theory, Feature Learning, Three-Layer Neural Network, Depth Separation, Gradient Descent, Representation Learning
Compressor summary: This paper explores how three-layer neural networks have better feature learning capabilities than two-layer networks, and provides theoretical and empirical evidence for this claim.
Yuzhou Cao, Hussein Mozannar, Lei Feng, Hongxin Wei, Bo An
https://openreview.net/forum?id=fPAAgjISu0
Keywords: Classification, Learning to Defer, Probability Estimation
Compressor summary: The paper proposes a new loss function for learning to defer in machine learning classifiers that avoids uncalibrated estimates and improves classification accuracy.
Blake Bordelon, Cengiz Pehlevan
https://openreview.net/forum?id=fKwG6grp8o
Keywords: Deep Learning Theory, Feature Learning, Dynamics, Ensembles
Compressor summary: The paper studies how finite width affects feature learning neural networks, showing how it influences kernel fluctuations, prediction variance, and online learning in different network depths and training regimes.
Shutong Ding, Jingya Wang, Yali Du, Ye Shi
https://openreview.net/forum?id=fKVEMNmWqU
Keywords: Reinforcement Learning, Hard Constraint, Generalized Reduced Gradient
Compressor summary: The paper proposes a reduced policy optimization (RPO) algorithm that combines reinforcement learning with generalized reduced gradient method to handle general hard constraints in continuous control tasks, and introduces three new benchmarks to test its performance.
Peiyu Yu, Yaxuan Zhu, Sirui Xie, Xiaojian Ma, Ruiqi Gao, Song-Chun Zhu, Ying Nian Wu
https://openreview.net/forum?id=fKQEmHoLb6
Keywords: Energy-Based Model, Denoising Diffusion Probabilistic Model, MCMC
Compressor summary: The paper introduces a diffusion-based amortization method for latent space EBMs to improve sampling quality and generation performance, and provides theoretical and experimental evidence for its effectiveness.
Nikita Gushchin, Alexander Kolesov, Alexander Korotin, Dmitry P. Vetrov, Evgeny Burnaev
https://openreview.net/forum?id=fHyLsfMDIs
Keywords: Optimal transport, Schrödinger Bridge, Entropy regularized OT, Neural Networks, Unpaired Learning
Compressor summary: The authors propose a new neural algorithm for computing optimal transport between distributions based on a dynamic Schrödinger Bridge problem, which is faster and more flexible than previous methods.
Nika Haghtalab, Chara Podimata, Kunhe Yang
https://openreview.net/forum?id=fHsBNNDroC
Keywords: calibration, Stackelberg games, learning in repeated games, strategic agents, best response, strategic classification, Stackelberg Security Games
Compressor summary: The paper proposes Calibrated Stackelberg Games, a generalization of standard Stackelberg Games where an agent does not directly observe the principal's actions but responds to calibrated forecasts, and shows that this model can capture real-life applications and achieve optimal utility.
Hyun Dong Lee, Andrew Warrington, Joshua I Glaser, Scott Linderman
https://openreview.net/forum?id=fFJThJ94rY
Keywords: switching, autoregressive, low-rank tensor, time-series, probabilistic, neural, neuroscience, behavioral, arhmm, slds
Compressor summary: SALT models are a new probabilistic approach to time-series analysis that balance the benefits of ARHMMs and SLDSs by using low-rank factorization to control parameters and capture long-range dependencies.
Haoran He, Chenjia Bai, Kang Xu, Zhuoran Yang, Weinan Zhang, Dong Wang, Bin Zhao, Xuelong Li
https://openreview.net/forum?id=fAdMly4ki5
Keywords: multi-task reinforcement learning, diffusion models, planning, data synthesis
Compressor summary: The paper proposes a diffusion model that can handle multiple tasks with diverse and multimodal data in offline settings, using Transformer backbones and prompt learning for planning and synthesis.
Mufeng Tang, Helen Barron, Rafal Bogacz
https://openreview.net/forum?id=f8zIs2IB6Q
Keywords: Predictive coding, sequential memory, hippocampus
Compressor summary: The authors propose a novel predictive coding-based model for sequential memory called temporal predictive coding, which can accurately memorize and retrieve sequential inputs with a biologically plausible neural implementation and exhibits properties consistent with neuroscience theories.
Xiaotong Luo, Yuan Xie, Yanyun Qu
https://openreview.net/forum?id=f7wFwPJwBe
Keywords: image super-resolution, long-tail distribution, re-sampling, integrated gradient
Compressor summary: The paper proposes a bi-sampling method for single image super-resolution, which balances the training data and improves the model's ability to reconstruct hard regions.
Justin Domke, Robert M. Gower, Guillaume Garrigos
https://openreview.net/forum?id=f71xXsoG1v
Keywords: optimization, variational inference
Compressor summary: The text discusses how to provide theoretical guarantees for black-box variational inference by improving gradient estimators with unusual noise bounds and a composite non-smooth objective.
Katie Z Luo, Zhenzhen Liu, Xiangyu Chen, Yurong You, Sagie Benaim, Cheng Perng Phoo, Mark Campbell, Wen Sun, Bharath Hariharan, Kilian Q Weinberger
https://openreview.net/forum?id=f6rQJ83ycb
Keywords: Self Driving, Self-Supervised Object Discovery, Reward Ranked Finetuning
Compressor summary: The paper proposes a machine learning method for detecting objects from LiDAR points using heuristics as human feedback, leading to faster and more accurate results than previous methods.
Shuchen Xue, Mingyang Yi, Weijian Luo, Shifeng Zhang, Jiacheng Sun, Zhenguo Li, Zhi-Ming Ma
https://openreview.net/forum?id=f6a9XVFYIo
Keywords: Diffusion Model Sampler, Multi-step SDE Solver
Compressor summary: The paper proposes SA-Solver, an efficient stochastic method for solving diffusion SDEs and generating high-quality data using a variance-controlled approach.
Dvir Samuel, Rami Ben-Ari, Nir Darshan, Haggai Maron, Gal Chechik
https://openreview.net/forum?id=f56xMRb7Vt
Keywords: diffusion models, few-shot learning, long-tail learning
Compressor summary: The paper proposes a new method for interpolating between text-to-image diffusion seeds that considers their norm values, which improves the generation of rare concepts and performance on few-shot and long-tail learning tasks.
Noam Wies, Yoav Levine, Amnon Shashua
https://openreview.net/forum?id=f3JNQd7CHM
Keywords: in-context, PAC, language models, foundation models, LLMs
Compressor summary: This paper proposes a theoretical framework for in-context learning, showing how it can efficiently identify and learn tasks from input examples without modifying large language models' weights.
Thomas Steinke, Milad Nasr, Matthew Jagielski
https://openreview.net/forum?id=f38EY21lBw
Keywords: Differential privacy, membership inference attacks, privacy auditing
Compressor summary: The proposed auditing scheme for differentially private machine learning systems uses parallelism and statistical generalization to assess privacy with minimal assumptions and fewer training runs than existing methods.
Milan Ganai, Zheng Gong, Chenning Yu, Sylvia Lee Herbert, Sicun Gao
https://openreview.net/forum?id=f2U4HCY8bg
Keywords: Constraints, Safety, Hamilton Jacobi Reachability, Deep Reinforcement Learning, Robotics
Compressor summary: The paragraph introduces a new framework called RESPO that optimizes for rewards and safety in reinforcement learning and shows its effectiveness on various environments.
Xingbo Du, Chonghua Wang, Ruizhe Zhong, Junchi Yan
https://openreview.net/forum?id=f0Jj3C3Pnp
Keywords: global routing, generative models
Compressor summary: The paper proposes a novel hub-based approach for global routing in VLSI systems, using deep generative models for hub generation and an actor-critic model for pin-hub connection, which improves efficiency and quality over existing methods.
Anthony Fuller, Koreen Millard, James R Green
https://openreview.net/forum?id=ezqI5WgGvY
Keywords: Remote Sensing, Earth Observation, Self-supervised learning, Multimodal
Compressor summary: CROMA is a framework that uses self-supervised learning to learn rich unimodal and multimodal representations from spatially aligned multimodal data in remote sensing, outperforming the current SoTA multispectral model on various benchmarks.
Ruijie Zheng, Xiyao Wang, Yanchao Sun, Shuang Ma, Jieyu Zhao, Huazhe Xu, Hal Daumé III, Furong Huang
https://openreview.net/forum?id=ezCsMOy1w9
Keywords: Deep Reinforcement Learning, Visual Reinforcement Learning, Online Visual RL, Offline Visual RL, Action Representation
Compressor summary: $\texttt{TACO}$ is a temporal contrastive learning method that improves sample efficiency in reinforcement learning by learning concurrent state and action representations.
Alane Suhr, Yoav Artzi
https://openreview.net/forum?id=ez6Cb0ZGzG
Keywords: continual learning, interaction, instruction following, user feedback, natural language processing, language grounding, situated interaction, collaboration
Compressor summary: The paper presents an approach that trains an agent from user feedback in natural language and binary form, using contextual bandit learning to improve instruction execution accuracy and show robustness and equivalence to supervised data.
Alessio Mazzetto, Eli Upfal
https://openreview.net/forum?id=exiXmAfuDK
Keywords: statistical learning, learning theory, machine learning, supervised learning, non-stationary, transfer learning, distribution drift
Compressor summary: The paper presents a method for learning from data with unknown distribution changes, without needing information about the change magnitude, and shows its effectiveness in binary classification and linear regression tasks.
Yi-Kai Zhang, Ting-Ji Huang, Yao-Xiang Ding, De-Chuan Zhan, Han-Jia Ye
https://openreview.net/forum?id=exg62lfHrB
Keywords: Pre-trained Model Ranking, Transfer Learning
Compressor summary: Model Spider is a method that uses vector representations to efficiently select and enrich Pre-Trained Models for various tasks, including visual models and Large Language Models.
Yingjie Liu, Xuan Liu, Hui Yu, XUAN TANG, Xian Wei
https://openreview.net/forum?id=exPzwOhBgx
Keywords: dictionary learning, attention, transformer, computer vision, point cloud
Compressor summary: The authors propose a novel attention module called Dic-Attn, which uses dictionary learning to decompose and reconstruct input data, enabling efficient visual attention construction for computer vision tasks.
Nicholas Roberts, Xintong Li, Dyah Adila, Sonia Cromp, Tzu-Heng Huang, Jitian Zhao, Frederic Sala
https://openreview.net/forum?id=exGOXqxR0L
Keywords: structured prediction, learning on graphs, partially observed label spaces, high cardinality label spaces
Compressor summary: The proposed technique Loki improves the performance of machine learning models by using external or self-derived metrics to adapt to new classes without additional training, achieving significant improvements over existing methods.
Zekun Qi, Muzhou Yu, Runpei Dong, Kaisheng Ma
https://openreview.net/forum?id=etd0ebzGOG
Keywords: Text to shape generation, 3D shape generation, Efficient inference, Representation Learning
Compressor summary: The proposed Voxel-Point Progressive Representation (VPP) method improves 3D generation quality and efficiency by combining structured voxel representation and sparse point representation, enabling diverse and high-fidelity 3D shape creation for various downstream tasks.
Mingzhou Liu, Xinwei Sun, Lingjing Hu, Yizhou Wang
https://openreview.net/forum?id=etYk6TeO2q
Keywords: causal discovery, time series, subsampling, proxy variables
Compressor summary: The paper proposes a new method to infer causality from time series data with missing measurements, using proxies for hidden variables.
Banghua Zhu, Mingyu Ding, Philip Jacobson, Ming Wu, Wei Zhan, Michael Jordan, Jiantao Jiao
https://openreview.net/forum?id=esy7pkZmKn
Keywords: semi-supervised learning, self-training, auto-labeling, self-labeling, doubly robust
Compressor summary: The paper proposes a semi-supervised learning algorithm called doubly-robust self-training that adapts to the quality of pseudo-labels and outperforms the standard self-training method in image classification and 3D object detection tasks.
Meena Jagadeesan, Nikhil Garg, Jacob Steinhardt
https://openreview.net/forum?id=eqyhjLG5Nr
Keywords: content creator incentives, Nash equilibria, specialization, economic aspects of recommender systems
Compressor summary: The paragraph discusses how algorithmic recommender systems affect producer behavior and content diversity, and presents a model that analyzes the conditions for specialization and its impact on profitability in such systems.
Yongqiang Chen, Wei Huang, Kaiwen Zhou, Yatao Bian, Bo Han, James Cheng
https://openreview.net/forum?id=eozEoAtjG8
Keywords: Out-of-Distribution Generalization, Feature Learning, Invariant Risk Minimization
Compressor summary: The paragraph discusses how ERM can learn both spurious and invariant features, but tends to favor spurious ones if they are stronger, making it harder for models to generalize to out-of-distribution data. It proposes Feature Augmented Training (FeAT) to improve feature learning for OOD tasks.
Haim Kaplan, Yishay Mansour, Shay Moran, Kobbi Nissim, Uri Stemmer
https://openreview.net/forum?id=eoDNaH3pfB
Keywords: Differential privacy, online learning
Compressor summary: The authors propose an interactive variant of joint differential privacy that can handle online processes more effectively than traditional forms, allowing for private online classification with less error.
Jun Yin, Chaozhuo Li, Hao Yan, Jianxun Lian, Senzhang Wang
https://openreview.net/forum?id=enfx8HM4Rp
Keywords: Intrinsic Interpretability, Graph Neural Networks, Pre-training and Fine-tuning
Compressor summary: The Pre-training Interpretable Graph Neural Network (π-GNN) is a new model that learns universal patterns of graph structure and local interactions to provide transparent predictions by identifying influential parts of the input graph.
Ibrahim Alabdulmohsin, Xiaohua Zhai, Alexander Kolesnikov, Lucas Beyer
https://openreview.net/forum?id=en4LGxpd9E
Keywords: Vision transformer, scaling laws, compute-optimal model design, vision
Compressor summary: The paper presents SoViT, a shape-optimized vision transformer that achieves competitive results with much larger models while using less compute resources, and evaluates its performance across various tasks.
Muhammad Jehanzeb Mirza, Leonid Karlinsky, Wei Lin, Horst Possegger, Mateusz Kozinski, Rogerio Feris, Horst Bischof
https://openreview.net/forum?id=elPtHcfjpH
Keywords: VL Models
Compressor summary: The paper proposes a label-free method to improve zero-shot visual classification using an unlabeled image collection and auto-generated texts describing the categories, achieving significant performance improvements and outperforming some few-shot prompting methods.
Haggai Agmon, Yoram Burak
https://openreview.net/forum?id=ekMLUoC2sq
Keywords: Theoretical Neuroscience, Computational Neuroscience, Recurrent Neural Networks, Attractor models
Compressor summary: The paragraph discusses how synaptic weight adjustment can reduce interference in recurrent neural networks that store multiple continuous variables in working memory, leading to better memory storage.
Victor Letzelter, Mathieu Fontaine, Mickael Chen, Patrick Perez, Slim Essid, Gaël Richard
https://openreview.net/forum?id=eibTaY6qGI
Keywords: Multiple Choice Learning, Audio processing.
Compressor summary: Resilient Multiple Choice Learning (rMCL) is a new method for estimating conditional distributions in regression settings with multiple targets per input, using a novel scoring scheme based on Voronoi tessellations to preserve diversity in predictions.
Xiaolong Zou, Zhikun Chu, Qinghai Guo, Jie Cheng, Bo Ho, Si Wu, Yuanyuan Mi
https://openreview.net/forum?id=eeeqORvJbf
Keywords: temporal sequence processing, temporal order structure, tree-structured attractor
Compressor summary: The paragraph discusses how recurrent neural circuits can learn to represent the abstract order structure of temporal sequences, allowing for flexible and robust processing of temporal sequences, and suggests that understanding this mechanism could help develop brain-inspired algorithms.
Zhiao Huang, Feng Chen, Yewen Pu, Chunru Lin, Hao Su, Chuang Gan
https://openreview.net/forum?id=ecRaDicXxw
Keywords: Differentiable physics; Soft body manipulation
Compressor summary: DiffVL is a method that lets non-experts create soft-body manipulation tasks using vision and natural language, making it easier for differential physics solvers to find solutions.
Fabian Paischer, Thomas Adler, Markus Hofmarcher, Sepp Hochreiter
https://openreview.net/forum?id=ebMPmx5mr7
Keywords: Reinforcement Learning, Language Models, History Compression, Partial Observability, Foundation Models, Interpretability, Explainable AI
Compressor summary: The authors propose a novel memory mechanism for reinforcement learning agents in partially observable environments that uses CLIP to associate visual inputs with language tokens, which are then fed to a pretrained language model to provide a human-readable representation of the past, improving interpretability and performance on tasks requiring memory.
Jun Wu, Lisa Ainsworth, Andrew Leakey, Haixun Wang, Jingrui He
https://openreview.net/forum?id=eZbqD9BoXe
Keywords: graph learning, transfer learning, Gaussian process
Compressor summary: GraphGP is a framework for adaptively transferring knowledge across graphs with different assumptions, addressing challenges in transferable graph learning by using a novel graph structure-aware neural network and showing its effectiveness through experiments.
Yifan Zhang, Qijian Zhang, Junhui Hou, Yixuan Yuan, Guoliang Xing
https://openreview.net/forum?id=eYCGrGdKf3
Keywords: 3D object detection, 3D point cloud
Compressor summary: The paper introduces UPIDet, a new cross-modal 3D object detector for autonomous vehicles that leverages image domain information to improve LiDAR-based detectors' performance and achieves top rank in the KITTI benchmark.
Ran Ran, Nuo Xu, Tao Liu, Wei Wang, Gang Quan, Wujie Wen
https://openreview.net/forum?id=eXubleMT0q
Keywords: Cryptographic inference, Graph Convolutional Network, Parallel Packing
Compressor summary: The authors propose Penguin, a ciphertext packing technique for enhancing the efficiency of homomorphic encryption-based graph convolutional network inference on encrypted data.
Van-Anh Nguyen, Long Tung Vuong, Hoang Phan, Thanh-Toan Do, Dinh Phung, Trung Le
https://openreview.net/forum?id=eX6xDto3Ed
Keywords: Bayesian, sharpness-aware, posterior
Compressor summary: The paper proposes a method to improve the generalization ability of Bayesian Neural Networks by making the posterior distribution aware of the model's sharpness/flatness.
Luhuan Wu, Brian L. Trippe, Christian A Naesseth, David Blei, John Patrick Cunningham
https://openreview.net/forum?id=eWKqr1zcRv
Keywords: diffusion models; conditional sampling; sequential monte carlo methods; generative models; protein design
Compressor summary: The Twisted Diffusion Sampler (TDS) is a sequential Monte Carlo algorithm that improves conditional generation by incorporating heuristic approximations without compromising exactness, achieving better results in simulation and protein design tasks.
Zangwei Zheng, Xiaozhe Ren, Fuzhao Xue, Yang Luo, Xin Jiang, Yang You
https://openreview.net/forum?id=eW233GDOpm
Keywords: large language models, inference optimization, batch processing
Compressor summary: The paper proposes an efficient inference pipeline for large language models that groups queries with similar response lengths into micro-batches, improving throughput without sacrificing effectiveness.
Kartik Chandra, Tony Chen, Tzu-Mao Li, Jonathan Ragan-Kelley, Joshua B. Tenenbaum
https://openreview.net/forum?id=eVrmcOvJV4
Keywords: cognitive science, cogsci, inverse planning, Bayesian inference, theory of mind, Monte Carlo, inverse reinforcement learning
Compressor summary: The paper proposes a Monte Carlo algorithm that models how humans infer complex sequences of events from static comic book panels or other dynamic scenes, by connecting its inference problem to Monte Carlo path tracing in computer graphics.
Zhongcong Xu, Jianfeng Zhang, Jun Hao Liew, Jiashi Feng, Mike Zheng Shou
https://openreview.net/forum?id=eUf0CaS5AP
Keywords: Human Avatar, 3D-aware GAN
Compressor summary: XAGen is a 3D generative model for human avatars that can control facial expressions, jaw poses, hand poses, and other expressive attributes, achieving high realism and diversity.
Ziye Ma, Javad Lavaei, Somayeh Sojoudi
https://openreview.net/forum?id=eU6P4aUdCA
Keywords: non-convex optimization, low-rank matrix optimization, matrix sensing, implicit bias, tensor, over-parametrization
Compressor summary: The paper investigates how gradient descent helps generalization in machine learning models by inducing implicit regularization for tensor optimization in the lifted matrix sensing framework.
Jongseok Park, Kyungmin Bin, Gibum Park, Sangtae Ha, Kyunghan Lee
https://openreview.net/forum?id=eTp4RetK74
Keywords: Deep Neural Network, Deep Learning, Parallel Execution Algorithm, Parallelization, Deep Learning Parallelism, Dynamic, Asynchronous, Scheduling, Dynamic Scheduling, Dynamic Execution, tile, tiling, dataflow, dataflow graph, tile-based dataflow graph, opportunistic parallelism
Compressor summary: ASPEN is a novel parallel computation solution for DNNs that removes operator barriers, exposes parallelism across operators, and achieves high resource utilization and memory reuse, outperforming TorchScript and TVM by up to 4.3$\times$.
Yuanshi Liu, Cong Fang, Tong Zhang
https://openreview.net/forum?id=eTMHsUp3Ii
Keywords: Langevin, Dimension dependence, Acceleration
Compressor summary: The paper presents a double randomization method for fast high-dimensional sampling of log-concave distributions with composite structures, achieving dimension-independent convergence guarantees and outperforming previous results by a $d^{1/3}$ factor.
Veit David Wild, Sahra Ghalebikesabi, Dino Sejdinovic, Jeremias Knoblauch
https://openreview.net/forum?id=eTHawKFT4h
Keywords: Wasserstein gradient flow, generalised variational inference, deep ensembles, Bayesian deep learning, variational Bayes
Compressor summary: The paper presents a unified theory linking Bayesian, variational Bayesian, and ensemble methods for uncertainty quantification in deep learning by reformulating the non-convex optimization problem as a convex one using Wasserstein gradient flows.
Bhaskar Mukhoty, Velibor Bojkovic, William de Vazelhes, Xiaohan Zhao, Giulia De Masi, Huan Xiong, Bin Gu
https://openreview.net/forum?id=eTF3VDH2b6
Keywords: Spiking Neural Network, Zeroth Order, Surrogate Gradient
Compressor summary: The paragraph describes a new technique for training spiking neural networks that uses zeroth-order methods, surrogate approximations of the Heaviside function, and efficient backpropagation to achieve better accuracy and faster training.
Ronald Xie, Kuan Pang, Sai W Chung, Catia Perciani, Sonya MacParland, BO WANG, Gary Bader
https://openreview.net/forum?id=eT1tMdAUoc
Keywords: BLEEP, Histology, H&E, Gene Expression Prediction, Spatial Transcriptomics, Contrastive Learning
Compressor summary: BLEEP is a framework that predicts gene expression from histology images, providing insights into tissue architecture and potentially reducing the time and cost of diagnosis and research.
Daniel Halpern, Rachel Li, Ariel D. Procaccia
https://openreview.net/forum?id=eT1QOsssRB
Keywords: social choice, strategyproof, voting
Compressor summary: The paper explores how relaxing strategyproofness criteria for selecting winners in ranked voting scenarios can still lead to impossibility theorems, and argues that plurality rule is a promising choice due to its strategyproofness under certain belief models.
Urte Adomaityte, Gabriele Sicuro, Pierpaolo Vivo
https://openreview.net/forum?id=eR7PrfJe9o
Keywords: Classification, Gaussian Mixture Model, Superstatistics, Empirical Risk Minimization, Replica theory, Power-law distribution
Compressor summary: The paper investigates how a mixture of two clouds of data points with random variances learns in high dimensions under different assumptions and distributions, testing Gaussian universality claims and analysing generalisation and separability.
Zhanke Zhou, Jiangchao Yao, Jiaxu Liu, Xiawei Guo, quanming yao, LI He, Liang Wang, Bo Zheng, Bo Han
https://openreview.net/forum?id=ePkLqJh5kw
Keywords: Robust link prediction, Edge noise
Compressor summary: The text proposes a method called Robust Graph Information Bottleneck (RGIB) to improve link prediction on graphs by extracting reliable supervision signals and avoiding representation collapse under edge noise.
Nabeel Seedat, Jonathan Crabbé, Zhaozhi Qian, Mihaela van der Schaar
https://openreview.net/forum?id=eP6cDDwBNC
Keywords: data-centric AI, data characterization, data quality
Compressor summary: TRIAGE is a novel framework for characterizing training data in regression tasks, using conformal predictive distributions to score samples as under-, over-, or well-estimated by the model, and enabling data sculpting/filtering and dataset selection improvements.
Haoting Zhang, Jinghai He, Rhonda Righter, Zuo-Jun Shen, Zeyu Zheng
https://openreview.net/forum?id=eNhW9UnlGG
Keywords: contextual bandit, Gaussian process, neural network
Compressor summary: The paper proposes a neural network-accompanied Gaussian process (NN-AGP) model for solving contextual decision-making problems with high accuracy and uncertainty quantification, and proves its theoretical guarantees and shows its empirical performance.
Daesol Cho, Seungjae Lee, H. Jin Kim
https://openreview.net/forum?id=eMR57voMz1
Keywords: Curriculum learning, Out-of-distribution disagreement, Underspecification, Outcome-directed RL
Compressor summary: The paper proposes a new curriculum RL method called D2C that uses diversification of classifiers and bipartite matching to explore and conquer uninformed search problems without access to domain knowledge or environment characteristics.
Leon Klein, Andreas Krämer, Frank Noe
https://openreview.net/forum?id=eLH2NFOO1B
Keywords: Normalizing Flows, Flow Matching, Equivariance, Boltzmann Generators, Molecular Dynamics, Optimal Transport
Compressor summary: Equivariant flow matching is a new training method for normalizing flows in physics that uses optimal transport and exploits physical symmetries to improve sampling efficiency and scalability.
Keji He, Chenyang Si, Zhihe Lu, Yan Huang, Liang Wang, Xinchao Wang
https://openreview.net/forum?id=eKFrXWb0sT
Keywords: Vision-and-Language Navigation; High-Frequency; Data Augmentation
Compressor summary: The paper proposes a new technique to improve visual-textual matching in vision-and-language navigation tasks by using high-frequency information and data augmentation, leading to better performance with simple and efficient models.
Jiayuan Mao, Tomás Lozano-Pérez, Joshua B. Tenenbaum, Leslie Pack Kaelbling
https://openreview.net/forum?id=eJZ5vJEaaa
Keywords: Planning, Relational Neural Network, Circuit Complexity
Compressor summary: The paper analyzes how relational neural networks can be used to learn goal-conditioned policies for planning problems, depending on the number of objects and planning horizon.
Jiakang Yuan, Bo Zhang, Xiangchao Yan, Botian Shi, Tao Chen, Yikang LI, Yu Qiao
https://openreview.net/forum?id=eIFZtkshgH
Keywords: 3D Object Detection, 3D Pre-training, Autonomous Driving
Compressor summary: The paper introduces a semi-supervised point cloud pre-training approach for autonomous driving that improves generalizability across various tasks and benchmarks.
Zhaolong Du, Shasha Mao, Yimeng Zhang, Shuiping Gou, Licheng Jiao, Lin Xiong
https://openreview.net/forum?id=eGoE9CVRPc
Keywords: Video Analysis, Multiple-Instance Learning, Representation learning
Compressor summary: The authors propose RGMIL, a method for video analysis with limited annotations, which uses Regressor-Guided Pooling (RGP) to produce discriminative instance-level representations and achieve near ideal performance compared to supervised models.
Zhiyu Lin, Yifei Gao, Yunfan Yang, Jitao Sang
https://openreview.net/forum?id=eE5L1RkxW0
Keywords: visual models, robustness, frequency domain, long-tailed distribution
Compressor summary: The paper proposes a new strategy to improve visual models' robustness and accuracy by addressing the under-fitting behavior on high-frequency components (HFC) in images, which is caused by limited information content in HFC.
Guy Hacohen, Daphna Weinshall
https://openreview.net/forum?id=eDDZh8C4W4
Keywords: Deep Active learning, Low budget, High budget, Deep learning
Compressor summary: The paper proposes a practical method for selecting the best active learning query strategy based on the problem characteristics and available budget, and demonstrates its effectiveness in computer vision tasks.
Junfeng Fang, Wei Liu, Yuan Gao, Zemin Liu, An Zhang, Xiang Wang, Xiangnan He
https://openreview.net/forum?id=eD534mPhAg
Keywords: Post-hoc Explainability, Explanation Evaluation, Graph Neural Network, Robustness Analysis
Compressor summary: The paper proposes OAR and SimOAR, novel evaluation metrics for explaining graph neural networks that address the out-of-distribution issue and improve credibility by measuring adversarial robustness.
Jerry Yao-Chieh Hu, Donglin Yang, Dennis Wu, Chenwei Xu, Bo-Yu Chen, Han Liu
https://openreview.net/forum?id=eCgWNU2Imw
Keywords: Hopfield Models; Modern Hopfield Networks; Sparse Attention; Memory Networks
Compressor summary: The paper introduces a sparse version of the modern Hopfield model with improved theoretical properties and empirical performance.
Antonín Vobecký, Oriane Siméoni, David Hurych, Spyros Gidaris, Andrei Bursuc, Patrick Perez, Josef Sivic
https://openreview.net/forum?id=eBXM62SqKY
Keywords: open-vocabulary segmentation, voxel occupancy prediction, semantic segmentation, autonomous driving, language-image alignment
Compressor summary: The paper presents a new model for predicting 3D semantic occupancy maps from 2D images using self-supervised learning without 3D annotations.
Zhiqi Bu, Yu-Xiang Wang, Sheng Zha, George Karypis
https://openreview.net/forum?id=e8i7OaPj0q
Keywords: deep learning, differential privacy, optimization, hyper-parameter tuning
Compressor summary: Automatic clipping is a technique that simplifies differential privacy training for deep learning models by eliminating the need to tune a clipping threshold and achieving similar or better performance than existing methods.
Naoki Egami, Musashi Hinck, Brandon M. Stewart, Hanying Wei
https://openreview.net/forum?id=e8RZwixcE4
Keywords: Computational Social Science, Large Language Models, Statistical Inference, Causal Inference
Compressor summary: The authors propose a new algorithm (DSL) that uses imperfect annotations from large language models for computational social science research, ensuring unbiased and accurate statistical analysis by combining them with a smaller number of high-quality labels.
Lazar Atanackovic, Alexander Tong, BO WANG, Leo J Lee, Yoshua Bengio, Jason Hartford
https://openreview.net/forum?id=e7MK5Vq44Q
Keywords: Bayesian Structure Learning, Generative Flow Networks, Single-cell, Dynamical Systems
Compressor summary: The paper proposes a novel approach for inferring gene regulatory networks using RNA velocity techniques and Generative Flow Networks, addressing both the challenges of cyclicity and uncertainty in the data.
Sida Wang
https://openreview.net/forum?id=e5srDjF9l7
Keywords: co-occurrences, unsupervised word translation, bilingual lexicon induction, robust statistics, unsupervised machine translation
Compressor summary: The authors develop a method for unsupervised word translation that uses high-dimensional signals instead of low-dimensional word vectors, achieving better results with less data and computation.
Michael Kleinman, Alessandro Achille, Stefano Soatto, Jonathan Kao
https://openreview.net/forum?id=e4XidX6AHd
Keywords: Common Information, Gacs-Korner, Variational Autoencoder
Compressor summary: The paragraph introduces a concept called common information that separates shared and unique aspects of two random variables, which can be estimated from data using variational autoencoders and measured against ground-truth factors.
Zhijing Jin, Yuen Chen, Felix Leeb, Luigi Gresele, Ojasv Kamal, Zhiheng LYU, Kevin Blin, Fernando Gonzalez Adauto, Max Kleiman-Weiner, Mrinmaya Sachan, Bernhard Schölkopf
https://openreview.net/forum?id=e2wtjx0Yqu
Keywords: Large Language Models, Causal Reasoning, Causal Inference, Benchmark Dataset, Natural Language Processing
Compressor summary: The paper introduces a new natural language processing task and dataset to assess whether large language models can perform causal inference based on formal rules.
Beepul Bharti, Paul Yi, Jeremias Sulam
https://openreview.net/forum?id=e2aCgjtjMR
Keywords: fairness, sensitive attributes, equalized odds, missing data, proxies
Compressor summary: This paper proposes a new method for controlling fairness in machine learning models without access to sensitive attributes by using post-processing corrections and provides theoretical guarantees and experimental validation.
Ishaan Gulrajani, Tatsunori Hashimoto
https://openreview.net/forum?id=e2MCL6hObn
Keywords: diffusion, language, model
Compressor summary: The authors develop and release a diffusion-based language model (Plaid 1B) that surpasses GPT-2 124M in likelihood on benchmarks and produces coherent samples.
Kai Tan, Pierre C Bellec
https://openreview.net/forum?id=e1oe8F2tjV
Keywords: High-dimensional statistics, statistical inference, multi-class classification, asymptotic normality, multinomial logistic regression
Compressor summary: The paper explores how to accurately estimate and test the importance of features in high-dimensional multinomial logistic models.
Wang Qinsi, Jinghan Ke, Zhi Liang, Sihai Zhang
https://openreview.net/forum?id=e1l4ZYprQH
Keywords: Neural Architecture Search
Compressor summary: MathNAS is a novel NAS framework that predicts network performance training-free by calculating the performances of all possible building blocks first, reducing the complexity of evaluation and achieving state-of-the-art results on large-scale datasets.
Tuan Dinh, Jinman Zhao, Samson Tan, Renato Negrinho, Leonard Lausen, Sheng Zha, George Karypis
https://openreview.net/forum?id=e1WgjvFGWp
Keywords: language model of code; code completion; language model; software engineering; machine learning for code
Compressor summary: The paragraph discusses code completion with buggy code contexts and introduces two datasets to study the problem, showing that current Code-LLMs struggle with this scenario and there is room for improvement in mitigating bugs.
Ruian Wang, Zixiong Wang, Yunxiao Zhang, Shuangmin Chen, Shiqing Xin, Changhe Tu, Wenping Wang
https://openreview.net/forum?id=e0tt2G8hqf
Keywords: implicit neural representation, signed distance function, shape operator
Compressor summary: The paper introduces a method to learn the Signed Distance Function (SDF) from point clouds without normals, which improves gradient accuracy and reduces ghost geometry in surface reconstruction.
Daogao Liu, Arun Ganesh, Sewoong Oh, Abhradeep Guha Thakurta
https://openreview.net/forum?id=e0pRF9tOtm
Keywords: Differential Privacy, Non-convex optimization, Stationary points, Exponential Mechanism
Compressor summary: The paper proposes a novel framework for non-convex optimization under differential privacy constraint that uses two types of gradient oracles to improve accuracy and identifies second-order stationary points.
Woojin Cho, Kookjin Lee, Donsub Rim, Noseong Park
https://openreview.net/forum?id=dzqKAM2sKa
Keywords: Scientific machine learning, Physics-informed neural networks, Meta learning, Hypernetworks
Compressor summary: The study suggests using lightweight low-rank physics-informed neural networks with meta-learning to solve partial differential equations for various input parameters quickly and effectively.
Yuzhe Lu, Yilong Qin, Runtian Zhai, Andrew Shen, Ketong Chen, Zhenlin Wang, Soheil Kolouri, Simon Stepputtis, Joseph Campbell, Katia P. Sycara
https://openreview.net/forum?id=dz5X8hnfJc
Keywords: Distribution Shift, OOD Error Prediction, Optimal Transport, Deep Learning
Compressor summary: The paper proposes a new method for estimating model performance on out-of-distribution data using optimal transport theory and an empirically-motivated variant that improves accuracy and robustness.
Yushan Zhang, Johan Edstedt, Bastian Wandt, Per-Erik Forssen, Maria Magnusson, Michael Felsberg
https://openreview.net/forum?id=dybrsuNAB9
Keywords: Scene flow, point clouds, transformers
Compressor summary: The paper proposes a simple one-shot global matching method for scene flow estimation from point clouds, which achieves state-of-the-art results on various benchmarks and outperforms previous methods.
Anurag Ajay, Seungwook Han, Yilun Du, Shuang Li, Abhi Gupta, Tommi S. Jaakkola, Joshua B. Tenenbaum, Leslie Pack Kaelbling, Akash Srivastava, Pulkit Agrawal
https://openreview.net/forum?id=dyXNh5HLq3
Keywords: Foundation Models, Composition, Hierarchical Planning
Compressor summary: The HiP model combines language, vision, and action models to create symbolic plans for long-horizon tasks involving table-top manipulation.
Owen Lewis Howell, David Klee, Ondrej Biza, Linfeng Zhao, Robin Walters
https://openreview.net/forum?id=dxVN2fZjx6
Keywords: Equivarient Machine Learning, Pose Prediction, Computer Vision
Compressor summary: The text discusses how to design neural networks for 3D object recognition from 2D images, considering constraints and consistency properties, and achieving state-of-the-art results on some tasks.
Zhibin Duan, Lv Zhiyi, Chaojie Wang, Bo Chen, Bo An, Mingyuan Zhou
https://openreview.net/forum?id=dxPcdEeQk9
Keywords: Generative Model, Memory-augmented Generative Model
Compressor summary: The authors propose a variational structured memory module (VSM) to enhance existing generative models for few-shot generation by mimicking human memory mechanisms.
Zachary Teed, Lahav Lipson, Jia Deng
https://openreview.net/forum?id=dwfHbm8g66
Keywords: SLAM, Simultaneous Localization and Mapping, Visual Odometry, Structure from motion, SfM
Compressor summary: Deep Patch Visual Odometry (DPVO) is a new system that uses sparse patch-based matching instead of dense flow for monocular visual odometry, achieving better accuracy and efficiency than previous methods.
Max Torop, Aria Masoomi, Davin Hill, Kivanc Kose, Stratis Ioannidis, Jennifer Dy
https://openreview.net/forum?id=dwIeEhbaD0
Keywords: Interpretability, Feature Interactions, Stein's Lemma
Compressor summary: SmoothHess is a method for estimating feature interactions in ReLU neural networks by using Stein's Lemma and efficient sampling, without needing network modifications or changing the architecture.
Dayou Yu, Weishi Shi, Qi Yu
https://openreview.net/forum?id=du0hvEpgj8
Keywords: active learning, active testing
Compressor summary: The paper proposes a novel active testing while learning (ATL) framework that integrates active learning with active testing, reducing data annotation costs and improving model performance using an "active feedback" mechanism.
Shengpu Tang, Jenna Wiens
https://openreview.net/forum?id=dsH244r9fA
Keywords: healthcare, reinforcement learning, offline RL, off-policy evaluation, counterfactuals
Compressor summary: The paper proposes a semi-offline evaluation framework for reinforcement learning that uses counterfactual trajectories annotated by humans to reduce bias and variance in off-policy evaluation.
Peter Nickl, Lu Xu, Dharmesh Tailor, Thomas Möllenhoff, Mohammad Emtiyaz Khan
https://openreview.net/forum?id=dqS1GuoG2V
Keywords: model interpretability, model understanding, bayesian learning, robustness, adaptive learning
Compressor summary: The Memory-Perturbation Equation (MPE) simplifies the study of model sensitivity to training data perturbations, using Bayesian principles, and shows its usefulness in predicting generalization to unseen test data.
Tahseen Rabbani, Marco Bornstein, Furong Huang
https://openreview.net/forum?id=dpdbbN7AKr
Keywords: distributed learning, locality-sensitive hashing, recommender systems, compression
Compressor summary: The paper proposes a new family of hash functions for locality-sensitive hashing that enables privacy, personalization, and memory efficiency for pruning dense hidden layers in large-scale recommender networks.
Yingying Fan, Yu Wu, Bo Du, Yutian Lin
https://openreview.net/forum?id=doWqIXcRlq
Keywords: Weakly-Supervised Audio-Visual Video Parsing, Language Guided Segment-Level Label Denoising, Dynamic Re-weighting
Compressor summary: The paper proposes a language-based method for weakly-supervised audio-visual video parsing that tackles segment-level label noise and outperforms existing approaches.
Sayan Bhattacharya, Martin Costa, Silvio Lattanzi, Nikos Parotsidis
https://openreview.net/forum?id=dnGEPkmnzO
Keywords: clustering, k-median, k-means, dynamic algorithms, amortized analysis
Compressor summary: The paper presents a fast and approximate dynamic algorithm for the k-median and k-means problems on metric spaces, compares it to previous methods, and provides a lower bound analysis.
Jonathan Schmidt, Philipp Hennig, Jörg Nick, Filip Tronarp
https://openreview.net/forum?id=dnB71DMyDD
Keywords: Gaussian, filtering, smoothing, bayesian, state-space models, dynamic-low-rank, high-dimensional, spatio-temporal, Gaussian processes, regression, low rank, state estimation
Compressor summary: The paper proposes a new low-rank Gaussian filtering and smoothing method that is faster and more accurate than existing ensemble-based methods for high-dimensional dynamical systems.
Mateusz Olko, Michał Zając, Aleksandra Nowak, Nino Scherrer, Yashas Annadani, Stefan Bauer, Łukasz Kuciński, Piotr Miłoś
https://openreview.net/forum?id=dmD63sv0TZ
Keywords: causal discovery, experimental design, active learning, neural networks
Compressor summary: The paper proposes a method called GIT that uses gradient estimators to minimize the number of required experiments for acquiring interventional data and inferring causal structures from data.
Tao Lin, Yiling Chen
https://openreview.net/forum?id=dlDFakG6kJ
Keywords: information aggregation, sample complexity, distribution learning, Bayesian forecast aggregation
Compressor summary: A Bayesian model for forecast aggregation requires exponentially more samples when experts' signals depend on each other, but only linearly more samples if their signals are independent given the event outcome.
Zijie Li, Dule Shu, Amir Barati Farimani
https://openreview.net/forum?id=djyn8Q0anK
Keywords: Efficient attention, Neural PDE solver
Compressor summary: FactFormer is a new model that uses an axial factorized kernel integral to improve the efficiency and stability of applying Transformer models to problems with many grid points, such as simulating 2D and 3D fluid dynamics.
Qizhang Li, Yiwen Guo, Wangmeng Zuo, Hao Chen
https://openreview.net/forum?id=dikH9tdPi2
Keywords: adversarial examples, black-box attack, adversarial transferability
Compressor summary: The ILPD method improves intermediate-level adversarial attacks by crafting effective and strong perturbations in one stage of optimization, leading to better performance against various victim models on ImageNet and CIFAR-10.
Vivek Bharadwaj, Osman Asif Malik, Riley Murray, Laura Grigori, Aydin Buluc, James Demmel
https://openreview.net/forum?id=deaHiTb6Cu
Keywords: Tensor Decomposition, Leverage Scores, Randomized Linear Algebra, Sketching, Khatri-Rao Product, Sparse Tensors
Compressor summary: The paper introduces a fast sampler for the Khatri-Rao product that can handle large matrices and outperforms existing methods in linear least-squares problems.
Matthew Farrugia-Roberts
https://openreview.net/forum?id=ddKCg3OhGw
Keywords: theory, neural network theory, structural redundancy, functional equivalence, functional equivalence class, partial identifiability, parameter canonicalisation, parameter space, piecewise-linear, connectivity
Compressor summary: The paper characterizes unit redundancies and reducible functional equivalence classes in a neural network architecture using an algorithm and shows that these classes are path-connected sets with a small maximum diameter.
Eli Chien, Wei-Ning Chen, Chao Pan, Pan Li, Ayfer Ozgur, Olgica Milenkovic
https://openreview.net/forum?id=dd3KNayGFz
Keywords: Graph Neural Networks, Differential Privacy, Multigranular Topology Protection
Compressor summary: The authors propose a new framework called Graph Differential Privacy (GDP) for graph learning that ensures provably private model parameters and predictions, and compare it to existing Differentially Private Decoupled Graph Convolutions (DPDGCs).
Fiona Lippert, Bart Kranstauber, E. Emiel van Loon, Patrick Forré
https://openreview.net/forum?id=dcw7qRUuD8
Keywords: probabilistic inference, graphical models, spatiotemporal dynamical systems, state-space models
Compressor summary: The paper proposes a computationally efficient method for estimating and learning state variables in graph-structured models with unknown dynamics and limited data by combining deep learning and Gaussian Markov random fields.
Berken Utku Demirel, Christian Holz
https://openreview.net/forum?id=dbVRDk2wt7
Keywords: Contrastive learning, Time-series, Augmentation
Compressor summary: The paper proposes a new data augmentation method for time-series tasks that connects samples to find order in the latent space and improves performance on three tasks, including heart rate estimation, human activity recognition, and cardiovascular disease detection.
Ian Osband, Zheng Wen, Seyed Mohammad Asghari, Vikranth Dwaracherla, Morteza Ibrahimi, Xiuyuan Lu, Benjamin Van Roy
https://openreview.net/forum?id=dZqcC1qCmB
Keywords: Uncertainty, Deep Learning, Neural Networks
Compressor summary: The epinet is an architecture that improves the uncertainty estimation of conventional neural networks by supplementing them with modest computation, enabling them to outperform large ensembles.
Shanyun Gao, Raghavendra Addanki, Tong Yu, Ryan A. Rossi, Murat Kocaoglu
https://openreview.net/forum?id=dYeUvLUxBQ
Keywords: time-series causal discovery, constraint-based causal discovery
Compressor summary: The paper proposes a method for discovering causal relations from non-stationary time series using a structural causal model that accounts for periodic changes, and tests it on simulated and real-world data.
Carl Hvarfner, Erik Orm Hellsten, Frank Hutter, Luigi Nardi
https://openreview.net/forum?id=dX9MjUtP1A
Keywords: Bayesian Optimization, Bayesian Active Learning, Gaussian Processes
Compressor summary: The authors propose two methods to improve Gaussian processes for Bayesian optimization and active learning by better learning hyperparameters and incorporating self-correction.
Yubin Shi, Yixuan Chen, Mingzhi Dong, Xiaochen Yang, Dongsheng Li, Yujiang Wang, Robert P. Dick, Qin Lv, Yingying Zhao, Fan Yang, Tun Lu, Ning Gu, Li Shang
https://openreview.net/forum?id=dWDEBW2raJ
Keywords: Modular Adaptive Training, Efficient Training, Over-parameterized Model, Neural Tangent Kernel.
Compressor summary: This paper introduces modular neural tangent kernel (mNTK) to study learning dynamics of over-parameterized models, proposes Modular Adaptive Training (MAT) to improve efficiency and performance by selectively updating modules based on their mNTK eigenvalues, and shows that MAT reduces computational costs and outperforms baselines.
Kevin Ellis
https://openreview.net/forum?id=dVnhdm9MIg
Keywords: Cognitive science, Bayesian, Language model, Induction, Psychology, Reasoning
Compressor summary: The model uses natural language hypotheses and Bayesian reasoning to learn a broad range of human-like concepts efficiently.
Zeming Chen, Gail Weiss, Eric Mitchell, Asli Celikyilmaz, Antoine Bosselut
https://openreview.net/forum?id=dUAcAtCuKk
Keywords: natural language processing, multi-hop reasoning, knowledge memorisation
Compressor summary: RECKONING teaches language models to reason more robustly by encoding contextual knowledge into their parameters before answering questions, improving performance and generalization.
TUNG QUOC LE, Rémi Gribonval, Elisa Riccietti
https://openreview.net/forum?id=dTj5tH94xv
Keywords: Topology, best approximation property, closedness, function space, sparse neural networks
Compressor summary: The paper investigates the existence of an optimal solution for sparse ReLU neural networks and shows that it is not always guaranteed, while providing conditions for its existence.
Xiaohan Lin, Liyuan Li, Boxin Shi, Tiejun Huang, Yuanyuan Mi, Si Wu
https://openreview.net/forum?id=dSRyKIYRnP
Keywords: Continuous attractor neural network; Excitation inhibition balance; Brain-inspired algorithms; Object tracking;
Compressor summary: The study explores how a neural circuit can have both attractor states and irregular firings by using two sets of synapses with different strengths and speeds, leading to improved performance in a tracking problem.
Qing Li, Huifang Feng, Kanle Shi, Yue Gao, Yi Fang, Yu-Shen Liu, Zhizhong Han
https://openreview.net/forum?id=dR6p49RYLq
Keywords: Point Clouds, Normal Estimation, Neural Gradient
Compressor summary: The authors propose a new neural network approach to estimate oriented normals from point clouds without using ground truth normals, achieving robust and accurate results on benchmarks.
Akifumi Wachi, Wataru Hashimoto, Xun Shen, Kazumune Hashimoto
https://openreview.net/forum?id=dQLsvKNwZC
Keywords: Reinforcement Learning, Safety Exploration
Compressor summary: MASE is a meta-algorithm for safe exploration in reinforcement learning that combines an unconstrained RL algorithm with an uncertainty quantifier to guarantee safety while penalizing unsafe actions.
Amit Dhurandhar, Karthikeyan Natesan Ramamurthy, Kartik Ahuja, Vijay Arya
https://openreview.net/forum?id=dOxm4FnMFu
Keywords: Explainable AI, Game theory, Invariance
Compressor summary: The paper proposes a novel method for explaining black-box models using game theory and invariant risk minimization, which can produce high fidelity, stable, and unidirectional explanations with simple and efficient training.
Robin San Roman, Yossi Adi, Antoine Deleforge, Romain Serizel, Gabriel Synnaeve, Alexandre Défossez
https://openreview.net/forum?id=dOanKg3jKS
Keywords: diffusion, audio, compression
Compressor summary: The proposed high-fidelity multi-band diffusion model generates any audio type from low-bitrate discrete representations, outperforming other methods in perceptual quality at equal bit rate.
Jing Zhang, Chi Zhang, Wenjia Wang, Bingyi Jing
https://openreview.net/forum?id=dLmDPVv19z
Keywords: Offline Reinforcement Learning, GAN, Flow Model, Policy Control
Compressor summary: CPED is a new method for offline RL that uses a flow-GAN model to estimate behavior policy density, allowing it to safely explore and achieve better performance than existing methods.
Jann Spiess, Guido Imbens, Amar Venugopal
https://openreview.net/forum?id=dL0GM9Wwtq
Keywords: Double descent, interpolating regression, synthetic control, causal inference
Compressor summary: The authors study how over-parameterized models in causal inference, such as high-dimensional linear regression and synthetic control with many control units, can perform better than simpler ones, and provide a unified theoretical perspective on their improvement.
Peter Yongho Kim, Junbeom Kwon, Sunghwan Joo, Sangyoon Bae, Donggyu Lee, Yoonho Jung, Shinjae Yoo, Jiook Cha, Taesup Moon
https://openreview.net/forum?id=dKeWh6EzBB
Keywords: fMRI, Swin Transformer, 4D, neuroscience
Compressor summary: SwiFT is a Swin Transformer architecture that learns brain dynamics directly from fMRI volumes using a 4D window multi-head self-attention mechanism and absolute positional embeddings, achieving better performance than existing models on predicting sex, age, and cognitive intelligence.
Zhangsihao Yang, Mengwei Ren, Kaize Ding, Guido Gerig, Yalin Wang
https://openreview.net/forum?id=dK0Ew3kkVf
Keywords: Keypoints, Medical image, self-supervised learning, transformer, segmentation
Compressor summary: Keypoint-augmented fusion layer improves medical image segmentation by enhancing CNN features with long-range spatial self-attention, and incorporates global and local self-supervised pretraining for better representations.
Amir Feder, Yoav Wald, Claudia Shi, Suchi Saria, David Blei
https://openreview.net/forum?id=dJZ3MvDw86
Keywords: Counterfactually Augmented Data, Invariant Learning, Out-of-distribution Generalization, Clinical NLP
Compressor summary: The authors propose using counterfactual data augmentation to improve the robustness of text classifiers in cases where the label is spuriously correlated with an attribute, and show improved OOD accuracy on medical diagnosis tasks.
Kyurae Kim, Jisu Oh, Kaiwen Wu, Yian Ma, Jacob R. Gardner
https://openreview.net/forum?id=dHQ2av9NzO
Keywords: black-box variational inference, stochastic gradient descent, Bayesian inference, variational inference, probabilistic machine learning, Bayesian machine learning, variational Bayes
Compressor summary: The paper shows that using a specific algorithm called proximal stochastic gradient descent can improve the performance of black-box variational inference, a method for Bayesian inference, compared to other common approaches.
Haoxuan Qu, Xiaofei Hui, Yujun Cai, Jun Liu
https://openreview.net/forum?id=dHF3Im8Aic
Keywords: Deep learning, Open-set object recognition, Large models, Training-free
Compressor summary: The paper proposes a method called Large Model Collaboration (LMC) that uses multiple pre-trained models to improve open-set object recognition without additional training and by extracting implicit knowledge from the models.
Erik Miehling, Rahul Nair, Elizabeth M. Daly, Karthikeyan Natesan Ramamurthy, Robert Nelson Redmond
https://openreview.net/forum?id=dFtpRphNb3
Keywords: fairness, cookies, recommender systems
Compressor summary: The text discusses how user consent on cookie sharing affects the accuracy of personalized ads and recommender systems, and suggests new notions of fairness to address privacy concerns.
Jiarui Hu, Mao Mao, Hujun Bao, Guofeng Zhang, Zhaopeng Cui
https://openreview.net/forum?id=dFSeZm6dTC
Keywords: Collaborative SLAM; Neural Point Field; Keyframe-based SLAM; Pose Graph Optimization
Compressor summary: The paper introduces a collaborative implicit neural SLAM system using RGB-D image sequences, with a novel 3D scene representation and learning strategy that improves accuracy and cooperation.
Junwoo Cho, Seungtae Nam, Hyunmo Yang, Seok-Bae Yun, Youngjoon Hong, Eunbyung Park
https://openreview.net/forum?id=dEySGIcDnI
Keywords: partial differential equations, scientific machine learning, physics-informed neural networks, fluid dynamics
Compressor summary: The proposed method, SPINN, improves the performance of PINNs by reducing computational costs and increasing accuracy in solving multi-dimensional PDEs using a per-axis basis network architecture and forward-mode automatic differentiation.
Zhongyi Cai, Ye Shi, Wei Huang, Jingya Wang
https://openreview.net/forum?id=dEDdRWunxU
Keywords: Federated Learning, Data Heterogeneity, Model Cooperation, Mutual Learning, Knowledge Transfer
Compressor summary: Fed-CO$_2$ is a framework for distributed learning that handles data heterogeneity issues like label and feature skew using cooperation between online and offline models, improving performance over existing algorithms.
Massil HIHAT, Stéphane Gaïffas, Guillaume Garrigos, Simon Bussy
https://openreview.net/forum?id=dDk6URGRXP
Keywords: online convex optimization, inventory control, newsvendor, online learning, regret analysis
Compressor summary: The paper presents an online algorithm, MaxCOSD, for inventory control problems with general demands, losses, and dynamics, and requires non-degeneracy assumptions on the demand process for learning.
Jonathan Lee, Annie Xie, Aldo Pacchiano, Yash Chandak, Chelsea Finn, Ofir Nachum, Emma Brunskill
https://openreview.net/forum?id=dCYBAGQXLo
Keywords: decision making, reinforcement learning, in-context learning, bandits, transformers, offline reinforcement learning, exploration, reinforcement learning theory
Compressor summary: The paper introduces Decision-Pretrained Transformer (DPT), a method for supervised pretraining of transformers to solve decision-making problems using in-context learning, and demonstrates its surprising capabilities in solving RL problems, generalizing to new tasks, and adapting strategies.
Chiyu Ma, Brandon Zhao, Chaofan Chen, Cynthia Rudin
https://openreview.net/forum?id=dCAk9VlegR
Keywords: deep learning, interpretability, prototype-based neural network, case-based reasoning
Compressor summary: ProtoConcepts is a method for interpretable image classification using multiple image patches to represent prototypical concepts, improving visual explanations and understanding.
Mengcheng Lan, Xinjiang Wang, Yiping Ke, Jiaxing Xu, Litong Feng, Wayne Zhang
https://openreview.net/forum?id=dB4lvScPIj
Keywords: unsupervised semantic segmentation; self-supervised learning; smoothness prior
Compressor summary: The paper proposes a new method called SmooSeg for unsupervised semantic segmentation that uses smoothness prior to simplify the task, improving performance on three datasets.
Junyu Zhang, Daochang Liu, Shichao Zhang, Chang Xu
https://openreview.net/forum?id=dAbGv5Jz5U
Keywords: diffusion models, contrastive loss, discretization error, contrastive sampling chain
Compressor summary: The text explains a method that combines contrastive loss and score matching to reduce discretization error in diffusion models, improving image quality and sample efficiency.
Zhaocheng Zhu, Xinyu Yuan, Mikhail Galkin, Sophie Xhonneux, Ming Zhang, Maxime Gazeau, Jian Tang
https://openreview.net/forum?id=dAJrxQz1lk
Keywords: Knowledge Graph Reasoning, Path-based Methods, Scalability, A* Algorithm
Compressor summary: A\*Net is a scalable path-based method for knowledge graph reasoning that learns a priority function to select important nodes and edges, achieving competitive performance with less computation.
Eldar Kurtic, Elias Frantar, Dan Alistarh
https://openreview.net/forum?id=d8j3lsBWpV
Keywords: LLMs, pruning, compression, inference
Compressor summary: ZipLM is a novel compression method for large language models that achieves high accuracy and speedup, outperforming prior techniques in various settings.
Yunhao Ge, Hong-Xing Yu, Cheng Zhao, Yuliang Guo, Xinyu Huang, Liu Ren, Laurent Itti, Jiajun Wu
https://openreview.net/forum?id=d86B6Mdweq
Keywords: 3D Copy-Paste, Object Insertion, Monocular 3D Object Detection
Compressor summary: The paper presents a method to insert virtual objects into real scenes for monocular 3D object detection, ensuring physical plausibility in locations, appearances, and shadows.
Siyuan Sun, Hongyang Gao
https://openreview.net/forum?id=d85pPNBHLt
Keywords: Few shot learning, Meta Learning
Compressor summary: Meta-AdaM is a meta-learned adaptive optimizer that uses weight-update history and momentum for faster few-shot learning, outperforming existing methods on benchmark datasets.
Hai Zhang, Hang Yu, Junqiao Zhao, Di Zhang, Chang Huang, Hongtu Zhou, Xiao Zhang, Chen Ye
https://openreview.net/forum?id=d7a5TpePV7
Keywords: model-based reinforcement learning, model shift, model bias, fine-tuning, performance difference bound
Compressor summary: The paper proposes USB-PO, an algorithm that adapts model updates to improve MBRL performance while accounting for model shift and bias, achieving state-of-the-art results on multiple benchmark tasks.
Yutong Xie, Mingze Yuan, Bin Dong, Quanzheng Li
https://openreview.net/forum?id=d6LShzSTOP
Keywords: unsupervised learning, image denoising, score function
Compressor summary: The paper introduces a new unsupervised learning approach for single image denoising that can handle various noise models and outperforms current methods in complex cases.
Gongfan Fang, Xinyin Ma, Xinchao Wang
https://openreview.net/forum?id=d4f40zJJIS
Keywords: Diffusion Model, Network Pruning, Model Compression, Efficient Deep Learning
Compressor summary: Diff-Pruning is an efficient compression method that reduces FLOPs and training time for diffusion models, preserving their generative behavior.
Peiran Dong, Song Guo, Junxiao Wang, Bingjie WANG, Jiewei Zhang, Ziming Liu
https://openreview.net/forum?id=d4X0QWS2Ln
Keywords: Diffusion models, test-time refusal, concept negation, safety in generative models
Compressor summary: The paper introduces ProtoRe, a framework to improve concept negation for generating refusals in image synthesis by using CLIP's knowledge to identify negative concepts and purify outputs.
XinRan Xie, Man-Jie Yuan, Xuetong Bai, Wei Gao, Zhi-Hua Zhou
https://openreview.net/forum?id=d47iuwOt3j
Keywords: classification, random forests, privacy-preserving machine learng, data encrytion
Compressor summary: This paper proposes a new encryption method that preserves the Gini impurity of data for random forests, using modified binary search trees and homomorphic encryption.
Meyer Scetbon, Michal Klein, Giovanni Palla, marco cuturi
https://openreview.net/forum?id=d2WsCmoITF
Keywords: Optimal Transport, Unbalanced
Compressor summary: The paper proposes new methods to improve optimal transport in machine learning by addressing its computational and modelling limitations, and applies them to solve real-world spatial transcriptomics problems.
Aodong Li, Chen Qiu, Marius Kloft, Padhraic Smyth, Maja Rudolph, Stephan Mandt
https://openreview.net/forum?id=d1wjMBYbP1
Keywords: deep anomaly detection, zero-shot learning, batch normalization
Compressor summary: The paper proposes ACR, a method that helps deep anomaly detectors adapt to new data distributions without training data, achieving good zero-shot performance on tabular and image data.
Aseem Baranwal, Kimon Fountoulakis, Aukosh Jagannath
https://openreview.net/forum?id=d1knqWjmNt
Keywords: graph neural networks, message passing, bayesian inference, node classification, contextual stochastic block model
Compressor summary: The paper proposes an optimal node classification method for sparse feature-decorated graphs using a message-passing graph neural network and compares its performance with existing methods on a statistical model.
Mingyu Xu, Zheng Lian, Bin Liu, Jianhua Tao
https://openreview.net/forum?id=d0VItRE2ZH
Keywords: Out-of-distribution Detection
Compressor summary: The paper introduces a new technique called Variational Rectified Activation (VRA) that improves out-of-distribution detection in machine learning models by simulating suppression and amplification operations using piecewise functions, and shows its effectiveness on multiple benchmark datasets.
Lucas Gnecco Heredia, Muni Sreenivas Pydi, Laurent Meunier, benjamin negrevergne, Yann Chevaleyre
https://openreview.net/forum?id=d0IEd3VgBh
Keywords: adversarial attacks, robustness, adversarial, attacks, deep learning, randomization, randomized ensembles
Compressor summary: The paper investigates how randomization affects adversarial robustness in classifiers and shows conditions for randomized ensembles to outperform deterministic ones, as well as examples of deterministic classifiers that outperform probabilistic ones.
Xu Yang, Yongliang Wu, Mingzhuo Yang, Haokun Chen, Xin Geng
https://openreview.net/forum?id=czwZnNf60r
Keywords: Image Caption; Few-shot Prompt; Vision Language Model;
Compressor summary: The authors explore different strategies for configuring image-text pairs in vision-language tasks, finding that they can significantly improve performance in image captioning compared to random sampling.
Max W. Y. Lam, Qiao Tian, Tang Li, Zongyu Yin, Siyuan Feng, Ming Tu, Yuliang Ji, Rui Xia, Mingbo Ma, Xuchen Song, Jitong Chen, Yuping Wang, Yuxuan Wang
https://openreview.net/forum?id=cxazQGSsQa
Keywords: Music Generation, Language Model, Diffusion Model, MusicLM
Compressor summary: MeLoDy is a new music generation model that uses less computing power than MusicLM while maintaining high quality and adaptability.
Xiaoxiao Sun, Nidham Gazagnadou, Vivek Sharma, Lingjuan Lyu, Hongdong Li, Liang Zheng
https://openreview.net/forum?id=cx9a4Xvb3l
Keywords: Privacy Assessment, Reconstructed Images, Evaluation Metrics, Human Perception
Compressor summary: The paper studies how well hand-crafted metrics reflect human judgement of model privacy leakage in reconstructed images and proposes a learning-based measure called SemSim to better evaluate semantic similarity between original and reconstructed images.
Rui Yang, Lin Song, Yanwei Li, Sijie Zhao, Yixiao Ge, Xiu Li, Ying Shan
https://openreview.net/forum?id=cwjh8lqmOL
Keywords: multimodality, foundation models, tool usage
Compressor summary: The paper proposes GPT4Tools, a method to enable open-source LLMs to use multi-modal tools efficiently and effectively using instruction-following datasets and Low-Rank Adaptation optimization.
Vinod Raman, UNIQUE SUBEDI, Ambuj Tewari
https://openreview.net/forum?id=cwBeRBe9hq
Keywords: Multilabel Ranking, PAC Learning, Online Learning
Compressor summary: This paper studies how easy it is to learn multilabel ranking problems with relevance-score feedback in different settings and introduces two classes of ranking losses based on learnability.
Huy Nguyen, TrungTin Nguyen, Nhat Ho
https://openreview.net/forum?id=cto6jIIbMZ
Keywords: Mixture of Experts, Maximum Likelihood Estimation, Voronoi Loss Function, Algebraic Geometry.
Compressor summary: The text discusses challenges in estimating parameters for softmax gating Gaussian mixture of experts models and proposes new Voronoi loss functions to address them, connecting convergence rates to solvability problems of polynomial equations.
Hanqi Yan, Lingjing Kong, Lin Gui, Yuejie Chi, Eric Xing, Yulan He, Kun Zhang
https://openreview.net/forum?id=cslnCXE9XA
Keywords: Causal Representation Learning, Identifiability, Counterfactual Generation, Latent variable models, Disentanglement.
Compressor summary: The paper proposes a method to generate counterfactual images or texts by identifying content and style latent variables that vary across domains, using theoretical insights from relative sparsity of influences, and achieving state-of-the-art performance in unsupervised style transfer tasks.
Minshuo Chen, Yu Bai, H. Vincent Poor, Mengdi Wang
https://openreview.net/forum?id=csdEeUn0ve
Keywords: Delayed and missing observations, MDPs, efficient regret bounds
Compressor summary: The paper studies how reinforcement learning can be efficient in control systems when agents have delayed or missing state observations and provides regret bounds and comparisons with full observability.
Zhuoman Liu, Bo Yang, Yan Luximon, Ajay Kumar, Jinxi Li
https://openreview.net/forum?id=crZlhMnfeO
Keywords: implicit shape representations, multi-view consistency, novel view synthesis
Compressor summary: The paper introduces RayDF, a fast and accurate method for representing 3D shapes using ray-based neural functions that maintain multi-view geometry consistency, outperforming existing coordinate- and ray-based approaches.
Abhishek Sinha, Ativ Joshi, Rajarshi Bhattacharjee, Cameron N Musco, Mohammad Hajiesmaili
https://openreview.net/forum?id=crNAh1EZKo
Keywords: Online Learning, Bandit Algorithms, Learning Theory
Compressor summary: The paper proposes an efficient online resource allocation policy that achieves sublinear approximate regret in a fairness problem with a surprising phase transition phenomenon at the critical exponent $\alpha=\frac{1}{2}.$
Zhi Li, Yifan Liu, Yin Zhang
https://openreview.net/forum?id=cr99foBDPV
Keywords: data augmentation, cross-modal
Compressor summary: Back-Modality is a new way to improve data augmentation by transforming data between different modalities and then back again, making it easier to use various techniques for different types of data.
Haris Aziz, Evi Micha, Nisarg Shah
https://openreview.net/forum?id=cpUuSV8kRw
Keywords: peer review; group fairness; core; stable
Compressor summary: The paper proposes a group fairness concept called the core for large AI conferences, which ensures that every community is treated fairly in the peer review process, and develops an efficient algorithm to find optimal assignments using real data from CVPR and ICLR.
Bernardo Fichera, Viacheslav Borovitskiy, Andreas Krause, Aude Billard
https://openreview.net/forum?id=co4p15OMoc
Keywords: Gaussian process, manifolds, manifold learning, uncertainty, regression, graph Laplacian
Compressor summary: The paper proposes a new Gaussian process regression method that can infer implicit structure from data without explicit input of a manifold, potentially improving its performance in high-dimensional settings.
Weikang BIAN, Zhaoyang Huang, Xiaoyu Shi, Yitong Dong, Yijin Li, Hongsheng Li
https://openreview.net/forum?id=cnpkzQZaLU
Keywords: Optical Flow; Video Correspondence; Computer Vision;
Compressor summary: The paper introduces Context-PIPs, a framework that improves point trajectory accuracy in videos by using spatial context features from the source and target points.
Saghar Adler, Vijay Subramanian
https://openreview.net/forum?id=cm53OBkctM
Keywords: Thompson Sampling, Reinforcement Learning, Queueing theory
Compressor summary: The paper proposes a Bayesian algorithm using Thompson sampling for optimal control of countably-infinite state-space Markov Decision Processes, and shows its applicability in two queueing models.
Arjun Somayazulu, Changan Chen, Kristen Grauman
https://openreview.net/forum?id=clKbFMt29V
Keywords: Audio-Visual learning, Visual Acoustic Matching
Compressor summary: The paper proposes a self-supervised method for acoustic matching that uses only target scene images and audio to re-synthesize audio in a different environment, outperforming existing methods.
Ruiying Lu, YuJie Wu, Long Tian, Dongsheng Wang, Bo Chen, Xiyang Liu, Ruimin Hu
https://openreview.net/forum?id=clJTNssgn6
Keywords: Anomaly Detection, Transformer, Vector Quantization, Unsupervised Anomaly Detection
Compressor summary: The paper proposes a hierarchical vector quantized prototype-oriented Transformer for unsupervised image anomaly detection that preserves normal patterns as discrete prototypes and uses optimal transport to evaluate abnormal scores.
Samuel Hurault, Ulugbek Kamilov, Arthur Leclaire, Nicolas Papadakis
https://openreview.net/forum?id=clCELP8zFb
Keywords: Plug-and-Play, Poisson Inverse Problems, Bregman distance, Proximal Gradient Descent, nonconvex and nonsmooth optimization, Poisson inverse problems
Compressor summary: The paper proposes a new method called Bregman Score Denoiser for solving Poisson inverse problems using Plug-and-Play algorithms that better capture the problem's smoothness properties.
Emile van Krieken, Thiviyan Thanapalasingam, Jakub M. Tomczak, Frank Van Harmelen, Annette Ten Teije
https://openreview.net/forum?id=chlTA9Cegc
Keywords: Neurosymbolic Learning, Generative Modeling, Approximate Inference
Compressor summary: A-NeSI is a new framework for probabilistic neurosymbolic learning that allows scalable approximate inference, symbolic explanations of predictions, and guarantees logical constraint satisfaction while maintaining performance.
Shirui Chen, Linxing Preston Jiang, Rajesh P. N. Rao, Eric Todd SheaBrown
https://openreview.net/forum?id=ch1buUOGa3
Keywords: neural coding, probabilistic sampling, neural dynamics, recurrent neural network
Compressor summary: The paper explores how recurrent neural circuits can sample from arbitrary probability distributions, which could improve Bayesian brain models.
Xiaolong Wang, Runsen Xu, Zhuofan Cui, Zeyu Wan, Yu Zhang
https://openreview.net/forum?id=cgiP4cMBP9
Keywords: Fine-Grained Cross-View Geo-Localization, Homography Estimation
Compressor summary: The paper presents a new method for accurately aligning ground images with satellite images using a differentiable spherical transform and a correlation-aware homography estimator, achieving sub-pixel resolution and meter-level GPS accuracy, and outperforming existing techniques in localization error reduction.
Xiaohui Chen, Yinkai Wang, Yuanqi Du, Soha Hassoun, Liping Liu
https://openreview.net/forum?id=cezKbXsT3V
Keywords: Transformer, Self-supervised Learning, Normalization
Compressor summary: The paper proposes a simple modification to self-supervised training methods for transformers by using separate normalization layers for tokens and the [CLS] symbol, which improves downstream task performance in various domains.
Dung Thuy Nguyen, Tuan Minh Nguyen, Anh Tuan Tran, Khoa D Doan, KOK SENG WONG
https://openreview.net/forum?id=cemEOP8YoC
Keywords: Backdoor Attacks, Federated Learning, Durability, Imperceptibility, Stealthiness
Compressor summary: The paper proposes a novel backdoor attack method for federated learning that is stealthy, efficient, and durable, and bypasses existing defenses.
Dor Tsur, Ziv Goldfeld, Kristjan Greenewald
https://openreview.net/forum?id=ce9B2x3zQa
Keywords: CCA, dimensionality reduction, information theory, mutual information, neural estimation, slicing
Compressor summary: The paper introduces a new method called max-sliced mutual information (mSMI) that combines the strengths of CCA and Shannon's mutual information for quantifying dependencies in high-dimensional data.
Lei Zhang, Ji-Fu Li, Wei Wang
https://openreview.net/forum?id=ce59j806df
Keywords: Domain Generalization, Semi-Supervised Learning, Out-of-Distribution Detection, Deep Learning
Compressor summary: The CWAEE method is a semi-supervised domain generalization approach that uses class-wise adaptive thresholds and Fourier Transformation to handle unknown classes in unlabeled data, achieving better performance on real-world datasets.
Aoyang Qin, Feng Gao, Qing Li, Song-Chun Zhu, Sirui Xie
https://openreview.net/forum?id=cdlmsnQkZ9
Keywords: Sequential Decision Making, Generative Model, Imitation Learning
Compressor summary: The paper proposes a deep generative model for imitation learning in non-Markovian settings, using energy-based policy prior and maximum likelihood estimation for inference and planning.
Haoran Ye, Jiarui Wang, Zhiguang Cao, Helan Liang, Yong Li
https://openreview.net/forum?id=cd5D1DD923
Keywords: Neural Combinatorial Optimization, Ant Colony Optimization, Evolutionary algorithm, Meta-heuristic, Deep reinforcement learning, Learned heuristic measure, Neural local search, Generalization
Compressor summary: DeepACO is a framework that uses deep reinforcement learning to improve Ant Colony Optimization algorithms without manual design of heuristics, achieving better results on various Combinatorial Optimization Problems.
Jianhong Bai, Zuozhu Liu, Hualiang Wang, Ruizhe Chen, Lianrui Mu, Xiaomeng Li, Joey Tianyi Zhou, YANG FENG, Jian Wu, Haoji Hu
https://openreview.net/forum?id=cczH4Xl7Zo
Keywords: Generalized Category Discovery, Open-world Recognition, Long-tail Learning, Contrastive Learning
Compressor summary: The paper introduces a novel framework, BaCon, that tackles the realistic task of distribution-agnostic generalized category discovery (DA-GCD) in computer vision, addressing data imbalance and open-endedness with contrastive learning and pseudo-labeling.
Farnood Salehi, Tunc Ozan Aydin, André Gaillard, Guglielmo Camporese, Yuxuan Wang
https://openreview.net/forum?id=cay8LnKSro
Keywords: sin activation, image prediction, image resampling, monte-carlo denoising, knowledge distillation
Compressor summary: The authors propose MetaSin activation, a novel ensemble function that improves performance and reliability in training sin networks for visual data tasks like denoising and resampling.
Jiawei Fan, Chao Li, Xiaolong Liu, Meina Song, Anbang Yao
https://openreview.net/forum?id=caUhYUVsLl
Keywords: Knowledge distillation, semantic segmentation, contrastive learning
Compressor summary: Af-DCD is a new contrastive distillation method for semantic segmentation that uses masked feature mimicking and clever feature partitions to train efficient student models without data augmentation or memory buffer.
Farzad Pourkamali, Nicolas Macris
https://openreview.net/forum?id=ca2QmdOlIh
Keywords: Matrix factorization, Bayesian inference, rotation invariant estimators, random matrix theory, spherical integrals, replica method
Compressor summary: The paper proposes Rotation Invariant Estimators for matrix factorization with additive noise, which are conjectured to be optimal in high dimensions based on a Bayesian approach using random matrix theory, spherical integrals, and the replica method.
Revan MacQueen, James R. Wright
https://openreview.net/forum?id=cZVBRg59eb
Keywords: Algorithmic Game Theory, Self-Play, Regret-Minimization, Multi-agent RL, Multiplayer Games, General-Sum Games
Compressor summary: The paragraph discusses how self-play can generate strategies with performance guarantees in some multiplayer games if they approximate constant-sum subgames with bounded subgame stability.
Cuong Tran, Ferdinando Fioretto
https://openreview.net/forum?id=cZS5X3PLOR
Keywords: Privacy; data minimization
Compressor summary: The paper explores whether using only a small fraction of features for inference in high-stakes domains is sufficient to achieve accurate predictions, and proposes an efficient algorithm to determine which attributes are needed for each individual.
Yishi Xu, Jianqiao Sun, Yudi Su, Xinyang Liu, Zhibin Duan, Bo Chen, Mingyuan Zhou
https://openreview.net/forum?id=cYkSt7jqlx
Keywords: Few-shot generative model; topic modeling;
Compressor summary: The paragraph describes a method for adapting word embeddings based on contextual information to improve topic modeling in low-resourced tasks and achieve better performance than existing methods.
Seungyong Moon, Junyoung Yeom, Bumsoo Park, Hyun Oh Song
https://openreview.net/forum?id=cUuXVaMmmv
Keywords: reinforcement learning, hierarchical reinforcement learning, contrastive learning, procedurally generated environments
Compressor summary: The paper proposes an improved model-free approach using PPO and contrastive learning to discover hierarchical achievements in procedurally generated environments effectively.
Isay Katsman, Eric Ming Chen, Sidhanth Holalkere, Anna Asch, Aaron Lou, Ser-Nam Lim, Christopher De Sa
https://openreview.net/forum?id=cRzt1umRNx
Keywords: neural network, riemannian, manifold, resnet
Compressor summary: The authors extend ResNets to work on general Riemannian manifolds, achieving better performance than existing manifold neural networks for graphs with hierarchical structures or manifold-valued data.
David Xing Wu, Anant Sahai
https://openreview.net/forum?id=cRGINXQWem
Keywords: overparameterized, multiclass, classification, theory, generalization, interpolation, bi-level, Gaussian model
Compressor summary: This paper analyzes the generalization performance of a linear model for multiclass and multi-label classification with growing data dimensions, resolves a conjecture from a previous paper, and introduces a new variant of the Hanson-Wright inequality for these problems.
Xiao Zang, Miao Yin, Jinqi Xiao, Saman Zonouz, Bo Yuan
https://openreview.net/forum?id=cQdc9Dyk4i
Keywords: graph neural network, deep learning
Compressor summary: GraphMP is a neural motion planner that improves graph information extraction and search processing, achieving better path quality and planning speed than existing methods in various environments.
Laixi Shi, Gen Li, Yuting Wei, Yuxin Chen, Matthieu Geist, Yuejie Chi
https://openreview.net/forum?id=cOQH8YO255
Keywords: distributionally robust reinforcement learning, robust Markov decision processes, sample complexity
Compressor summary: The paper studies how robust reinforcement learning is to different uncertainty sets and shows that learning with total variation is more efficient than standard MDPs.
Zhijian Duan, Haoran Sun, Yurong Chen, Xiaotie Deng
https://openreview.net/forum?id=cNb5hkTfGC
Keywords: Automated Mechanism Design, Auction Design, Affine Maximizer Auctions, Deep Learning, Game Theory
Compressor summary: AMenuNet is a scalable neural network that designs dominant strategy incentive compatible auctions by constructing allocation menus from bidder and item representations.
Jialu Li, Mohit Bansal
https://openreview.net/forum?id=cNObl6QQEH
Keywords: Vision-and-Language Navigation, diffusion models, image inpainting for panorama generation
Compressor summary: PanoGen is a method to create diverse and realistic 3D environments based on text descriptions, which can improve Vision-and-Language Navigation performance and help agents generalize better to new situations.
Ziyan Wang, Hao Wang
https://openreview.net/forum?id=cMUBkkTrMo
Keywords: probabilistic methods, imbalanced regression, variational inference
Compressor summary: The paper introduces a probabilistic deep learning model, VIR, that performs well in imbalanced regression and naturally estimates uncertainty using data borrowing and reweighting techniques.
Seohong Park, Dibya Ghosh, Benjamin Eysenbach, Sergey Levine
https://openreview.net/forum?id=cLQCCtVDuW
Keywords: reinforcement learning
Compressor summary: The paper proposes a hierarchical algorithm for goal-conditioned RL from offline data that leverages the structure of reaching distant goals through subgoals, and shows its effectiveness on various benchmarks.
Srinivasan A, Vojtěch Havlíček, Louis Schatzki
https://openreview.net/forum?id=cGeLeh995N
Keywords: Quantum Computing, Statistical Learning, Quantum learning theory, Entanglement
Compressor summary: The paper investigates how entangled, separable, and statistical measurements affect learning in the quantum statistical query model. It reveals different trade-offs depending on the task and introduces a new quantum dimensionality concept to lower bound the complexity of learning various quantum states and functions. It also gives unconditional separation results for error mitigation methods.
Qijian Zhang, Junhui Hou, Yohanes Yudhi Adikusuma, Wenping Wang, Ying He
https://openreview.net/forum?id=cGdGh3Mp2W
Keywords: geodesic distance, implicit representation, 3D geometry
Compressor summary: The paper introduces NeuroGF, a neural implicit representation of geodesics on 3D mesh models that can efficiently answer queries and encode both geometry and geodesics in a unified format.
Lucrezia Valeriani, Diego Doimo, Francesca Cuturello, Alessandro Laio, Alessio ansuini, Alberto Cazzaniga
https://openreview.net/forum?id=cCYvakU5Ek
Keywords: Representations, transformers, geometry, interpretability
Compressor summary: Large transformers have similar representation evolution across various data types and can be used to identify layers with high semantic content.
Hyunsoo Lee, Minsoo Kang, Bohyung Han
https://openreview.net/forum?id=cBS5CU96Jq
Keywords: Diffusion, Image-to-Image Translation
Compressor summary: The paper proposes a text-driven image-to-image translation method that edits regions of interest in a source image based on a modifying text and preserves other parts, using a novel score function and mixup technique for better guidance and fusion.
Keegan Harris, Chara Podimata, Steven Wu
https://openreview.net/forum?id=cBIPcZKFdw
Keywords: strategic classification, strategic learning, apple tasting, bandit feedback, learning with incentives
Compressor summary: The text presents algorithms for decision-making under incentives and one-sided feedback, aiming to minimize strategic regret compared to the optimal policy. One algorithm works for stochastic agents and the other for adversarial ones, with different performance guarantees. The algorithms also apply to bandit feedback settings.
Daniel Y Fu, Simran Arora, Jessica Grogan, Isys Johnson, Sabri Eyuboglu, Armin W Thomas, Benjamin Frederick Spector, Michael Poli, Atri Rudra, Christopher Re
https://openreview.net/forum?id=cB0BImqSS9
Keywords: structured matrices, transformers, efficiency
Compressor summary: Monarch Mixer (M2) is a new architecture that uses Monarch matrices, a sub-quadratic primitive, to scale in sequence length and model dimension, achieving high performance and efficiency in various domains.
Shiyan Chen, Jiyuan Zhang, Yajing Zheng, Tiejun Huang, Zhaofei Yu
https://openreview.net/forum?id=cAyLnMxiTl
Keywords: spike camera, neuromorphic vision sensors, motion deblurring, high speed imaging
Compressor summary: The paper proposes a new method to remove motion blur from high-speed scenes using spike camera data and traditional RGB images, improving visual quality over existing methods.
Saurabh Sihag, Gonzalo Mateos, Corey McMillan, Alejandro Ribeiro
https://openreview.net/forum?id=cAhJF87GN0
Keywords: graph neural networks, brain age, Alzheimer's disease, interpretability, explainability, computational neuroscience
Compressor summary: The paper proposes a framework using coVariance neural networks for predicting brain age based on cortical thickness features, which allows for anatomical interpretability in identifying regions related to elevated brain age gap in Alzheimer's disease.
Yibo Yang, Stephan Eckstein, Marcel Nutz, Stephan Mandt
https://openreview.net/forum?id=cAaTbLa3ad
Keywords: information theory, rate-distortion function, optimal transport
Compressor summary: The paper proposes a new method to estimate the rate-distortion function using optimal transport, which learns the support of the optimal reproduction distribution and has advantages over neural network methods in terms of computational effort and tuning.
Marina Knittel, Max Springer, John P Dickerson, MohammadTaghi Hajiaghayi
https://openreview.net/forum?id=cAPMmCl2f3
Keywords: Fair machine learning, hierarchical clustering, clustering
Compressor summary: The paper introduces a new algorithm for fair hierarchical clustering that breaks the polynomial-approximate barrier and achieves low cost.
Murat Kocaoglu
https://openreview.net/forum?id=cANkPsVtsw
Keywords: causal discovery
Compressor summary: The paper proposes a new constraint-based causal discovery algorithm, the k-PC algorithm, which uses conditional independence tests with an upper bound on the conditioning set size and graphically characterizes the resulting k-Markov equivalence class of causal graphs for more robust causal discovery in small data settings.
Jiawei Du, Qin Shi, Joey Tianyi Zhou
https://openreview.net/forum?id=c9fXCzR5fK
Keywords: Dataset distillation, gradients matching
Compressor summary: The paper introduces Sequential Subset Matching (SeqMatch), a new method for dataset distillation that adaptively optimizes synthetic data to improve performance by addressing the coupling issue caused by static optimization.
Masoud Moravej Khorasani, Erik Weyer
https://openreview.net/forum?id=c8nIdZ5HJJ
Keywords: Stochastic Multi-armed bandit, Online Learning, Upper Confidence Bound
Compressor summary: MARS is a data-dependent UCB algorithm for multi-armed bandits that doesn't need scaling, uses maximum average of randomly sampled rewards, and performs similarly to $\psi$-UCB without correction factors.
Ping Guo, Xiangpeng Wei, Yue Hu, Baosong Yang, Dayiheng Liu, Fei Huang, jun xie
https://openreview.net/forum?id=c5dRV9tA3K
Keywords: cross-lingual pretraining;language-agnostic representation
Compressor summary: The paper introduces Emma-X, a pre-training algorithm to learn cross-lingual universals using non-parallel data, and shows its effectiveness on xrete benchmark tasks.
Hojoon Lee, Hanseul Cho, Hyunseung Kim, Daehoon Gwak, Joonkee Kim, Jaegul Choo, Se-Young Yun, Chulhee Yun
https://openreview.net/forum?id=c5WOU7p4ES
Keywords: Reinforcement Learning, Sharpness Minimization, Generalization, Plasticity, Deep Learning
Compressor summary: The PLASTIC algorithm, which balances input and label plasticity in off-policy reinforcement learning, enhances sample efficiency by preserving the model's ability to adapt to changing data and goals.
Aditya Bhaskara, Sepideh Mahabadi, Ali Vakilian
https://openreview.net/forum?id=c4Xc0uTLXW
Keywords: volumetric spanner, well-conditioned basis, determinant maximization, minimum volume enclosing ellipsoid
Compressor summary: A volumetric spanner is a subset of points that can express all other points with small coefficients, and this paper provides almost optimal bounds on its size and a simple construction method for all $\ell_p$ norms, with applications to finding coresets for the MVEE problem.
Matan Schliserman, Tomer Koren
https://openreview.net/forum?id=c2eedxSlPJ
Keywords: Convex optimization, Gradient Descent, separable data, generalization bounds, Stochastic Gradient Descent.
Compressor summary: This paper investigates the generalization abilities of unregularized gradient methods in separable linear classification, provides tighter risk bounds depending on the tail decay rate of the loss function, and simplifies the proof technique compared to previous work.
Chaoqi Yang, M Brandon Westover, Jimeng Sun
https://openreview.net/forum?id=c2LZyTyddi
Keywords: biological signal, transformer, cross-data learning, in-the-wild learning
Compressor summary: The paper introduces Biosignal Transformer, a flexible encoder for biosignals that can handle different formats and enable cross-data learning, outperforming baselines and improving performance on seizure detection task.
Miles Turpin, Julian Michael, Ethan Perez, Samuel R. Bowman
https://openreview.net/forum?id=bzs4uPLXvi
Keywords: Natural language processing, large language models, XAI, explainability
Compressor summary: CoT explanations from LLMs can be misleading due to added biasing features, impacting accuracy and trust in the models.
Li Fan, Ruida Zhou, Chao Tian, Cong Shen
https://openreview.net/forum?id=bzXpQUnule
Keywords: Federated bandits, contextual bandits, regret analysis
Compressor summary: FedSupLinUCB is an order-optimal algorithm for federated linear bandits with adversarial finite action sets, achieving low regret and controlled communication cost in asynchronous and synchronous cases.
Jing-Cheng Pang, Xinyu Yang, Si-Hang Yang, Xiong-Hui Chen, Yang Yu
https://openreview.net/forum?id=bx0SDRVDzF
Keywords: Reinforcement learning, instruction-following, autonomous agent
Compressor summary: TALAR simplifies natural language instructions into task-related language for efficient policy training in reinforcement learning.
Jinhui HOU, Zhiyu Zhu, Junhui Hou, Hui LIU, Huanqiang Zeng, Hui Yuan
https://openreview.net/forum?id=bv9mmH0LGF
Keywords: Image enhancement, diffusion models
Compressor summary: The paper proposes a diffusion-based method for low-light image enhancement, using curvature and uncertainty regularization to improve detail preservation, contrast, and noise suppression, outperforming existing methods.
Matan Levy, Rami Ben-Ari, Nir Darshan, Dani Lischinski
https://openreview.net/forum?id=bt7pQ7o7zG
Keywords: Image Retrieval, Multi-modal learning
Compressor summary: The paragraph introduces ChatIR, a chat-based image retrieval system that uses dialog with the user to clarify their search intent and leverages Large Language Models to generate follow-up questions, achieving significant gains in image retrieval compared to existing approaches.
Fan Feng, Sara Magliacane
https://openreview.net/forum?id=bsNslV3Ahe
Keywords: multi-object RL, compositional generalization, factored representations
Compressor summary: DAFT-RL is a framework that improves object-centric representation learning and generalization in reinforcement learning by factorizing dynamics and rewards according to object attributes.
Shoubin Yu, Jaemin Cho, Prateek Yadav, Mohit Bansal
https://openreview.net/forum?id=brOMKBEGXP
Keywords: Video Question Answering, Video Localization, Image-Language Model
Compressor summary: The text describes a new video question answering method called SeViLA that uses a single image-language model to perform both keyframe localization and answer prediction, improving performance on various benchmarks.
Andy Shih, Suneel Belkhale, Stefano Ermon, Dorsa Sadigh, Nima Anari
https://openreview.net/forum?id=bpzwUfX1UP
Keywords: diffusion models, parallel sampling
Compressor summary: ParaDiGMS is a new method that speeds up diffusion model sampling by running multiple denoising steps in parallel, achieving 2-4x faster sampling without sacrificing quality.
Yuting Hu, Jiajie Li, Florian Klemme, Gi-Joon Nam, Tengfei Ma, Hussam Amrouch, Jinjun Xiong
https://openreview.net/forum?id=bprclnHNvm
Keywords: Graph Neural Networks, Integrated Circuits, Circuit Timing Analysis, Physics-guided Deep Learning
Compressor summary: The paper proposes SyncTREE, a tree-based graph neural network that uses AI to improve the speed and accuracy of timing analysis for interconnects in integrated circuits.
Johann Brehmer, Joey Bose, Pim De Haan, Taco Cohen
https://openreview.net/forum?id=bpmM6SkDUy
Keywords: Planning, Diffusion models, Equivariance, Equivariant generative models
Compressor summary: EDGI is an algorithm that uses a new equivariant diffusion model to efficiently plan and learn in structured environments with various symmetries.
Yilun Du, Sherry Yang, Bo Dai, Hanjun Dai, Ofir Nachum, Joshua B. Tenenbaum, Dale Schuurmans, Pieter Abbeel
https://openreview.net/forum?id=bo8q5MRcwy
Keywords: sequential decision making, general-purpose agent, video diffusion
Compressor summary: The paper explores using text-guided image synthesis to create more general-purpose agents that can plan and execute various tasks by generating realistic videos of their actions.
Manjie Xu, Guangyuan Jiang, Wei Liang, Chi Zhang, Yixin Zhu
https://openreview.net/forum?id=bo5oIoL95U
Keywords: Visual Reasoning, Abductive Reasoning, Active Reasoning
Compressor summary: The authors introduce Conan, an interactive environment for assessing active reasoning abilities in vision-language models, which differ from humans' ability to explore and reason with incomplete information.
Xudong Wang, Shufan Li, Konstantinos Kallidromitis, Yusuke Kato, Kazuki Kozuka, Trevor Darrell
https://openreview.net/forum?id=bn4qZxltsH
Keywords: Universal Image Segmentation, Hierarchical, Open-vocabulary
Compressor summary: HIPIE is a model that uses text descriptions and hierarchical representation to perform open-vocabulary image segmentation, achieving state-of-the-art results in various image comprehension tasks.
Jacob Lindbäck, Zesen Wang, Mikael Johansson
https://openreview.net/forum?id=bmdnWIuypV
Keywords: optimal transport, domain adaptation, splitting methods, gpu computations
Compressor summary: The algorithm efficiently solves regularized optimal transport problems using a Douglas-Rachford splitting technique with global convergence guarantees, low cost, and GPU parallelization, making it fast for various applications like domain adaptation and generative modeling.
Zhenhailong Wang, Ansel Blume, Sha Li, Genglin Liu, Jaemin Cho, Zineng Tang, Mohit Bansal, Heng Ji
https://openreview.net/forum?id=blm1pqiOXe
Keywords: video-language model, action knowledge benchmarking, action understanding, temporal understanding
Compressor summary: The paper introduces ActionBench, a benchmark to assess video-language models' action knowledge, and Paxion, a framework that improves their performance by training them with DVDM objective.
Xiaowen Jiang, Sebastian U Stich
https://openreview.net/forum?id=blC2kbzvNC
Keywords: Convex Optimization, SGD, Adaptive Methods, Variance Reduction, Polyak Stepsize, Line-Search
Compressor summary: The authors propose new variants of SPS and SLS that achieve optimal rates in various settings and a novel VR method that works well with them.
Morris Alper, Hadar Averbuch-Elor
https://openreview.net/forum?id=bfmSc1ETT9
Keywords: multimodal learning, computer vision, NLP, cognitive science
Compressor summary: This paper investigates whether sound symbolism, the idea that certain sounds are associated with specific meanings across languages, is reflected in vision-and-language models like CLIP and Stable Diffusion, and finds strong evidence that they do show this pattern.
Yash Sanjay Bhalgat, Iro Laina, Joao F. Henriques, Andrea Vedaldi, Andrew Zisserman
https://openreview.net/forum?id=bbbbbov4Xu
Keywords: Neural Radiance Fields, Instance Segmentation, Metric Learning, Clustering, 3D Computer Vision
Compressor summary: The paper presents a novel 3D instance segmentation method that uses 2D pre-trained models and a slow-fast clustering objective function to handle large numbers of objects in complex scenes, achieving state-of-the-art results on several datasets.
Markus Utke, Ulrike Schmidt-Kraepelin
https://openreview.net/forum?id=bbL20Oupi4
Keywords: liquid democracy, directed trees, parameterized markov chain, matrix tree theorem, axiomatic method
Compressor summary: The paragraph discusses a new voting system that combines representative and direct democracy, and two rules that balance anonymity and copy-robustness, which are shown to be equivalent using a polynomial-time algorithm with applications in other fields.
Tara Akhound-Sadegh, Laurence Perreault-Levasseur, Johannes Brandstetter, Max Welling, Siamak Ravanbakhsh
https://openreview.net/forum?id=ba4boN3W1n
Keywords: PDE, Lie point symmetry, Symmetry, Neural PDE solver, PINNs
Compressor summary: The text discusses how incorporating PDE symmetries into physics-informed neural networks can improve their ability to learn solutions and neighboring solutions efficiently.
Kalle Kujanpää, Joni Pajarinen, Alexander Ilin
https://openreview.net/forum?id=bY0c46ZtXa
Keywords: Planning, Subgoal search, Reinforcement learning, Hierarchical Imitation Learning, Hierarchical planning, Hierarchical reinforcement learning
Compressor summary: The paper proposes complete subgoal search, an efficient hybrid method that combines high-level and low-level actions to achieve completeness in solving complex planning problems.
Hannah Dröge, Zorah Lähner, Yuval Bahat, Onofre Martorell Nadal, Felix Heide, Michael Moeller
https://openreview.net/forum?id=bXvmnpCMmq
Keywords: low rank, permutation, kissing number, matrix factorization, assigment problem
Compressor summary: The paper proposes an efficient method for approximating large permutation matrices using low-rank matrix factorization and reducing memory costs, enabling the solution of previously infeasible problems.
Shenao Zhang, Boyi Liu, Zhaoran Wang, Tuo Zhao
https://openreview.net/forum?id=bUgqyyNo8j
Keywords: Reinforcement Learning, Model-Based Reinforcement Learning, Policy Gradient
Compressor summary: The paper analyzes the challenges of applying ReParameterization Policy Gradient Methods to long-term reinforcement learning problems and proposes a spectral normalization method to improve gradient estimation and reduce variance.
AJAY KUMAR JAISWAL, Shiwei Liu, Tianlong Chen, Zhangyang Wang
https://openreview.net/forum?id=bU9hwbsVcy
Keywords: Pre-trained Models, Sparsity, Emergence, Transformers, Pruning
Compressor summary: The paper studies sparse patterns in large pre-trained transformers, proposes a concept of essential sparsity with a sharp dropping point, and finds that self-supervised learning triggers stronger emergent sparsification than supervised learning.
Woojun Kim, Yongjae Shin, Jongeui Park, Youngchul Sung
https://openreview.net/forum?id=bTidcHIK2t
Keywords: deep reinforcement learning, primacy bais, reset, deep ensemble learning
Compressor summary: The paper proposes a new reset method for deep reinforcement learning that uses deep ensembles to address primacy bias, overfitting, and performance collapse issues while improving sample efficiency.
Wenhao Ding, Laixi Shi, Yuejie Chi, Ding Zhao
https://openreview.net/forum?id=bTL5SNOpfa
Keywords: reinforcement learning, robustness, causality, spurious correlation
Compressor summary: The paper proposes a new robust reinforcement learning framework that avoids learning spurious correlations caused by unobserved confounders and shows improved performance in real-world tasks.
Yingyi Chen, Qinghua Tao, Francesco Tonin, Johan Suykens
https://openreview.net/forum?id=bRyduWAAVT
Keywords: Self-attention, primal-dual representations, SVD, kernel method, asymmetry, transformer
Compressor summary: The paper introduces Primal-Attention, a new attention mechanism based on asymmetric Kernel Singular Value Decomposition (KSVD), which improves efficiency and promotes low-rank property in self-attention.
Hisham Husain, Vu Nguyen, Anton van den Hengel
https://openreview.net/forum?id=bRlEwWd7Vy
Keywords: Bayesian Optimization, Distributionally Robust Optimization, φ-divergences
Compressor summary: The authors propose a new algorithm for Bayesian Optimization that is robust against data-shift in various divergences and achieve sublinear regret bounds with experimental validation.
Zun Wang, Guoqing Liu, Yichi Zhou, Tong Wang, Bin Shao
https://openreview.net/forum?id=bPJmu1PbZD
Keywords: Machine learning force field, graph neural network, many-body interactions
Compressor summary: The quintuple network (QuinNet) is a graph neural network that efficiently represents many-body interactions up to five-body interactions with high accuracy for molecular dynamics simulations.
Xinyi Chen, Elad Hazan
https://openreview.net/forum?id=bOQNd7tWAp
Keywords: online learning, control, hyperparameter optimization
Compressor summary: The authors propose a novel meta-optimization approach based on control theory to learn the best optimization algorithm online, overcoming nonconvexity challenges and achieving regret guarantees.
Asic Q Chen, Ruian Shi, Xiang Gao, Ricardo Baptista, Rahul G Krishnan
https://openreview.net/forum?id=bNXVRJjmOl
Keywords: generative models, density estimation, normalizing flows, binary matrix factorization, causal inference
Compressor summary: The Structured Neural Network (StrNN) injects structure into neural networks by masking pathways based on binary matrix factorization, enabling learning functions with desired independencies and improving data-efficient generative modeling and causal effect estimation.
An Zhang, Leheng Sheng, Zhibo Cai, Xiang Wang, Tat-Seng Chua
https://openreview.net/forum?id=bNNIf8F9OU
Keywords: Collaborative filtering, Contrastive loss, Recommendation, Generalization ability
Compressor summary: The paragraph discusses the challenges of using contrastive learning in collaborative filtering and proposes a new loss function called Adversarial InfoNCE that improves generalization ability for out-of-distribution tasks.
Tycho F.A. van der Ouderaa, Alexander Immer, Mark van der Wilk
https://openreview.net/forum?id=bNIHdyunFC
Keywords: learning layer-wise relaxed equivariances bayesian symmetry discovery marginal likelihood
Compressor summary: The authors propose a method to learn flexible symmetry constraints in neural networks from data using gradients and differentiable Laplace approximations, which improves generalization performance on image classification tasks.
Aleksandr Podkopaev, Aaditya Ramdas
https://openreview.net/forum?id=bN1ZBSOV2f
Keywords: two-sample testing, independence testing, testing by betting, sequential testing
Compressor summary: The text discusses sequential nonparametric two-sample and independence testing using prediction-based betting strategies that work well on structured high-dimensional data and adapt to changing data distributions.
Balhae Kim, Hyungi Lee, Juho Lee
https://openreview.net/forum?id=bM6mynsusR
Keywords: Bayesian pseudocoresets, Function space variational inference
Compressor summary: The paper proposes a new method to create a compact synthetic dataset for scalable Bayesian inference that works on function space instead of weight space, improving uncertainty quantification and robustness.
Chenghu Du, junyin Wang, Shuqing Liu, Shengwu Xiong
https://openreview.net/forum?id=bLB4vTwSbC
Keywords: parser-free virtual try-on, self-cycle consistency, human analysis and understanding, fashion synthesis, Markov Random Field
Compressor summary: The paragraph discusses a new parser-free virtual try-on network (USC-PFN) that improves garment deformation modeling and alignment with the human body using self-cycle consistency and Markov Random Field, achieving state-of-the-art performance.
Pritam Sarkar, Ahmad Beirami, Ali Etemad
https://openreview.net/forum?id=bKqrWLCMrX
Keywords: computer vision, self-supervised learning, video self-supervised learning, natural distribution shift, video learning, out-of-distribution generalization
Compressor summary: The paper investigates how six self-supervised video methods handle different types of distribution shifts and reveals their strengths and weaknesses in various scenarios.
Theo Joseph Adrai, Guy Ohayon, Michael Elad, Tomer Michaeli
https://openreview.net/forum?id=bJJY9TFfe0
Keywords: Computer Vision, Image Restoration, Deep Learning, Perceptual Quality
Compressor summary: The algorithm improves image restoration by using optimal transport in the latent space of a variational auto-encoder to control the trade-off between perceptual quality and MSE, and it works well with few training images.
Ting Li, Jianguo Li, Zhanxing Zhu
https://openreview.net/forum?id=bISkJSa5Td
Keywords: Neural CDE, Time-series forecasting, Latent Dynamic
Compressor summary: Neural Lad is a new neural ODE framework that addresses issues in existing models by incorporating local change and seasonality in latent dynamics and achieving better performance in various time series forecasting tasks.
Federico Bergamin, Pablo Moreno-Muñoz, Søren Hauberg, Georgios Arvanitidis
https://openreview.net/forum?id=bHS7qjLOAy
Keywords: Riemannian geometry, Laplace approximation, Approximate inference, Bayesian neural networks
Compressor summary: The paper proposes a Riemannian Laplace approximation for Bayesian neural networks that adapts to the true posterior's shape and improves performance across tasks.
Samuel Pfrommer, Brendon G. Anderson, Julien Piet, Somayeh Sojoudi
https://openreview.net/forum?id=bH4LVNVXUo
Keywords: asymmetric certified robustness, input-convex neural networks
Compressor summary: The paper introduces a new neural network architecture for machine learning models to resist asymmetric adversarial attacks, which induce false negatives, and proves its effectiveness theoretically and empirically.
Kun Yi, Qi Zhang, Wei Fan, Hui He, Liang Hu, Pengyang Wang, Ning An, Longbing Cao, Zhendong Niu
https://openreview.net/forum?id=bGs1qWQ1Fx
Keywords: multivariate time series forecasting, fourier space
Compressor summary: The authors propose a novel graph neural network architecture, Fourier Graph Neural Network (FouderGNN), that uses a hypervariate graph data structure to capture spatiotemporal dynamics in multivariate time series forecasting and outperforms existing methods with lower complexity.
Zhitong Gao, Shipeng Yan, Xuming He
https://openreview.net/forum?id=bGcdjXrU2w
Keywords: ood detection, semantic segmentation, anomaly segmentation, test-time adaptation
Compressor summary: The paper proposes a dual-level framework to detect out-of-distribution images that can handle both domain shift and semantic shift, improving OOD detection performance on various benchmarks.
CHEN SHENGYUAN, YUNFENG CAI, Huang Fang, Xiao Huang, Mingming Sun
https://openreview.net/forum?id=bETvUctiTR
Keywords: Neuro-Symbolic Reasoning, Knowledge graph embedding, Probabilistic soft logic
Compressor summary: DiffLogic is a neuro-symbolic reasoning framework that uses a tailored filter to select essential triples and probabilistic soft logic to assess agreement among rules, weights, and observed triples for effective and efficient knowledge graph reasoning.
Enrico Giudice, Jack Kuipers, Giusi Moffa
https://openreview.net/forum?id=bBIHqoZ3OR
Keywords: Bayesian networks, structure learning, graphical models, gaussian processes, Bayesian inference, MCMC sampling, importance sampling
Compressor summary: The paragraph explains how Gaussian Process Networks (GPNs) model continuous distributions with flexible Graphical Models, but their Bayesian structure learning is hard to compute, so the authors propose a Monte Carlo method that samples from the posterior distribution and performs well in simulations.
Pratik Karmakar, Debabrota Basu
https://openreview.net/forum?id=bAI21VEMvM
Keywords: Differential Privacy, Model Extraction Attacks, Active Sampling, Max-Information Attack
Compressor summary: The authors propose a novel black-box model extraction attack called Marich that uses active sampling to select queries from public data and create an accurate replica of the target ML model.
Sarah Hooper, Mayee F Chen, Khaled Kamal Saab, Kush Bhatia, Curtis Langlotz, Christopher Re
https://openreview.net/forum?id=b8xowIlZ7v
Keywords: Medical imaging, segmentation, classification
Compressor summary: The paper compares classification and segmentation models for medical image analysis, proposes methods to use segmentation models for classification, and discusses the benefits of switching from segmentation to classification in terms of performance, robustness, and interpretability.
Zhengyang Geng, Ashwini Pokle, J Zico Kolter
https://openreview.net/forum?id=b6XvK2de99
Keywords: Deep Equilibrium Models, Diffusion Models, Distillation, Generative Models
Compressor summary: The paper introduces the GET model, which distills diffusion models directly from noise to image using an offline training approach with better performance than existing one-step methods.
Rachel Ward, Tamara G. Kolda
https://openreview.net/forum?id=b6FeLpKKjl
Keywords: matrix factorization; gradient descent; global convergence; concentration; optimization
Compressor summary: The paper studies alternating gradient descent with fixed step size for asymmetric matrix factorization, showing that a specific number of iterations is enough to reach an approximate solution with high probability, starting from a certain random initialization, and experiments confirm the theoretical results.
Chris Lin, Ian Connick Covert, Su-In Lee
https://openreview.net/forum?id=b60wLlkBta
Keywords: explainable artificial intelligence, interpretable machine learning, feature attributions, removal-based feature attributions, robustness
Compressor summary: The paper studies how removal-based feature attribution methods can be more robust to input and model perturbations and provides a unified analysis with theoretical and empirical results.
Feynman T. Liang, Liam Hodgkinson, Michael W. Mahoney
https://openreview.net/forum?id=b5R8mbqo9Q
Keywords: probabilistic programming, static analysis, heavy tails, monte carlo, mcmc, variational inference
Compressor summary: The paper proposes a method to improve tail behavior accuracy in probabilistic models using noise in neural networks, based on an algebra using the generalized Gamma distribution.
Aarush Gupta, Junli Cao, Chaoyang Wang, Ju Hu, Sergey Tulyakov, Jian Ren, Laszlo Attila Jeni
https://openreview.net/forum?id=b2wSODM7iG
Keywords: light field, neural radiance field, novel view synthesis
Compressor summary: The authors propose a new method for real-time image synthesis on mobile devices using the light slab representation, which improves both quality and speed compared to existing neural light field methods.
Thomas Pethick, Wanyun Xie, Volkan Cevher
https://openreview.net/forum?id=b1JPBGJhUi
Keywords: Minimax optimization, Lookahead, Generative adversarial networks, Stability, Nonconvex-nonconcave, Cohypomonotone
Compressor summary: The paper proposes a new optimization method called RAPP that uses linear interpolation to stabilize neural network training, and extends it to constrained and regularized settings, as well as to Lookahead algorithms, with experiments on generative adversarial networks showing its benefits.
Michael Jacob Feldman, David Donoho
https://openreview.net/forum?id=b1BhHjBxsx
Keywords: tensor PCA, spectral algorithms, random matrix theory
Compressor summary: The paper studies different ways to convert noisy tensors into matrices and uses spectral methods to recover the low-rank signal, proving that some methods work better than others in certain conditions.
Noam Razin, Tom Verbin, Nadav Cohen
https://openreview.net/forum?id=ayZpFoAu5c
Keywords: Graph Neural Networks, Expressivity, Interactions, Edge Sparsification
Compressor summary: The paper proposes a measure to quantify interaction strength in graph neural networks (GNNs) and introduces an edge sparsification method that preserves interaction ability while reducing computation.
Aldo Pacchiano, Jonathan Lee, Emma Brunskill
https://openreview.net/forum?id=axmY49ahVI
Keywords: regret, model selection, planning, static, lower bound
Compressor summary: The paper proposes two experiment planning strategies for function approximation in contextual bandits using large datasets of contexts but no rewards, and analyzes their optimality guarantees and differences from adaptive learning.
Gergely Flamich
https://openreview.net/forum?id=axRMkinASf
Keywords: channel simulation, relative entropy coding, reverse channel coding, rejection sampling, Poisson process
Compressor summary: The paper presents a new algorithm (GPRS) for efficient one-shot channel simulation in one-dimensional problems with unimodal density ratio and shows its superior performance over existing methods.
Jingyuan Xu, Weiwei Liu
https://openreview.net/forum?id=awbWWO0nb6
Keywords: Learning Theory
Compressor summary: The paper investigates how adaptive algorithms may overfit the test dataset in multiclass classification, with bounds on the overfitting bias.
Nora Belrose, David Schneider-Joseph, Shauli Ravfogel, Ryan Cotterell, Edward Raff, Stella Biderman
https://openreview.net/forum?id=awIpKpwTwF
Keywords: interpretability, fairness, concept erasure, representation, adversarial, robustness
Compressor summary: LEACE is a method to remove specific features from representations while minimizing changes, and it can be applied to language models for fairness and interpretability tasks.
Yuchao Qin, Mihaela van der Schaar, Changhee Lee
https://openreview.net/forum?id=aw1vLo7TE7
Keywords: active sensing, value of information, risk-averse learning
Compressor summary: The paper proposes RAS, a novel risk-averse active sensing approach for cost-efficient acquisition of patient covariates in healthcare to enable timely and accurate outcome predictions.
Chengzhi Cao, Chao Yang, Ruimao Zhang, Shuang Li
https://openreview.net/forum?id=avuRopYsCg
Keywords: Logic rule, human actions, sports analyze
Compressor summary: The paper proposes a model that uses spatial-temporal logic rules to capture human movement patterns based on intentions and environmental factors, with potential applications in sports analytics, robotics, and autonomous cars.
Avani Gupta, Saurabh Saini, P J Narayanan
https://openreview.net/forum?id=arkmhtYLL6
Keywords: Human Centered Concepts, ML interpretability, XAI based Model Improvement, Debiasing
Compressor summary: The authors propose CAVs for ante-hoc training with Concept Loss and Concept Distillation to make models more sensitive and robust to concepts, which enhances interpretability and knowledge transfer.
Yiming Lei, Zilong Li, Yangyang Li, Junping Zhang, Hongming Shan
https://openreview.net/forum?id=apjOYp3mOa
Keywords: Language-image consistency, prompt learning, image classification, CNN interpretation
Compressor summary: The paper proposes LICO, a model that uses linguistic prompts and optimal transport theory to generate more accurate and interpretable attention maps for image classification tasks.
Cameron Omid Smith, Yilun Du, Ayush Tewari, Vincent Sitzmann
https://openreview.net/forum?id=apFDDJOYf5
Keywords: Pose Estimation, Scene Flow Estimation, Scene Representation Learning, Computer Vision, Neural Implicit Representations, Neural Radiance Fields, View Synthesis, Self-Supervised Representation Learning
Compressor summary: The authors propose a self-supervised method for learning 3D neural scenes from videos by jointly estimating camera poses and reconstructing the scene in a single forward pass.
Paul Pu Liang, Zihao Deng, Martin Q. Ma, James Zou, Louis-Philippe Morency, Russ Salakhutdinov
https://openreview.net/forum?id=alLs7EtRJP
Keywords: multimodal learning, contrastive learning, self-supervised learning, information theory
Compressor summary: FactorCL is a new method for learning multimodal representations that captures both shared and unique information relevant to downstream tasks using factorization, maximizing mutual information lower bounds, minimizing mutual information upper bounds, and data augmentations.
Zikang Tian, Ruizhi Chen, Xing Hu, Ling Li, Rui Zhang, Fan Wu, Shaohui Peng, Jiaming Guo, Zidong Du, Qi Guo, Yunji Chen
https://openreview.net/forum?id=aky0dKv9ip
Keywords: Multi-Agent Reinforcement Learning, Transfer Learning, Zero-Shot Generalization
Compressor summary: The paper proposes a novel framework called DT2GS that decomposes MARL tasks into generalizable subtasks using a scalable encoder and an adaptive semantic module to achieve zero-shot generalization and task transferability.
Amin Ghiasi, Ali Shafahi, Reza Ardekani
https://openreview.net/forum?id=ajnThDhuq6
Keywords: Adaptive weight decay, adversarial robustness, weight decay, robust overfitting, overfitting, adversarial attacks, noisy label
Compressor summary: The paper proposes adaptive weight decay that adjusts the regularization strength during training to improve adversarial robustness without extra data or tuning.
Kulin Shah, Sitan Chen, Adam Klivans
https://openreview.net/forum?id=aig7sgdRfI
Keywords: Mixtures of Gaussians, score-based generative models, provable learning of score, Expectation-Maximization, DDPM generative model
Compressor summary: The paper proves that gradient descent on a specific diffusion model can efficiently learn Gaussian mixture models with different levels of initialization and separation.
Lucy Xiaoyang Shi, Yunfan Jiang, Jake Grigsby, Linxi Fan, Yuke Zhu
https://openreview.net/forum?id=afKnrwJBAl
Keywords: Transformers, In-context Learning, Reinforcement Learning, Robotics
Compressor summary: CEC is a new algorithm that improves Transformer agents' learning efficiency and generalization by sequentially structuring online learning trials and mixed-quality demonstrations, resulting in better performance and strong generalization across episodes.
Mark McDonnell, Dong Gong, Amin Parvaneh, Ehsan Abbasnejad, Anton van den Hengel
https://openreview.net/forum?id=aec58UfBzA
Keywords: continual learning, class incremental learning, domain incremental learning, pre-trained models, parameter-efficient transfer learning
Compressor summary: The paper proposes a new approach for continual learning with pre-trained models that uses random projectors and class-prototype accumulation to avoid forgetting without rehearsal memory.
Laura Manduchi, Moritz Vandenhirtz, Alain Ryser, Julia E Vogt
https://openreview.net/forum?id=adq0oXb9KM
Keywords: hierarchical clustering, hierarchical VAE, representation learning, VAE, deep clustering
Compressor summary: TreeVAE is a new model that learns a tree structure to encode latent variables and discover hidden clusters in data, while improving generative performance and allowing conditional sampling.
Yu-Jie Zhang, Zhen-Yu Zhang, Peng Zhao, Masashi Sugiyama
https://openreview.net/forum?id=ad3JNoR2np
Keywords: Covariate Shift, Density Ratio Estimation, Online Convex Optimization, Dynamic Regret, Logistic Regression
Compressor summary: The paper studies continuous covariate shift in machine learning, proposes an online density ratio estimation method to adaptively train predictors, and provides theoretical and empirical guarantees.
Dapeng Hu, Jian Liang, Jun Hao Liew, Chuhui Xue, Song Bai, Xinchao Wang
https://openreview.net/forum?id=ackajXqei2
Keywords: Unsupervised Domain Adaptation; Model Selection; Hyperparameter Selection; Unsupervised Validation;
Compressor summary: MixVal is a novel model selection method for unsupervised domain adaptation that uses mixed target samples with pseudo labels to evaluate the performance of different UDA models on unlabeled target data.
Shyam Sudhakaran, Miguel González-Duque, Matthias Freiberger, Claire Glanois, Elias Najarro, Sebastian Risi
https://openreview.net/forum?id=aa8KsqfTPa
Keywords: Large Language Models, Procedural Content Generation, Open-endedness, Novelty Search
Compressor summary: MarioGPT is a text-prompted, fine-tuned GPT2 model that generates diverse Super Mario Bros levels and can be controlled by user input to address challenges in procedural content generation.
Nora Schneider, Shirin Goshtasbpour, Fernando Perez-Cruz
https://openreview.net/forum?id=aZ9hvpnp0k
Keywords: Data Augmentation, Regression, Deep Learning
Compressor summary: The paper introduces a new data augmentation method for nonlinear regression called Anchor Data Augmentation, which uses causality and distributionally robust Anchor regression to generate more training examples and improve prediction accuracy.
Edward Raff, James Holt
https://openreview.net/forum?id=aZ44Na3l9p
Keywords: reproducibility; multiple instance learning
Compressor summary: The paragraph discusses how five deep-MIL models do not follow the standard assumption of Multiple Instance Learning (MIL) and may lead to incorrect predictions in various applications like healthcare and cyber security.
Xiaoming Shi, Siqiao Xue, Kangrui Wang, Fan Zhou, James Y. Zhang, JUN ZHOU, Chenhao Tan, Hongyuan Mei
https://openreview.net/forum?id=aW9BqtRQkh
Keywords: event sequences, irregular time series, event prediction, large language model, reasoning, few-shot prompting
Compressor summary: The paper proposes LAMP, a framework that uses a large language model for abductive reasoning to improve event prediction performance on real-world datasets.
Chengsen Wang, Zirui Zhuang, Qi Qi, Jingyu Wang, Xingyu Wang, Haifeng Sun, Jianxin Liao
https://openreview.net/forum?id=aW5bSuduF1
Keywords: Anomaly Detection, Time Series, Diffusion, Transformer
Compressor summary: D$^3$R is an anomaly detection network for unstable data that uses dynamic decomposition and reconstruction to handle drift and achieve better performance than existing methods.
Mohammad Reza Karimi Jaghargh, Ya-Ping Hsieh, Andreas Krause
https://openreview.net/forum?id=aRBa0lSxEB
Keywords: Non-Convex Sampling, Langevin Dynamics, Dynamical Systems
Compressor summary: The paragraph discusses a novel framework for addressing non-convex sampling challenges in machine learning using tools from dynamical systems theory to improve convergence guarantees.
Jose Javier Gonzalez Ortiz, John Guttag, Adrian V Dalca
https://openreview.net/forum?id=aN0llPIbdg
Keywords: hypernetworks, amortized learning, computer vision, rescaling, convolutional neural networks, pareto efficiency
Compressor summary: The authors propose Scale-Space HyperNetworks (SSHN), a method that learns a spectrum of CNNs with varying internal rescaling factors, allowing for better accuracy-efficiency trade-offs in medical image analysis tasks.
Borjan Geshkovski, Cyril Letrouit, Yury Polyanskiy, Philippe Rigollet
https://openreview.net/forum?id=aMjaEkkXJx
Keywords: Transformers, Self-Attention, Clustering, Interacting Particle Systems, Continuous Time
Compressor summary: The paragraph discusses how learned representations in Transformers form clusters and limiting objects depending on the value matrix spectrum, and proves convergence to low-rank Boolean matrices in one dimension.
Jihyun Lee, Junbong Jang, Donghwan Kim, Minhyuk Sung, Tae-Kyun Kim
https://openreview.net/forum?id=aMTiwdK3y8
Keywords: 4D representation, hand reconstruction, implicit representation
Compressor summary: FourierHandFlow is a novel method for learning spatio-temporally continuous representations of human hands from RGB videos using Fourier series and implicit shape priors.
Kenneth Li, Oam Patel, Fernanda Viégas, Hanspeter Pfister, Martin Wattenberg
https://openreview.net/forum?id=aLLuYpn83y
Keywords: Large Language Model, AI Safety
Compressor summary: ITI is a technique that improves the truthfulness of large language models by shifting model activations during inference based on a learned set of directions.
Thomas Schmied, Markus Hofmarcher, Fabian Paischer, Razvan Pascanu, Sepp Hochreiter
https://openreview.net/forum?id=aIpGtPwXny
Keywords: Reinforcement Learning, Transformer, Decision Transformer, Multi-task learning, Continual learning, NLP, Fine-tuning, Prompt Tuning, Parameter efficient Fine-tuning
Compressor summary: The paper investigates catastrophic forgetting in reinforcement learning, compares various fine-tuning methods, and proposes a novel method called Learning-to-Modulate (L2M) that preserves pre-training skills while adapting to new tasks.
Jonathan Kelner, Frederic Koehler, Raghu Meka, Dhruv Rohatgi
https://openreview.net/forum?id=aIUnoHuENG
Keywords: theory, sparse linear regression, feature adaptation, lasso
Compressor summary: The paper presents an algorithm that adapts the Lasso to handle sparse linear regression with correlated design and ill-conditioned covariates, achieving near-optimal sample complexity.
Juan M Cardenas, Ben Adcock, Nick Dexter
https://openreview.net/forum?id=aINqoP32cb
Keywords: active learning, regression, arbitrary data, leverage scores, Christoffel functions, generative models, Magnetic Resonance Imaging (MRI), Physics-Informed Neural Networks (PINNs)
Compressor summary: The paper presents a general active learning framework for regression problems that can handle different types of data and shows its effectiveness in scientific computing applications.
Qi Xu, Yuyuan Gao, Jiangrong Shen, Yaxin Li, Xuming Ran, Huajin Tang, Gang Pan
https://openreview.net/forum?id=aGZp61S9Lj
Keywords: Spiking neural networks (SNNs), Recurrent spiking neural network (RSNN), Dynamic Vision Sensor (DVS), Spiking convolutional block attention module (SCBAM)
Compressor summary: The paper proposes a recurrent spiking neural network with an attention module to process spatio-temporal patterns in time series data by combining both spatial and temporal features efficiently.
Ashwinkumar Badanidiyuru, Badih Ghazi, Pritish Kamath, Ravi Kumar, Ethan Jacob Leeman, Pasin Manurangsi, Avinash V Varadarajan, Chiyuan Zhang
https://openreview.net/forum?id=aG6xOP9QY7
Keywords: label differential privacy
Compressor summary: We propose a new family of label randomizers for training _regression_ models under the constraint of label differential privacy (DP). In particular, we leverage the trade-offs between bias and variance to construct better label randomizers depending on a privately estimated prior distribution over the labels. We demonstrate that these randomizers achieve state-of-the-art privacy-utility trade-offs on several datasets, highlighting the importance of reducing bias when training neural networks with label DP. We also provide theoretical results shedding light on the structural properties of the optimal unbiased randomizers.
Stefan Stojanovic, Yassir Jedra, Alexandre Proutiere
https://openreview.net/forum?id=aDLmRMb0K9
Keywords: Low-rank matrix estimation; low rank bandits; low rank MDP; spectral methods
Compressor summary: The paper explores matrix estimation methods with low entry-wise prediction error in low-rank reinforcement learning problems, and presents two new algorithms with near-optimal performance.
Gabriele Farina, Charilaos Pipis
https://openreview.net/forum?id=aCOKUvqHtD
Keywords: online learning, algorithmic game theory, extensive form games, correlated equilibrium, swap regret, linear swap regret
Compressor summary: The paragraph discusses how no-linear-swap regret learners can achieve sublinear regret in sequential games by considering all linear transformations of the mixed strategy space, and shows this leads to a stronger notion of hindsight rationality than previous ones.
Rahul Mazumder, Haoyue Wang
https://openreview.net/forum?id=a2svOXTVgO
Keywords: decision tree, CART
Compressor summary: The paper studies the convergence rate and error bound of CART, a machine-learning model, under a regression setting and introduces sufficient conditions for the SID condition.
Matthew Jagielski, Milad Nasr, Katherine Lee, Christopher A. Choquette-Choo, Nicholas Carlini, Florian Tramèr
https://openreview.net/forum?id=a2Yg9Za6Rb
Keywords: model distillation, membership inference, privacy, dark knowledge
Compressor summary: The paper studies how well knowledge distillation protects privacy, and shows that it is not very effective against membership inference attacks, especially if the student and teacher models are similar or the attacker can tamper with the teacher data.
Matthew Douglas Hoffman, Du Phan, david dohan, Sholto Douglas, Tuan Anh Le, Aaron T Parisi, Pavel Sountsov, Charles Sutton, Sharad Vikram, Rif A. Saurous
https://openreview.net/forum?id=a147pIS2Co
Keywords: Large language models, latent-variable models, control variates, chain-of-thought, MCMC
Compressor summary: The paragraph discusses a new method to improve large language models' performance by combining chain-of-thought prompting and supervised fine-tuning, using a Markov-chain Monte Carlo expectation-maximization algorithm.
Pedro O. Pinheiro, Joshua Rackers, joseph Kleinhenz, Michael Maser, Omar Mahmood, Andrew Martin Watkins, Stephen Ra, Vishnu Sresht, Saeed Saremi
https://openreview.net/forum?id=Zyzluw0hC4
Keywords: generative model, molecule generation, drug discovery
Compressor summary: The paragraph describes a new method called VoxMol that generates 3D molecules as atomic densities on grids by training a neural network to denoise noisy molecule distributions.
Tian Qin, Tian-Zuo Wang, Zhi-Hua Zhou
https://openreview.net/forum?id=ZwQJRXLjVm
Keywords: decision-making, structural rehearsal model, Bayesian inference, probabilistic graphical model
Compressor summary: The paper introduces a rehearsal learning framework that helps find decisions to avoid undesired outcomes using structural rehearsal models, probabilistic graphical models, and a Bayesian approach with a risk quantification bound.
Shengran Hu, Jeff Clune
https://openreview.net/forum?id=ZvDmna23r3
Keywords: Reinforcement learning, Imitation Learning, AI Safety, Interpretability
Compressor summary: Thought Cloning is a novel Imitation Learning framework that trains AI agents to not only mimic human behaviors but also the thoughts behind them, improving performance, safety, and interpretability.
Ruijia Wang, YiWu Sun, Yujie Luo, Shaochuan Li, Cheng Yang, Xingyi Cheng, Hui Li, Chuan Shi, Le Song
https://openreview.net/forum?id=ZuaVKlWdD2
Keywords: complex structure prediction, rigid docking, protein docking, antibody-antigen docking
Compressor summary: BiDock is a novel rigid protein docking model that uses both sequence and structure information to predict inter-protein distances for better predictions.
Alexandr Andoni, Piotr Indyk, Sepideh Mahabadi, Shyam Narayanan
https://openreview.net/forum?id=Zt9RzHjSEy
Keywords: Differential Privacy, Near Neighbor Search, Locality Sensitive Hashing, Data Structures, Range Query
Compressor summary: The paper proposes an efficient algorithm for range counting under differential privacy that balances additive and multiplicative errors without depending on the dimension.
Yanchao Tan, Zihao Zhou, Hang Lv, Weiming Liu, Carl Yang
https://openreview.net/forum?id=ZrG8kTbt70
Keywords: Arributed graph, unsupervised graph learning, language models, representation learning
Compressor summary: The text describes an approach to create unsupervised graph embeddings using language models and random walks, which can handle complex attributes and structures in real-world graphs and improve performance on various downstream tasks.
Josh Alman, Jiehao Liang, Zhao Song, Ruizhe Zhang, Danyang Zhuo
https://openreview.net/forum?id=ZqSx5vXOgC
Keywords: Training neural network, Dynamic activated neuron detection, Sparsity, Fine-grained complexity, Data structure
Compressor summary: This paper proposes a preprocessing method to reduce the time needed to process each data point in deep neural networks by storing weight-data correlation in a tree structure.
Haobo Jiang, Mathieu Salzmann, Zheng Dang, Jin Xie, Jian Yang
https://openreview.net/forum?id=Znpz1sv4IP
Keywords: 6D object pose estimation, Point cloud registration, Diffusion probabilistic model
Compressor summary: The paper presents a new point cloud registration method using an SE(3) diffusion model for accurate 6D object pose estimation in real-world scenarios.
Zekun Li, Shiyang Li, Xifeng Yan
https://openreview.net/forum?id=ZmeAoWQqe0
Keywords: irregularly sampled time series, vision transformer, healthcare, time series classification
Compressor summary: The paper proposes a new method to classify irregularly sampled time series by converting them into images and using pre-trained vision transformers, which shows strong performance and robustness against missing data.
Sebastian Flennerhag, Tom Zahavy, Brendan O'Donoghue, Hado van Hasselt, András György, Satinder Singh
https://openreview.net/forum?id=ZmSg4f16uo
Keywords: meta-learning, online optimisation, convex optimisation
Compressor summary: The paper explores how gradient-based meta-learning relates to convex optimization, proves convergence rates for meta learning with momentum, and shows that the Bootstrapped Meta-Gradient method captures optimism in meta-learning.
Zhe Zeng, Guy Van den Broeck
https://openreview.net/forum?id=Zi1KKzh5Aj
Keywords: Bayesian Model Averaging, Weighted Model Integration, Bayesian Deep Learning, Collapsed Inference
Compressor summary: The paper introduces a novel inference method for Bayesian neural networks that improves both uncertainty estimation and predictive performance by using collapsed samples and volume computation solvers.
Yi Zhang, Xiaoyu Shi, Dasong Li, Xiaogang Wang, Jian Wang, Hongsheng Li
https://openreview.net/forum?id=ZgVJvaAS2h
Keywords: image restoration, diffusion model, denoising, deblurring, JPEG restoration
Compressor summary: The paper introduces a conditional framework for image restoration using diffusion probabilistic models that integrates guidance from a lightweight UNet and produces high-quality results on challenging tasks.
Pierre Marion, Raphaël Berthier
https://openreview.net/forum?id=ZfFR4d5gUM
Keywords: neural networks, non-convex optimization, gradient flow, convergence proof, two-timescale algorithm
Compressor summary: The study analyzes how shallow neural networks trained with different step sizes converge to a global optimum in a simple setting, showing that stochastic gradient descent works well when the stepsizes follow a two-timescale regime.
Jiaqi Xue, Mengxin Zheng, Ting Hua, Yilin Shen, Yepeng Liu, Ladislau Bölöni, Qian Lou
https://openreview.net/forum?id=ZejTutd7VY
Keywords: Large Language Model, Trojan Attack, Adversary Attack, Prompt Injection, GPT-4, Black-box
Compressor summary: The paper introduces TrojLLM, a framework that can generate stealthy triggers to manipulate the outputs of large language models like GPT-3 and GPT-4 for malicious purposes.
Guillaume Wang, Lénaïc Chizat
https://openreview.net/forum?id=ZeRiLBvIps
Keywords: Gradient methods, min-max optimization, spectral analysis, last-iterate convergence
Compressor summary: The paper studies how gradient methods for two-player zero-sum differentiable games converge to local Nash equilibria and shows that they are faster when the game has partial curvature, which is related to the average eigenvalues of the symmetric part of the Jacobian.
Yifan Yang, Peiyao Xiao, Kaiyi Ji
https://openreview.net/forum?id=ZdxGmJGKOo
Keywords: Federated bilevel optimization, federated hypergradient, communication efficiency, system-level heterogeneity, linear speedup
Compressor summary: The paper introduces SimFBO, a simple and efficient federated bilevel optimization framework, and its variant ShroFBO, which improve convergence speed, communication efficiency, and robustness to heterogeneity in machine learning and edge computing.
Lasse Vuursteen, Botond Szabo, Aad van der Vaart, Harry van Zanten
https://openreview.net/forum?id=ZcuFDaMTYw
Keywords: testing, meta-analysis, p-values, e-values, optimal, combining trials
Compressor summary: The paper investigates the power and limitations of combining test statistics from independent trials in high-dimensional models, and introduces methods to improve meta-analysis.
Danyang Zhang, Lu Chen, Situo Zhang, Hongshen Xu, Zihan Zhao, Kai Yu
https://openreview.net/forum?id=ZcJa1R6j3v
Keywords: Learning from Experiences, LLM, Reinforcement Learning, Decision Making, Experience Memory
Compressor summary: The paragraph describes a novel evolvable LLM-based agent framework called Rememberer, which uses a long-term memory to learn from past experiences and improve its performance in different tasks without fine-tuning the parameters of the LLM, achieving superior results on two RL task sets.
Jianqin Luo, Zhexiong Wan, yuxin mao, Bo Li, Yuchao Dai
https://openreview.net/forum?id=ZZgfS1DbmO
Keywords: optical flow, point trajectories, continuous motion, neural ordinary differential equation
Compressor summary: The paper introduces a new method to represent continuous and dense motion in videos using B-splines and neural ODEs, and proposes a synthetic dataset and evaluation metrics to measure its performance.
Hyeonjeong Ha, Minseon Kim, Sung Ju Hwang
https://openreview.net/forum?id=ZZWg9jJQ1j
Keywords: neural architecture search, generalization, efficiency, zero-cost proxy
Compressor summary: The paper proposes a novel method to find optimal neural architectures that can learn generalizable features and robustness against various perturbations using a lightweight and efficient proxy.
David Abel, Andre Barreto, Benjamin Van Roy, Doina Precup, Hado van Hasselt, Satinder Singh
https://openreview.net/forum?id=ZZS9WEWYbD
Keywords: Continual Reinforcement Learning, Reinforcement Learning, Lifelong Reinforcement Learning, Continual Learning
Compressor summary: The paper defines the continual reinforcement learning problem as a setting where agents never stop learning and always search for better policies through an implicit process, and shows that traditional views of multi-task reinforcement learning and continual supervised learning are special cases of this definition.
Kush Bhatia, Avanika Narayan, Christopher De Sa, Christopher Re
https://openreview.net/forum?id=ZXbgVm3PSt
Keywords: In-context learning, task-agnostic methods, large language models
Compressor summary: The paper proposes TART, a method to improve large language models' reasoning abilities using synthetic tasks, which enhances their in-context learning performance across different models, sizes, tasks, and modalities.
Arun Jambulapati, Jerry Li, Christopher Musco, Kirankumar Shiragur, Aaron Sidford, Kevin Tian
https://openreview.net/forum?id=ZViPzk1sUI
Keywords: preconditioning, semidefinite programming, numerical linear algebra, linear regression, semi-random models
Compressor summary: The authors present a framework and algorithms for finding approximate optimal preconditioners and solving linear systems faster than previous methods.
Mert Yuksekgonul, Linjun Zhang, James Zou, Carlos Guestrin
https://openreview.net/forum?id=ZVRG3toCTT
Keywords: trustworthy machine learning, reliable machine learning, uncertainty
Compressor summary: The authors study how atypical inputs or classes affect model reliability, accuracy, and uncertainty, and show incorporating atypicality improves performance in various tasks, including skin lesion classification.
Grégoire Mialon, Quentin Garrido, Hannah Lawrence, Danyal Rehman, Yann LeCun, Bobak Kiani
https://openreview.net/forum?id=ZULq9QV8rH
Keywords: Self-supervised learning, partial differential equations, Lie symmetries, data augmentation
Compressor summary: The text discusses using self-supervised learning to learn representations of partial differential equations (PDEs) from heterogeneous data, improving both coefficient regression and neural solver performance.
Arun Jambulapati, Kevin Tian
https://openreview.net/forum?id=ZRBGwpeewz
Keywords: Optimization, optimal transport, linear programming, semidefinite programming
Compressor summary: The paper explores area convexity to create a fast first-order algorithm for solving box-simplex games and improve the complexity of other combinatorial optimization problems.
Jinhyuk Lee, Zhuyun Dai, Sai Meher Karthik Duddu, Tao Lei, Iftekhar Naim, Ming-Wei Chang, Vincent Y Zhao
https://openreview.net/forum?id=ZQzm0Z47jz
Keywords: information retrieval, document retrieval, natural language processing
Compressor summary: XTR simplifies multi-vector retrieval by rethinking token retrieval and improves ranking efficiency without sacrificing performance on information retrieval benchmarks.
Chen Xu, Xiuyuan Cheng, Yao Xie
https://openreview.net/forum?id=ZQMlfNijY5
Keywords: Normalizing flow, invertible neural networks, JKO scheme
Compressor summary: JKO-iFlow is a neural ODE flow network for efficient sampling and likelihood estimation in high dimensions, which avoids SDE trajectories and score matching, reducing memory load and improving accuracy.
Canzhe Zhao, Ruofeng Yang, Baoxiang Wang, Xuezhou Zhang, Shuai Li
https://openreview.net/forum?id=ZPtzwr2SwJ
Keywords: adversarial low-rank mdps
Compressor summary: This paper studies low-rank MDPs with changing losses and proposes POLO, an oracle-efficient algorithm with a sublinear regret guarantee, which is the first of its kind for this problem setting.
Haithem Turki, Michael Zollhöfer, Christian Richardt, Deva Ramanan
https://openreview.net/forum?id=ZPj7ey5fXa
Keywords: view synthesis, 3d reconstruction, scene representation, 3d deep learning
Compressor summary: The paper introduces a simple modification to grid-based Neural Radiance Fields that improves rendering quality and reduces error rates by using different spatial grid resolutions at render time.
Guanghui Yu, Wei Tang, Saumik Narayanan, Chien-Ju Ho
https://openreview.net/forum?id=ZOKhtz2Z9X
Keywords: Information design; Human behavior; Behavioral experiments
Compressor summary: The authors study how to design optimal information policies using a neural network (HAIDNet) that adapts to different human behavior patterns, and show its effectiveness in simulations and real experiments.
Wonhyeok Choi, Mingyu Shin, Sunghoon Im
https://openreview.net/forum?id=ZNBblMEP16
Keywords: Monocular 3D object detection, Autonomous driving, Recognition, Regression, Metric learning
Compressor summary: The text describes a new method for monocular 3D object detection that uses metric learning and an auxiliary head to improve depth estimation without increasing model size or inference time.
Zheng Chen, Yan-Pei Cao, Yuan-Chen Guo, Chen Wang, Ying Shan, Song-Hai Zhang
https://openreview.net/forum?id=ZKVxABGJ6r
Keywords: neural rendering, neural radiance field, novel view synthesis, panorama, 360-degree image
Compressor summary: PanoGRF is a method to generate realistic views of virtual environments using wide-baseline panoramas by incorporating 360-degree scene priors and depth information.
Zongsheng Yue, Jianyi Wang, Chen Change Loy
https://openreview.net/forum?id=ZIyAHaLlsn
Keywords: Super-resolution; Diffusion model; Efficient
Compressor summary: A novel diffusion model for image super-resolution reduces the number of steps needed, improves efficiency, and maintains or exceeds performance compared to existing methods.
Xing Wei, Anjia Cao, Funing Yang, Zhiheng Ma
https://openreview.net/forum?id=ZIfhYAE2xg
Keywords: Dataset Distillation, Dataset Condensation, Sparse Coding, Dictionary Learning
Compressor summary: The SPEED framework uses dictionary learning, sparse coding, and feature recurrence to create synthetic datasets that capture the essential information of large and diverse original datasets.
Yiyou Sun, Zhenmei Shi, Yixuan Li
https://openreview.net/forum?id=ZITOHWeAy7
Keywords: open-world learning, clustering, spectral analysis
Compressor summary: The paper presents a graph-theoretic framework for open-world semi-supervised learning and proposes a new algorithm, SORL, that provides theoretical guarantees and empirical performance on clustering known and novel classes.
Christian Fiedler, Michael Herty, Sebastian Trimpe
https://openreview.net/forum?id=ZGElmTRk3w
Keywords: Reproducing Kernel Hilbert Spaces, Kernel Methods, Mean Field Limit, Interacting Particle Systems, Support Vector Machines, Statistical Learning Theory
Compressor summary: The paper studies how machine learning kernels and Support Vector Machines behave when the number of input variables goes to infinity, providing new theoretical tools and insights for large-scale problems.
Jack Lanchantin, Shubham Toshniwal, Jason E Weston, Arthur Szlam, Sainbayar Sukhbaatar
https://openreview.net/forum?id=ZFwNdsDCRL
Keywords: Memory, Reasoning, Language Models
Compressor summary: The authors propose Self-Notes, a method for improving large language models' multi-step reasoning and memory retention by allowing them to deviate from the input context and write down their thoughts.
Jacob Granley, Tristan Fauvel, Matthew Chalk, Michael Beyeler
https://openreview.net/forum?id=ZED5wdGous
Keywords: Brain Computer Interfaces, BCI, Stimulus Encoding, Visual Prostheses, Bayesian Optimization, Preferential Bayesian Optimization, Human-in-the-loop Optimization, Sensory Neuroprostheses, Neuroprostheses, Patient-Specific Optimization, Latent Space Bayesian Optimization
Compressor summary: The paragraph discusses a new approach to optimize stimuli for neuroprosthetic devices that improves the quality of restored vision by combining deep learning and Bayesian optimization.
Kaifu Wang, Efthymia Tsamoura, Dan Roth
https://openreview.net/forum?id=ZD65F3x1jU
Keywords: weak supervision, partial label learning, neuro-symbolic learning, latent structural learning
Compressor summary: The paper introduces a new problem called multi-instance Partial Label Learning and provides the first theoretical analysis of its learnability, error bounds, and experiments.
Aniket Rege, Aditya Kusupati, Sharan Ranjit S, Alan Fan, Qingqing Cao, Sham M. Kakade, Prateek Jain, Ali Farhadi
https://openreview.net/forum?id=ZBzYWP2Gpl
Keywords: Semantic Search, Approximate Nearest Neighbor Search, Large-scale search, Representation Learning
Compressor summary: The paper proposes AdANNS, a framework that uses adaptive representations with varying capacities to improve accuracy-compute trade-offs in web-scale search systems based on Matryoshka Representations.
Anton Xue, Rajeev Alur, Eric Wong
https://openreview.net/forum?id=ZBxycYCuEL
Keywords: Feature Attribution, Smoothing, Explainable, Interpretable, Provable Guarantees
Compressor summary: The paper proposes a smoothing method called Multiplicative Smoothing (MuS) that improves the stability of feature attribution methods for machine learning models by making the model sufficiently Lipschitz.
Chaoyue Liu, Dmitriy Drusvyatskiy, Misha Belkin, Damek Davis, Yian Ma
https://openreview.net/forum?id=ZBB8EFO7ma
Keywords: Polyak-Lojasiewicz condition, SGD, interpolation, fast convergence
Compressor summary: This paper introduces a regularity condition for interpolation regime in deep learning, allowing stochastic gradient method to have the same worst-case complexity as deterministic gradient method using only one sampled gradient per iteration.
Richard Gao, Michael Deistler, Jakob H. Macke
https://openreview.net/forum?id=ZARAiV25CW
Keywords: simulation-based inference, generalized bayesian inference, neural network, machine learning for science
Compressor summary: ACE is a new method that uses neural networks to approximate the cost function for generalized Bayesian inference, enabling efficient and robust parameter estimation for complex simulators, especially when the model is misspecified.
Jeffrey Li, Jieyu Zhang, Ludwig Schmidt, Alexander Ratner
https://openreview.net/forum?id=Z8TjsPFBSx
Keywords: Weak Supervision, Semi-supervised Learning, Learning From Limited Labels
Compressor summary: The paper proposes a simple design space to study how semi-supervised learning techniques can improve programmatic weak supervision, finding that simpler methods often perform as well as complex ones, except in certain cases where smaller end models or fewer labeled examples are beneficial.
Seokil Ham, Jungwuk Park, Dong-Jun Han, Jaekyun Moon
https://openreview.net/forum?id=Z7Cz9un2Fy
Keywords: Multi-exit Neural Network, Adversarial Training, Knowledge Distillation, Adversarial Transferability
Compressor summary: NEO-KD is a new adversarial training strategy for multi-exit neural networks that improves robustness against attacks by using neighbor and exit-wise orthogonal knowledge distillation.
Sepidehsadat Hosseini, Mohammad Amin Shabani, Saghar Irandoust, Yasutaka Furukawa
https://openreview.net/forum?id=Z764QxwETf
Keywords: Diffusion, Jigsaw, puzzle solving
Compressor summary: The paper proposes a neural architecture using Diffusion Models for solving jigsaw puzzles of room layouts, and introduces new datasets for training an end-to-end system that outperforms other methods.
Saumya Gupta, Yikai Zhang, Xiaoling Hu, Prateek Prasanna, Chao Chen
https://openreview.net/forum?id=Z6eexoCy7W
Keywords: Topological Representation, Discrete Morse Theory, Structural Uncertainty, Image Segmentation
Compressor summary: This paper proposes a method to estimate uncertainty in the units of topological structures for segmentation tasks, using tools from topological data analysis and a joint prediction model.
Yuxin Wen, John Kirchenbauer, Jonas Geiping, Tom Goldstein
https://openreview.net/forum?id=Z57JrmubNl
Keywords: Diffusion Model, Watermark, Privacy and Security
Compressor summary: Tree-Ring Watermarking is a novel technique for fingerprinting diffusion models' outputs by embedding a pattern in the initial noise vector, making it invisible to humans and resistant to common image manipulations.
Yunzhe Qi, Yikun Ban, Tianxin Wei, Jiaru Zou, Huaxiu Yao, Jingrui He
https://openreview.net/forum?id=Z2L7F0nekb
Keywords: Meta Learning, Contextual Bandits
Compressor summary: The paper introduces BASS, a novel task scheduling framework for meta-learning that optimizes the task schedule based on the meta-model's status and balances exploitation and exploration.
Angela Yuan, Chris Junchi Li, Gauthier Gidel, Michael Jordan, Quanquan Gu, Simon Shaolei Du
https://openreview.net/forum?id=Z28nPtAVxx
Keywords: Stochastic variational inequalities, convex-concave separable saddle-point optimization, extragradient-based algorithm, Nesterov's acceleration, scheduled restarting, scaling reduction
Compressor summary: The paper proposes a new algorithm, AG-EG, for solving stochastic monotone variational inequalities with a separable structure and proves its optimal convergence rate.
Tianjun Ke, Haoqun Cao, Zenan Ling, Feng Zhou
https://openreview.net/forum?id=Z1W0u3Cr74
Keywords: Few-shot learning, Gaussian processes, Conditional conjugate
Compressor summary: The paper proposes a redesigned logistic-softmax likelihood for Bayesian meta-learning that controls the confidence level and improves uncertainty estimates, leading to better results on few-shot classification tasks.
Tolga Ergen, Mert Pilanci
https://openreview.net/forum?id=Z1Aj59LoZD
Keywords: Convex optimization, deep learning theory, path norm, group sparsity, polynomial-time training, ReLU networks, parallel architectures, global optimality, computational complexity
Compressor summary: The authors study deep neural networks and introduce an analytic method to reveal hidden convexity in their optimization landscape, leading to a parsimonious convex model that can be trained efficiently.
Rufeng Xiao, Yuze Ge, Rujun Jiang, Yifan Yan
https://openreview.net/forum?id=Z16jo3d6OD
Keywords: rank-based loss, ADMM, nonconvex nonsmooth optimization, conditional Value-at-Risk, human-aligned risk, ranked range loss
Compressor summary: The paper presents a new optimization algorithm for rank-based loss, which is used in various machine learning models, and shows its effectiveness and efficiency through experiments.
Đorđe Žikelić, Mathias Lechner, Abhinav Verma, Krishnendu Chatterjee, Thomas A Henzinger
https://openreview.net/forum?id=Yx8Sw2H5Q7
Keywords: Verification, Compositional learning
Compressor summary: The authors propose a method to learn neural network policies for stochastic environments with formal guarantees, using SpectRL's logical specifications and reach-avoid supermartingales.
Zijian Zhou, Oluwatosin Alabi, Meng Wei, Tom Vercauteren, Miaojing Shi
https://openreview.net/forum?id=YwgA3avHrP
Keywords: Surgical Instrument Segmentation, Vision Language Models, Text Promptable Segmentation
Compressor summary: The paper presents a text promptable surgical instrument segmentation approach that uses vision-language models and multiple text prompts to improve accuracy and adaptability in minimally invasive surgeries.
YIXUAN ZHANG, Quyu Kong, Feng Zhou
https://openreview.net/forum?id=Yvpenkym8A
Keywords: Spatio-temporal Point Processes, Deep Kernel, Covariate, Integration-free
Compressor summary: The study introduces a new model called Deep Kernel Mixture Point Processes (DKMPP) that uses a deep kernel to better capture relationships between events and multimodal covariate data, and improves efficiency by using score matching methods.
Gongjie Zhang, Jiahao Lin, Shuang Wu, Yilin Song, Zhipeng Luo, Yang Xue, Shijian Lu, Zuoguan Wang
https://openreview.net/forum?id=YvO5yTVv5Y
Keywords: Online HD Map Construction, Map Vectorization, Autonomous Driving, Evaluation Metric, Rasterization, Differentiable Rasterization, Bird's-Eye-View Perception
Compressor summary: The paper introduces MapVR, a novel framework that applies rasterization to vectorized maps for more accurate perception and safer autonomous driving.
Kun-Yu Lin, Jia-Run Du, Yipeng Gao, Jiaming Zhou, Wei-Shi Zheng
https://openreview.net/forum?id=YsZTDcIQwQ
Keywords: video understanding and analysis, video domain generalization
Compressor summary: The paper introduces Spatial-Temporal Diversification Network (STDN), a model that improves video domain generalization by discovering diverse spatial-temporal cues in videos, to defend against heavy reliance on domain-specific cues.
Hu Yu, Jie Huang, Lingzhi Li, Man Zhou, Feng Zhao
https://openreview.net/forum?id=YsYKv95jy9
Keywords: Fractional Fourier Transform, image restoration
Compressor summary: The paper introduces a new spatial-frequency analysis tool, FRFT, which provides comprehensive unified perspectives for image processing and proposes a simple yet effective operator, MFRFC, that significantly improves performance on various computer vision tasks.
Michael Crawshaw, Yajie Bao, Mingrui Liu
https://openreview.net/forum?id=Yq6GKgN3RC
Keywords: federated learning, client subsampling, nonconvex optimization, relaxed smoothness, data heterogeneity, lower bound
Compressor summary: EPISODE++ is a new algorithm for Federated Learning with client subsampling and data heterogeneity that achieves linear speedup, reduces communication rounds, and works well on RNNs for text classification.
Benjamin L. Edelman, Surbhi Goel, Sham M. Kakade, eran malach, Cyril Zhang
https://openreview.net/forum?id=Ypbke6biDm
Keywords: deep learning, feature learning, parity, grokking, lottery tickets, scaling
Compressor summary: This work studies how algorithm choices affect resource tradeoffs in deep learning for sparse parity learning, and shows that width and sparsity improve sample efficiency.
Cheng-Ju Ho, Chen-Hsuan Tai, Yen-Yu Lin, Ming-Hsuan Yang, Yi-Hsuan Tsai
https://openreview.net/forum?id=YoghyvSG0H
Keywords: Semi-supervised learning, 3D object detection, diffusion model
Compressor summary: The paper proposes Diffusion-SS3D, a method that improves semi-supervised 3D object detection by using a diffusion model to denoise corrupted labels and enhance pseudo-label generation in a teacher-student framework.
Xin Li, Dongze Lian, Zhihe Lu, Jiawang Bai, Zhibo Chen, Xinchao Wang
https://openreview.net/forum?id=YmEDnMynuO
Keywords: Efficient transfer learning, vision-language model, adapter-style tuning
Compressor summary: GraphAdapter is an efficient tuning strategy for vision-language models that leverages dual-modality knowledge graphs to capture task-specific structure and inter-class relationships, improving performance on downstream tasks.
Shalev Lifshitz, Keiran Paster, Harris Chan, Jimmy Ba, Sheila A. McIlraith
https://openreview.net/forum?id=YkBDJWerKg
Keywords: minecraft, instruction following, foundation models, sequence models, reinforcement learning, sequential decision making, goal conditioned reinforcement learning, text conditioned reinforcement learning, transformers, deep learning
Compressor summary: The text introduces STEVE-1, a cost-effective AI model that follows instructions in Minecraft by adapting pretrained models and using self-supervised learning techniques.
Hee-Youl Kwak, Dae-Young Yun, Yongjune Kim, Sang-Hyo Kim, Jong-Seon No
https://openreview.net/forum?id=Yj3lFEyfnl
Keywords: Error-floor, Low-density parity-check codes, Boosting learning, Training shcedule, weight sharing, Neural decoders, Min-sum
Compressor summary: The authors propose new training methods for neural min-sum decoders to eliminate the error-floor effect in low-density parity-check codes, achieving better error-floor performance than other methods without extra hardware costs.
Yimin Fan, Fahim Dalvi, Nadir Durrani, Hassan Sajjad
https://openreview.net/forum?id=YiwMpyMdPX
Keywords: Neuron interpretation, NLP, Interpretability, Machine Learning
Compressor summary: The paper proposes a voting-based framework to compare different neuron interpretation methods and finds that they agree on important neurons and focus on last layer representations.
Ilias Diakonikolas, Jelena Diakonikolas, Daniel Kane, Puqian Wang, Nikos Zarifis
https://openreview.net/forum?id=YifKp5b15e
Keywords: PAC Learning, Random Classification Noise
Compressor summary: The paper studies learning general halfspaces with random classification noise and shows an information-computation gap between algorithmic and statistical query complexity results.
Parnian Kassraie, Nicolas Emmenegger, Andreas Krause, Aldo Pacchiano
https://openreview.net/forum?id=YiRX7nQ77Q
Keywords: bandits, model selection, online learning
Compressor summary: ALEXP is a model selection algorithm for bandit optimization that uses online learning with emulated full-information feedback to achieve exponentially improved regret compared to existing methods.
Evangelos Ntavelis, Aliaksandr Siarohin, Kyle Olszewski, Chaoyang Wang, Luc Van Gool, Sergey Tulyakov
https://openreview.net/forum?id=YhAZqWhOnS
Keywords: 3D Generation, Diffusion Models
Compressor summary: The proposed 3D autodecoder approach generates high-quality 3D assets from 2D images or monocular videos using a latent space that embeds properties learned from the target data, achieving superior results compared to existing methods on various benchmarks.
Kiarash Banihashem, MohammadTaghi Hajiaghayi, Suho Shin, Aleksandrs Slivkins
https://openreview.net/forum?id=YeP8osxOht
Keywords: multi-armed bandits, greedy algorithm, social learning, myopic behavior, learning failures, algorithmic game theory
Compressor summary: The paper examines how social learning agents collectively follow a multi-armed bandit protocol but fail to explore due to myopic behaviors.
Yonatan Dukler, Alessandro Achille, Hao Yang, Varsha Vivek, Luca Zancato, Benjamin Bowman, Avinash Ravichandran, Charless Fowlkes, Ashwin Swaminathan, Stefano Soatto
https://openreview.net/forum?id=Ydxnan4P2G
Keywords: Efficient learning, Compute-efficient deep learning, Deep Learning Theory, class-incremental-learning, downstream adaptation
Compressor summary: InCA is a transfer learning framework that uses lightweight cross-attention modules to adapt large models efficiently and identify useful representations for downstream tasks.
Jingwen Fu, Zhizheng Zhang, Dacheng Yin, Yan Lu, Nanning Zheng
https://openreview.net/forum?id=YdfcKb4Wif
Keywords: Generalization, Learning Trajectory
Compressor summary: The paper examines how the learning path of DNNs affects their ability to generalize, proposing a new bound that considers trajectory complexity and training set diversity.
Zhejun Zhang, Alexander Liniger, Christos Sakaridis, Fisher Yu, Luc Van Gool
https://openreview.net/forum?id=YcmGuwdLoU
Keywords: Motion Prediction, Autonomous Driving, Transformer
Compressor summary: The paper introduces a novel attention mechanism (KNARPE) and a hierarchical framework (HPTR) for motion prediction in autonomous driving systems, improving efficiency and scalability while achieving superior performance on public benchmarks.
Qining Zhang, Lei Ying
https://openreview.net/forum?id=Yc9bqbnrbs
Keywords: stochastic multi-armed bandits, regret optimal best arm identification, commitment
Compressor summary: The paper proposes ROBAI, a method for solving a stochastic MAB problem with dual objectives of quick optimal arm identification and reward maximization in a limited number of rounds.
Haotong Qin, Lei Ke, Xudong Ma, Martin Danelljan, Yu-Wing Tai, Chi-Keung Tang, Xianglong Liu, Fisher Yu
https://openreview.net/forum?id=YbYQ0JEQ80
Keywords: Video Matting, Model Binarization, Deep Learning
Compressor summary: BiMatting is a binarized video matting model that overcomes representation degradation and redundant computations, achieving high accuracy, efficiency, and resource savings for real-time video applications.
Timo Schick, Jane Dwivedi-Yu, Roberto Dessi, Roberta Raileanu, Maria Lomeli, Eric Hambro, Luke Zettlemoyer, Nicola Cancedda, Thomas Scialom
https://openreview.net/forum?id=Yacmpz84TH
Keywords: Language Models, Zero-Shot Learning, Tool Use, APIs
Compressor summary: The paper introduces Toolformer, a language model that learns to use external APIs to improve its performance on various tasks.
Jingyuan Sun, Mingxiao Li, Yunhao Zhang, Marie-Francine Moens, Zijiao Chen, Shaonan Wang
https://openreview.net/forum?id=YZSLDEE0mw
Keywords: Neural decoding, brain machine interface, mind reader, visual reconstruction, vision decoding
Compressor summary: The paper introduces a two-phase fMRI representation learning framework that uses denoising and image reconstruction to improve visual decoding from neural responses, achieving superior results compared to previous methods.
Ziang Li, Mengda Yang, Yaxin Liu, Juan Wang, Hongxin Hu, Wenzhe Yi, Xiaoyang Xu
https://openreview.net/forum?id=YZGWhs1H7F
Keywords: deep learning, split inference, data reconstruction attack
Compressor summary: GLASS is a novel GAN-based attack that effectively reconstructs private data from Split Inference models, while GLASS++ enhances its stability and both are resistant to common defense mechanisms.
Anqi Mao, Mehryar Mohri, Yutao Zhong
https://openreview.net/forum?id=YZ7ip645Ra
Keywords: structured prediction, consistency, learning theory, natural language processing
Compressor summary: The authors study surrogate losses for structured prediction and propose new ones with better learning guarantees and efficient algorithms.
Rajat Vadiraj Dwaraknath, Tolga Ergen, Mert Pilanci
https://openreview.net/forum?id=YWsPN0EMZr
Keywords: neural tangent kernel, NTK, ReLU activations, neural networks, gated ReLU, convex optimization, kernel, multiple kernel learning, MKL, group lasso, iterative reweighting, group norm
Compressor summary: This paper connects neural network training with gradient flow to multiple kernel learning using a gated ReLU network and shows how to improve the weights for better predictions.
Jeffrey Ouyang-Zhang, Daniel Jesus Diaz, Adam Klivans, Philipp Kraehenbuehl
https://openreview.net/forum?id=YWSOpYjyG4
Keywords: stability, proteins, biology, physical
Compressor summary: The authors propose a deep learning algorithm called Mutate Everything that predicts the effect of mutations on protein stability quickly and accurately, using existing models as its base.
Dan Qiao, Yu-Xiang Wang
https://openreview.net/forum?id=YVMc3KiWBQ
Keywords: Differential privacy, offline reinforcement learning, reinforcement learning theory
Compressor summary: The paper proposes offline reinforcement learning algorithms with differential privacy guarantees that protect sensitive information while maintaining good learning performance and utility.
Sizhe Yang, Yanjie Ze, Huazhe Xu
https://openreview.net/forum?id=YV1MYtj2AR
Keywords: visual reinforcement learning, visual generalization
Compressor summary: The paper proposes MoVie, a method to improve visual reinforcement learning agents' ability to generalize to unseen views without explicit rewards, achieving significant improvements in 18 real-world robotics tasks.
Emaad Khwaja, Yun S. Song, Aaron Agarunov, Bo Huang
https://openreview.net/forum?id=YSMLVffl5u
Keywords: text-to-image, protein localization, protein engineering, transformers
Compressor summary: CELL-E 2 is a transformer that can generate and design protein subcellular localization images and sequences from amino acid sequences, capturing spatial complexity and producing NLS.
Yang Sui, Xin HE, Yang Bai
https://openreview.net/forum?id=YSFQRVkkl0
Keywords: Over-parameterization, SVM, Sparsity, Lasso
Compressor summary: The paper proposes a regularization-free algorithm for high-dimensional SVMs using over-parameterization and Nesterov's method, which improves computational efficiency and achieves near-oracle convergence rate.
Yining Hong, Haoyu Zhen, Peihao Chen, Shuhong Zheng, Yilun Du, Zhenfang Chen, Chuang Gan
https://openreview.net/forum?id=YQA28p7qNz
Keywords: 3D Visual Reasoning, 3D Large Language Model
Compressor summary: The paper introduces 3D-LLMs, a new type of language models that can perform various 3D tasks using point clouds and features. The paper collects large amounts of 3D-language data and uses a novel training method to achieve superior performance on several benchmarks.
Zicheng Liu, Siyuan Li, Ge Wang, Lirong Wu, Cheng Tan, Stan Z. Li
https://openreview.net/forum?id=YPQg2RTFD8
Keywords: mixup, data augmentation, classification, data efficiency
Compressor summary: Decoupled mixup is an efficient data augmentation method that smooths decision boundaries and mines discriminative features without extra computation, outperforming dynamic mixup methods.
Matthew Lyon, Paul Armitage, Mauricio A Álvarez
https://openreview.net/forum?id=YPHIrNKI0d
Keywords: Diffusion MRI, super-resolution, image synthesis, conditional image synthesis, continuous convolution, parametric continuous convolution
Compressor summary: The paragraph introduces a novel method for improving diffusion MRI resolution using a parametric continuous convolution network (PCCNN) that performs well and requires fewer parameters than existing models.
Christopher T.H Teo, Milad Abdollahzadeh, Ngai-man Cheung
https://openreview.net/forum?id=YOZaej0ZC7
Keywords: Fairness, Generative models, GAN, Calibration
Compressor summary: This paper presents CLEAM, a new framework to measure fairness in generative models that reduces errors and has minimal overhead.
Suraj Srinivas, Sebastian Bordt, Himabindu Lakkaraju
https://openreview.net/forum?id=YMMlHBSQdC
Keywords: robustness, generative models, perceptually aligned gradients, bayes optimality, manifold hypothesis
Compressor summary: This paper explains how perceptually-aligned gradients (PAGs) in robust computer vision models arise from off-manifold robustness and identifies three regimes of robustness affecting both alignment and accuracy.
Chenwei Wu, Holden Lee, Rong Ge
https://openreview.net/forum?id=YLOJ4aKAka
Keywords: language model representation, downstream performance, deep learning theory
Compressor summary: The paper investigates how pre-trained language models' representations depend on downstream tasks' properties and structures, and proposes an "anchor vector" concept to improve performance transfer understanding.
Yuqi Chen, Kan Ren, Yansen Wang, Yuchen Fang, Weiwei Sun, Dongsheng Li
https://openreview.net/forum?id=YJDz4F2AZu
Keywords: Irregular Time Series Modeling, Transformer, Neural Ordinary Differential Equation
Compressor summary: ContiFormer combines Neural ODEs and Transformers to model continuous-time dynamics on irregular time series, capturing both input relationships and dynamic changes.
PRASENJIT DEY, Srujana Merugu, Sivaramakrishnan R Kaveri
https://openreview.net/forum?id=YI4bn6aAmz
Keywords: Ordinal Classification, Conformal Predictions, Unimodal modelling
Compressor summary: The paper proposes a framework to adapt conformal prediction methods for ordinal classification, using a non-parametric approach to generate contiguous sets with guaranteed coverage and minimal cardinality.
Yao Ni, Piotr Koniusz
https://openreview.net/forum?id=YFW6MVGVTn
Keywords: Image Generation, limited dataset, Generative Adversarial Networks
Compressor summary: NICE is a novel approach that uses adaptive noise to modulate the discriminator's latent features, preventing overfitting and improving GAN training stability on limited data.
Feiyang Wu, Jingyang Ke, Anqi Wu
https://openreview.net/forum?id=YFSrf8aciU
Keywords: Machine Learning, Reinforcement Learning, Inverse Reinforcement Learning, Markov Decision Process, stochastic optimization, complexity analysis
Compressor summary: This paper presents a novel inverse reinforcement learning method that works under an average-reward setting and has lower complexity than previous methods, along with experiments on various tasks.
Alessio Russo, Alexandre Proutiere
https://openreview.net/forum?id=YEtstXIpP3
Keywords: reinforcement learning; best policy identification; model free; exploration; sample complexity
Compressor summary: The paper proposes a new model-free exploration method for Reinforcement Learning that can find near-optimal policies more quickly than existing methods.
Adel Nabli, Eugene Belilovsky, Edouard Oyallon
https://openreview.net/forum?id=YE04aRkeZb
Keywords: Decentralized Optimization for Deep Learning, Asynchronous Optimization, Distributed Training, Data-Parallel
Compressor summary: The paper introduces a new gossip-based optimization algorithm called $\textbf{A}^2\textbf{CiD}^2$ that speeds up distributed training of deep learning models by using continuous local momentum and reducing idle time.
Stratis Tsirtsis, Manuel Gomez Rodriguez
https://openreview.net/forum?id=YDCpf85eXc
Keywords: Counterfactual reasoning, Markov decision process, Structural causal model, A* search
Compressor summary: The paper presents a new method to analyze continuous state environments for retrospective analysis of treatment decisions using causal inference and reinforcement learning techniques.
Jiayu Chen, Vaneet Aggarwal, Tian Lan
https://openreview.net/forum?id=Y8p3ThNDmK
Keywords: Reinforcement Learning, Unsupervised Skill Discovery, Determinantal Point Process, Options
Compressor summary: The paper proposes a new algorithm that combines diversity and coverage in unsupervised option discovery using Determinantal Point Process, and shows its effectiveness on Mujoco and Atari tasks.
Xinyi Xu, Thanh Lam, Chuan-Sheng Foo, Bryan Kian Hsiang Low
https://openreview.net/forum?id=Y6IGTNMdLT
Keywords: Model Valuation, Dirichlet Abstraction, Shapley Value
Compressor summary: The paper proposes a new method for valuing machine learning models in AI marketplaces, called model Shapley, which is fair and can be used to predict the values of many models efficiently.
Zhongxiang Dai, Gregory Kang Ruey Lau, Arun Verma, Yao Shu, Bryan Kian Hsiang Low, Patrick Jaillet
https://openreview.net/forum?id=Y44NurSDjq
Keywords: quantum bandits, kernelized bandits
Compressor summary: The Q-GP-UCB algorithm optimizes non-linear reward functions with quantum computing, achieving better regret upper bounds than classical methods and previous quantum approaches.
Yefei He, Luping Liu, Jing Liu, Weijia Wu, Hong Zhou, Bohan Zhuang
https://openreview.net/forum?id=Y3g1PV5R9l
Keywords: Diffusion models, Post-training quantization, Mixed precision
Compressor summary: The paragraph discusses a new post-training quantization method for diffusion models that reduces computational costs and improves sample quality by correcting quantization noise and adjusting bitwidths for each denoising step.
Yuki Kawana, Tatsuya Harada
https://openreview.net/forum?id=Y3NjoeO4Q1
Keywords: articulated objects, shape reconstruction, 3D reconstruction
Compressor summary: The paper presents a novel method for reconstructing multiple man-made articulated objects from a single RGBD image using part-level representation and proposes three techniques to overcome challenges, achieving better shape reconstruction and kinematics estimation on synthetic and real data.
Arvind Venkat Mahankali, Jeff Z. HaoChen, Kefan Dong, Margalit Glasgow, Tengyu Ma
https://openreview.net/forum?id=Y2hnMZvVDm
Keywords: Nonconvex Optimization, Mean-Field Analysis, Beyond NTK, Deep Learning Theory
Compressor summary: The paper analyzes projected gradient flow on two-layer neural networks without modifications and shows they achieve better sample complexity than kernel methods.
Sara Sangalli, Ertunc Erdil, Ender Konukoglu
https://openreview.net/forum?id=Y2VQWfi7Vc
Keywords: human-ai collaboration system, optimization
Compressor summary: The paper proposes a new loss function for training deep neural networks that considers both model accuracy and expert load, aiming to improve classification and reduce human intervention in critical applications.
Mohammad Pedramfar, Christopher John Quinn, Vaneet Aggarwal
https://openreview.net/forum?id=Y1sJJW3pID
Keywords: Stochastic optimization, submodular maximization, Frank-Wolfe algorithm
Compressor summary: The paper proposes a unified method to maximize continuous DR-submodular functions across various settings and oracle access types, improving or matching existing results in most cases.
Paul Rosa, Viacheslav Borovitskiy, Alexander Terenin, Judith Rousseau
https://openreview.net/forum?id=Y18r0xWkSh
Keywords: Gaussian processes, posterior contraction, manifolds, kernels
Compressor summary: The paper investigates whether intrinsic models of Gaussian processes on Riemannian manifolds lead to better performance than Euclidean ones, by proving optimal contraction rates for both types and showing empirical results that support the theoretical findings.
Yuchuan Tian, Hanting Chen, Tianyu Guo, Chao Xu, Yunhe Wang
https://openreview.net/forum?id=Y17N9B0vXn
Keywords: Weight Pruning, Matrix Rank
Compressor summary: The paper introduces RPG, a pruning method for CNNs that maintains the ranks of sparse weights in an adversarial manner to improve performance on edge devices with high sparsity.
Bonwoo Lee, Jeongyoun Ahn, Cheolwoo Park
https://openreview.net/forum?id=XzTM9gVRT4
Keywords: Data privacy, Functional local differential privacy, Gaussian mechanism, Minimax risks, Statistical utility
Compressor summary: This paragraph discusses how functional differential privacy (DP), a type of local DP, balances data privacy and utility by analyzing its performance in estimating mean and density, and suggests it as a reliable standard for local DP.
Matthew Chang, Aditya Prakash, Saurabh Gupta
https://openreview.net/forum?id=Xyj46OxEhK
Keywords: Inpainting, Diffusion, Robot Learning, Egocentric Vision
Compressor summary: This paper presents a method to separate human hand and environment in egocentric videos using a diffusion model and attention, improving performance on various robotics tasks.
Haoran Yang, Xiangyu Zhao, Yicong Li, Hongxu Chen, Guandong Xu
https://openreview.net/forum?id=XyAP8ScqLV
Keywords: graph contrastive learning, prompt tuning, recommendation system
Compressor summary: CPTPP is a novel framework for graph learning that uses personalized prompts to bridge the gap between pre-training and downstream tasks, achieving better results and modeling user preferences more effectively.
Denis Kuznedelev, Eldar Kurtic, Elias Frantar, Dan Alistarh
https://openreview.net/forum?id=Xy7DoWSNZX
Keywords: neural network pruning, vision transformer, sparsity, model compression
Compressor summary: The Correlation Aware Pruner (CAP) is a new framework that significantly improves the compressibility of state-of-the-art computer vision models by handling complex weight correlations and providing efficient post-compression recovery.
Huikang Liu, Xiao Li, Anthony Man-Cho So
https://openreview.net/forum?id=Xxllzjt6T5
Keywords: Manifold optimization, Riemannian subgradient method, rotation synchronization
Compressor summary: ReSync is a Riemannian subgradient algorithm for synchronizing robust rotations in various engineering applications, with strong theoretical guarantees and experimentally proven effectiveness.
Puhua Jiang, Mingze Sun, Ruqi Huang
https://openreview.net/forum?id=XvfEYqEbIb
Keywords: shape registration; functional maps; unsupervised learning
Compressor summary: The paper introduces a learning-based method for registering non-rigid shapes without correspondence supervision using deep functional maps and a trained orientation regressor, achieving state-of-the-art results on several benchmarks.
Liang Yang, Runjie Shi, Qiuliang Zhang, Bingxin Niu, Zhen Wang, Xiaochun Cao, Chuan Wang
https://openreview.net/forum?id=XvGQ6F3sG8
Keywords: Graph neural network, Self-supervised learning, Low-Rank recovery
Compressor summary: This paper proposes Low-Rank Decomposition-based methods for graph neural networks to address issues with propagation-based approaches and improve local and long-distance node representation.
Weixi Feng, Wanrong Zhu, Tsu-Jui Fu, Varun Jampani, Arjun Reddy Akula, Xuehai He, S Basu, Xin Eric Wang, William Yang Wang
https://openreview.net/forum?id=Xu8aG5Q8M3
Keywords: Large Language Models, Compositional Image Generation, 3D scene synthesis
Compressor summary: The paragraph discusses LayoutGPT, a method that uses LLMs to generate layouts from text inputs and enhances the visual planning skills of LLMs, achieving superior performance in text-to-image generation and 3D indoor scene synthesis.
Anton Voronov, Mikhail Khoroshikh, Artem Babenko, Max Ryabinin
https://openreview.net/forum?id=Xs6Xwc0Glj
Keywords: text-to-image generation, diffusion models, early stopping
Compressor summary: The researchers propose a simple early stopping criterion that speeds up text-to-image adaptation by eight times without sacrificing quality.
Jiachen T. Wang, Saeed Mahloujifar, Tong Wu, Ruoxi Jia, Prateek Mittal
https://openreview.net/forum?id=XrqqPDAsRE
Keywords: Differential Privacy; Privacy Accounting
Compressor summary: The paper introduces a new differential privacy approach (EVR) that verifies and releases query outputs based on an estimated privacy parameter, using Monte Carlo techniques to improve accuracy and efficiency in privacy accounting.
Doyup Lee, Chiheon Kim, Minsu Cho, Wook-Shin Han
https://openreview.net/forum?id=XqcXf7ix5q
Keywords: implicit neural representations, representation learning, neural fields
Compressor summary: The proposed framework combines a transformer encoder with a locality-aware INR decoder to learn generalizable implicit neural representations that capture fine-grained details in spatial and spectral aspects, improving performance on downstream tasks like image generation.
Haoran Chen, Xintong Han, Zuxuan Wu, Yu-Gang Jiang
https://openreview.net/forum?id=Xq2s5yxzd2
Keywords: multi source unsupervised domain adaptation; transfer learning; computer vision
Compressor summary: The text introduces Multi-Prompt Alignment (MPA), a method for unsupervised domain adaptation that uses prompt learning and auto-encoding to align source and target domains, achieving state-of-the-art results.
Seokin Seo, HyeongJoo Hwang, Hongseok Yang, Kee-Eung Kim
https://openreview.net/forum?id=XpmJNP8BVA
Keywords: Imitation learning, Information leakage, Causal Confusion
Compressor summary: The paper proposes a new regularization method called Past Action Leakage Regularization (PALR) for behavior cloning in offline imitation learning, which uses conditional independence to reduce the leakage of past actions into observation histories and improves performance.
Ilias Diakonikolas, Daniel Kane, Lisheng Ren, Yuxin Sun
https://openreview.net/forum?id=Xp68yXQiRk
Keywords: Non-Gaussian Component Analysis
Compressor summary: The paper investigates whether the chi-squared condition is necessary for proving SQ lower bounds for Non-Gaussian Component Analysis and shows that it is not.
Zeyinzi Jiang, Chaojie Mao, Ziyuan Huang, Ao Ma, Yiliang Lv, Yujun Shen, Deli Zhao, Jingren Zhou
https://openreview.net/forum?id=XmpthbaJql
Keywords: Parameter-efficient Transfer Learning, Memory-efficient Transfer Learning, Residual Network, Vision Transformer, Foundation Model
Compressor summary: Res-Tuning is a new tuning paradigm that separates tuners from the backbone, allowing for flexible combination of various tuning strategies and improved efficiency in foundation models.
Alexander Borzunov, Max Ryabinin, Artem Chumachenko, Dmitry Baranchuk, Tim Dettmers, Younes Belkada, Pavel Samygin, Colin Raffel
https://openreview.net/forum?id=XmN7ZNbUAe
Keywords: volunteer computing, distributed deep learning, distributed inference, efficient inference, large language models
Compressor summary: The authors propose methods to efficiently run large language models on geodistributed devices, addressing reliability and load-balancing issues, and demonstrating their approach with Petals, a decentralized system that speeds up inference and fine-tuning of LLMs.
Sudhanshu Chanpuriya, Ryan A. Rossi, Anup Rao, Tung Mai, Nedim Lipka, Zhao Song, Cameron N Musco
https://openreview.net/forum?id=XlvsieCnAX
Keywords: graph, network, embeddings, arboricity, factorization, model, community, nonnegative
Compressor summary: The paragraph describes a new graph factorization model that captures both similar and dissimilar links between nodes and can represent low arboricity graphs with good performance on community detection and link prediction tasks.
Xin Zheng, Miao Zhang, Chunyang Chen, Quoc Viet Hung Nguyen, Xingquan Zhu, Shirui Pan
https://openreview.net/forum?id=XkcufOcgUc
Keywords: graph neural networks (GNNs), graph condensation, training trajectory meta-matching, graph neural feature score
Compressor summary: The paper introduces SFGC, a new method for reducing the size of large-scale graphs by synthesizing small-scale graph-free data that implicitly encodes topology structure information without explicit graph structures, and evaluates its effectiveness and generalization ability through two components: meta-matching and graph neural feature score.
Antonio Norelli, Marco Fumero, Valentino Maiorca, Luca Moschella, Emanuele Rodolà, Francesco Locatello
https://openreview.net/forum?id=XjOj3ZmWEl
Keywords: Representation learning, Multimodal models, Analogy, Sparsity, Nonparametric, Relative representations, Language, Semiotics
Compressor summary: The paper proposes a new method to create a common space between visual and language domains without training, using single-domain encoders and fewer image-text pairs, and shows its advantages in interpretability, fast deployment, and zero-shot transfer ability.
Zhongxiang Dai, Quoc Phong Nguyen, Sebastian Shenghong Tay, Daisuke Urano, Richalynn Leong, Bryan Kian Hsiang Low, Patrick Jaillet
https://openreview.net/forum?id=Xj4LJiXvlX
Keywords: Bayesian optimization, Gaussian processes, AI4Science
Compressor summary: The BTS-RED framework proposes three algorithms to tackle trade-offs in experimental design problems with large, heteroscedastic noise by adaptively choosing replications and balancing average performance and variability.
Botao WANG, Jia Li, Yang Liu, Jiashun Cheng, Yu Rong, Wenjia Wang, Fugee Tsung
https://openreview.net/forum?id=XhNlBvb4XV
Keywords: Pseudo labeling, Graph data, Error analysis, Cautious
Compressor summary: Pseudo labeling is a technique to expand labeled data for graph learning models, but it can introduce errors; this paper analyzes these errors and proposes a cautious methodology to improve performance.
Khaled Eldowa, Emmanuel Esposito, Tommaso Cesari, Nicolò Cesa-Bianchi
https://openreview.net/forum?id=XfYpIaKDb6
Keywords: Online learning, Feedback graphs, Multiarmed bandits
Compressor summary: The paper presents an improved upper and lower bound for the regret of online learning with different types of feedback graphs, using FTRL with $q$-Tsallis entropy and new techniques for time-varying graphs.
Hui En Pang, Zhongang Cai, Lei Yang, Qingyi Tao, Zhonghua Wu, Tianwei Zhang, Ziwei Liu
https://openreview.net/forum?id=XfKnoW4Zef
Keywords: Whole-body, SMPLX Model, Human Pose and Shape Estimation, Human Mesh Recovery
Compressor summary: The paper proposes a new framework to improve whole-body pose and shape estimation by addressing challenges related to bounding box quality, robustness to augmentations, and pixel alignment.
NIRANJAN DAMERA VENKATA, Chiranjib Bhattacharyya
https://openreview.net/forum?id=XetXfkYZ6i
Keywords: optimal stopping, recurrent neural networks, probabilistic graphical models, policy gradient methods
Compressor summary: The paper introduces an OSPG algorithm that uses RNNs to optimize value functions in non-Markovian optimal stopping problems without recursion, reducing the curse of non-Markovianity and the need for costly Monte Carlo simulations.
Zheng Chen, Yulun Zhang, Ding Liu, Bin Xia, Jinjin Gu, Linghe Kong, Xin Yuan
https://openreview.net/forum?id=XeMryhpniy
Keywords: image deblurring, diffusion model
Compressor summary: The Hierarchical Integration Diffusion Model (HI-Diff) is a new image deblurring method that uses a compact latent space and multiple scales to improve efficiency, accuracy, and generalization compared to previous diffusion models.
Shuyi Li, Michael O'Connor, Shiwei Lan
https://openreview.net/forum?id=XddoUFpjkP
Keywords: Functional Regularization, Besov Process, $Q$-Exponential Distribution, Elliptic Contour Distribution
Compressor summary: The paper introduces a new stochastic process, Q-exponential (Q-EP), for high-dimensional optimization and modeling that generalizes the $q$-exponential distribution and improves upon Gaussian process in terms of penalty strength and flexibility.
Tianyu Pang, Cheng Lu, Chao Du, Min Lin, Shuicheng YAN, Zhijie Deng
https://openreview.net/forum?id=XcQzXeF7fX
Keywords: Diffusion Probabilistic Models, Model Calibration
Compressor summary: The paper proposes a calibration method for diffusion probabilistic models that reduces the score matching loss and increases the model likelihood, with experiments on various datasets to validate the approach.
Cheng-Hao Tu, Hong-You Chen, Zheda Mai, Jike Zhong, Vardaan Pahuja, Tanya Berger-Wolf, Song Gao, Charles Stewart, Yu Su, Wei-Lun Chao
https://openreview.net/forum?id=XbVnNXaIQY
Keywords: Fine-tuning, Transfer learning, Domain adaptation, Continual learning, Robustness, Personalization
Compressor summary: The paragraph discusses a learning problem involving adapting a model to a new domain using limited data and presents challenges and solutions to improve classification accuracy for all classes.
Aditya Vora, Akshay Gadi Patil, Hao Zhang
https://openreview.net/forum?id=XbInLmYLDr
Keywords: Multi-view Neural 3D Reconstruction, Sparse and Disparate Views, Neural Rendering, Volume Rendering
Compressor summary: DiViNet is a method that uses neural templates to reconstruct 3D surface details from few RGB images by regularizing the ill-posed problem.
Sang Keun Choe, Sanket Vaibhav Mehta, Hwijeen Ahn, Willie Neiswanger, Pengtao Xie, Emma Strubell, Eric Xing
https://openreview.net/forum?id=Xazhn0JoNx
Keywords: meta learning, bilevel optimization, large-scale learning, implicit differentiation
Compressor summary: SAMA is a meta learning framework that improves scalability, reduces computational burden, and supports adaptive optimizers while achieving state-of-the-art results in text and image classification tasks.
Yue Yu, Xiao Wang, Mengmei Zhang, Nian Liu, Chuan Shi
https://openreview.net/forum?id=Xasl21tSOf
Keywords: Graph Contrastive Learning, Graph Neural Networks, Bound Propagation
Compressor summary: The paper proposes a metric called "node compactness" and a regularization method named PrOvable Training (POT) to improve Graph Contrastive Learning by addressing its imbalanced training across nodes.
Seungjoo Shin, Jaesik Park
https://openreview.net/forum?id=XY6BnwIh4q
Keywords: neural radiance fields, inverse rendering, binarization
Compressor summary: The paper introduces BiRF, a compact way to store radiance fields using binary encoding and 2D-3D grids, which outperforms existing efficient models in static scene reconstruction with minimal storage space.
Rui Wang, Pei Pei Li, Huaibo Huang, Chunshui Cao, Ran He, Zhaofeng He
https://openreview.net/forum?id=XXagS1RQH0
Keywords: Ordinal Classification, Representation Learning, Vision-Language, Prompt Learning
Compressor summary: The paper introduces L2RCLIP, a method that uses language-driven ordering alignment and vision-language models to improve ordinal classification tasks such as age estimation, HCI classification, and aesthetic assessment.
Franziska Boenisch, Christopher Mühl, Adam Dziedzic, Roy Rinberg, Nicolas Papernot
https://openreview.net/forum?id=XXPzBhOs4f
Keywords: privacy, machine learning, differential privacy, DP-SGD, individualized privacy
Compressor summary: The paper proposes Individualized DP-SGD (IDP-SGD), a variant of Differentially Private Stochastic Gradient Descent that allows for tailored privacy budgets based on individual user preferences, improving privacy-utility trade-offs.
Alaa Maalouf, Murad Tukan, Noel Loo, Ramin Hasani, Mathias Lechner, Daniela Rus
https://openreview.net/forum?id=XWYv4BNShP
Keywords: Dataset Distillation, Size and Approximation Error
Compressor summary: The paper analyzes the theoretical limitations and guarantees of dataset distillation using kernel ridge regression methods and provides the first proof of existence of small distilled datasets with corresponding excess risk for shift-invariant kernels.
Byeongchan Lee, Sehyun Lee
https://openreview.net/forum?id=XUu2GloTXb
Keywords: representation learning, self-supervised learning, contrastive learning
Compressor summary: The paper introduces a novel method for self-supervised representation learning using an asymmetric network architecture, which stabilizes training, boosts performance, and works well with small batch sizes and no predictor.
Elisa Nguyen, Minjoon Seo, Seong Joon Oh
https://openreview.net/forum?id=XSCYxDp3yE
Keywords: training data attribution, interpretability, explainability, data-driven xai
Compressor summary: This paper introduces a Bayesian perspective on training data attribution techniques for deep models, highlighting the challenges of noise and suggesting to only use TDA for predictions consistently influenced by specific training data.
Hao Wang, Luxi He, Rui Gao, Flavio Calmon
https://openreview.net/forum?id=XRy4YQYLe0
Keywords: information theory, fair machine learning
Compressor summary: The text discusses two types of discrimination in ML models and provides methods to measure and reduce them, especially when dealing with missing data.
Jonathan Pilault, Mahan Fathi, Orhan Firat, Christopher Pal, Pierre-Luc Bacon, Ross Goroshin
https://openreview.net/forum?id=XRTxIBs2eu
Keywords: State Space Models, Efficient Transformers, Long Range Language Modeling, Language Modeling
Compressor summary: The Block-State Transformer combines an SSM sublayer for long-range context and a Block Transformer sublayer for short-term sequence representation, improving language modeling perplexity and speed on parallelizable architectures.
Guan-Horng Liu, Tianrong Chen, Evangelos Theodorou, Molei Tao
https://openreview.net/forum?id=XPWEtXzlLy
Keywords: diffusion models, constrained generation, constrained manifold, mirror map, watermarked generation, generation privacy
Compressor summary: The paper proposes Mirror Diffusion Models, a new type of diffusion models that can generate data on constrained sets without losing tractability, by using a mirror map in a standard Euclidean space and embedding invisible information for safety and privacy purposes.
Lihe Yang, Xiaogang Xu, Bingyi Kang, Yinghuan Shi, Hengshuang Zhao
https://openreview.net/forum?id=XOotfgPiUF
Keywords: learning from synthetic, semantic segmentation, generative models
Compressor summary: FreeMask uses synthetic images generated by generative models to ease the data collection and annotation process for semantic segmentation, achieving comparable or better performance than real images.
Kavosh Asadi, Shoham Sabach, Yao Liu, Omer Gottesman, Rasool Fakoor
https://openreview.net/forum?id=XOCbdqxAR2
Keywords: Reinforcement Learning, Temporal Difference Learning, Value Function Optimization, Convergence
Compressor summary: The paper analyzes how the temporal-difference learning algorithm works as an iterative optimization process, identifies factors that affect its convergence or divergence, and proves its convergence in different settings, explaining its success in reinforcement learning.
Yanfang Xue, Pengfei Fang, Jinyue Tian, Shipeng Zhu, hui xue
https://openreview.net/forum?id=XNBeTgYcAq
Keywords: Spectral kernel; complex-valued networks
Compressor summary: The CosNet is a generalized spectral kernel network that preserves the inherent complex-valued representation in time-sequential data and improves the analysis of time-varying statistical characteristics by combining complex-valued spectral kernels with neural networks.
Haoxing Chen, Zhuoer Xu, Zhangxuan Gu, jun lan, 行 郑, Yaohui Li, Changhua Meng, Huijia Zhu, Weiqiang Wang
https://openreview.net/forum?id=XKeSauhUdJ
Keywords: Diffusion model, text editing, self-supervied learning
Compressor summary: The paper proposes DiffUTE, a self-supervised text editing diffusion model that can draw multilingual characters and edit images realistically using web data.
Rachael Hwee Ling Sim, Yehong Zhang, Trong Nghia Hoang, Xinyi Xu, Bryan Kian Hsiang Low, Patrick Jaillet
https://openreview.net/forum?id=XKP3mAsNHd
Keywords: Incentives, Privacy, Shapley fairness, Collaborative machine learning, data valuation, reward, sufficient statistics
Compressor summary: The authors propose a method for collaborative machine learning that incentivizes data sharing while preserving privacy by using differential privacy and adjusting sufficient statistics.
Litu Rout, Negin Raoof, Giannis Daras, Constantine Caramanis, Alex Dimakis, Sanjay Shakkottai
https://openreview.net/forum?id=XKBFdYwfRo
Keywords: Inverse Problems, Posterior Sampling, Latent Diffusion Model, Stable Diffusion, Sample Recovery
Compressor summary: The framework uses pre-trained latent diffusion models to solve linear inverse problems and shows superior performance in various tasks like image restoration and enhancement.
Arpit Bansal, Eitan Borgnia, Hong-Min Chu, Jie S. Li, Hamid Kazemi, Furong Huang, Micah Goldblum, Jonas Geiping, Tom Goldstein
https://openreview.net/forum?id=XH3ArccntI
Keywords: Generative Models, Computer Vision, Diffusion Models
Compressor summary: Deterministic image degradations can be used to create a family of generative diffusion models without the need for added noise.
Yukun Huang, Jianan Wang, Ailing Zeng, He CAO, Xianbiao Qi, Yukai Shi, Zheng-Jun Zha, Lei Zhang
https://openreview.net/forum?id=XGXL1E8Yyo
Keywords: Avatar Generation, 3D Content Creation, NeRF, Diffusion Model
Compressor summary: DreamWaltz is a framework that generates and animates complex 3D avatars from text guidance using novel techniques like SDS, 3D-aware skeleton conditioning, and image priors.
Jungbin Kim, Insoon Yang
https://openreview.net/forum?id=XFE6zpevLc
Keywords: convex optimization, accelerated gradient methods
Compressor summary: The paper introduces a new way to analyze optimization methods using functional analysis tools that leads to faster convergence rates and reveals a link between function values and gradient norms.
Fengzhuo Zhang, Vincent Tan, Zhaoran Wang, Zhuoran Yang
https://openreview.net/forum?id=XF923QPCGw
Keywords: mean-field approximation, graphon games, multi-agent reinforcement learning
Compressor summary: The paper explores the existence and efficient learning of Nash Equilibria in regularized Graphon Mean-Field Games, extending previous results and proposing new algorithms for discrete-time learning under weak monotonicity conditions.
Ege Beyazit, Jonathan Kozaczuk, Bo Li, Vanessa Wallace, Bilal H Fadlallah
https://openreview.net/forum?id=XEUc1JegGt
Keywords: Tabular Deep Learning, Spectral Bias, Neural Networks
Compressor summary: The paper proposes a new neural network layer that helps improve deep learning performance on irregularly shaped tabular data by learning low-frequency representations of the input features.
Jinggang Chen, Junjie Li, Xiaoyang Qu, Jianzong Wang, Jiguang Wan, Jing Xiao
https://openreview.net/forum?id=XEBzQP3e7B
Keywords: out-of-distribution detection, distribution shifts, attribution gradients
Compressor summary: The paper proposes a method called GAIA that uses gradient abnormality to detect out-of-distribution examples in deep neural networks, improving their reliability and safety.
Yichi Zhang, Ankush Garg, Yuan Cao, Lukasz Lew, Behrooz Ghorbani, Zhiru Zhang, Orhan Firat
https://openreview.net/forum?id=XAyPlfmWpu
Keywords: neural network quantization, binarized transformer, machine translation, scaling law
Compressor summary: The paper introduces BMT, a novel binarization technique for Transformers that achieves the same quality as float models while being much smaller, by using additional LayerNorms and residual connections to address one-bit weight variance issues.
Sen Lin, Daouda Sow, Kaiyi Ji, Yingbin Liang, Ness Shroff
https://openreview.net/forum?id=X9Vjq9Fuhq
Keywords: Bilevel Optimization, Time-Varying Functions, Single-Loop, Sublinear Bilevel Local Regret
Compressor summary: SOBOW is an efficient online bilevel optimizer with window averaging that works well for streaming data and time-varying functions, achieving sublinear regret.
Jie Xu, Shuo Chen, Yazhou Ren, Xiaoshuang Shi, Heng Tao Shen, Gang Niu, Xiaofeng Zhu
https://openreview.net/forum?id=X8dbFcAox2
Keywords: Multi-view learning, Contrastive learning, Representation degeneration, Self-supervised learning
Compressor summary: SEM is a new contrastive learning framework for multi-view scenarios that adaptively strengthens useful views and weakens unreliable ones, while improving representation quality with reconstruction regularization.
Mohammad Reza Karimi Jaghargh, Ya-Ping Hsieh, Andreas Krause
https://openreview.net/forum?id=X6mwdEVYvc
Keywords: Stochastic Approximation, Mean-Field Dynamics, Dynamical Systems, Neural Networks, Sampling
Compressor summary: The paper proposes a new framework to connect discrete-time interacting particle systems and their mean-field limits, improving the analysis of these systems in machine learning tasks.
Karthik Valmeekam, Matthew Marquez, Sarath Sreedharan, Subbarao Kambhampati
https://openreview.net/forum?id=X6dEqXIsEW
Keywords: Large Language Models, Planning, LLMs for autonomous and heuristic planning guidance
Compressor summary: The paper investigates how well large language models (LLMs) can generate plans autonomously and as heuristic guidance for other planners, finding limited success in the former but more promise in the latter.
Ao Sun, Pingchuan Ma, Yuanyuan Yuan, Shuai Wang
https://openreview.net/forum?id=X6TBBsz9qi
Keywords: EXplainable AI, Machine Learning, Computer Vision
Compressor summary: The paper introduces Explain Any Concept (EAC), a method that uses the Segment Anything Model (SAM) to explain deep neural network decisions with any concept, improving human understanding of computer vision tasks.
Sebastian Zeng, Florian Graf, Roland Kwitt
https://openreview.net/forum?id=X6Eapo5paw
Keywords: Variational Bayesian inference, stochastic differential equation, homogeneous spaces, geometric Euler-Maruyama, time series
Compressor summary: The authors study variational Bayesian inference for latent stochastic differential equations (SDEs) on the unit sphere, which simplifies the learning process and achieves good results on time series tasks.
Qizhang Li, Yiwen Guo, Wangmeng Zuo, Hao Chen
https://openreview.net/forum?id=X5MH7iut9K
Keywords: adversarial examples, adversarial transferability, black-box attack
Compressor summary: The paragraph introduces TA-Bench, a benchmark for transfer-based methods to attack black-box DNNs, which evaluates 30+ methods on 10 popular victim models using ImageNet data.
Drew Linsley, Ivan F Rodriguez Rodriguez, Thomas FEL, Michael Arcaro, Saloni Sharma, Margaret Livingstone, Thomas Serre
https://openreview.net/forum?id=X4mmXQ4Nxw
Keywords: neural system identification, behavioral alignment, neural object recognition
Compressor summary: The authors find that as deep neural networks (DNNs) improve at object recognition, they become worse models of inferotemporal (IT) neuron responses to images, and propose a method to align DNN representations with human vision using the neural harmonizer.
Fateme Jamshidi, Sina Akbari, Negar Kiyavash
https://openreview.net/forum?id=X3IeHRD0zf
Keywords: causal inference, conditional independence, context-specific independence relations, imitability
Compressor summary: The paper explores the benefits of using context-specific independence information in causal imitation learning and provides a necessary and sufficient graphical criterion for it, as well as an algorithm that considers both CSI and data.
William Joseph Swartworth, Deanna Needell, Rachel Ward, Mark Kong, Halyun Jeong
https://openreview.net/forum?id=X25L5AjHig
Keywords: catastrophic forgetting, linear systems
Compressor summary: The paper derives theoretical limits on how much information is forgotten in continual learning for linear tasks and introduces a new characterization of numerical ranges of products of projections.
James Queeney, Mouhacine Benosman
https://openreview.net/forum?id=X0CIxqYc4Z
Keywords: deep reinforcement learning, model uncertainty, safety, risk-averse, distributionally robust
Compressor summary: The paper presents a deep reinforcement learning method for safe decision making in uncertain environments using risk-averse distortion risk measures, without minimax optimization, and shows its effectiveness on continuous control tasks with safety constraints.
Weipu Zhang, Gang Wang, Jian Sun, Yetian Yuan, Gao Huang
https://openreview.net/forum?id=WxnrX42rnS
Keywords: deep learning, reinforcement learning, model-based reinforcement learning, world model, learning in imagination, transformer, variational autoencoders, sequence modeling
Compressor summary: STORM is a world model architecture that combines Transformers and variational autoencoders to improve sequence modeling and generation in model-based reinforcement learning, achieving a record performance on the Atari 100k benchmark with less interaction experience and faster training.
Leonardo Cotta, Gal Yehuda, Assaf Schuster, Chris J. Maddison
https://openreview.net/forum?id=WwP2JaXAtB
Keywords: Invariant Learning, Geometric Deep Learning, Set Representations, Graph Representations, Expressive Power, Randomized Algorithms
Compressor summary: The paper proposes Randomized Linear Classifiers, a probabilistic model that can approximate any smooth function and preserve invariances while using fewer resources than deterministic neural networks.
Huiyang Shao, Qianqian Xu, Zhiyong Yang, Peisong Wen, Gao Peifeng, Qingming Huang
https://openreview.net/forum?id=WsmBcJarWW
Keywords: AUC, Cost Learning, Bilevel, machine learning
Compressor summary: The paper proposes a new cost-sensitive learning framework that integrates cost distribution into the AUC metric using a bilevel optimization method and shows its effectiveness in handling long-tail datasets.
Hitesh Sapkota, Dingrong Wang, ZHIQIANG TAO, Qi Yu
https://openreview.net/forum?id=WrRG0C1Vo5
Keywords: sparse network training, model calibration
Compressor summary: The paper proposes a new method to improve the reliability of sparse network training by using Distributionally Robust Optimization (DRO) to create an ensemble of diverse and complementary sub-networks that capture different data distributions and reduce overconfidence.
Biao Jiang, Xin Chen, Wen Liu, Jingyi Yu, Gang YU, Tao Chen
https://openreview.net/forum?id=WqiZJGNkjn
Keywords: 3d motion, motion generation, human motion synthesis, text-driven, text-to-motion
Compressor summary: The paper introduces MotionGPT, a unified model for language and human motion that can handle various motion-related tasks, by treating human motion as a specific language.
Shin'ya Yamaguchi, Daiki Chijiwa, Sekitoshi Kanai, Atsutoshi Kumagai, Hisashi Kashima
https://openreview.net/forum?id=WpuBEtrn0t
Keywords: Deep Learning, Generative Models, Generative Data Augmentation, Regularization, Meta-Learning
Compressor summary: The paper proposes a method called meta generative regularization (MGR) to improve generative data augmentation for deep learning by using synthetic samples to regularize feature extractors instead of training classifiers, which can avoid performance degradation and boost accuracy, especially on smaller datasets.
Alan Jeffares, Tennison Liu, Jonathan Crabbé, Mihaela van der Schaar
https://openreview.net/forum?id=WpGLxnOWhn
Keywords: Deep Ensembles, Deep Learning
Compressor summary: Joint optimization of ensemble loss in deep ensembles leads to artificial diversity that fails to generalize, causing a larger performance gap.
Jongmin Lee, Ernest K. Ryu
https://openreview.net/forum?id=Wn82NbmvJy
Keywords: Value Iteration, Reinforcement Learning, Reinforcement Learning Theory, Dynamic Programming, Acceleration, Anchoring mechanism
Compressor summary: Anc-VI is an accelerated value iteration method that reduces Bellman error faster than standard VI by using an anchoring mechanism.
Ayush Sekhari, Karthik Sridharan, Wen Sun, Runzhe Wu
https://openreview.net/forum?id=WmqYhqvz5i
Keywords: Contextual Bandit, Imitation Learning, Learning from Expert Feedback, Theory
Compressor summary: The paper proposes an algorithm for preference-based contextual bandits and imitation learning that minimizes regret and query complexity, and can outperform a sub-optimal expert with fewer queries.
Zirui Zhao, Wee Sun Lee, David Hsu
https://openreview.net/forum?id=Wjp1AYB8lH
Keywords: Embodied Task Planning, Large Language Models, Human-Robot Interaction
Compressor summary: The paper proposes an LLM-MCTS algorithm that combines a commonsense world model from LLMs with MCTS search, improving task planning efficiency and performance on complex tasks.
Xingyuan Zhang, Philip Becker-Ehmck, Patrick van der Smagt, Maximilian Karl
https://openreview.net/forum?id=WjlCQxpuxU
Keywords: Imitation Learning, World Models, Latent Variable Model, Transfer Learning, Variational Inference
Compressor summary: The paper proposes AIME, a method for imitating expert behaviors using world models without needing additional environment interactions or training.
Huayang Li, Tian Lan, Zihao Fu, Deng Cai, Lemao Liu, Nigel Collier, Taro Watanabe, Yixuan Su
https://openreview.net/forum?id=WjgCRrOgip
Keywords: language modeling, text generation, natural language processing
Compressor summary: This paper proposes a simple explanation and solution for the neural text degeneration problem by showing that penalizing repetitions in training data reduces this issue.
Junsheng Zhou, Baorui Ma, Wenyuan Zhang, Yi Fang, Yu-Shen Liu, Zhizhong Han
https://openreview.net/forum?id=WjWifKqmcG
Keywords: LiDAR Point Clouds, 2D images, Cross-modality registration, Matching
Compressor summary: The paper proposes a method to improve cross-modality registration between 2D images and 3D point clouds using a structured latent space, a triplet network, and a differentiable probabilistic PnP solver, achieving better results on KITTI and nuScenes datasets.
Sihan Zeng, Thinh T. Doan, Justin Romberg
https://openreview.net/forum?id=WjDj6W872v
Keywords: Reinforcement learning, superlevel sets, minimax optimization, robust reinforcement learning
Compressor summary: The paper explores new properties of policy optimization landscapes in reinforcement learning and uses them to derive minimax theorems for robust reinforcement learning problems under adversarial rewards.
Adel Javanmard, Vahab Mirrokni
https://openreview.net/forum?id=WfsWy59bX2
Keywords: high-dimensional regression, generalization error, asymptotic analysis, Convex Gaussian Minimax Theorem, regularization
Compressor summary: This paper explores how using anonymous cluster centers in personalized recommendation systems can protect user privacy and sometimes improve model performance, by analyzing the generalization error using the Convex Gaussian Minimax Theorem.
Chang Lu, Chandan K. Reddy, Ping Wang, Yue Ning
https://openreview.net/forum?id=Wff6DWFY2W
Keywords: ICD Coding, Contrastive Learning, NLP, Healthcare, Text Categorization, Pre-training
Compressor summary: The authors propose an algorithm to segment clinical notes into sections and use contrastive pre-training and masked section training to improve ICD coding performance with limited data.
Arthur Pellegrino, N Alex Cayco Gajic, Angus Chadwick
https://openreview.net/forum?id=WcoX8eJJjI
Keywords: Recurrent Neural Networks, Computational Neuroscience, Neural Data Analysis, Tensor, Learning
Compressor summary: The study investigates how low rank structures in recurrent neural networks (RNNs) evolve during learning and their implications for population connectivity and learning dynamics in both biological and artificial neural networks.
Kalpesh Krishna, Yixiao Song, Marzena Karpinska, John Frederick Wieting, Mohit Iyyer
https://openreview.net/forum?id=WbFhFvjjKj
Keywords: AI-generated text detection, text detection, paraphrasing, attacks, retrieval, defenses, large language models, LLMs
Compressor summary: The authors develop a paraphrase generation model (DIPPER) that evades several AI text detection algorithms and propose a defense method based on searching for semantically-similar texts in a database.
Yiqun Duan, Charles Zhou, Zhen Wang, Yu-Kai Wang, Chin-teng Lin
https://openreview.net/forum?id=WaLI8slhLw
Keywords: EEG; Neural Encoding; Brain Computer Interface
Compressor summary: DeWave is a novel framework that translates brain signals into natural language without relying on eye-tracking or event markers, improving EEG-to-text translation accuracy.
Nicolas Zucchet, Robert Meier, Simon Schug, Asier Mujika, Joao Sacramento
https://openreview.net/forum?id=Wa1GGPqjUn
Keywords: online learning, linear recurrent units, temporal credit assignment, biologically-plausible learning, local learning rules, neuromorphic computing
Compressor summary: The paper introduces an efficient online learning algorithm for recurrent neural networks that leverages independent modules, enabling them to learn long-range dependencies and opening new possibilities in neuromorphic computing.
Jacobus G.M. van der Linden, Mathijs de Weerdt, Emir Demirović
https://openreview.net/forum?id=WYYpxVsKpR
Keywords: optimal decision trees, dynamic programming, separability
Compressor summary: The authors propose a dynamic programming framework for global optimization of decision trees that can handle separable objectives and constraints, improving scalability and performance over general-purpose solvers.
Davoud Ataee Tarzanagh, Yingcong Li, Xuechen Zhang, Samet Oymak
https://openreview.net/forum?id=WXc8O8ghLH
Keywords: attention mechanism, implicit bias, margin maximization, nonconvex optimization, prompt tuning
Compressor summary: The paper analyzes the attention mechanism in transformer models and proves that it converges to a max-margin solution for optimal token selection.
Yuyang Deng, Ilja Kuzborskij, Mehrdad Mahdavi
https://openreview.net/forum?id=WVmus8NWE8
Keywords: Multi-source domain adaptation; minimax optimization; learning theory
Compressor summary: The paper proposes algorithms to efficiently learn from heterogeneous data sources and adapt to different target distributions with theoretical guarantees.
Yifan Pu, Weicong Liang, Yiduo Hao, Yuhui Yuan, Yukang Yang, Chao Zhang, Han Hu, Gao Huang
https://openreview.net/forum?id=WUott1ZvRj
Keywords: Object Detection
Compressor summary: The Rank-DETR method improves DETR-based object detectors by enhancing ranking accuracy and reducing false positives, leading to better localization and detection performance.
Jing Li, Quanxue Gao, QIANQIAN WANG, Ming Yang, Wei Xia
https://openreview.net/forum?id=WRtlsxA5h7
Keywords: Multi-view clustering, tensor Schatten p-norm, non-negative matrix factorization.
Compressor summary: The paragraph describes a novel multi-view clustering method based on orthogonal non-negative tensor factorization, which considers both within-view and between-view information and achieves good performance on various datasets.
Salva Rühling Cachay, Bo Zhao, Hailey James, Rose Yu
https://openreview.net/forum?id=WRGldGm5Hz
Keywords: AI for science, diffusion models, scientific machine learning, probabilistic forecasting
Compressor summary: The authors present a dynamics-informed diffusion model that leverages temporal data dynamics for forecasting complex systems and demonstrate its advantages over Gaussian noise-based models.
ZAIXI ZHANG, Zepu Lu, Zhongkai Hao, Marinka Zitnik, Qi Liu
https://openreview.net/forum?id=WPdGRRJaPb
Keywords: Graph Representation Learning, AI for Science
Compressor summary: FAIR is a framework that co-designs protein pocket sequence and 3D structure for ligand binding by refining residue types, backbone coordinates, and sidechain atoms in an efficient full-shot manner.
Akshaykumar G Gattani, Sharath Raghvendra, Pouyan Shirzadian
https://openreview.net/forum?id=WPbIAdB6aQ
Keywords: Euclidean bipartite matching, exact algorithms, primal dual method
Compressor summary: The paper presents a new algorithm for computing minimum-cost bipartite matching between two sets of points, which has faster execution time than previous methods, especially for stochastic point sets with real-valued coordinates and any dimension.
Morteza Boroun, Erfan Yazdandoost Hamedani, Afrooz Jalilzadeh
https://openreview.net/forum?id=WO1kHC5Lfz
Keywords: Saddle Point Problem, Projection-free method
Compressor summary: The paper proposes efficient single-loop projection-free methods for constrained saddle point problems using regularization and nested approximation techniques, achieving convergence guarantees in limited iterations.
Lorenzo Perini, Jesse Davis
https://openreview.net/forum?id=WK8LQzzHwW
Keywords: Anomaly Detection, Learning with Rejection, Unsupervised Learning
Compressor summary: The paper proposes a method to improve anomaly detection by rejecting predictions with high uncertainty using a constant threshold on a stability metric, while providing theoretical guarantees and empirical results.
Siming Lan, Rui Zhang, Qi Yi, Jiaming Guo, Shaohui Peng, Yunkai Gao, Fan Wu, Ruizhi Chen, Zidong Du, Xing Hu, Xishan Zhang, Ling Li, Yunji Chen
https://openreview.net/forum?id=WIrZh2XxLT
Keywords: reinforcement learning, multi-task learning, contrastive learning
Compressor summary: The paper proposes CMTA, a method for multi-task reinforcement learning that uses contrastive learning to prevent conflicts within tasks and temporal attention to combine shared modules at a finer level than tasks, improving performance on Meta-World benchmark tasks.
Kaiwen Zha, Peng Cao, Jeany Son, Yuzhe Yang, Dina Katabi
https://openreview.net/forum?id=WHedsAeatp
Keywords: regression, representation learning, continuity
Compressor summary: Rank-N-Contrast (RNC) is a framework that learns continuous representations for regression by contrasting samples based on their rankings in the target space, improving performance, robustness, efficiency, and generalization.
Qiang Ding, Yixuan Cao, Ping Luo
https://openreview.net/forum?id=WBq6Q4ml04
Keywords: selective classification, uncertainty estimation, ensemble learning
Compressor summary: The paper analyzes why ensemble methods improve prediction reliability, focusing on top-ambiguity samples where member models differ, and proves that ensembles have lower selective risk under certain conditions.
Joshua John Horacsek, Usman Alim
https://openreview.net/forum?id=WBXYGBQXiB
Keywords: Computer Vision and Pattern Recognition
Compressor summary: The paper introduces the lattice tensor data structure to enable machine learning on non-Cartesian domains using standard algorithms and a software library.
Scott Alexander Cameron, Arnu Pretorius, Stephen J. Roberts
https://openreview.net/forum?id=WAd5ZRdFoc
Keywords: PINNs, physics informed neural networks, geometric deep learning, neural operator, PDEs
Compressor summary: The paper proposes a neural network that takes triangular meshes as input and solves partial differential equations quickly, without retraining or fixed geometry representation.
Bin Huang, Jiaqian Yu, Yiwei Chen, Siyang Pan, Qiang Wang, Zhi Wang
https://openreview.net/forum?id=W9pJx9sFCh
Keywords: Backdoor Attack, Visual Object Tracking, Deep Learning, Poison-Only
Compressor summary: This paper proposes a poison-only backdoor attack on visual object tracking that attaches a trigger pattern to the background of video frames, which degrades the performance of VOT trackers.
Yucheng Shi, Mengnan Du, Xuansheng Wu, Zihan Guan, Jin Sun, Ninghao Liu
https://openreview.net/forum?id=W6U2xSbiE1
Keywords: backdoor defense, black-box defense, diffusion model
Compressor summary: The paper proposes a novel defense framework called Zero-shot Image Purification (ZIP) to defend against backdoor attacks on black-box models by applying a linear transformation and using a pre-trained diffusion model to recover the purified image.
Chenyu Zheng, Guoqiang Wu, Chongxuan Li
https://openreview.net/forum?id=W5Clq1bSrR
Keywords: generative data augmentation, algorithmic stability, non-i.i.d. learning
Compressor summary: Generative data augmentation can improve classification performance in various tasks by obtaining fake labeled examples from a trained conditional generative model, but its effect has not been well studied theoretically; this paper establishes a stability bound and shows how it helps learning in different settings.
Zhuoping Zhou, Davoud Ataee Tarzanagh, Bojian Hou, Boning Tong, Jia Xu, Yanbo Feng, Qi Long, Li Shen
https://openreview.net/forum?id=W3cDd5xlKZ
Keywords: Fairness, Canonical Correlation Analysis, Riemannian Optimization, Pareto Optimization
Compressor summary: The paper proposes a framework for minimizing unfairness in Canonical Correlation Analysis by balancing correlation levels across groups while maintaining accuracy and evaluating its effectiveness on various datasets.
Jiacheng Chen, Ruizhi Deng, Yasutaka Furukawa
https://openreview.net/forum?id=W2ZBLdfa16
Keywords: Structured Reconstruction, Floorplan Reconstruction, HD Map Construction, Diffusion Models
Compressor summary: PolyDiffuse is a new algorithm that uses Diffusion Models to reconstruct structured data like polygonal shapes from sensor data, overcoming challenges related to denoising ambiguity and initial noise choice.
Lalit Pandey, Samantha Marie Waters Wood, Justin Newell Wood
https://openreview.net/forum?id=W23ZTdsabj
Keywords: vision transformer, newborn, controlled rearing, object recognition, data hungry
Compressor summary: ViTs and chicks both learn view-invariant object recognition in impoverished visual environments, challenging the assumption that ViTs are more data hungry than brains.
Bingbin Liu, Jordan T. Ash, Surbhi Goel, Akshay Krishnamurthy, Cyril Zhang
https://openreview.net/forum?id=VzmpXQAn6E
Keywords: Transformers, language models, hallucinations, long-range dependencies, generalization, extrapolation, out-of-distribution
Compressor summary: The text discusses attention glitches, a phenomenon where large language models sometimes make errors in reasoning, and introduces flip-flop language modeling as a way to study and potentially improve these models' extrapolative abilities.
Jose Blanchet, Miao Lu, Tong Zhang, Han Zhong
https://openreview.net/forum?id=VzLBMkc7tB
Keywords: distributionally robust offline reinforcement learning, double pessimism, general function approximation
Compressor summary: The paper introduces Doubly Pessimistic Model-based Policy Optimization ($\texttt{P}^2\texttt{MPO}$), a novel algorithm framework for distributionally robust offline reinforcement learning that combines flexible model estimation and doubly pessimistic policy optimization to overcome distribution shift and perturbation.
Wenhai Wang, Zhe Chen, Xiaokang Chen, Jiannan Wu, Xizhou Zhu, Gang Zeng, Ping Luo, Tong Lu, Jie Zhou, Yu Qiao, Jifeng Dai
https://openreview.net/forum?id=Vx1JadlOIt
Keywords: Large Vision-Language Model, Detection, Image Caption
Compressor summary: The paper introduces VisionLLM, an LLM-based framework that treats images as foreign languages and uses language instructions to perform open-ended vision tasks with high task customization and competitive performance on COCO detection benchmark.
Qi Han, Yuxuan Cai, Xiangyu Zhang
https://openreview.net/forum?id=VvnfMeC3gQ
Keywords: architecture design, representation learning, masked image modeling, self-supervised learning
Compressor summary: RevColV2 is a new architecture for masked image modeling that keeps the entire autoencoder during pre-training and fine-tuning, maintaining disentangled low-level and semantic information and achieving competitive performance on various vision tasks.
Atli Kosson, Martin Jaggi
https://openreview.net/forum?id=Vtqymej1tA
Keywords: multiplication-free, neural architectures, piecewise linear networks, piecewise affine networks, efficient training, efficient arithmetics
Compressor summary: The authors propose a method to replace matrix multiplications and non-linearities with piecewise affine approximations, achieving fully multiplication-free training of neural networks without sacrificing performance.
Ida Momennejad, Hosein Hasanbeig, Felipe Vieira Frujeri, Hiteshi Sharma, Nebojsa Jojic, Hamid Palangi, Robert Ness, Jonathan Larson
https://openreview.net/forum?id=VtkGvGcGe3
Keywords: Large Language Models, LLM evaluation, model comparison, GPT-4, graph analysis, cognitive science, cognitive map, hippocampus, planning, multi-step planning, reasoning, community graph
Compressor summary: The authors propose CogEval, a systematic evaluation protocol for cognitive abilities in LLMs, and find that current LLMs have limited planning ability and fail to understand cognitive maps.
Yang Yu, Qi Liu, Kai Zhang, Yuren Zhang, Chao Song, Min Hou, Yuqing Yuan, ZHIhao Ye, ZAIXI ZHANG, Sanshi Lei Yu
https://openreview.net/forum?id=VsbrdJpwpT
Keywords: User Model Pre-training, Data Augmentation, Contrastive Learning
Compressor summary: AdaptSSR is a new pretext task for user modeling that improves performance by capturing similarity orders between implicitly and explicitly augmented views and other users' views, while adjusting the similarity constraint based on estimated similarities.
Pulkit Verma, Rushang Karia, Siddharth Srivastava
https://openreview.net/forum?id=VqclD6Nfaj
Keywords: Sequential Decision Making, Interpretable Models, Relational Model Learning, Black-Box Agents, Symbolic Descriptions
Compressor summary: The paper proposes an active-learning method to learn a probabilistic model of black-box AI systems with sequential decision-making capabilities, allowing users to understand and assess their safety in stochastic settings.
Beining Yang, Kai Wang, Qingyun Sun, Cheng Ji, Xingcheng Fu, Hao Tang, Yang You, Jianxin Li
https://openreview.net/forum?id=VqIWgUVsXc
Keywords: data-efficient learning, graph generation, graph neural networks
Compressor summary: The paper proposes a new method, SGDD, to preserve the original structure information of large-scale graphs when generating synthetic ones, improving performance in cross-architecture settings and specific tasks while reducing storage costs.
Mostafa Dehghani, Basil Mustafa, Josip Djolonga, Jonathan Heek, Matthias Minderer, Mathilde Caron, Andreas Peter Steiner, Joan Puigcerver, Robert Geirhos, Ibrahim Alabdulmohsin, Avital Oliver, Piotr Padlewski, Alexey A. Gritsenko, Mario Lucic, Neil Houlsby
https://openreview.net/forum?id=VpGFHmI7e5
Keywords: Vision Transformer, variable aspect ratio, flexible inference, efficient training
Compressor summary: NaViT is a flexible and efficient computer vision model that can handle inputs of different resolutions and aspect ratios without resizing them, leading to improved performance and robustness.
Anurag Ghosh, Vaibhav Balloli, Akshay Nambi, Aditya Singh, Tanuja Ganu
https://openreview.net/forum?id=VpCjozUOM2
Keywords: approximate execution framework; real time perception; latency-accuracy tradeoffs
Compressor summary: The paper introduces Chanakya, a learned framework that optimizes real-time perception by automatically balancing accuracy and latency tradeoffs in decisions induced by intrinsic and extrinsic factors.
Sadhika Malladi, Tianyu Gao, Eshaan Nichani, Alex Damian, Jason D. Lee, Danqi Chen, Sanjeev Arora
https://openreview.net/forum?id=Vota6rFhBQ
Keywords: language models, fine-tuning, zeroth order optimization, memory efficiency
Compressor summary: MeZO is a memory-efficient zeroth-order optimizer that can fine-tune large language models with less memory and GPU resources than backpropagation, achieving comparable performance on various tasks.
Tianyuan Teng, Li Kevin Wenliang, Hang Zhang
https://openreview.net/forum?id=VnfeOjR73Q
Keywords: human representation of uncertainty; Bayesian inference; bounded rationality; inductive bias; Chinese Restaurant Process
Compressor summary: This study examines how humans learn to represent environmental uncertainty, finding that they tend to overestimate the number of clusters in a distribution, possibly due to cognitive limitations.
Grace Luo, Lisa Dunlap, Dong Huk Park, Aleksander Holynski, Trevor Darrell
https://openreview.net/forum?id=Vm1zeYqwdc
Keywords: semantic correspondence, hypercolumns, diffusion models, generative model representations
Compressor summary: The paper proposes Diffusion Hyperfeatures, a method to consolidate feature maps from diffusion models into per-pixel descriptors for semantic keypoint correspondence tasks on real and synthetic images.
Lenart Treven, Jonas Hübotter, Bhavya Sukhija, Florian Dorfler, Andreas Krause
https://openreview.net/forum?id=VkhvDfY2dB
Keywords: Reinforcement Learning, Optimal Control, Continuous Time
Compressor summary: The paper presents a reinforcement learning algorithm that uses nonlinear ODEs to model continuous-time dynamics and shows how to choose when and how often to observe the system for optimal exploration.
Chen Zhang, Xiaofeng Cao, Weiyang Liu, Ivor Tsang, James Kwok
https://openreview.net/forum?id=VkUNovXoxx
Keywords: Nonparametric machine teaching, Multiple learners
Compressor summary: The paper proposes a new framework called Multi-learner Nonparametric Teaching (MINT) that enables teachers to teach multiple students simultaneously more efficiently than traditional single-learner methods, especially when students can communicate with each other.
Garrett Bingham, Risto Miikkulainen
https://openreview.net/forum?id=ViFTWelHVZ
Keywords: automl, activation function, surrogate modeling, fisher information matrix, eigenvalues, optimization, umap, imagenet
Compressor summary: The paper introduces new datasets, methods, and findings to improve the design and optimization of activation functions in neural networks, including a surprising sigmoidal discovery.
Haochen Wang, Junsong Fan, Yuxi Wang, Kaiyou Song, Tong Wang, Zhaoxiang Zhang
https://openreview.net/forum?id=VhcsIxVEd9
Keywords: Self-Supervised Learning, Vision Transformer, Visual Representation Learning
Compressor summary: DropPos is a self-supervised pretext task that improves location awareness in Vision Transformers by reconstructing dropped positions with visual appearance and attention.
Oscar Li, James Harrison, Jascha Sohl-Dickstein, Virginia Smith, Luke Metz
https://openreview.net/forum?id=VhbV56AJNt
Keywords: Evolution Strategies, unrolled computation graph, online gradient estimation, variance reduction, stochastic gradient estimation
Compressor summary: The paper introduces Noise-Reuse Evolution Strategies (NRES), an unbiased online evolution strategies method for machine learning that has lower variance, faster convergence, and better parallelism than existing methods.
Jiaming Qiu, Xiongtao Dai
https://openreview.net/forum?id=VhLU3pStsl
Keywords: metric learning, manifold learning, local metric, dissimilarity, geometry
Compressor summary: The paper proposes a new local regression approach to learn the Riemannian metric tensor from similarity measures between data points, and provides theoretical convergence rates and empirical results on various datasets.
Shihao Zhao, Dongdong Chen, Yen-Chun Chen, Jianmin Bao, Shaozhe Hao, Lu Yuan, Kwan-Yee K. Wong
https://openreview.net/forum?id=VgQw8zXrH8
Keywords: computer vision, diffusion model, text-to-image generation
Compressor summary: The paper introduces Uni-ControlNet, a framework that allows for flexible and composable control of text-to-image diffusion models using different local and global controls with minimal fine-tuning and model size costs.
Xiwen Wang, Jiaxi Ying, Daniel P. Palomar
https://openreview.net/forum?id=Vfp8sDST4g
Keywords: MTP2 Gaussian Graphical Model, High-dimensional precision matrix estimation, Bridge-block decomposition.
Compressor summary: The paper proposes a method to learn large Gaussian graphical models by decomposing the problem into smaller sub-problems and using bridge concepts, which reduces computational complexity and improves existing algorithms' performance.
Ori Press, Steffen Schneider, Matthias Kuemmerer, Matthias Bethge
https://openreview.net/forum?id=VfP6VTVsHc
Keywords: test time adaptation, continual adaptation, benchmarking, imagenet-c, imagenet classification, robustness, continual learning, imagenet benchmark
Compressor summary: The Continually Changing Corruptions (CCC) benchmark reveals that most Test-Time Adaptation (TTA) methods perform poorly over long timescales and are beaten by a simple resetting strategy.
Jiankai Sun, Yiqi Jiang, Jianing Qiu, Parth Talpur Nobel, Mykel Kochenderfer, Mac Schwager
https://openreview.net/forum?id=VeO03T59Sh
Keywords: Uncertainty, Conformal Prediction, Dynamics Model
Compressor summary: PlanCP is a method that uses conformal prediction to quantify and reduce the uncertainty of diffusion dynamics models for trajectory prediction in robotic applications, improving performance on offline reinforcement learning and continuous planning tasks.
Bo Xue, Yimu Wang, Yuanyu Wan, Jinfeng Yi, Lijun Zhang
https://openreview.net/forum?id=Vbm5UCaYeh
Keywords: linear bandits, heavy-tailed, truncated, mean of medians
Compressor summary: The paper proposes two new algorithms for generalized linear bandits with heavy-tailed rewards and shows they have better performance than existing methods in terms of regret bounds.
Wei Dong, Dawei Yan, Zhijun Lin, Peng Wang
https://openreview.net/forum?id=VbYdaK8ek0
Keywords: computer vision; vision transformer; visual adapter; transfer learning
Compressor summary: The study proposes a new method called Adapter Re-Composing (ARC) to efficiently adapt pre-trained models for downstream tasks by reusing and sharing parameters in adapter design.
Shuli Jiang, Pranay Sharma, Gauri Joshi
https://openreview.net/forum?id=VacSQpbI0U
Keywords: distributed vector mean estimation, communication efficiency, cross-client correlation
Compressor summary: Rand-Proj-Spatial is a new method for communication-efficient distributed vector mean estimation that uses projection to a random subspace and correlation information to outperform existing techniques like Rand-$k$-Spatial.
François Rozet, Gilles Louppe
https://openreview.net/forum?id=VUvLSnMZdX
Keywords: data assimilation, score-based, generative modeling, posterior inference, dynamical systems
Compressor summary: The paper introduces score-based data assimilation, a method for inferring state trajectories in stochastic dynamical systems by learning a generative model that can handle long time horizons and complex dynamics without relying on transition dynamics.
Yong-Hyun Park, Mingi Kwon, Jaewoong Choi, Junghyo Jo, Youngjung Uh
https://openreview.net/forum?id=VUlYp3jiEI
Keywords: diffusion models, semantic image editing, differential geometry
Compressor summary: This paper analyzes the geometric structure of diffusion models' latent space and shows how it enables image editing through traversal without additional training.
Amur Ghose, Apurv Gupta, Yaoliang Yu, Pascal Poupart
https://openreview.net/forum?id=VQ1heZKSLQ
Keywords: Adversarial, Batch normalization, Robustness, Geometric, radial
Compressor summary: The text explains how intermediate latents from batch normalized deep image recognition models can be used to create adversarial examples without labels, and discusses the security implications of this finding.
Giulio Franzese, Giulio Corallo, Simone Rossi, Markus Heinonen, Maurizio Filippone, Pietro Michiardi
https://openreview.net/forum?id=VPrir0p5b6
Keywords: Hilbert spaces, Diffusion models, Stochastic Partial Differential Equations
Compressor summary: Functional Diffusion Processes (FDPs) are a new type of generative model that works with continuous data in function spaces, requiring less complex network architectures and achieving high-quality image generation.
Shangyuan LIU, Linglingzhi Zhu, Anthony Man-Cho So
https://openreview.net/forum?id=VPTZVVP4tm
Keywords: Graph Signal Processing, Spectral Template, Network Inference, Optimization, Linearized ADMM
Compressor summary: The paper introduces LogSpecT, a novel graph learning model that is always feasible and has recovery guarantees, and proposes an efficient algorithm with convergence guarantees to learn graphs from stationary signals.
Yuxin Wen, Neel Jain, John Kirchenbauer, Micah Goldblum, Jonas Geiping, Tom Goldstein
https://openreview.net/forum?id=VOstHxDdsN
Keywords: Diffusion Model, Generative AI, Prompt Discovery
Compressor summary: The paragraph discusses an optimization method for hard text prompts that can be used in various applications and allows users to easily generate and mix image concepts without prior knowledge.
Insu Jeon, Minui Hong, Junhyeog Yun, Gunhee Kim
https://openreview.net/forum?id=VNyKBipt91
Keywords: Personalized Federated Learning, Variational Dropout, Meta-Learning, Bayesian Neural Network
Compressor summary: MetaVD is a novel Bayesian meta-learning approach that predicts dropout rates per client and improves Federated Learning in non-IID data settings by personalizing models and reducing communication costs.
Aditya Hemant Shahane, Saripilli Venkata Swapna Manjiri, Ankesh Jain, Sandeep Kumar
https://openreview.net/forum?id=VNjJAWjuEU
Keywords: Analog design optimization, Analog synthesis, Graph Neural Networks, EDA, Graph learning, Optimization
Compressor summary: The paper introduces GCX, a framework that uses graph structure learning and graph neural networks to create a surrogate model for efficient exploration of the analog circuit design space in a semi-supervised learning framework.
ShuLin Xu, Yifan Sun, Faen Zhang, Anqi Xu, Xiu-Shen Wei, Yi Yang
https://openreview.net/forum?id=VMz5GhfxgV
Keywords: Fine-grained learning, Coarse-to-fine learning, Hyperbolic space, Hierarchical margin
Compressor summary: The paper proposes a novel hyperbolic space method for fine-grained recognition tasks that leverages hierarchical cosine margins to improve discriminative ability and achieves state-of-the-art results on five benchmark datasets.
Xin Li, Sima Behpour, Thang Doan, Wenbin He, Liang Gou, Liu Ren
https://openreview.net/forum?id=VMAgvbBBts
Keywords: Unsupervised prompt learning, UP-DP, Data preselection
Compressor summary: This study proposes UP-DP, an unsupervised prompt learning method that improves data pre-selection by combining vision and text features from foundation models like BLIP-2, achieving up to 20% performance gain and showing generalizability across datasets.
Johannes Kirschner, Alireza Bakhtiari, Kushagra Chandak, Volodymyr Tkachuk, Csaba Szepesvari
https://openreview.net/forum?id=VLnEFGu9V7
Keywords: sequential decision-making, decision-estimation coefficient, regret minimization, bandits, reinforcement learning, partial monitoring
Compressor summary: The paper presents a new algorithm (Anytime-E2D) that balances exploration and exploitation in sequential decision-making, based on re-parametrizing the existing DEC method, and demonstrates its effectiveness in high-dimensional linear bandits.
Yifei Zhou, Juntao Ren, Fengyu Li, Ramin Zabih, Ser-Nam Lim
https://openreview.net/forum?id=VKbEO2eh5w
Keywords: contrastive learning, pre-trained visual-language models, zero-shot learning, test-time augmentation
Compressor summary: The paper introduces Distribution Normalization (DN), which improves test-time performance in visual-language contrastive learning by approximating negative samples during inference without retraining or fine-tuning.
Jiaxin Zhang, Zhuohang Li, Kamalika Das, Sricharan Kumar
https://openreview.net/forum?id=VIaw1XHb4G
Keywords: multi-fidelity optimization, cost-effective learning, exploration-exploitation query, limited annotation budgets
Compressor summary: IMFL is a framework for developing small domain-specific language models using a mix of low-cost automatic and high-quality human annotations, achieving superior performance with limited budgets.
Eden Saig, Inbal Talgam-Cohen, Nir Rosenfeld
https://openreview.net/forum?id=VGLXjbTSYa
Keywords: Delegation, Algorithmic Contract Design, Moral Hazard, Learning Curves
Compressor summary: The paper proposes a framework for incentive-aware delegation of machine learning tasks using performance-based contracts that balance budget and accuracy, connecting classic statistical theory with modern learning curves and scaling laws.
Mingli Zhu, Shaokui Wei, Hongyuan Zha, Baoyuan Wu
https://openreview.net/forum?id=VFhN15Vlkj
Keywords: Backdoor Defense, Backdoor Learning, Trustworthy AI
Compressor summary: The proposed method filters trigger information from poisoned samples using a lightweight neural polarizer inserted as an intermediate layer in the backdoored model, requiring less clean data and being more efficient than other defense methods.
Esmaeil Seraj, Jerry Yuyang Xiong, Mariah L Schrum, Matthew Gombolay
https://openreview.net/forum?id=VCOZaczCHg
Keywords: Learning from Demonstration, Multi-Robot Systems, Teaching Robot Teams
Compressor summary: MixTURE is a novel Multi-Agent Learning from Demonstration framework that learns both collaborative tasks and inter-agent communication from human expert data, reducing human workload and improving usability.
Vincent Froese, Christoph Hertrich
https://openreview.net/forum?id=VAQp2EnZeW
Keywords: Computational Complexity, Neural Network, Rectified Linear Unit, Empirical Risk Minimization, Parameterized Complexity
Compressor summary: The paper investigates the complexity of training two-layer neural networks with different activation functions in various input dimensions and neuron counts, answering several open questions and proving some fixed-parameter tractability results.
Kumar Ashutosh, Santhosh Kumar Ramakrishnan, Triantafyllos Afouras, Kristen Grauman
https://openreview.net/forum?id=VAC7aB6qSG
Keywords: Instructional Videos, Task Graph, Keystep Recognition
Compressor summary: The paragraph describes a method for automatically discovering task graphs from how-to videos to improve keystep recognition in novel videos.
Xiangchen Song, Weiran Yao, Yewen Fan, Xinshuai Dong, Guangyi Chen, Juan Carlos Niebles, Eric Xing, Kun Zhang
https://openreview.net/forum?id=V8GHCGYLkf
Keywords: Unsupervised learning, Temporal disentanglement, Nonlinear ICA, Identifiability theory
Compressor summary: The paper proposes NCTRL, a method for identifying time-delayed causal influences in nonstationary sequential data without using auxiliary information or simplifying latent dynamics.
Julius von Kügelgen, Michel Besserve, Wendong Liang, Luigi Gresele, Armin Kekić, Elias Bareinboim, David Blei, Bernhard Schölkopf
https://openreview.net/forum?id=V87gZeSOL4
Keywords: Causal representation learning, identifiability, theory, nonparametric, interventions, multi-environment
Compressor summary: The paper proposes a method to learn latent causal variables and their relations from high-dimensional functions, using multiple datasets from unknown interventions and showing identifiability conditions without restrictive assumptions or partial knowledge.
Martin Gonzalez, Nelson Fernandez, Thuy Vinh Dinh Tran, Elies Gherbi, Hatem Hajri, Nader Masmoudi
https://openreview.net/forum?id=V6IgkYKD8P
Keywords: Diffusion Probabilistic Models, Exponential SDE methods, Image Generation, Generative Models
Compressor summary: SEEDS are stochastic exponential derivative-free solvers that improve and generalize Exponential Integrator approaches for fast sampling from Diffusion Probabilistic Models without requiring derivatives or training, achieving optimal quality and speed.
Jae Sung Park, Jack Hessel, Khyathi Chandu, Paul Pu Liang, Ximing Lu, Peter West, Youngjae Yu, Qiuyuan Huang, Jianfeng Gao, Ali Farhadi, Yejin Choi
https://openreview.net/forum?id=V5eG47pyVl
Keywords: multimodal, commonsense reasoning, instruction tuning, large language model
Compressor summary: The Localized Visual Commonsense model allows users to specify regions within images as input, enabling precise within-image reasoning and improving multimodal tasks with a reference-as-input interface.
Tiansheng Huang, Sihao Hu, Ka-Ho Chow, Fatih Ilhan, Selim Furkan Tekin, Ling Liu
https://openreview.net/forum?id=V5cQH7JbGo
Keywords: Federated learning, backdoor defense, isolated subspace training.
Compressor summary: Lockdown is a novel isolated subspace training method that effectively defends against backdoor attacks in federated learning, while reducing communication and model complexity.
Abhinav Kumar, Amit Deshpande, Amit Sharma
https://openreview.net/forum?id=V5Oh7Aqfft
Keywords: Spurious Correlation, Out of Distribution Generalization
Compressor summary: The paper proposes a method to automatically identify and remove spurious attributes in classification datasets, improving generalization and reducing dependence on noisy factors.
Stefania Ionescu, Yuhao Du, Kenneth Joseph, Aniko Hannak
https://openreview.net/forum?id=V5FNSilWiC
Keywords: matching markets, strategic behaviour, ML-based forecasting, recommender systems, adversarial attacks, agent-based modelling
Compressor summary: The paper introduces adversarial interaction attacks, where agents return to a matching market and act non-optimally with their matches to disrupt future predictions, and shows how these attacks can benefit returning agents and increase inequality.
Dror Freirich, Tomer Michaeli, Ron Meir
https://openreview.net/forum?id=V4hqq2NGTW
Keywords: Kalman filter, estimation theory, causal filtering, signal processing, distortion-perception tradeoff
Compressor summary: The text discusses the challenge of achieving perfect perceptual-quality in reconstructing temporal signals from corrupted or missing data under a causal filtering constraint, which may require ignoring new information and increasing MSE.
Gehua Ma, Runhao Jiang, Rui Yan, Huajin Tang
https://openreview.net/forum?id=V4YeOvsQfu
Keywords: neuroscience, neural coding, sensory neuroscience, visual coding, SNN, spiking neural networks, generative model, latent variable model, cognitive computational neuroscience, computational neuroscience
Compressor summary: TeCoS-LVM is a new spiking neural network model that can capture temporal dependencies in natural stimuli and produce realistic spike activities, outperforming current methods.
Hussein Mozannar, Jimin J Lee, Dennis Wei, Prasanna Sattigeri, Subhro Das, David Sontag
https://openreview.net/forum?id=V2yFumwo5B
Keywords: human-ai, collaboration, onboarding, region-discovery, LLM, data description
Compressor summary: The paper proposes learning natural language rules for human-AI collaboration based on data regions discovered by an algorithm.
Lesia Semenova, Harry Chen, Ronald Parr, Cynthia Rudin
https://openreview.net/forum?id=Uzi22WryyX
Keywords: Rashomon Set, Simplicity, Interpretable Machine Learning, Model Selection, Model Multiplicity
Compressor summary: This paragraph discusses how data noise affects the performance of different models and introduces a measure called pattern diversity to quantify the difference in predictions between them.
Alessandro Epasto, Vahab Mirrokni, Shyam Narayanan, Peilin Zhong
https://openreview.net/forum?id=UzUhiKACmS
Keywords: differential Privacy, k-means, k-median, clustering, distance-based privacy
Compressor summary: The paper proposes constant-approximate algorithms for Euclidean clustering with Distance-based privacy, which protect exact locations and outperform existing methods.
Zekun Li, Baolin Peng, Pengcheng He, Michel Galley, Jianfeng Gao, Xifeng Yan
https://openreview.net/forum?id=UvIN8oQ4uI
Keywords: Black-box Large Language Models, Directional Stimulus Prompting, Hint, Reinforcement learning, Prompt optimization
Compressor summary: The paragraph introduces Directional Stimulus Prompting, a framework that uses a small policy model to generate hints for large language models, guiding them towards specific outputs and improving their performance on various tasks.
Injae Kim, Minhyuk Choi, Hyunwoo J. Kim
https://openreview.net/forum?id=UvBwXdL95b
Keywords: neural radiance field, pose estimation
Compressor summary: UP-NeRF optimizes NeRF with unconstrained images without pose priors by using surrogate tasks, a separate module for transient occluders, and improved pose estimation and depth supervision.
Xingchen Wan, Pierre Osselin, Henry Kenlay, Binxin Ru, Michael A Osborne, Xiaowen Dong
https://openreview.net/forum?id=UuNd9A6noD
Keywords: graphs, Bayesian optimisation, scalability
Compressor summary: The paper proposes a new Bayesian optimization framework that adapts to graph-structured functions and improves sample efficiency for large-scale graphs using local modeling and suitable kernels.
Arthur da Cunha, Francesco D'Amore, Emanuele Natale
https://openreview.net/forum?id=UqYrYB3dp5
Keywords: lottery ticket hypothesis, convolutional neural network, network pruning, structured pruning, random subset sum
Compressor summary: The paper introduces a new mathematical tool to prove that random neural networks have efficient subnetworks, which can help understand the role of over-parameterization in deep learning.
Shaohan Huang, Li Dong, Wenhui Wang, Yaru Hao, Saksham Singhal, Shuming Ma, Tengchao Lv, Lei Cui, Owais Khan Mohammed, Barun Patra, Qiang Liu, Kriti Aggarwal, Zewen Chi, Johan Bjorck, Vishrav Chaudhary, Subhojit Som, Xia Song, Furu Wei
https://openreview.net/forum?id=UpN2wfrLec
Keywords: multimodal large language model
Compressor summary: KOSMOS-1 is a Multimodal Large Language Model that can perceive various modalities, learn in context, and follow instructions, achieving impressive performance on language and perception tasks without gradient updates or finetuning.
Jiarui Feng, Lecheng Kong, Hao Liu, Dacheng Tao, Fuhai Li, Muhan Zhang, Yixin Chen
https://openreview.net/forum?id=UlJcZoawgU
Keywords: Graph neural network, expressive power, Folklore Weisfeiler-Lehman test.
Compressor summary: The paper proposes $(k,t)$-FWL+ and Neighborhood$^2$-FWL (N$^2$-FWL), which are more powerful and flexible graph neural network frameworks than MPNNS, with better performance on various tasks.
Michael Hassid, Tal Remez, Tu Anh Nguyen, Itai Gat, Alexis Conneau, Felix Kreuk, Jade Copet, Alexandre Défossez, Gabriel Synnaeve, Emmanuel Dupoux, Roy Schwartz, Yossi Adi
https://openreview.net/forum?id=UlHueVjAKr
Keywords: LLM, speech, generative, GSLM
Compressor summary: TWIST trains SpeechLMs using a warm-start from pretrained textual models, leading to better performance than cold-start methods, and introduces new benchmarks for evaluation.
Krunoslav Lehman Pavasovic, Alain Durmus, Umut Simsekli
https://openreview.net/forum?id=UkPeUXML7s
Keywords: SGD, heavy-tails, wasserstein convergence
Compressor summary: The paper explores why stochastic gradient descent (SGD) shows heavy-tailed behavior in practical settings and investigates the role of offline SGD's stationary distribution in achieving power-law tails as more data points are added.
Yu-Kun Qiu, Guohao Xu, Wei-Shi Zheng
https://openreview.net/forum?id=UkAGqeWTuL
Keywords: 3D reconstruction
Compressor summary: IOAR is a 3D scene reconstruction method that uses a coarse-to-fine strategy to distinguish outer-surface, inner-surface, and surface voxels, resulting in more precise mesh predictions.
Atsuyuki Miyai, Qing Yu, Go Irie, Kiyoharu Aizawa
https://openreview.net/forum?id=UjtiLdXGMC
Keywords: out-of-distribution detection, vision-language foundation model, prompt learning
Compressor summary: The authors propose LoCoOp, a novel prompt learning method for few-shot OOD detection that uses CLIP's local features to remove ID-irrelevant information and achieve superior performance over existing methods.
Youpeng Zhao, Yaodong Yang, Zhenbo Lu, Wengang Zhou, Houqiang Li
https://openreview.net/forum?id=UgomCjCWjC
Keywords: Safe Multi-agent Reinforcement Learning, constrained policy optimisation, first-order optimisation
Compressor summary: MAFOCOPS is a novel safety-aware method for multi-agent systems in MARL, which effectively addresses high performance and enforcing safety constraints by solving a constrained optimization problem in the policy space.
Yash Jain, Harkirat Behl, Zsolt Kira, Vibhav Vineet
https://openreview.net/forum?id=UgSSOpqvPI
Keywords: mixture-of-experts, moe, object detection, mixture of datasets, multiple datasets
Compressor summary: The paper proposes a method called DAMEX that improves object detection by learning to route different datasets to their corresponding experts, achieving state-of-the-art results on various datasets and scenarios.
Jose Blanchet, Haoxuan Chen, Yiping Lu, Lexing Ying
https://openreview.net/forum?id=UdrybSp67L
Keywords: Information-theoretic Lower Bounds, Sobolev Embedding Theorem, Quadrature Rule
Compressor summary: The paper explores how machine learning-based control variates can reduce the variance of Monte Carlo sampling for estimating moments of a Sobolev function, and shows that they can improve the estimation rate under certain conditions and not in the presence of rare events.
David Yu-Tung Hui, Aaron Courville, Pierre-Luc Bacon
https://openreview.net/forum?id=UdaTyy0BNB
Keywords: deep reinforcement learning, Q-Learning, TD-Learning with function approximation, extreme value theory, maximum-likelihood estimation, moment-matching
Compressor summary: Double Gumbel Q-Learning is a new algorithm that handles heteroscedastic noise in Deep Q-Learning and performs well on various control tasks.
Allen Nie, Yuhui Zhang, Atharva Amdekar, Christopher J Piech, Tatsunori Hashimoto, Tobias Gerstenberg
https://openreview.net/forum?id=UdByCgCNdr
Keywords: cognitive science, causal reasoning, moral reasoning, dataset, language models
Compressor summary: The paragraph discusses a study that examines how well large language models make causal and moral judgments about text-based scenarios compared to human participants, using a dataset of stories from cognitive science papers annotated with factors influencing judgments.
Jing Yu Koh, Daniel Fried, Ruslan Salakhutdinov
https://openreview.net/forum?id=Uczck6TlSZ
Keywords: multimodal, vision-and-language, language models
Compressor summary: The proposed method fuses large language models with image encoder and decoder models for multimodal capabilities including image retrieval, novel image generation, and multimodal dialogue.
Lingjing Kong, Biwei Huang, Feng Xie, Eric Xing, Yuejie Chi, Kun Zhang
https://openreview.net/forum?id=Uc5yyiytR1
Keywords: Causal discovery, causal representation learning, latent variable models, causal structure learning, causal identifiability.
Compressor summary: This paper presents a method for identifying causal structures and latent variables in nonlinear latent hierarchical models using novel identifiability guarantees and an estimation procedure.
Xichen Ye, Xiaoqiang Li, Songmin Dai, Tong Liu, Yan Sun, Weiqin Tong
https://openreview.net/forum?id=Uafbv4rfJc
Keywords: noisy label learning, robust loss function, multiclass classification, computer vision
Compressor summary: The authors propose a new robust passive loss function called Normalized Negative Loss Functions (NNLFs) that improves the Active Passive Loss (APL) framework by focusing more on memorized clean samples and outperform state-of-the-art methods.
Hailin Zhang, Yujing Wang, Qi Chen, Ruiheng Chang, Ting Zhang, Ziming Miao, Yingyan Hou, Yang Ding, Xupeng Miao, Haonan Wang, Bochen Pang, Yuefeng Zhan, Hao Sun, Weiwei Deng, Qi Zhang, Fan Yang, Xing Xie, Mao Yang, Bin CUI
https://openreview.net/forum?id=UZlAjSnmvB
Keywords: document retrieval, model-based index, dense retrieval, residual quantization
Compressor summary: The paper proposes MEVI, a differentiable index enhanced by deep generative models, which improves vector index search efficiency while maintaining performance on academic benchmarks.
Tingyu Weng, Jun Xiao, Haiyong Jiang
https://openreview.net/forum?id=UYl9IIsjq7
Keywords: 3D point clouds, 3D recognition, part-based representation, unsupervised class discovery
Compressor summary: The paper proposes DNIK, a method to discover novel 3D classes by decomposing them into known parts and using a part relation encoding module (PRE) for better recognition.
Alexander Modell, Ian Gallagher, Emma Ceccherini, Nick Whiteley, Patrick Rubin-Delanchy
https://openreview.net/forum?id=UXtLrsG4Rf
Keywords: dynamic networks, representation learning, spectral methods
Compressor summary: The paragraph introduces a new method to represent network data over time using continuous trajectories that satisfy structural and temporal coherence, with estimation theory providing error control and smoothing trade-offs.
Derek Lim, Joshua Robinson, Stefanie Jegelka, Haggai Maron
https://openreview.net/forum?id=UWd4ysACo4
Keywords: Eigenvectors, spectral, geometry, universal approximation, graph, equivariance, invariance
Compressor summary: The authors propose novel sign equivariant neural network architectures for tasks such as building orthogonally equivariant models and learning node positional encodings, which they show have better performance than sign invariant models.
Apoorva Sharma, Sushant Veer, Asher Hancock, Heng Yang, Marco Pavone, Anirudha Majumdar
https://openreview.net/forum?id=URrUpcp6Qh
Keywords: Conformal Prediction, PAC Bayes, Generalization Theory
Compressor summary: The authors propose a framework for optimizing the efficiency of set-valued predictions based on PAC-Bayes theory, which allows using the entire calibration dataset and provides test-time coverage and efficiency guarantees.
Xiran Fan, Chun-Hao Yang, Baba C. Vemuri
https://openreview.net/forum?id=URAZeoIC1q
Keywords: Large-margin clssifier, Hyperbolic space, Horosphere, SVM, Geodesically convex, Global optimility, Busemann function
Compressor summary: The paper introduces a new large margin classifier with horospherical decision boundaries for hyperbolic spaces, which optimizes geodesically convex problems and outperforms existing methods.
Dorian Baudry, Fabien Pesquerel, Rémy Degenne, Odalric-Ambrym Maillard
https://openreview.net/forum?id=UPo8vlZ0wQ
Keywords: Multi-Armed Bandits
Compressor summary: The paper proposes methods to approximate the infimum KL in non-parametric stochastic bandits, reducing computational and memory costs while maintaining regret guaranties.
Jerry Tang, Meng Du, Vy A. Vo, Vasudev Lal, Alexander Huth
https://openreview.net/forum?id=UPefaFqjNQ
Keywords: fMRI, neuroscience, encoding models, multimodal transformers, language, vision
Compressor summary: The study used multimodal transformers to train encoding models that can transfer across fMRI responses to stories and movies, revealing shared semantic dimensions in language and vision processing.
Ming-Kun Xie, Jia-Hao Xiao, Hao-Zhe Liu, Gang Niu, Masashi Sugiyama, Sheng-Jun Huang
https://openreview.net/forum?id=UOB1UgPjuG
Keywords: Semi-supervised multi-label learning, pseudo labeling.
Compressor summary: The paper proposes Class-Aware Pseudo-Labeling (CAP), which uses class-aware thresholds to control pseudo-label assignment and improve semi-supervised multi-label learning by aligning pseudo-label distribution with the true one.
Suyash Gupta, Dominik Rothenhaeusler
https://openreview.net/forum?id=UKtjq3dIs0
Keywords: Distributional Stability, Distributional Robustness, Distributional Shifts, Generalizability
Compressor summary: The paper proposes a new measure of instability for statistical parameters that accounts for both distributional changes and directional shifts, and demonstrates its usefulness in transfer learning and improving estimation accuracy.
Weiwei Sun, Lingyong Yan, Zheng Chen, Shuaiqiang Wang, Haichao Zhu, Pengjie Ren, Zhumin Chen, Dawei Yin, Maarten de Rijke, Zhaochun Ren
https://openreview.net/forum?id=UKd6dpVGdu
Keywords: Information Retrieval, Document Retrieval, Generative Retrieval
Compressor summary: The paper proposes a novel document tokenization learning method, GenRet, which learns to encode document semantics into docids for generative retrieval models.
Teng Xiao, Huaisheng Zhu, Zhengyu Chen, Suhang Wang
https://openreview.net/forum?id=UK8mA3DRnb
Keywords: Contrastive Learning, Graph Representation Learning
Compressor summary: GraphACL is a simple algorithm for contrastive learning on graph-structured data that works well on both homophilic and heterophilic graphs without relying on prefabricated augmentations or homophily assumptions.
Nuoya Xiong, Yihan Du, Longbo Huang
https://openreview.net/forum?id=UJ9o8wbB5U
Keywords: safe reinforcement learning, step-wise violation, reinforcement learning theory
Compressor summary: The paper proposes two algorithms for safe reinforcement learning problems with step-wise violation constraints, one for when safe actions are known and another for when they are not, achieving near-optimal performance in both violation and regret.
Matteo Pagliardini, Daniele Paliotta, Martin Jaggi, François Fleuret
https://openreview.net/forum?id=UINHuKeWUa
Keywords: self-attention, large language models, transformers
Compressor summary: The paper proposes FlashAttention improvements that allow for more dynamic sparse attention patterns with no computational overhead and significant runtime speedup, while maintaining perplexity and increasing training speed for longer sequences in transformer language models.
Riccardo Giuseppe Margiotta, Sebastian Goldt, Guido Sanguinetti
https://openreview.net/forum?id=UHwmoJYwSV
Keywords: Adversarial attacks, data poisoning, online learning, optimal control, teacher-student setup, solvable model
Compressor summary: The paper investigates how attackers can manipulate online learning dynamics by perturbing data labels and studies the effects on learners with different architectures using both theory and experiments.
Sizhe Wei, Yuxi Wei, Yue Hu, Yifan Lu, Yiqi Zhong, Siheng Chen, Ya Zhang
https://openreview.net/forum?id=UHIDdtxmVS
Keywords: Collaborative Perception; BEV Flow; Time Asynchronization
Compressor summary: CoBEVFlow is a system to align asynchronous messages in collaborative perception using bird's eye view flow, handling irregular and continuous time stamps without discretization, and improving performance on real-world datasets.
Xueyan Zou, Jianwei Yang, Hao Zhang, Feng Li, Linjie Li, Jianfeng Wang, Lijuan Wang, Jianfeng Gao, Yong Jae Lee
https://openreview.net/forum?id=UHBrWeFWlL
Keywords: Generaic segmentation, interactive segmentation, referring segmentation, multi-modality prompting.
Compressor summary: SEEM is an interactive model for image segmentation that unifies different prompts, learns from segmentation history, and encodes text queries and mask labels semantically.
Junho Song, Keonwoo Kim, Jeonglyul Oh, Sungzoon Cho
https://openreview.net/forum?id=UFW67uduJd
Keywords: Multivariate time series, Anomaly detection
Compressor summary: MEMTO is a memory-guided Transformer that uses a novel memory module and bi-dimensional deviation-based detection criterion to achieve high anomaly detection performance on real-world multivariate time series data.
Heng Wang, Shangbin Feng, Tianxing He, Zhaoxuan Tan, Xiaochuang Han, Yulia Tsvetkov
https://openreview.net/forum?id=UDqHhbqYJV
Keywords: large language models, graph reasoning, structured reasoning
Compressor summary: The authors introduce NLGraph, a natural language benchmark for graph-based problem solving, and evaluate GPT-3/4 on it, finding that LLMs have preliminary graph reasoning abilities but are brittle in complex scenarios and improve with instruction-based prompting approaches.
Giovanni De Felice, John Y Goulermas, Vladimir Gusev
https://openreview.net/forum?id=UBUWFEwn7p
Keywords: Time Series, Kernel methods, NVAR processes, Dynamical systems, Reservoir Computing
Compressor summary: The authors propose a new kernel for time series analysis that is based on Nonlinear Vector AutoRegressive processes, simplifies hyperparameter setting, and performs well in classification tasks.
Yang Qin, Yuan Sun, Dezhong Peng, Joey Tianyi Zhou, Xi Peng, Peng Hu
https://openreview.net/forum?id=UBBeUjTja8
Keywords: Cross-modal learning, Image-text matching, Noisy correspondence.
Compressor summary: The text describes a new framework called CRCL that improves image-text matching by addressing challenges caused by noisy correspondence between visual and textual modalities using active complementary learning and self-refining correction.
Zitai Wang, Qianqian Xu, Zhiyong Yang, Yuan He, Xiaochun Cao, Qingming Huang
https://openreview.net/forum?id=UAow2kPsYP
Keywords: Imbalanced Learning, Re-weighting, Logit Adjustment, Genralization Analysis
Compressor summary: The paragraph discusses imbalanced real-world datasets and proposes a novel technique called data-dependent contraction to analyze modified losses for such datasets, along with a fine-grained generalization bound, a learning algorithm, and empirical validation.
Zeyang Zhang, Xin Wang, Ziwei Zhang, Guangyao Shen, Shiqi Shen, Wenwu Zhu
https://openreview.net/forum?id=UAFa5ZhR85
Keywords: Graph Neural Architecture Search, Unsupervised Learning, Self-supervised Learning
Compressor summary: The paper proposes a novel model for unsupervised graph neural architecture search, which can discover optimal architectures by disentangling latent graph factors and using self-supervised training and contrastive search.
Wenchong He, Zhe Jiang, Tingsong Xiao, Zelin Xu, Shigang Chen, Ronald Fick, MILES D MEDINA, Christine Angelini
https://openreview.net/forum?id=U9zRgpgdFI
Keywords: Spatial representation learning, transformer, quadtree, efficiency
Compressor summary: The paper introduces a new transformer model for massive spatial point data, addressing challenges such as long-range dependencies, non-uniform distributions, and high computational costs, by using a hierarchical structure, coarse approximation, and uncertainty estimation.
Yiwei Lu, Yaoliang Yu, Xinlin Li, Vahid Partovi Nia
https://openreview.net/forum?id=U6fp6IUBdr
Keywords: Neural network quantization, Model compression, Conditional gradient algorithm
Compressor summary: The paper introduces ProxConnect++, a generalization of BinaryConnect with forward-backward quantizers that enables principled binarization methods for neural networks without relying on heuristics or training tricks, and shows its effectiveness in image classification tasks.
Dongsheng Ding, Chen-Yu Wei, Kaiqing Zhang, Alejandro Ribeiro
https://openreview.net/forum?id=U6bhCLSPun
Keywords: Constrained Markov decision processes, policy gradient primal-dual methods, non-convex saddle-point problem, last-iterate convergence, entropy regularization, optimistic gradient
Compressor summary: The paper proposes two new methods for solving constrained Markov decision processes (MDPs) that have better convergence properties than existing Lagrangian-based policy search methods.
Chengliang Liu, Jie Wen, Yabo Liu, Chao Huang, Zhihao Wu, Xiaoling Luo, Yong Xu
https://openreview.net/forum?id=U4pFV192JQ
Keywords: Incomplete Multi-view Weak Multi-label Learning, Multi-view learning, Multi-label Classification
Compressor summary: The paper proposes a new deep neural network framework for incomplete multi-view weak multi-label learning, which decouples single-channel view-level representation into shared and view-specific representations, and uses cross-channel contrastive loss and label-guided graph regularization to enhance the features and preserve geometric structure.
Chen Tianqi, Weixiang Xu, Weihan Chen, Peisong Wang, Jian Cheng
https://openreview.net/forum?id=U4WTG06Yu3
Keywords: Winograd Convolution, Quantization
Compressor summary: The paper introduces PAW, which optimizes transformation procedures for Winograd convolution with post-training quantization and FSQ, a hardware-friendly method that balances range differences in the domain, improving accuracy and efficiency.
Guoyuan An, Ju-hyeong Seon, Inkyu An, Yuchi Huo, Sung-eui Yoon
https://openreview.net/forum?id=U1Kr8FTyhQ
Keywords: Landmarks retrieval, non-fine-tuning, spatial verification, explainable AI, hypothesis and test
Compressor summary: The paper proposes a new image retrieval method that uses a topological model instead of a spatial one to improve performance and overcome limitations in recognizing features, while retaining high explainability and being lightweight.
Hamza Tahir Chaudhry, Jacob A Zavatone-Veth, Dmitry Krotov, Cengiz Pehlevan
https://openreview.net/forum?id=Tz2uONpgpy
Keywords: Sequence Recall, Dense Associative Memory, Memory Capacity, Hopfield Networks, Biological Motor Control
Compressor summary: The authors propose a new sequence memory model that can store longer sequences by using nonlinear interactions and a novel recall rule, and discuss its biological implications for motor neuroscience.
Diego Martinez-Taboada, Aaditya Ramdas, Edward Kennedy
https://openreview.net/forum?id=TyLjNSbSOe
Keywords: kernel treatment effect, causal inference, maximum mean discrepancy
Compressor summary: The authors propose a new kernel-based test for distributional effects of binary treatments that is computationally efficient and has valid type-I error.
Dami Choi, Yonadav G Shavit, David Duvenaud
https://openreview.net/forum?id=TwLHB8sKme
Keywords: Large Scale Learning, ML Security, AI Governance
Compressor summary: The authors propose and explore efficient strategies for verifying the provenance of large neural models using "Proof-of-Training-Data" protocols that can detect various attacks, such as those based on harmful or beneficial data sources.
Chao Chen, Zhihang Fu, Kai Liu, Ze Chen, Mingyuan Tao, Jieping Ye
https://openreview.net/forum?id=TtCPFN5fhO
Keywords: Out-of-Distribution Detection, Parameter Sensitivity, Parameter Pruning, Neuron Pruning
Compressor summary: The paper proposes an efficient method called OPNP to detect out-of-distribution samples in machine learning models without requiring extra training data or resources, and demonstrates its effectiveness on various tasks and architectures.
Haiteng Zhao, Shengchao Liu, Chang Ma, Hannan Xu, Jie Fu, Zhi-Hong Deng, Lingpeng Kong, Qi Liu
https://openreview.net/forum?id=Tt6DrRCgJV
Keywords: Instruction, Molecule, Zero Shot, Graph, Language Model
Compressor summary: GIMLET is a model that uses language instructions and graph structures to accomplish molecule-related tasks without expensive lab experiments, outperforming existing methods in instruction-based zero-shot learning.
Cheng Cheng, Lin Song, Ruoyi Xue, Hang Wang, Hongbin Sun, Yixiao Ge, Ying Shan
https://openreview.net/forum?id=Ts0d8PvTeB
Keywords: Few-shot Learning; Vision-Language Model Adaption
Compressor summary: CLIP is a powerful vision-language pre-training method for image recognition, but it needs offline fine-tuning and can overfit. The proposed Meta-Adapter improves CLIP's few-shot learning capabilities online, without extra fine-tuning, achieving high efficiency and performance.
Xi Ye, Qiaochu Chen, Isil Dillig, Greg Durrett
https://openreview.net/forum?id=TqW5PL1Poi
Keywords: Reasoning, Chain-of-thought, Logical Reasoning, Arithmetic Reasoning, Prompting, In-Context Learning, Large Language Model
Compressor summary: The paper proposes SatLM, which uses an LLM to generate a declarative task specification and an automated theorem prover to derive the answer, improving reasoning capabilities for constraint solving problems.
Chunlin Sun, Linyu Liu, Xiaocheng Li
https://openreview.net/forum?id=TnTDiCppx5
Keywords: Uncertainty Quantification, Contextual LP, Robust Optimization, Distributionally Robust Optimization
Compressor summary: The paper proposes a method called predict-then-calibrate for solving optimization problems with contextual information, which separates the prediction and uncertainty quantification steps and allows for more flexible machine learning models and improved performance.
Clement Benard, Brian Staber, Sébastien Da Veiga
https://openreview.net/forum?id=TjgG4UT62W
Keywords: Bayesian inference, Markov chain Monte Carlo, kernelized Stein discrepancy, Stein thinning, kernel methods
Compressor summary: Stein thinning is a post-processing method for MCMC outputs that minimizes kernelized Stein discrepancy, but has some drawbacks; the paper analyzes these issues and proposes an improved algorithm with theoretical guarantees and experimental results.
Maximilien Dreveton, Felipe Schreiber Fernandes, Daniel R. Figueiredo
https://openreview.net/forum?id=TjJJmcHw9p
Keywords: community detection, stochastic block model, bregman divergence
Compressor summary: The text describes a new approach to cluster nodes in networks based on their connections and attributes, using an information-theoretic criterion and an iterative algorithm that performs better than existing methods.
Blake Bordelon, Paul Masset, Henry Kuo, Cengiz Pehlevan
https://openreview.net/forum?id=Tj0eXVPnRX
Keywords: Reinforcement Learning, Statistical Mechanics, Stochastic Gradient Descent
Compressor summary: The paper uses statistical physics concepts to study the effects of parameter choice and feature representation on the learning dynamics of reinforcement learning models with linear function approximators and sparse feedback.
Marcus Triplett, Marta Agnieszka Gajowa, Hillel Adesnik, Liam Paninski
https://openreview.net/forum?id=TiFMYdQiqp
Keywords: Neuroscience, neural stimulation, optogenetics, calcium imaging
Compressor summary: Bayesian target optimization is a novel computational approach that reduces off-target stimulation in two-photon optogenetics by modeling neural responses and optimizing laser powers and locations for precise stimulation of desired activity patterns.
Liliang Ren, Yang Liu, Shuohang Wang, Yichong Xu, Chenguang Zhu, ChengXiang Zhai
https://openreview.net/forum?id=TfbzX6I14i
Keywords: Sequence Modeling, Modularity, Sparsity, Attention Mechanism, State Space Model, Mixture of Experts, Neural Network, Transformer
Compressor summary: Sparse Modular Activation (SMA) improves sequence modeling efficiency by allowing neural networks to dynamically activate sub-modules, enabling linear inference complexity and better quality-efficiency trade-offs.
David Mizrahi, Roman Bachmann, Oguzhan Fatih Kar, Teresa Yeo, Mingfei Gao, Afshin Dehghan, Amir Zamir
https://openreview.net/forum?id=TegmlsD8oQ
Keywords: multimodal learning, multitask learning, representation learning, transfer learning, foundation models, generative models, computer vision
Compressor summary: The paper introduces 4M, a multimodal training scheme that trains a unified Transformer encoder-decoder on various input/output modalities, achieving versatile models for computer vision tasks.
Jon Donnelly, Srikar Katta, Cynthia Rudin, Edward P Browne
https://openreview.net/forum?id=TczT2jiPT5
Keywords: Rashomon Effect, Variable Importance, XAI, Stability, Interpretable Machine Learning
Compressor summary: The authors propose a new variable importance framework that quantifies the importance of a variable across all good models and is stable across data distribution, which can handle complex simulation setups and estimate true importance for real-world case studies like predicting HIV load.
Zheng Chang, Shuchen Weng, Peixuan Zhang, Yu Li, Si Li, Boxin Shi
https://openreview.net/forum?id=TcmjewOAd1
Keywords: Colorization, Language-based generation, Diffusion model
Compressor summary: The paper proposes a model that can colorize images using natural language descriptions of any level of detail, handling ambiguity and preserving local structures, while achieving better results than previous methods.
Zachary Coalson, Gabriel Ritter, Rakesh B Bobba, Sanghyun Hong
https://openreview.net/forum?id=TcG8jhOPdv
Keywords: Efficient Methods for NLP; Multi-exit Language Models; Adversarial Slowdown
Compressor summary: The paper evaluates the robustness of multi-exit language models to an adversarial slowdown attack called WAFFLE, which generates natural text perturbations that reduce computational savings and require further research on efficient yet robust models.
Junjiao Tian, Yen-Cheng Liu, James Smith, Zsolt Kira
https://openreview.net/forum?id=Tb7np0MInj
Keywords: fine-tuning, transfer learning, regularization
Compressor summary: The paper introduces Fast Trainable Projection, a new projection-based fine-tuning algorithm that improves efficiency and scalability for robust fine-tuning of pre-trained models in various vision tasks.
Yuzhang Shang, Zhihang Yuan, Yan Yan
https://openreview.net/forum?id=TZtw5YgxTE
Keywords: Dataset Distillation
Compressor summary: MIM4DD uses mutual information to measure and optimize the shared information between synthetic and real datasets in dataset distillation, improving their performance.
Hoomaan Maskan, Konstantinos C. Zygalakis, Alp Yurtsever
https://openreview.net/forum?id=TXq8PCRSoY
Keywords: Nesterov's accelerated gradient, gradient descent, Lyapunov function, gradient norm minimization, rate-matching, stochastic variance reduction, stochastic gradient descent, noisy gradient
Compressor summary: The text discusses a novel variational approach to unconstrained minimization that leads to faster convergence, interprets Nesterov's method as a high-resolution ODE discretization, and proposes a stochastic method for noisy gradients.
Hamish Flynn, David Reeb, Melih Kandemir, Jan Peters
https://openreview.net/forum?id=TXoZiUZywf
Keywords: Linear bandits, confidence sequences, martingales, convex optimization, cumulative regret, regret analysis
Compressor summary: The paper proposes new algorithms with better regret guarantees for the stochastic linear bandit problem using tailored confidence sequences for reward function estimation and action selection.
Zhang-Wei Hong, Aviral Kumar, Sathwik Karnik, Abhishek Bhandwaldar, Akash Srivastava, Joni Pajarinen, Romain Laroche, Abhishek Gupta, Pulkit Agrawal
https://openreview.net/forum?id=TW99HrZCJU
Keywords: offline reinforcement learning, reinforcement learning, sampling, experience replay
Compressor summary: Our method improves offline RL by using importance sampling weights to emulate data from an optimal policy, addressing the challenge of distributional mismatch due to imbalanced datasets.
Junghyun Lee, Hanseul Cho, Se-Young Yun, Chulhee Yun
https://openreview.net/forum?id=TW3ipYdDQG
Keywords: streaming, PCA, memory-limited, fair representation, online learning
Compressor summary: The paper proposes a new notion of fair principal component analysis (PCA) called PAFO-learnability and a memory-efficient algorithm for fair streaming PCA called fair noisy power method (FNPM).
Leonard Papenmeier, Luigi Nardi, Matthias Poloczek
https://openreview.net/forum?id=TVD3wNVH9A
Keywords: Bayesian optimization, global optimization, Gaussian process, combinatorial optimization, high-dimensional
Compressor summary: Bounce is a new algorithm for optimizing black-box functions with mixed and combinatorial input spaces, which outperforms existing methods in high-dimensional problems.
Bariscan Bozkurt, Cengiz Pehlevan, Alper Tunga Erdogan
https://openreview.net/forum?id=TUGoUNkccV
Keywords: Correlative information maximization, Biologically-plausible learning, Multi-compartment neural model
Compressor summary: The authors propose a new approach for signal propagation in biological neural networks that addresses concerns about the plausibility of conventional artificial neural networks and backpropagation algorithm, and provides a natural resolution to the weight symmetry problem.
Jiahui Lei, Congyue Deng, Bokui Shen, Leonidas Guibas, Kostas Daniilidis
https://openreview.net/forum?id=TTkklyFv7e
Keywords: 3D articulated objects, diffusion models, generative models
Compressor summary: NAP is a 3D deep generative model that synthesizes articulated objects using an innovative articulation tree/graph parameterization and diffusion-denoising probabilistic model.
Amelie Royer, Tijmen Blankevoort, Babak Ehteshami Bejnordi
https://openreview.net/forum?id=TSuq3debnD
Keywords: multitask, multidomain, optimization, population based training
Compressor summary: The paragraph discusses how training a single model on multiple inputs and outputs improves efficiency and transfer learning, but optimizing such networks is challenging, and proposes a population-based training method to find optimal scalarization weights.
Yunkai Gao, Rui Zhang, Jiaming Guo, Fan Wu, Qi Yi, Shaohui Peng, Siming Lan, Ruizhi Chen, Zidong Du, Xing Hu, Qi Guo, Ling Li, Yunji Chen
https://openreview.net/forum?id=TStMZH3Xqx
Keywords: offline meta-reinforcement learning, offline reinforcement learning, meta-reinforcement learning
Compressor summary: The paper proposes a new approach called CSRO to address the context shift problem in offline meta-reinforcement learning by reducing the influence of policy in context during both training and testing phases using mutual information representation learning and non-prior context collection strategy.
Minyoung Kim, Timothy Hospedales
https://openreview.net/forum?id=TRuqrVsmZK
Keywords: Parameter-efficient Foundation model fine-tuning, Bayesian methods, Stochastic-Gradient MCMC
Compressor summary: The paper proposes a new Bayesian sparse fine-tuning algorithm that uses a Laplace prior to select which parameters of foundation models need to be updated for specific tasks, achieving significant improvement over existing methods on NLP and vision benchmarks.
Mingzhen Sun, Weining Wang, Zihan Qin, Jiahui Sun, Sihan Chen, Jing Liu
https://openreview.net/forum?id=TRbklCR2ZW
Keywords: Video Generation, Video Autoencoder, Diffusion Probabilistic Model
Compressor summary: The GLOBER method generates coherent videos using a novel non-autoregressive approach that first creates global features and then synthesizes video frames based on them.
Tal Amir, Steven J. Gortler, Ilai Avni, Ravina Ravina, Nadav Dym
https://openreview.net/forum?id=TQlpqmCeMe
Keywords: Equivariant Neural Networks, Universal approximation, Geometric deep learning, multiset learning, injective multiset functions, learning on measures. WL test
Compressor summary: This paper shows that using non-polynomial activation in neural networks creates injective multiset functions, filling a gap between theory and practice, and also provides new results for graph neural networks and multiset functions approximation.
Qiong Wu, Wei Yu, Yiyi Zhou, Shubin Huang, Xiaoshuai Sun, Rongrong Ji
https://openreview.net/forum?id=TPeAmxwPK2
Keywords: vision and language, parameter and computation efficient transfer learning
Compressor summary: The paper proposes a dynamic architecture skipping approach to achieve efficient transfer learning for vision-language models by observing module significance with reinforcement learning and reducing redundant parameters and computation.
Yijian Qin, Xin Wang, Ziwei Zhang, Hong Chen, Wenwu Zhu
https://openreview.net/forum?id=TOxpAwp0VE
Keywords: graph neural network, neural architecture search, multi-task learning
Compressor summary: The paper proposes a novel method called MTGC3 that can automatically design optimal graph neural network architectures for multiple tasks simultaneously by learning their collaborative relationships and using a task-wise curriculum training strategy.
Xuehao Ding, Dongsoo Lee, Joshua Brendan Melander, George Sivulka, Surya Ganguli, Stephen Baccus
https://openreview.net/forum?id=TNLO8KNFFZ
Keywords: neural coding, theoretical neuroscience, stochastic methods, neural networks
Compressor summary: The study uses a neural network model of salamander retinal cells to show that retinal noise correlations limit information transmission in natural scenes and that population coding benefits from complementary coding.
Louis Sharrock, Lester Mackey, Christopher Nemeth
https://openreview.net/forum?id=TNAGFUcSP7
Keywords: Bayesian Inference, Particle Based Variational Inference, Sampling, Wasserstein Gradient Descent, Coin Betting, Constrained Domains
Compressor summary: The paper presents new particle-based algorithms for sampling on constrained domains that don't require learning rate adjustments and are based on convex optimization ideas and probability measure space perspective.
Chuofan Ma, Yi Jiang, Xin Wen, Zehuan Yuan, XIAOJUAN QI
https://openreview.net/forum?id=TKjX41IP7n
Keywords: Open-vocabulary Object Detection; Object-level Vision-Language Pretraining
Compressor summary: The paper introduces CoDet, a new method for open-vocabulary object detection that improves region-word alignment by grouping images with shared caption concepts and leveraging visual similarities.
Yaodong Yu, Sam Buchanan, Druv Pai, Tianzhe Chu, Ziyang Wu, Shengbang Tong, Benjamin David Haeffele, Yi Ma
https://openreview.net/forum?id=THfl8hdVxH
Keywords: white-box deep neural networks, representation learning, transformer, sparse coding
Compressor summary: The paper argues that representation learning aims to compress and transform data distributions using low-dimensional Gaussian mixtures, and shows how transformer blocks achieve this using sparse rate reduction as a unified objective.
Lijia Zhou, Zhen Dai, Frederic Koehler, Nathan Srebro
https://openreview.net/forum?id=TEpRn67828
Keywords: Uniform Convergence, Square-Root Lipschitz, Benign Overfitting, Minimal Norm Interpolation, Phase Retrieval, ReLU Regression, Matrix Sensing
Compressor summary: The paper provides uniform convergence guarantees for Gaussian data using Radamacher complexity and square-root-Lipschitz losses, generalizing previous smoothness-based results and covering non-smooth loss functions like phase retrieval and ReLU regression.
Edouard YVINEC, Arnaud Dapogny, Matthieu Cord, Kevin Bailly
https://openreview.net/forum?id=TDS3kqRteY
Keywords: deep learning, quantization, compression, acceleration, data-free
Compressor summary: REx is a data-free quantization method for DNNs that adapts to different devices and bit widths, improving accuracy-speed trade-offs and solving the outlier problem in large language models.
Mingjie Li, Yisen Wang, Zhouchen Lin
https://openreview.net/forum?id=TBOfDCX4Gz
Keywords: equilibirum models, neural networks
Compressor summary: The paper proposes GEQ, a new equilibrium model that uses Gaussian kernels to capture nonlinear feature dependencies in input data, improving performance over traditional OptEqs models.
Julien Niklas Siems, Konstantin Ditschuneit, Winfried Ripken, Alma Lindborg, Maximilian Schambach, Johannes Otterbach, Martin Genzel
https://openreview.net/forum?id=TAIYBdRb3C
Keywords: Interpretable Machine Learning, Generalized Additive Models, Concurvity, Multicollinearity, Regularization, Time-Series Forecasting, Interpretability
Compressor summary: The paper introduces a regularizer to improve interpretability of generalized additive models by reducing feature dependencies, and demonstrates its effectiveness on synthetic and real datasets.
Shengbang Tong, Erik Jones, Jacob Steinhardt
https://openreview.net/forum?id=T6iiOqsGOh
Keywords: safety, red-teaming, robustness, explainability, failures, multimodal models, vision-language, natural-language explanations
Compressor summary: MultiMon is a system that automatically finds and describes systematic failures in multimodal models using natural language.
Zhong-Qiu Wang, Shinji Watanabe
https://openreview.net/forum?id=T5h69frFF7
Keywords: Speech separation, microphone array processing, deep learning
Compressor summary: UNSSOR is an algorithm that uses over-determined mixtures to separate speakers in noisy recordings with deep neural networks and linear filters.
Hyemi Jang, Junsung Park, Dahuin Jung, Jaihyun Lew, Ho Bae, Sungroh Yoon
https://openreview.net/forum?id=T3SstRu5fq
Keywords: self-supervised image denoising, low-level vision
Compressor summary: PUCA is a novel J-invariant U-Net architecture for self-supervised image denoising that leverages patch-unshuffle/shuffle and dilated attention blocks to achieve state-of-the-art performance.
Yuhao Mao, Mark Niklas Mueller, Marc Fischer, Martin Vechev
https://openreview.net/forum?id=T2lM4ohRwb
Keywords: Certified Training, Certified Robustness, Adversarial Robustness, Robustness Verification
Compressor summary: TAPS is a new certified training method that combines IBP and PGD to optimize worst-case loss approximations, improving both certification and standard accuracies.
Alexandre Verine, benjamin negrevergne, Muni Sreenivas Pydi, Yann Chevaleyre
https://openreview.net/forum?id=SzYHu7EIwZ
Keywords: Generative Models, Precision, Recall, Optimization, f-Divergeces
Compressor summary: The paper proposes a new method for training generative models that optimizes a user-defined trade-off between image quality (precision) and diversity (recall), based on a novel family of $f$-divergences called PR-divergences.
Anshuk Uppal, Kristoffer Stensbo-Smidt, Wouter Boomsma, Jes Frellsen
https://openreview.net/forum?id=Sxu7xlUJGx
Keywords: Implicit models, Variational Inference, Bayesian Deep Learning, Bayesian Inference, Generative Modelling
Compressor summary: The paper proposes a new method for Bayesian models using neural samplers with implicit distributions to capture complex posteriors in high-dimensional spaces, enabling better performance in large Bayesian neural networks.
Dhawal Gupta, Yinlam Chow, Azamat Tulepbergenov, Mohammad Ghavamzadeh, Craig Boutilier
https://openreview.net/forum?id=SxXN3kNTsV
Keywords: Reinforcement Learning, Mixture of Experts, Dialogue Management
Compressor summary: The text discusses challenges and solutions for using reinforcement learning in developing conversational chatbots, focusing on combining mixture-of-expert language models to reduce action space and improve dialogue management.
Rong Wang, Wei Mao, Hongdong Li
https://openreview.net/forum?id=SxVHyYavHy
Keywords: hand-object pose estimation, physics simulation
Compressor summary: The paper proposes DeepSimHO, a method that combines deep learning with physics simulation to improve 3D pose estimation of hands interacting with objects in a single image.
Yannai Gonczarowski, Gregory Kehne, Ariel D. Procaccia, Ben Schiffer, Shirley Zhang
https://openreview.net/forum?id=Sv5bo2StIx
Keywords: computational social choice, statistics, distortion
Compressor summary: The paper explores how well voting rules perform on average with different preference distributions, proposing a new rule called binomial voting that works well across various situations.
Edward Raff, Amol Ashish Khanna, Fred Lu
https://openreview.net/forum?id=SuvDnzrKCo
Keywords: Sparsity, Differential Privacy, Regression
Compressor summary: The paper proposes an algorithm for training private regression models on sparse data, which can greatly reduce the runtime compared to existing methods.
Abraham David Smith, Michael J. Catanzaro, Gabrielle Angeloro, Nirav Patel, Paul Bendich
https://openreview.net/forum?id=SthlUe5xDP
Keywords: topological data analysis, persistent homology, convexity, AI safety, interpolation
Compressor summary: Topological parallax is a tool that compares AI models and datasets to ensure their geometric similarity, which is important for safety and robustness.
Quang Ho Nguyen, Truong Tuan Vu, Anh Tuan Tran, Khoi Nguyen
https://openreview.net/forum?id=StD4J5ZlI5
Keywords: Deep learning; Diffusion Models; Semantic Segmentation; Text-to-Image
Compressor summary: The paragraph describes a novel method for generating pixel-level semantic segmentation labels using the text-to-image generative model Stable Diffusion, which reduces the need for labor-intensive pixel-wise annotation and outperforms current approaches.
Bowen Tan, Yun Zhu, Lijuan Liu, Eric Xing, Zhiting Hu, Jindong Chen
https://openreview.net/forum?id=Srt1hhQgqa
Keywords: multi-task, large language models, pretrain model
Compressor summary: Cappy is a small pretrained scorer that improves the performance and efficiency of multi-task large language models without requiring their finetuning or access to their parameters.
Zirui Liu, Guanchu Wang, Shaochen Zhong, Zhaozhuo Xu, Daochen Zha, Ruixiang Tang, Zhimeng Jiang, Kaixiong Zhou, Vipin Chaudhary, Shuai Xu, Xia Hu
https://openreview.net/forum?id=SquMNyrk1O
Keywords: memory-efficient tuning, language model, transformers
Compressor summary: The authors propose a new unbiased estimator called \sas\ for reducing memory usage in large pre-trained language models, which can achieve significant memory reduction and improve downstream task performance.
Martino Bernasconi, Matteo Castiglioni, Alberto Marchesi, Mirco Mutti
https://openreview.net/forum?id=SqTUGq0R7j
Keywords: Bayesian Persuasion, MPD, information design, signaling
Compressor summary: The paper proposes a new class of optimal signaling schemes for Bayesian persuasion that consider the receiver's future rewards and can be computed efficiently using a promise-form representation.
Xidong Wu, Jianhui Sun, Zhengmian Hu, Aidong Zhang, Heng Huang
https://openreview.net/forum?id=SpStmVboGy
Keywords: Federated Learning, Non-Convex Optimization, Minimax Optimization
Compressor summary: The paper proposes FL algorithms for federated nonconvex minimax optimization problems, reducing communication complexity and achieving similar performance to centralized methods in some settings.
Maria Sofia Bucarelli, Matilde Fjeldsø Larsen, Chris Schwiegelshohn, Mads Toftrup
https://openreview.net/forum?id=Sp0yOBfelp
Keywords: Subspace Clustering, Learning Theory, Clustering, Error bounds
Compressor summary: The paper studies learning bounds for clustering problems with center-based and subspace-based objectives and provides near-optimal convergence rates for both cases.
Nicholas Franzese, Adam Dziedzic, Christopher A. Choquette-Choo, Mark R. Thomas, Muhammad Ahmad Kaleem, Stephan Rabanser, Congyu Fang, Somesh Jha, Nicolas Papernot, Xiao Wang
https://openreview.net/forum?id=SouroWC5Un
Keywords: collaborative learning, robust aggregation, secure machine learning
Compressor summary: The paper proposes a peer-to-peer learning scheme that is secure from malicious servers and robust to malicious clients, and transforms any compatible algorithm for aggregation of model updates in such settings.
Tianyu Xie, Cheng Zhang
https://openreview.net/forum?id=SoLebIqHgZ
Keywords: phylogenetic inference, autoregressive model, graph neural network, density estimation, variational inference
Compressor summary: The paper introduces ARTree, a deep autoregressive model for phylogenetic inference using graph neural networks, which can efficiently learn probabilistic models over tree topologies without relying on hand-engineered features.
Yuhao Wang, Enlu Zhou
https://openreview.net/forum?id=SjiLtmZETc
Keywords: Q-learning, risk-averse reinforcement learning, off-policy learning, Bayesian risk Markov decision process, distributionally robust Markov decision process
Compressor summary: The paper proposes a robust reinforcement learning framework using BRMDP and a multi-stage Q-learning algorithm that learns a risk-averse optimal policy with real environment observations and strong convergence.
Dogyoon Song, Kyunghee Han
https://openreview.net/forum?id=Sg3aCpWUQP
Keywords: Frechet regression, principal component regression, non-Euclidean, low-rank matrix, errors-in-variables analysis
Compressor summary: The paper proposes a new method that combines global Fr\'echet regression and principal component regression to improve efficiency and accuracy in non-Euclidean regression analysis by leveraging low-rank structure in the covariate matrix.
Marco Rando, Cesare Molinari, Lorenzo Rosasco, Silvia Villa
https://openreview.net/forum?id=SfdkS6tt81
Keywords: nonsmooth optimization;zeroth order optimization;nonsmooth zeroth-order
Compressor summary: O-ZD is a new structured finite-difference algorithm for non-smooth black-box optimization that uses smooth approximations of the target function and orthogonal directions to approximate its gradient efficiently.
Jihao Andreas Lin, Javier Antoran, Shreyas Padhy, David Janz, José Miguel Hernández-Lobato, Alexander Terenin
https://openreview.net/forum?id=Sf9goJtTCE
Keywords: Gaussian processes, scalable learning, posterior sampling, Bayesian optimization
Compressor summary: Stochastic gradient descent is a computationally efficient method for solving linear systems in Gaussian processes, achieving state-of-the-art performance and accurate predictions even when not converging quickly to the optimum.
Bingcong Li, Georgios B. Giannakis
https://openreview.net/forum?id=Sf3t6Bth4P
Keywords: generalization, optimization, neural networks
Compressor summary: VaSSO improves SAM by making adversaries more aggressive, leading to better generalization, stability, and robustness in neural network training.
Sarath Sreedharan, Michael Katz
https://openreview.net/forum?id=Sf17j2pkCU
Keywords: Planning, Reinforcement Learning, Exploration
Compressor summary: The paragraph discusses a new method for learning optimistic symbolic models that guide reinforcement learning agents in sparse reward settings, improving exploration speed through action generalization.
Shentong Mo, Enze Xie, Ruihang Chu, Lanqing HONG, Matthias Nießner, Zhenguo Li
https://openreview.net/forum?id=Se71ks7Mfz
Keywords: Diffusion Models, Transformers, 3D Shape Generation
Compressor summary: The paper proposes a novel Diffusion Transformer (DiT-3D) for generating high-quality 3D shapes from voxelized point clouds, which improves scalability and quality over existing U-Net methods and leverages pre-trained DiT-2D on ImageNet.
Ziteng Sun, Ananda Theertha Suresh, Jae Hun Ro, Ahmad Beirami, Himanshu Jain, Felix Yu
https://openreview.net/forum?id=SdYHLTCC5J
Keywords: autoregressive sampling; computation efficiency; optimal transport
Compressor summary: The authors propose SpecTr, a new autoregressive sampling algorithm that uses optimal transport with membership cost to improve the speed and quality of decoding from large language models.
Shuyang Sun, Weijun Wang, Andrew G. Howard, Qihang Yu, Philip Torr, Liang-Chieh Chen
https://openreview.net/forum?id=SaMrN9tnxE
Keywords: Panoptic segmentation, efficient models
Compressor summary: The paper introduces ReMaX, a technique that improves the training of mask transformers for efficient panoptic segmentation by adding relaxations to predictions.
Shijie Wang, Jianlong Chang, Haojie Li, Zhihui Wang, Wanli Ouyang, Qi Tian
https://openreview.net/forum?id=SaII5qMgKH
Keywords: Open-set Fine-grained Retrieval, Visual Attribute, Unknown Categories
Compressor summary: VAPNet is a novel network that learns visual attributes from known categories and integrates them into a retrieval model for unknown categories, without manual annotations.
Xingsi Dong, Si Wu
https://openreview.net/forum?id=SWU8YLlFVH
Keywords: Baysian brain, sampling-based inference, energy-based models, local learning, exponential-family
Compressor summary: The Hierarchical Exponential-family Energy-based (HEE) model is a new method for inferring and learning in the brain, which uses neural adaptation and allows fast computation and representation similar to biological vision systems.
Lijie Fan, Dilip Krishnan, Phillip Isola, Dina Katabi, Yonglong Tian
https://openreview.net/forum?id=SVjDiiVySh
Keywords: contrastive learning; CLIP; large language model
Compressor summary: LaCLIP is a method that enhances CLIP training by rewriting image text descriptions with diverse sentences while preserving original meanings, improving transfer performance without extra computation or memory.
Kaipeng Zheng, Huishuai Zhang, Weiran Huang
https://openreview.net/forum?id=SVUQX1W7RL
Keywords: Few-shot learning
Compressor summary: This paper shows that using Kendall's rank correlation instead of geometric similarity metrics improves few-shot learning performance across different methods and datasets.
Jiannan Xiang, Tianhua Tao, Yi Gu, Tianmin Shu, Zirui Wang, Zichao Yang, Zhiting Hu
https://openreview.net/forum?id=SVBR6xBaMl
Keywords: Language Model, World Model, Embodied Experience
Compressor summary: The paper proposes a method to improve large language models' abilities in reasoning and acting in the physical world by finetuning them with embodied experiences from a simulated environment and using techniques like EWC and LoRA to preserve generality.
Daolang Huang, Ayush Bharti, Amauri H Souza, Luigi Acerbi, Samuel Kaski
https://openreview.net/forum?id=STrXsSIEiq
Keywords: Simulation-based inference, model misspecification, likelihood-free inference, approximate Bayesian computation, neural posterior estimation
Compressor summary: The authors propose a general approach to handle model misspecification in simulation-based inference methods and demonstrate its superior performance on artificially and realistically misspecified models.
Chaofei Fan, Nick Hahn, Foram Kamdar, Donald Avansino, Guy H Wilson, Leigh Hochberg, Krishna V. Shenoy, Jaimie M. Henderson, Francis R Willett
https://openreview.net/forum?id=STqaMqhtDi
Keywords: brain-computer interface, self-training, continual online learning
Compressor summary: The paper proposes a method to improve intracortical brain-computer interfaces (iBCIs) for communication by enabling self-recalibration using large language models, achieving high performance and long-term stability in a human participant.
Pandeng Li, Chen-Wei Xie, Hongtao Xie, Liming Zhao, Lei Zhang, Yun Zheng, Deli Zhao, Yongdong Zhang
https://openreview.net/forum?id=SQouRKRIXY
Keywords: Video Moment Retrieval, Diffusion Model
Compressor summary: MomentDiff is a generative diffusion-based framework that learns to retrieve specific temporal segments of videos based on language descriptions, overcoming temporal location biases in existing methods.
Robert F Allison, Anthony Stephenson, Samuel F, Edward Pyzer-Knapp
https://openreview.net/forum?id=SQP1H9Jy8W
Keywords: Gaussian Processes, Bayesian Inference, Regression, Bayesian Nonparametrics, Kernel Methods
Compressor summary: GP nearest-neighbour prediction is accurate and computationally efficient even when data size is large, as parameter estimation and model assumptions become less important.
Kevin J Miller, Maria K Eckstein, Matthew Botvinick, Zeb Kurth-Nelson
https://openreview.net/forum?id=SOEF0i0G1z
Keywords: Cognitive modeling, neural networks, interpretability, disentangling, neuroscience, rodent behavior
Compressor summary: The authors propose an alternative approach to construct cognitive models from behavioral data using sparse and interpretable recurrent neural networks.
Sebastian Sanokowski, Wilhelm Franz Berghammer, Sepp Hochreiter, Sebastian Lehner
https://openreview.net/forum?id=SLx7paoaTU
Keywords: Combinatorial Optimization, Entropy Regularization, Graph Neural Networks, Statistical Mechanics
Compressor summary: Unsupervised learning methods using probabilistic approaches for combinatorial optimization may not perform well on difficult problems due to the assumption of independent solution variables; author propose Subgraph Tokenization, which represents a set of solution variables by one token, and annealed entropy regularization, which improves efficiency and stability.
Yizhe Zhang, Jiatao Gu, Zhuofeng Wu, Shuangfei Zhai, Joshua M. Susskind, Navdeep Jaitly
https://openreview.net/forum?id=SLwy8UVS8Y
Keywords: Text generation, diffusion model, NLP
Compressor summary: PLANNER combines latent semantic diffusion with autoregressive generation to produce fluent and controllable long-form text for various tasks.
Andres Potapczynski, Marc Anton Finzi, Geoff Pleiss, Andrew Gordon Wilson
https://openreview.net/forum?id=SLtNFERsHo
Keywords: Machine Learning, Numerical Linear Algebra, partial differential equations, Gaussian processes, equivariance, graph learning, spectral analysis
Compressor summary: CoLA is a framework for large-scale linear algebra problems in machine learning that combines a linear operator abstraction with compositional dispatch rules to automatically construct memory and runtime efficient numerical algorithms.
Tyler Kastner, Murat A Erdogdu, Amir-massoud Farahmand
https://openreview.net/forum?id=SLTQluG80x
Keywords: Reinforcement learning, Risk-Sensitive Reinforcement Learning, Model-Based Reinforcement Learning, Distributional Reinforcement Learning
Compressor summary: The paper explores learning optimal models for risk-sensitive reinforcement learning using distributional reinforcement learning and introduces two new notions of model equivalence.
Yanxiang Ma, Minjing Dong, Chang Xu
https://openreview.net/forum?id=SIE9N5nnHg
Keywords: Adversarial robustness; Randomized defense; Random parameters optimization
Compressor summary: The authors propose a new randomized defense mechanism for deep neural networks, called Constrained Trainable Random Weight (CTRW), which optimizes randomness parameters to balance natural performance and adversarial robustness, achieving better results on several datasets and benchmarks.
Kai Hu, Andy Zou, Zifan Wang, Klas Leino, Matt Fredrikson
https://openreview.net/forum?id=SHyVaWGTO4
Keywords: adversarial robustness, ImageNet, Lipschitz-based certification, ResNet, adversarial examples, ML security
Compressor summary: The paper proposes LiResNet, a new architecture and EMMA, a loss function, that improve the efficiency of calculating Lipschitz bounds and stabilize robust training for deep networks on large-scale datasets like ImageNet.
Nurendra Choudhary, Nikhil Rao, Chandan K. Reddy
https://openreview.net/forum?id=SHVwG9yOEk
Keywords: meta learning, hyperbolic networks, scalability, graph neural networks
Compressor summary: H-GRAM is a novel meta-learning method for hyperbolic neural networks that learns inductive biases from local subgraphs and transfers them to new tasks, improving generalization and scalability in few-shot settings.
Yifan Hu, Jie Wang, Yao Xie, Andreas Krause, Daniel Kuhn
https://openreview.net/forum?id=SHBksHKutP
Keywords: stochastic optimization, bilevel optimization, contextual stochastic optimization, Multilevel Monte Carlo
Compressor summary: CSBO is a new framework for bilevel optimization with contextual information that requires a double-loop gradient method based on MLMC to handle its challenges and has applications in various areas like meta-learning and federated learning.
Zihao Wang, Lin Gui, Jeffrey Negrea, Victor Veitch
https://openreview.net/forum?id=SGlrCuwdsB
Keywords: disentanglement; representation learning; text-controlled generative models; diffusion models
Compressor summary: The paper explores how text-guided generative models like Stable Diffusion represent and manipulate disentangled concepts in a subspace of their representation space.
Joe Suk, Samory Kpotufe
https://openreview.net/forum?id=SGerL9HMrp
Keywords: multi-armed bandits, non-stationary, contextual bandits, nonparametric, Lipschitz
Compressor summary: The paper studies nonparametric contextual bandits with changing reward functions and proposes a new concept of change that accounts for locality and significance, achieving adaptivity in the minimax regret rate.
Dongyang Fan, Celestine Mendler-Dünner, Martin Jaggi
https://openreview.net/forum?id=SGKbHXoLCI
Keywords: Collaborative training, decentralized learning, consensus reaching
Compressor summary: Our method helps agents improve their models by sharing unlabeled data and using trust weights to reach consensus on pseudo-labels, which enhances performance and reduces bad influence.
Zhihao Wu, Zhao Zhang, Jicong Fan
https://openreview.net/forum?id=SFfOt1oDsX
Keywords: graph neural network, kernel method
Compressor summary: Graph convolutional kernel machine (GCKM) is a framework for graph-based machine learning that uses kernel functions with graph convolution and has advantages over deep graph convolutional networks (GCN).
Yahong Yang, Haizhao Yang, Yang Xiang
https://openreview.net/forum?id=SE73LzWNjr
Keywords: VC-dimension, pseudo-dimension, Sobolev space, generalization error, nearly optimal approximation
Compressor summary: The paper studies how to accurately estimate the complexity of derivative functions in deep neural networks and applies these estimations to various physics-informed machine learning tasks.
Xitong Liang, Alberto Caron, Samuel Livingstone, Jim Griffin
https://openreview.net/forum?id=SCsJFNcSHQ
Keywords: Bayesian Networks, structure MCMC on graphs, Structure Learning, Random neighborhood samplers, Locally informed Metropolis-Hastings schemes
Compressor summary: PARNI-DAG is a new MCMC sampler that efficiently learns Directed Acyclic Graphs (DAGs) from observational data under causal sufficiency by using adaptive random neighborhood proposals and pre-tuning parameters based on a skeleton graph.
Yixin Liu, Kaize Ding, Qinghua Lu, Fuyi Li, Leo Yu Zhang, Shirui Pan
https://openreview.net/forum?id=SAzaC8f3cM
Keywords: Anomaly Detection, Graph Neural Networks, Explanation, Self-Interpretation
Compressor summary: SIGNET is a model that detects abnormal graphs and explains why they are abnormal using subgraphs from the input graph and its dual hypergraph.
Kai Liu, Zhihang Fu, Chao Chen, Sheng Jin, Ze Chen, Mingyuan Tao, Rongxin Jiang, Jieping Ye
https://openreview.net/forum?id=SA2KrosYjY
Keywords: out-of-distribution detection, vision-language models, category-extendable classification
Compressor summary: This paper introduces two contexts, perceptual and spurious, for detecting out-of-distribution samples in vision-language models using automatic prompt tuning and applies them to a novel detection method called CATEX.
Yesom Park, Taekyung Lee, Jooyoung Hahn, Myungjoo Kang
https://openreview.net/forum?id=S8hg5LpFvz
Keywords: Surface reconstruction, Signed distance function, Implicit neural representations, Point cloud
Compressor summary: The paper proposes a new surface reconstruction method using implicit neural representations, which learns a signed distance function from a point cloud and imposes physical constraints for better results.
Yibo Jiang, Bryon Aragam
https://openreview.net/forum?id=S8DFqgmEbe
Keywords: graphical models, directed acyclic graphs, causality, identifiability, causal representation learning, unknown interventions
Compressor summary: The paper presents methods for identifying latent causal graphs from interventions without making parametric assumptions or knowing the number of hidden variables, introducing new graphical concepts and characterizing the limits of edge orientations in a general setting.
Shuang Qiu, Ziyu Dai, Han Zhong, Zhaoran Wang, Zhuoran Yang, Tong Zhang
https://openreview.net/forum?id=S75ccNdOYG
Keywords: Markov game, Partial observation, Function approximation, Posterior sampling, Reinforcement Learning
Compressor summary: The paper presents new algorithms for competitive reinforcement learning in different game settings, using complexity measures and posterior sampling methods that handle partial observability and trade-off exploration and exploitation.
Hangfan Zhang, Jinyuan Jia, Jinghui Chen, Lu Lin, Dinghao Wu
https://openreview.net/forum?id=S6ajVZy6FA
Keywords: Backdoor Attack, Federated Learning
Compressor summary: The paper introduces a new backdoor attack, A3FL, that adapts the trigger to survive global training dynamics in federated learning.
Ziqian Zhong, Ziming Liu, Max Tegmark, Jacob Andreas
https://openreview.net/forum?id=S5wmbQc1We
Keywords: mechanistic interpretability, algorithmic phase transitions, arithmetic learning, neural network, transformer, ensemble
Compressor summary: The study shows that neural networks can learn different algorithms, including some novel ones, when trained on modular addition tasks, indicating the complexity of algorithm discovery in neural networks.
Sourya Basu, Pulkit Katdare, Prasanna Sattigeri, Vijil Chenthamarakshan, Katherine Rose Driggs-Campbell, Payel Das, Lav R. Varshney
https://openreview.net/forum?id=S4NN3OOiwP
Keywords: zero-shot learning, equivariant machine learning, equivariant fine-tuning, pretrained models
Compressor summary: λ-equitune is a method to improve equivariant outputs from non-equivariant neural networks by averaging features with importance weights learned from data, leading to better zero-shot and finetuned results than equitune on diverse applications and models.
Kaiwen Wang, Kevin Zhou, Runzhe Wu, Nathan Kallus, Wen Sun
https://openreview.net/forum?id=S3Y0VvegGm
Keywords: Reinforcement Learning Theory, Distributional Reinforcement Learning, Small-Loss Bounds, First-order regret
Compressor summary: The paper explains why distributional reinforcement learning (DistRL) is better than non-distributional RL by using small-loss bounds, proposes a DistCB algorithm with empirical results, and shows that pessimistic DistRL has novel small-loss PAC bounds in offline RL.
Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, Shashank Gupta, Bodhisattwa Prasad Majumder, Katherine Hermann, Sean Welleck, Amir Yazdanbakhsh, Peter Clark
https://openreview.net/forum?id=S37hOerQLB
Keywords: LLMs, Iterative Refinement, Feedback-driven Generation
Compressor summary: The Self-Refine approach improves the output of large language models by iteratively refining their own outputs based on feedback, without needing additional data or training.
Alexis Bellot, Alan Malek, Silvia Chiappa
https://openreview.net/forum?id=S2k5dBb91q
Keywords: Transportability, transfer learning, bandits
Compressor summary: The paper proposes a new bandit algorithm that uses causal models to learn from batch data and exploit invariances across related environments, leading to lower regret and faster learning.
James Chenhao Liang, Yiming Cui, Qifan Wang, Tong Geng, Wenguan Wang, Dongfang Liu
https://openreview.net/forum?id=S1KGaTSOTS
Keywords: Universal Model, Clustering
Compressor summary: ClusterFormer is a novel vision model that uses clustering with TransFormers, enabling explainable and transferable image processing for various tasks and outperforming specialized architectures.
Valerii Likhosherstov, Krzysztof Marcin Choromanski, Kumar Avinava Dubey, Frederick Liu, Tamas Sarlos, Adrian Weller
https://openreview.net/forum?id=S0xrBMFihS
Keywords: Gaussian kernel, softmax kernel
Compressor summary: The authors propose new parameterized random features to approximate Gaussian and softmax kernels, which reduce variance and improve performance in kernel methods and Transformers.
Bill Yuchen Lin, Yicheng Fu, Karina Yang, Faeze Brahman, Shiyu Huang, Chandra Bhagavatula, Prithviraj Ammanabrolu, Yejin Choi, Xiang Ren
https://openreview.net/forum?id=Rzk3GP1HN7
Keywords: interactive reasoning, text game, agents, action planning, large language models
Compressor summary: SwiftSage is a new agent framework that combines behavior cloning and large language models to improve action planning for complex reasoning tasks by using two modules: one for fast thinking and another for deliberate thought.
Aleksandr Beznosikov, Martin Takáč, Alexander Gasnikov
https://openreview.net/forum?id=Rvk1wdwz1L
Keywords: convex optimization, variational inequalities, similarity, local methods, compression, partial participation
Compressor summary: The paper proposes a novel triple synergy technique to reduce communication costs for solving large-scale variational inequality problems in machine learning, achieving the best theoretical guarantees and outperforming existing methods.
Jianing Zhu, Geng Yu, Jiangchao Yao, Tongliang Liu, Gang Niu, Masashi Sugiyama, Bo Han
https://openreview.net/forum?id=RuxBLfiEqI
Keywords: out-of-distribution detection, outlier exposure
Compressor summary: DivOE is a novel framework for OOD detection that diversifies auxiliary outliers and extrapolates them during training using a multi-step optimization method.
Xiyuan Yang, Wenke Huang, Mang Ye
https://openreview.net/forum?id=RteNLuc8D9
Keywords: federated learning, differential privacy, personalization
Compressor summary: The paper proposes FedDPA, a novel federated learning method that uses layer-wise Fisher information to flexibly personalize models while mitigating convergence difficulties caused by clipping operations in differential privacy.
Anindya Sarkar, Nathan Jacobs, Yevgeniy Vorobeychik
https://openreview.net/forum?id=Rs6pzz21U4
Keywords: Visual Active Search, Reinforcement Learning
Compressor summary: Visual active search (VAS) has been proposed as a modeling framework in which visual cues are used to guide exploration, with the goal of identifying regions of interest in a large geospatial area. Its potential applications include identifying hot spots of rare wildlife poaching activity, search-and-rescue scenarios, identifying illegal trafficking of weapons, drugs, or people, and many others. State of the art approaches to VAS include applications of deep reinforcement learning (DRL), which yield end-to-end search policies, and traditional active search, which combines predictions with custom algorithmic approaches. While the DRL framework has been shown to greatly outperform traditional active search in such domains, its end-to-end nature does not make full use of supervised information attained either during training, or during actual search, a significant limitation if search tasks differ significantly from those in the training distribution. We propose an approach that combines the strength of both DRL and conventional active search approaches by decomposing the search policy into a prediction module, which produces a geospatial distribution of regions of interest based on task embedding and search history, and a search module, which takes the predictions and search history as input and outputs the search distribution. In addition, we develop a novel meta-learning approach for jointly learning the resulting combined policy that can make effective use of supervised information obtained both at training and decision time. Our extensive experiments demonstrate that the proposed representation and meta-learning frameworks significantly outperform state of the art in visual active search on several problem domains.
Yijun Dong, Kevin Miller, Qi Lei, Rachel Ward
https://openreview.net/forum?id=RrdBNXBUIF
Keywords: Relational knowledge distillation, Semi-supervised learning, Spectral clustering, Sample complexity
Compressor summary: The paper provides a theoretical understanding and analysis of relational knowledge distillation in semi-supervised classification problems using spectral clustering and cluster-aware frameworks.
Zihan Chen, Howard Hao Yang, Tony Quek, Kai Fong Ernest Chong
https://openreview.net/forum?id=RqjQL08UFc
Keywords: Personalized federated learning, spectral bias, co-distillation, communication efficiency
Compressor summary: The text introduces spectral distillation, a new personalized federated learning method that leverages model spectrum information to improve performance and efficiency on diverse data settings.
Pengfei Wei, Lingdong Kong, Xinghua Qu, Yi Ren, zhiqiang xu, Jing Jiang, Xiang Yin
https://openreview.net/forum?id=Rp4PA0ez0m
Keywords: action recognition, unsupervised domain adaptation, video analysis
Compressor summary: This paper introduces a new method called TranSVAE to generate videos across different domains by disentangling static and dynamic information and removing spatial and temporal domain differences using constraints and adversarial learning.
Nicholas Rittler, Kamalika Chaudhuri
https://openreview.net/forum?id=RmxP5ZcQhC
Keywords: learning theory, active learning, multi-group learning
Compressor summary: The paper presents a novel active learning algorithm for learning from multiple groups with minimal label queries, and shows its advantages in some scenarios, as well as providing results for special cases of realizable groups and approximations for the general case.
Zhenyu Zhang, Ying Sheng, Tianyi Zhou, Tianlong Chen, Lianmin Zheng, Ruisi Cai, Zhao Song, Yuandong Tian, Christopher Re, Clark Barrett, Zhangyang Wang, Beidi Chen
https://openreview.net/forum?id=RkRrPp7GKO
Keywords: Large Language Models; Efficient Generative Inference
Compressor summary: The paper proposes a new memory management technique for large language models that reduces their cost by focusing on the most important tokens and evicting less relevant ones.
Nikki Lijing Kuang, Ming Yin, Mengdi Wang, Yu-Xiang Wang, Yian Ma
https://openreview.net/forum?id=RiyH3z7oIF
Keywords: Posterior Sampling, Reinforcement Learning Theory, Linear Markov Decision Processes, Delayed Feedback, Langevin Monte Carlo
Compressor summary: This paper proposes two optimistic value-based algorithms for reinforcement learning with delayed feedback, and analyzes their performance in terms of regret.
Yifang Chen, Yingbing Huang, Simon Shaolei Du, Kevin Jamieson, Guanya Shi
https://openreview.net/forum?id=RiwPYAMLur
Keywords: active learning, representation learning, robotics, theory
Compressor summary: The paper proposes a general framework for active representation learning that optimally chooses source tasks to sample from and shows its effectiveness in various domains with theoretical and empirical results.
Hao Dong, Ismail Nejjar, Han Sun, Eleni Chatzi, Olga Fink
https://openreview.net/forum?id=RiSMijlsLT
Keywords: Domain Generalization, Multi-modal Learning, Distribution Shift, Out-of-distribution Generalization
Compressor summary: SimMMDG is a multi-modal domain generalization framework that splits features into modality-specific and modality-shared components, uses supervised contrastive learning to encourage joint properties, imposes distance constraints for diversity, and employs cross-modal translation for regularization and missing-modality generalization.
Ethan Weinberger, Ian Connick Covert, Su-In Lee
https://openreview.net/forum?id=RhE01dqo8u
Keywords: Feature selection, contrastive analysis, computational biology, representation learning, information theory
Compressor summary: The paper introduces contrastive feature selection (CFS), a method for selecting features in the contrastive analysis setting, and shows that it outperforms existing methods on biomedical datasets.
Yu Pan, Ye Yuan, Yichun Yin, Zenglin Xu, Lifeng Shang, Xin Jiang, Qun Liu
https://openreview.net/forum?id=RgNXKIrWyU
Keywords: Model Growth, Efficient Training, Pretrained Model, Multi-linearity
Compressor summary: The paper proposes a method to accelerate training large models by linearly correlating each weight of the target model with all weights of the pretrained model, reducing computational and spatial complexity.
Federico Danieli, Miguel Sarabia, Xavier Suau, Pau Rodriguez, Luca Zappella
https://openreview.net/forum?id=RgD92idA32
Keywords: Acceleration, layer-parallelization, diffusion, Parallel Cyclic Reduction
Compressor summary: DeepPCR is a novel algorithm that parallelizes typically sequential operations in deep neural networks, such as forward and backward passes, to achieve significant speedups.
Ruihai Wu, Kai Cheng, Yan Zhao, Chuanruo Ning, Guanqi Zhan, Hao Dong
https://openreview.net/forum?id=Re2NHYoZ5l
Keywords: Visual Affordance for Robotics, Articulated Object Manipulation, Occlusion Handling
Compressor summary: The paper proposes an environment-aware affordance framework for home-assistant robots that learns from object-level actionable priors and various occlusions using a contrastive learning approach.
Rui Peng, Xiaodong Gu, Luyang Tang, Shihe Shen, Fanqi Yu, Ronggang Wang
https://openreview.net/forum?id=Rcit6V3vus
Keywords: Generalizable Neural Surface, Volume Rendering, Signed Distance Function
Compressor summary: The paper introduces GenS, a neural surface reconstruction model that generalizes well to new scenes and outperforms existing methods without 3D supervision or long-time optimizations.
Huiwon Jang, Jihoon Tack, Daewon Choi, Jongheon Jeong, Jinwoo Shin
https://openreview.net/forum?id=RZGtK2nDDJ
Keywords: Self-Supervised Learning, Modality-Agnostic Self-Supervised Learning, Meta-Learning, Masked Auto-Encoder
Compressor summary: The paper proposes MetaMAE, a meta-learning enhanced version of Masked Auto-Encoder, to improve self-supervised learning across diverse modalities.
Mo Tiwari, Ryan Kang, Donghyun Lee, Sebastian Thrun, Ilan Shomorony, Martin Jinye Zhang
https://openreview.net/forum?id=RWcfpmjlYm
Keywords: multi-armed bandits, clustering, k-medoids, best-arm identification
Compressor summary: BanditPAM++ is a faster and improved version of BanditPAM, which is a randomized $k$-medoids algorithm with state-of-the-art complexity and accuracy.
Dong Qiao, Chris Ding, Jicong Fan
https://openreview.net/forum?id=RW7rZ8Y3Bp
Keywords: clustering, federated learning, privacy
Compressor summary: Federated learning has a significant advantage in protecting information privacy. Many scholars proposed various secure learning methods within the framework of federated learning but the study on secure federated unsupervised learning especially clustering is limited. We in this work propose a secure kernelized factorization method for federated spectral clustering on distributed dataset. The method is non-trivial because the kernel or similarity matrix for spectral clustering is computed by data pairs, which violates the principle of privacy protection. Our method implicitly constructs an approximation for the kernel matrix on distributed data such that we can perform spectral clustering under the constraint of privacy protection. We provide a convergence guarantee of the optimization algorithm, reconstruction error bounds of the Gaussian kernel matrix, and the sufficient condition of correct clustering of our method. We also present some results of differential privacy. Numerical results on synthetic and real datasets demonstrate that the proposed method is efficient and accurate in comparison to the baselines.
Harit Vishwakarma, Heguang Lin, Frederic Sala, Ramya Korlakai Vinayak
https://openreview.net/forum?id=RUCFAKNDb2
Keywords: Auto Labeling, Active Learning, Selective Classification
Compressor summary: The paragraph discusses threshold-based auto-labeling (TBAL), a technique to reduce manual annotation in machine learning, and analyzes its sample complexity, quality, and potential pitfalls using theoretical and empirical methods.
Di Huang, Ziyuan Nan, Xing Hu, Pengwei Jin, Shaohui Peng, Yuanbo Wen, Rui Zhang, Zidong Du, Qi Guo, Yewen Pu, Yunji Chen
https://openreview.net/forum?id=RTRS3ZTsSj
Keywords: programming language, large language models, program synthesis, code generation, human-ai interaction
Compressor summary: ANPL is an interactive programming system that allows users to refine code generated by LLMs using structured decompositions and natural language specifications.
Lorenzo Loconte, Nicola Di Mauro, Robert Peharz, Antonio Vergari
https://openreview.net/forum?id=RSGNGiB1q4
Keywords: knowledge graph, knowledge graph embeddings, probabilistic circuits, probabilistic reasoning, tractable inference
Compressor summary: The authors propose generative circuit models for knowledge graph embedding that enable exact maximum-likelihood estimation, efficient sampling, and logical constraint satisfaction while maintaining performance for link prediction.
Zeyang Zhang, Xin Wang, Ziwei Zhang, Zhou Qin, Weigao Wen, Hui Xue', Haoyang Li, Wenwu Zhu
https://openreview.net/forum?id=RRUVZygUtr
Keywords: Dynamic Graph Neural Networks, Out-of-Distribution Generalization
Compressor summary: The paper proposes a new method, SILD, to handle distribution shifts in dynamic graphs by capturing and utilizing invariant and variant spectral patterns.
Bochuan Cao, Changjiang Li, Ting Wang, Jinyuan Jia, Bo Li, Jinghui Chen
https://openreview.net/forum?id=RRSltzPc7w
Keywords: Image Generation Godels, Latent Diffusion Models, Image Purifying
Compressor summary: IMPRESS is a platform that evaluates the effectiveness of imperceptible perturbations in protecting original images from unauthorized data usage by diffusion-based image generation models.
Giorgia Ramponi, Pavel Kolev, Olivier Pietquin, Niao He, Mathieu Lauriere, Matthieu Geist
https://openreview.net/forum?id=RPFd3D3P3L
Keywords: Mean-field games, Imitation Learning
Compressor summary: The paper studies imitation learning in mean-field games and introduces a new solution concept called the Nash imitation gap, showing that it is harder than single-agent imitation learning when the dynamics depend on the population distribution.
Zicheng Zhang, Bonan Li, Xuecheng Nie, Congying Han, Tiande Guo, Luoqi Liu
https://openreview.net/forum?id=RNVwm4BzXO
Keywords: diffusion model, video editing, text-to-video diffusion model
Compressor summary: The paper proposes a new EI$^2$ model to improve Text-to-Image diffusion models for video editing by addressing inconsistency issues caused by modules learning temporal information and using two attention mechanisms to enhance spatial and temporal features.
Yueh-Hua Wu, Xiaolong Wang, Masashi Hamaya
https://openreview.net/forum?id=RMeQjexaRj
Keywords: Offline Reinforcement Learning, Trajectory Stitching, Decision Transformer
Compressor summary: The Elastic Decision Transformer (EDT) improves upon Decision Transformer (DT) by better handling trajectory stitching and adapting history length for optimal performance, leading to competitive results with Q Learning-based methods.
Xiaoqian Wu, Yong-Lu Li, Jianhua Sun, Cewu Lu
https://openreview.net/forum?id=RJq9bVEf6N
Keywords: neuro-symbolic, visual reasoning, human activity understanding
Compressor summary: The paragraph discusses a new symbolic system for visual activity understanding that uses large language models to improve explainability, generalization, and data efficiency.
Deepak Narayanan, Keshav Santhanam, Peter Henderson, Rishi Bommasani, Tony Lee, Percy Liang
https://openreview.net/forum?id=RJpAz15D0S
Keywords: Systems for Machine Learning, Inference efficiency, Transformer models, Text generation APIs, Capability-efficiency tradeoffs
Compressor summary: The authors propose idealized runtime and cost models to measure inference efficiency for large language models, and compare ten LLMs from 2022 using these metrics.
Mehdi Azabou, Michael Jacob Mendelson, Nauman Ahad, Maks Sorokin, Shantanu Thakoor, Carolina Urzay, Eva L Dyer
https://openreview.net/forum?id=RInTOCEL3l
Keywords: animal behavior, behavioral neuroscience, self-supervised learning, multi-timescale
Compressor summary: The paper presents a multi-task representation learning model for animal behavior that predicts actions over future timesteps and incorporates short- and long-term dynamics, achieving success in various environments and ranking first in the MABe 2022 Multi-Agent Behavior challenge.
Tianyu Liu, Qitan Lv, Jie Wang, Shuling Yang, Hanzhu Chen
https://openreview.net/forum?id=RHDXkRPNQa
Keywords: inductive relation prediction, knowledge graph completion, knowledge graph reasoning
Compressor summary: REST is a novel GNN model for inductive relation prediction that initializes edge features only for the target link and uses RNN-based functions for edge-wise message passing to learn rule-induced subgraph representations.
Yufan Cai, Yun Lin, Chenyan Liu, Jinglian Wu, Yifan Zhang, Yiming Liu, Yeyun Gong, Jin Song Dong
https://openreview.net/forum?id=RFgv7cfMUy
Keywords: Code Summarization, Adaptation, Language Model
Compressor summary: Adacom is a novel approach to improve code comment generation by adapting deep learning models on the fly using contradictory training samples.
Junliang Li, Yajun Yang, Qinghua Hu, Xin Wang, Hong Gao
https://openreview.net/forum?id=RFE1eI0zNZ
Keywords: public opinion field effect, heterogeneous networks, representation learning, trending topic diffusion
Compressor summary: The paper proposes a new representation learning framework that considers public opinion field and social circle influence effects for trending topic diffusion analysis.
Fu Luo, Xi Lin, Fei Liu, Qingfu Zhang, Zhenkun Wang
https://openreview.net/forum?id=RBI4oAbdpm
Keywords: Neural Combinatorial Optimization, Generalization, Large scale problem, Heavy decoder
Compressor summary: The LEHD model is a novel neural combinatorial optimization method that can learn to generalize across problem sizes and scales, enabling it to solve large-scale real-world problems more effectively than previous methods.
Tongxin Li, Yiheng Lin, Shaolei Ren, Adam Wierman
https://openreview.net/forum?id=RACcp8Zbr9
Keywords: Time-varying MDP, Learning-augmented online algorithm, consistency and robustness tradeoff
Compressor summary: The paper investigates how using information about the advice generation process affects the tradeoff between consistency and robustness in single-trajectory time-varying MDPs with machine-learned Q-value advice, showing that it can lead to near-optimal performance.
Lei Ke, Mingqiao Ye, Martin Danelljan, Yifan liu, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu
https://openreview.net/forum?id=RA7ND878XP
Keywords: segment anything, zero-shot segmentation, high-quality segmentation
Compressor summary: The authors propose HQ-SAM, an improved version of the Segment Anything Model (SAM) that can accurately segment any object with minimal additional parameters and computation, while maintaining SAM's efficiency and zero-shot generalizability.
Anastasia Batsheva, Andrei Chertkov, Gleb Ryzhakov, Ivan Oseledets
https://openreview.net/forum?id=R9R7YDOar1
Keywords: Tensor Train, Black Box Optimization, Sampling, Optimal Control
Compressor summary: PROTES is a novel black-box optimization method that uses probabilistic sampling from a low-parametric tensor train format and performs better than popular discrete methods on complex problems.
Jincheng Cao, Ruichen Jiang, Nazanin Abolfazli, Erfan Yazdandoost Hamedani, Aryan Mokhtari
https://openreview.net/forum?id=R8GF0EsNsI
Keywords: Bilevel optimization, stochastic optimization
Compressor summary: The paper proposes new methods for solving stochastic bilevel optimization problems that improve the complexity of existing methods by using stochastic cutting planes and conditional gradient updates with variance reduction.
Eoin M. Kenny, Weipeng Fuzzy Huang
https://openreview.net/forum?id=R6wXP7txer
Keywords: Semifactual Explanation, Counterfactual Explanation, Explainable AI, Recourse, User Study
Compressor summary: The paper introduces semifactuals, a new type of XAI that optimizes positive outcomes without crossing decision boundaries, and shows they are more useful than counterfactuals for users who receive positive outcomes like loan approvals.
Yihong Sun, Bharath Hariharan
https://openreview.net/forum?id=R6qMmdl4qP
Keywords: monocular, depth estimation, dynamical scenes, motion segmentation, self-supervised
Compressor summary: Dynamo-Depth is a method that estimates depth, motion, and segmentation from unlabeled videos by first separating moving objects from static scenes.
Sebastien Lachapelle, Divyat Mahajan, Ioannis Mitliagkas, Simon Lacoste-Julien
https://openreview.net/forum?id=R6KJN1AUAR
Keywords: identifiability, nonlinear ICA, causal representation learning, disentanglement, object-centric representation learning, extrapolation
Compressor summary: The paragraph discusses how additive decoders can identify latent variables and generate "out-of-support" images for representation learning, with theoretical guarantees and applications in nonlinear ICA and OCRL methods.
Qitian Wu, Wentao Zhao, Chenxiao Yang, Hengrui Zhang, Fan Nie, Haitian Jiang, Yatao Bian, Junchi Yan
https://openreview.net/forum?id=R4xpvDTWkV
Keywords: graph transformers, large graphs
Compressor summary: The authors propose Simplified Graph Transformers (SGFormer), which uses one-layer attention to efficiently predict node properties on large graphs without complex models or pre-processing, achieving significant inference acceleration and scalability.
Gregor Bachmann, Sotiris Anagnostidis, Thomas Hofmann
https://openreview.net/forum?id=R45A8eKcax
Keywords: MLP, scaling-laws, inductive bias, DL theory
Compressor summary: This paper investigates the limits and performance of multi-layer perceptrons (MLPs) in vision tasks, showing that they improve with scale and can mimic modern models.
Shutong Ding, Tianyu Cui, Jingya Wang, Ye Shi
https://openreview.net/forum?id=R2rJq5OHdr
Keywords: Deep Equilibrium Models, Neural Ordinary Differential Equations, Homotopy Continuation
Compressor summary: The authors propose a new implicit model called HomoODE, which combines the advantages of DEQs and Neural ODEs, and show its effectiveness on image classification tasks using homotopy continuation.
Tianqin Li, Ziqi Wen, Yangfan Li, Tai Sing Lee
https://openreview.net/forum?id=QzcZb3fWmW
Keywords: neuroscience, computer vision, shape & texture bias
Compressor summary: The paper proposes that sparse coding can introduce shape bias into deep learning models, improving their robustness and structure generation, and provides code at a specified GitHub repository.
Fangxin Wang, Lu Cheng, Ruocheng Guo, Kay Liu, Philip S. Yu
https://openreview.net/forum?id=QxYzmYmQQe
Keywords: Equal Opportunity; Fair Machine Learning; Conformal Prediction; Uncertainty Quantification
Compressor summary: The authors propose Equal Opportunity of Coverage (EOC), a fair machine learning approach that ensures equal coverage rates for different groups and maintains population-level coverage, while using Binned Fair Quantile Regression to improve prediction interval width.
Wang Xinrui, wan wenhai, Chuanxing Geng, Shao-Yuan Li, Songcan Chen
https://openreview.net/forum?id=QwvaqV48fB
Keywords: positive and unlabeled learning, machine learning, deep learning, temporal point process, data imbalance
Compressor summary: The paper proposes a novel balanced resampling technique for PUL methods and uses temporal point processes to detect trends in scores, improving performance in real-world settings.
Zian Li, Xiyuan Wang, Yinan Huang, Muhan Zhang
https://openreview.net/forum?id=QwQ5HhhSNo
Keywords: geometric deep learning, expressiveness, equivariant neural networks, universality
Compressor summary: $k$-DisGNNs improve geometric deep learning by capturing high-order information, unifying existing models, and being universal function approximators.
Philip Sun, David Simcha, Dave Dopson, Ruiqi Guo, Sanjiv Kumar
https://openreview.net/forum?id=QvIvWMaQdX
Keywords: ann, quantization, mips, nearest neighbor search, retrieval
Compressor summary: SOAR is a new data indexing technique that uses multiple redundant representations and an orthogonality-amplified residual loss to improve approximate nearest neighbor search performance, speed, and memory efficiency.
Xingjian Bai, Christian Coester
https://openreview.net/forum?id=Qv7rWR9JWa
Keywords: sorting, learning-augmented algorithms, algorithms with predictions, adaptive sorting
Compressor summary: The paper explores how learning-augmented algorithms can use predictions to improve sorting efficiency, and designs new algorithms that adapt to prediction quality and achieve optimal comparison complexity.
Phillip Lippe, Bastiaan S. Veeling, Paris Perdikaris, Richard E Turner, Johannes Brandstetter
https://openreview.net/forum?id=Qv6468llWS
Keywords: Neural PDE Solvers, Neural Operators, Temporal Stability, Long-Horizon Modeling, Autoregressive Forecasting
Compressor summary: This paper introduces a new neural network model called PDE-Refiner that improves the accuracy and stability of solving time-dependent partial differential equations by using a multistep refinement process and spectral data augmentation, outperforming existing methods.
Shikai Fang, Xin Yu, Shibo Li, Zheng Wang, Robert Kirby, Shandian Zhe
https://openreview.net/forum?id=Qu6Ln7d9df
Keywords: Tensor Decomposition, streaming method, Bayesian model
Compressor summary: SFTL is a method that uses Gaussian processes to model the temporal evolution of factors in tensor decomposition for streaming data.
Stefan Matthes, Zhiwei Han, Hao Shen
https://openreview.net/forum?id=QrB38MAAEP
Keywords: Disentanglement, Contrastive Learning, Identifiability, Representation Learning, Nonlinear ICA
Compressor summary: This paper extends disentanglement guarantees and identifiability of latents to a broader family of contrastive learning methods and evaluates them on benchmark datasets.
Ziyad Benomar, Vianney Perchet
https://openreview.net/forum?id=QpZubU4yD9
Keywords: online algorithms, competitive ratio, learning augmented algorithms, scheduling, ski-rental, secretary
Compressor summary: The paper explores algorithms that can request a limited number of predictions at any time while solving classical problems in competitive analysis.
Ayush Sekhari, Karthik Sridharan, Wen Sun, Runzhe Wu
https://openreview.net/forum?id=QoeOVgayLp
Keywords: Selective Sampling, Imitation Learning, Learning from Expert Feedback, Theory, General purpose algorithms
Compressor summary: This paper proposes an interactive algorithm for Imitation Learning that uses selective sampling to query a noisy expert for feedback and achieves the best-known bounds for regret and queries in general function classes and multiple actions, with tight theoretical results.
Lukas Eisenmann, Zahra Monfared, Niclas Alexander Göring, Daniel Durstewitz
https://openreview.net/forum?id=QmPf29EHyI
Keywords: dynamical systems, bifurcations, Recurrent Neural Networks, attractors, training algorithm, BPTT, exploding and vanishing gradient problem, nonlinear dynamics, time series
Compressor summary: The authors propose a novel heuristic algorithm to detect bifurcations and fixed points in ReLU-based RNNs, which can help understand their dynamical behavior and the effects of parameter variations during training.
Haoxing Tian, Alex Olshevsky, Ioannis Paschalidis
https://openreview.net/forum?id=QlfGOVD5PO
Keywords: Reinforcement Learning, Actor-Critic, gradient splitting, neural network
Compressor summary: This paper demonstrates that actor-critic methods can converge using deep neural networks with any number of hidden layers, improving the connection between theory and practice in this area.
Marco Bagatella, Georg Martius
https://openreview.net/forum?id=QlbZabgMdK
Keywords: deep reinforcement learning, unsupervised reinforcement learning, goal-conditioned reinforcement learning, model-based planning
Compressor summary: Curiosity-driven exploration helps learn robust dynamics models, but extracting goal-conditioned behavior from unsupervised data is challenging; combining model-based planning and graph-based aggregation improves zero-shot goal-reaching performance.
Chunming He, Kai Li, Yachao Zhang, Guoxia Xu, Longxiang Tang, Yulun Zhang, Zhenhua Guo, Xiu Li
https://openreview.net/forum?id=QlHosp050r
Keywords: Concealed Object Segmentation, Weakly-Supervised Learning, Segment Anything Model
Compressor summary: The paper proposes a new WSCOS method that tackles intrinsic similarity and weak supervision challenges by using multi-scale feature grouping and SAM-based segmentation with strategies to improve supervision quality, achieving state-of-the-art performance.
Siwon Kim, Sangdoo Yun, Hwaran Lee, Martin Gubri, Sungroh Yoon, Seong Joon Oh
https://openreview.net/forum?id=QkLpGxUboF
Keywords: Personal identifiable information, Private data leakage, Large language model
Compressor summary: ProPILE is a tool that helps people check how much of their personal information might be leaked by large language models like OPT-1.3B trained on web data.
Liulei Li, Jianan Wei, Wenguan Wang, Yi Yang
https://openreview.net/forum?id=QjI36zxjbW
Keywords: Human-Object Interaction, Neuro-Symbolic Computing, Compositional Generalization
Compressor summary: LogicHOI is a new HOI detector that uses neural-logic reasoning and Transformer to infer feasible interactions between entities, taking into account affordances and proxemics, and achieves significant improvements over existing methods.
Ryan Rogers, Gennady Samorodnitsky, Steven Wu, Aaditya Ramdas
https://openreview.net/forum?id=QezJbfW01r
Keywords: differential privacy, brownian motion, composition, martingale
Compressor summary: The authors develop privacy filters that combine differentially private mechanisms with ex-post private mechanisms, such as noise reduction, for better performance in tasks requiring accuracy and privacy.
Kobbi Nissim, Uri Stemmer, Eliad Tsfadia
https://openreview.net/forum?id=QatZNssk7T
Keywords: Adaptive Data Analysis, Differential Privacy, Statistical Queries
Compressor summary: The paper investigates adaptive data analysis with a balanced adversary model, where the analyst does not know the underlying distribution and shows that it is hard to answer more than n^2 queries under standard cryptography assumptions.
Zhiqun Zuo, Mohammad Mahdi Khalili, Xueru Zhang
https://openreview.net/forum?id=QZo1cge4Tc
Keywords: Counterfactual fairness, Representation learning
Compressor summary: This paper introduces a new algorithm for training machine learning models that are fair under Counterfactual Fairness, a notion that requires individuals to be treated equally regardless of their social group, by using all available features instead of excluding those related to sensitive attributes.
Yang Cao, Yihan Zeng, Hang Xu, Dan Xu
https://openreview.net/forum?id=QW5ouyyIgG
Keywords: 3D vision, open-vocabulary perception, multi-modal learning, point cloud, 3D object detection
Compressor summary: The paper proposes CoDA, a unified framework for simultaneous novel object localization and classification in 3D scenes using both geometric and semantic priors, as well as cross-modal alignment between point cloud and image/text modalities.
Arthur Jacot
https://openreview.net/forum?id=QVpfk2C3Dm
Keywords: Feature Learning, Symmetry Learning, Theory of Deep Learning, Weight Decay
Compressor summary: The paper investigates how deep neural networks learn low-dimensional representations and balance complexity/irregularity, proving a conjectured bottleneck structure in the limit of infinite depth.
Hao ZHANG, Tianyuan DAI, Yanbo Xu, Yu-Wing Tai, Chi-Keung Tang
https://openreview.net/forum?id=QUkYZNhfc6
Keywords: NeRF Editing, NeRF Relighting, Face, Diffusion model, 3d synthesis, GAN inversion
Compressor summary: The paper introduces FaceDNeRF, a method to create high-quality 3D faces from single images with semantic editing and relighting capabilities, outperforming existing 2D editing approaches.
Jinjin Gu, Xianzheng Ma, Xiangtao Kong, Yu Qiao, Chao Dong
https://openreview.net/forum?id=QRWA5nTWuM
Keywords: Image Deraining, Generalization, Interpretation
Compressor summary: Despite being successful in labs, deep deraining networks struggle with real-world applications due to overfitting; simplifying training background images can improve their generalization performance.
Charles Guille-Escuret, Pau Rodriguez, David Vazquez, Ioannis Mitliagkas, Joao Monteiro
https://openreview.net/forum?id=QRAS5wSgEy
Keywords: Contrastive learning, OOD detection, adversarial detection, MMD, ImageNet-O, Anomaly detection, CIFAR-10.1
Compressor summary: The authors propose a novel self-supervised contrastive learning method (CADet) that can detect both unseen classes and adversarial perturbations in machine learning systems without needing labels or additional OOD samples.
Tao Wang, Sylvia Lee Herbert, Sicun Gao
https://openreview.net/forum?id=QQidjdmyPp
Keywords: Reinforcement learning, policy gradient, non-smooth landscape
Compressor summary: The paper proposes a framework to understand and handle non-smooth or fractal optimization landscapes in deep reinforcement learning, which can cause failures in policy gradient methods.
Lianghe Shi, Weiwei Liu
https://openreview.net/forum?id=QNUs3Ramad
Keywords: learning theory
Compressor summary: Gradual self-training with adversarial training improves both clean and adversarial accuracies in domains with intermediate adaptations, as it outperforms standard training when dealing with incorrect pseudo-labels.
Chuanruo Ning, Ruihai Wu, Haoran Lu, Kaichun Mo, Hao Dong
https://openreview.net/forum?id=QLllDwizVd
Keywords: articulated object manipulation, few-shot learning, visual affordance for robotics
Compressor summary: Where2Explore is a framework that helps robots learn how to manipulate novel object categories by efficiently exploring local geometries shared across different objects.
Yingtai Xiao, Guanlin He, Danfeng Zhang, Daniel Kifer
https://openreview.net/forum?id=QKSejqE8Vp
Keywords: differential privacy, marginals, matrix mechanism, scalability
Compressor summary: ResidualPlanner is an optimal and scalable matrix mechanism for confidentiality-protecting data release that can optimize various loss functions and handle large scale settings efficiently.
Adrián Javaloy, Pablo Sanchez Martin, Isabel Valera
https://openreview.net/forum?id=QIFoCI7ca1
Keywords: causality, causal inference, normalizing flows, identifiability, interventions, counterfactuals
Compressor summary: The authors present a method to identify and learn causal models from observational data using autoregressive normalizing flows, and demonstrate its applicability on real-world problems with mixed discrete-continuous data.
Ruihan Yang, Stephan Mandt
https://openreview.net/forum?id=QIBpzaDCAv
Keywords: generative model, diffusion model, image compression, computer vision
Compressor summary: The paper presents a lossy image compression framework using diffusion generative models that maps images to and from latent space, achieving good performance on various metrics.
Xiaomeng Hu, Pin-Yu Chen, Tsung-Yi Ho
https://openreview.net/forum?id=QGrkbaan79
Keywords: Large Language Models, Text Detection, Adversarial Learning, Paraphrase
Compressor summary: RADAR is a framework that uses adversarial learning to train a robust AI-text detector capable of detecting paraphrased texts generated by large language models, outperforming existing methods.
Jihoon Tack, Subin Kim, Sihyun Yu, Jaeho Lee, Jinwoo Shin, Jonathan Richard Schwarz
https://openreview.net/forum?id=QGmNMtK3pQ
Keywords: Meta-Learning, Efficient Meta-Learning, Neural Fields, Implicit Neural Representations, Data Pruning
Compressor summary: The paragraph describes an efficient optimization-based meta-learning technique for large-scale neural field training that reduces memory usage, improves model quality, and allows fast learning of high-quality neural fields across multiple modalities.
Samuel Lanthaler, T. Konstantin Rusch, Siddhartha Mishra
https://openreview.net/forum?id=QGQsOZcQ2H
Keywords: neural ODE, universal approximation, oscillator
Compressor summary: The paper introduces neural oscillators as a universal class for machine learning architectures and proves their ability to approximate continuous and causal operator mappings using forced harmonic oscillators with a nonlinear read-out.
Ali Behrouz, Farnoosh Hashemi, Sadaf Sadeghian, Margo Seltzer
https://openreview.net/forum?id=QG4nJBNEar
Keywords: Hypergraph Learning, Temporal Networks, Higher-order Temporal Motifs, Inductive Representation Learning
Compressor summary: CAT-Walk is a method to learn dynamic laws of temporal hypergraphs using a higher-order walk, SetWalk, and an adaptive pooling strategy, SetMixer, that achieves high performance in various tasks.
Mina Dalirrooyfard, Slobodan Mitrovic, Yuriy Nevmyvaka
https://openreview.net/forum?id=QDByreuQyk
Keywords: Differential Privacy, clustering, multiway cut, min cut, graph partitioning
Compressor summary: The authors study the complexity of maintaining differential privacy while finding minimal edge removals to disconnect two or more nodes in a graph, and develop efficient algorithms for this problem with near optimal trade-offs between privacy and efficiency.
Mintong Kang, Dawn Song, Bo Li
https://openreview.net/forum?id=QB7ot7p6j7
Keywords: adversarial attack, adversarial purification, adversarial robustness, diffusion model
Compressor summary: The paper proposes DiffAttack, an efficient method to break diffusion-based purification defenses in adversarial examples, improving on existing attacks by over 10% on ImageNet and over 20% on CIFAR-10.
Mengyu Wang, Henghui Ding, Jun Hao Liew, Jiajun Liu, Yao Zhao, Yunchao Wei
https://openreview.net/forum?id=Q9CNA7B7v2
Keywords: Refinement, Segmentation, Discrete Diffusion
Compressor summary: The paper introduces SegRefiner, a model-agnostic method that improves object masks using denoising diffusion steps and shows its superior performance on various segmentation tasks.
Zheyun Qin, Cheng Han, Qifan Wang, Xiushan Nie, Yilong Yin, Xiankai Lu
https://openreview.net/forum?id=Q6zd1hr7sD
Keywords: Point Cloud Segmentation, Prototypical Classifier, Unified Framework
Compressor summary: ProtoSEG is a prototype-based model that unifies semantic, instance, and panoptic segmentation tasks for point clouds using a Transformer architecture and achieves competitive results on several benchmarks.
Kyowoon Lee, Seongun Kim, Jaesik Choi
https://openreview.net/forum?id=Q5tuGgqJwt
Keywords: Offline Reinforcement Learning, Trajectory Optimization, Diffusion Models, Sequential Decision Making
Compressor summary: The paper proposes a method to improve the reliability of long-horizon plans generated by diffusion models using a restoration gap metric and an attribution map regularizer, which enables refinement of infeasible plans and provides explainability.
Hanting Chen, Yunhe Wang, Jianyuan Guo, Dacheng Tao
https://openreview.net/forum?id=Q5Eb6qIKux
Keywords: computer vision, foundation models.
Compressor summary: VanillaNet is a simple neural network architecture that achieves high performance in computer vision tasks while being efficient and easy to deploy.
Peter Macgregor
https://openreview.net/forum?id=Q3FXnCPZ1X
Keywords: spectral clustering, power method, spectral graph theory, graph algorithms
Compressor summary: The paper proposes a fast and accurate spectral clustering algorithm using fewer eigenvectors and evaluates its performance on various datasets.
Lai Wei, Muhammad Qasim Elahi, Mahsa Ghasemi, Murat Kocaoglu
https://openreview.net/forum?id=Q3CRHnttxW
Keywords: multi-armed bandits, causal Inference, sequential decision-making
Compressor summary: Structural causal bandit is an online decision-making framework using causal models to learn from unknown environments, balancing exploration and exploitation for maximizing rewards and minimizing regret.
YiFan Zhang, Qingsong Wen, Xue Wang, Weiqi Chen, Liang Sun, Zhang Zhang, Liang Wang, Rong Jin, Tieniu Tan
https://openreview.net/forum?id=Q25wMXsaeZ
Keywords: Time series forecasting, concept drift, online learning, online convex programming
Compressor summary: OneNet is an online time series forecasting algorithm that combines two models, one for temporal dependency and one for cross-variable dependency, using reinforcement learning to adjust their weights dynamically and improve performance against concept drift.
Guy Kornowski, Steve Hanneke, Aryeh Kontorovich
https://openreview.net/forum?id=Q0ntwxVtcy
Keywords: Hölder smoothness, average smoothness, bracketing numbers, generalization, risk bounds, metric space
Compressor summary: The paper generalizes a smoothness measure for functions and shows that using an average smoothness instead of a worst-case one leads to better learning rates in different settings.
Chen Sun, Calvin Luo, Xingyi Zhou, Anurag Arnab, Cordelia Schmid
https://openreview.net/forum?id=PzYAMXmIT3
Keywords: visual reasoning, self-supervised learning
Compressor summary: The authors propose a self-supervised method to compress video frames into tokens using a transformer network, and show that visual pretraining is crucial for end-to-end visual reasoning.
Aoxiang Zhang, Yu Ran, Weixuan Tang, Yuan-Gen Wang
https://openreview.net/forum?id=Pz8xvVCLNJ
Keywords: video quality assessment, adversarial attack, black-box, just noticeable difference
Compressor summary: The paper introduces a method to evaluate the robustness of NR-VQA models against adversarial attacks and proposes a patch-based random search technique for black-box attacks.
Hongjie Chen, Vincent Cohen-Addad, Tommaso d'Orsi, Alessandro Epasto, Jacob Imola, David Steurer, Stefan Tiegel
https://openreview.net/forum?id=Pya0kCEpDk
Keywords: differential privacy, stochastic block model, Gaussian mixture model, sum of squares
Compressor summary: The authors propose efficient private estimation algorithms that match non-private counterparts in high-dimensional settings and apply them to stochastic block models and mixtures of spherical Gaussians problems.
Jane Lee, Andre Wibisono, Manolis Zampetakis
https://openreview.net/forum?id=PxcWJqO3qj
Keywords: truncated statistics, robustness, exponential families, extrapolation
Compressor summary: The paper presents an estimation algorithm that can extrapolate from truncated data for log-concave exponential families, using Projected Stochastic Gradient Descent, which is simpler and more efficient than previous methods.
Minyang Hu, Hong Chang, Zong Guo, Bingpeng Ma, Shiguang Shan, Xilin CHEN
https://openreview.net/forum?id=Pvgxecj5aS
Keywords: Few-shot Learning, Meta-Learning, Task Relatedness, Task Adaptation Difficulty
Compressor summary: The paper proposes a metric called Task Attribute Distance (TAD) to measure the similarity between training and novel tasks in few-shot learning, and shows how it relates to adaptation difficulty.
Francesco Giannini, Stefano Fioravanti, Oguzhan Keskin, Alisia Maria Lupidi, Lucie Charlotte Magister, Pietro Lio, Pietro Barbiero
https://openreview.net/forum?id=Psnph85KYc
Keywords: universal algebra, interpretability, graph neural networks, concept-based models
Compressor summary: This paper uses AI to explore Universal Algebra's conjectures, creating datasets and a new neural layer for interpretable graph networks that can validate and generate conjectures.
Chunlin Yu, Ye Shi, Jingya Wang
https://openreview.net/forum?id=Psj0jHocm1
Keywords: Deep Clustering, Self-supervised learning, re-ranking
Compressor summary: The paragraph discusses a new method for deep clustering that uses online re-ranking to find more informative neighbors, cross-view consistency, and boundary filtering to handle noisy data and improve performance on benchmark datasets.
Lorenzo Noci, Chuning Li, Mufan Bill Li, Bobby He, Thomas Hofmann, Chris J. Maddison, Daniel M. Roy
https://openreview.net/forum?id=PqfPjS9JRX
Keywords: Deep Learning Theory, Covariance SDE, Attention Mechanism, Infinite-Depth-and-Width, Scaling Limit
Compressor summary: The paper studies how modifying a Softmax-based attention model with skip connections and width-dependent temperature can improve trainability and stability in deep learning models, leading to a new architecture called shaped Transformer.
Zhengxin Zhang, Yucheng Huang, Guanglin Duan, Qing Li, Dan Zhao, Yong Jiang, Lianbo Ma, Xi Xiao, Hengyang Xu
https://openreview.net/forum?id=Pplq1TRnma
Keywords: Network Security; Regular Expression; Knowledge Distillation; Machine Learning; Programmable Switch
Compressor summary: Metis converts regular expressions to neural networks for better accuracy and device deployment in networking tasks.
Ilias Diakonikolas, Daniel Kane, Jasper C.H. Lee, Ankit Pensia, Thanasis Pittas
https://openreview.net/forum?id=PpI7XvOXkF
Keywords: robust statistics, covariance estimation, list-decodable learning
Compressor summary: The paper presents a spectral algorithm for list-decodable Gaussian covariance estimation that works with unknown fractions of Gaussian samples and can also learn robust partial clusters of GMMs efficiently.
Utkarsh Ojha, Yuheng Li, Anirudh Sundara Rajan, Yingyu Liang, Yong Jae Lee
https://openreview.net/forum?id=Poj71ASubN
Keywords: knowledge distillation
Compressor summary: The paragraph discusses the need to understand what kind of knowledge is transferred in knowledge distillation techniques and how it affects a student network's properties beyond task performance.
Ruiqi Zhong, Peter Zhang, Steve Li, Jinwoo Ahn, Dan Klein, Jacob Steinhardt
https://openreview.net/forum?id=PnbCA4ylIc
Keywords: large language model, prompting, exploratory text analysis
Compressor summary: The authors introduce a new task (D5) that uses language models to automatically discover differences between two large corpora based on user-specified goals, and evaluate its performance on synthetic and real datasets.
Qizhi Pei, Kaiyuan Gao, Lijun Wu, Jinhua Zhu, Yingce Xia, Shufang Xie, Tao Qin, Kun He, Tie-Yan Liu, Rui Yan
https://openreview.net/forum?id=PnWakgg1RL
Keywords: protein-ligand docking
Compressor summary: The authors propose FABind, a fast and accurate model for predicting protein-ligand binding structures using a pocket prediction and docking approach that leverages ligand information.
Minhak Song, Chulhee Yun
https://openreview.net/forum?id=PnJaA0A8Lr
Keywords: non-convex optimization, trajectory alignment of GD, edge of stability, progressive sharpening, bifurcation theory
Compressor summary: The paper investigates how the sharpness of neural networks changes during training, and shows that different gradient descent trajectories align on a bifurcation diagram when the Edge of Stability occurs.
Anish Agarwal, Keegan Harris, Justin Whitehouse, Steven Wu
https://openreview.net/forum?id=PmqBJ02V1p
Keywords: adaptive data collection, principal component regression, error-in-variables regression, panel data, synthetic controls, synthetic interventions, causal inference
Compressor summary: The paper introduces time-uniform finite sample guarantees for online principal component regression in error-in-variables settings and uses it to estimate counterfactual treatment effects in panel data with adaptive interventions.
Huafeng Liu, Liping Jing, Jian Yu
https://openreview.net/forum?id=PmlNxZoXr4
Keywords: Neural processes, stability
Compressor summary: The paper introduces a method to improve neural processes by adding algorithmic stability, which leads to better accuracy and robustness.
Suhas Shrinivasan, Konstantin-Klemens Lurz, Kelli Restivo, George Denfield, Andreas S. Tolias, Edgar Y. Walker, Fabian H. Sinz
https://openreview.net/forum?id=Pl416tPkNv
Keywords: Neural Sampling Code, Probabilistic Inference, Bayesian Brain, Macaque V1, Natural Images, Population Recordings, Normalizing Flows, Probabilistic Models, Computational Neuroscience, Theoretical Neuroscience
Compressor summary: The authors propose a novel formalization of the Neural Sampling Code theory that allows fitting more flexible generative models to recorded neuronal activity, enabling quantitative evaluation and comparison of different models on natural images.
Riccardo Poiani, Nicole Nobili, Alberto Maria Metelli, Marcello Restelli
https://openreview.net/forum?id=PkKpTK7hJ6
Keywords: Reinforcement Learning, Policy Evaluation, Budget Optimization, Monte Carlo
Compressor summary: The paper proposes RIDO, an adaptive algorithm that splits the interaction budget into mini-batches and adjusts the length of trajectories to minimize the variance of policy return estimators in Monte Carlo Reinforcement Learning.
Michalis Titsias
https://openreview.net/forum?id=Pk9CdOZYRA
Keywords: MCMC, Langevin diffusion, preconditioning, Fisher information, adaptive MCMC, score function
Compressor summary: The paper proposes an optimal preconditioning for the Langevin diffusion that improves the performance of the Metropolis adjusted Langevin algorithm (MALA) in high dimensions, and compares it with other methods.
Jiahua Dong, Yu-Xiong Wang
https://openreview.net/forum?id=Pk49a9snPe
Keywords: neural radiance field, diffusion model, editing
Compressor summary: ViCA-NeRF is a method for 3D editing with text instructions that uses view-consistency and two types of regularization to ensure multi-view consistency and produce detailed results.
Shaochen Zhong, Zaichuan You, Jiamu Zhang, Sebastian Zhao, Zachary LeClaire, Zirui Liu, Daochen Zha, Vipin Chaudhary, Shuai Xu, Xia Hu
https://openreview.net/forum?id=Pjky9XG8zP
Keywords: pruning, structured pruning, adversarial robustness, grouped kernel pruning, CNN, one-shot
Compressor summary: The authors investigate the vulnerability of modern structured pruning methods to adversarial attacks and propose a new method that improves robustness while maintaining compression and acceleration benefits.
Bochen Lyu, Zhanxing Zhu
https://openreview.net/forum?id=PjBEUTVzoe
Keywords: implicit bias, gradient descent, stochastic gradient descent, linear networks
Compressor summary: The paper investigates the implicit bias of rank-1 linear networks and finds new insights on how gradient descent and stochastic gradient descent behave in over-parameterized regression problems, which could help understand deep learning better.
Zihui Xue, Kristen Grauman
https://openreview.net/forum?id=Pj6X6GqNy8
Keywords: fine-grained video understanding, egocentric video, self-supervised learning, temporal alignment
Compressor summary: The paragraph describes a new method called AE2 that learns action features invariant to viewpoints by aligning egocentric and exocentric videos, even when not captured at the same time or place, and shows its effectiveness on four datasets.
Adam Block, Ali Jadbabaie, Daniel Pfrommer, Max Simchowitz, Russ Tedrake
https://openreview.net/forum?id=PhFVF0gwid
Keywords: Imitation Learning, Control, Diffusion Models, Optimal Transport
Compressor summary: The paper presents a theory for imitating expert actions using generative models and low-level controllers to ensure stability and match the distribution of the expert trajectories.
Yilin Lyu, Liyuan Wang, Xingxing Zhang, Zicheng Sun, Hang Su, Jun Zhu, Liping Jing
https://openreview.net/forum?id=Ph65E1bE6A
Keywords: Continual Learning, Batch Normalization, Recency Bias, Catastrophic Forgetting
Compressor summary: The paper analyzes the limitations of Batch Normalization in continual learning and proposes Adaptive Balance of BN, which adapts to task-wise contributions and balances BN statistics, achieving significant performance gains on benchmarks.
Dhawal Gupta, Yash Chandak, Scott M. Jordan, Philip S. Thomas, Bruno Castro da Silva
https://openreview.net/forum?id=PfpAQuyZCB
Keywords: Reinforcement Learning, Behavior Alignment, Implicit Gradient, Bi-level Optimization
Compressor summary: The paper introduces a new framework that uses a bi-level objective to learn behavior alignment reward functions that integrate designer's heuristics with environment's primary rewards for guiding reinforcement learning agents efficiently and robustly.
Dan Friedman, Alexander Wettig, Danqi Chen
https://openreview.net/forum?id=Pe9WxkN8Ff
Keywords: mechanistic interpretability, transformers
Compressor summary: The authors propose a method for training mechanistically interpretable Transformers by converting them into human-readable programs and show their effectiveness on various tasks, including NLP problems.
Mohammad Jalali, Cheuk Ting Li, Farzan Farnia
https://openreview.net/forum?id=PdZhf6PiAb
Keywords: Generative Models; Evaluation in Learning; Information Measures
Compressor summary: The paper proposes an information-theoretic method to measure the number of modes in multi-modal image datasets using R\'enyi Kernel Entropy and evaluates state-of-the-art generative models with it.
Yash Gupta, Runtian Zhai, Arun Suggala, Pradeep Kumar Ravikumar
https://openreview.net/forum?id=PcNpL9Q39p
Keywords: Responsible AI, fairness, DRO, robustness
Compressor summary: The authors propose a general framework called Responsible AI (RAI) games to study societal effects of AI, and present two classes of algorithms for solving them.
Prateek Jaiswal, Harsha Honnappa, Vinayak Rao
https://openreview.net/forum?id=PcKHQFsvel
Keywords: Variational Bayes, Loss Calibration, Bayesian Statistics, Variational Inference, Statistical Theory
Compressor summary: The authors propose a new variational Bayesian method for risk-sensitive decision-making when the posterior distribution is difficult to compute, and study its theoretical and practical properties.
Naman Deep Singh, Francesco Croce, Matthias Hein
https://openreview.net/forum?id=Pbpk9jUzAi
Keywords: adversarial robustness, deep learning, vision transformers, convnext
Compressor summary: The paper compares adversarial training on ImageNet for ViTs and ConvNeXts using different architectural modifications and shows that these changes affect robustness in different ways depending on the threat model.
Quentin Delfosse, Hikaru Shindo, Devendra Singh Dhami, Kristian Kersting
https://openreview.net/forum?id=PbMBfRpVgU
Keywords: Reinforcement Learning, First-Order-Logic, Symbolic Abstraction, Interpretable Reinforcement Learning, Logic Reinforcement Learning
Compressor summary: NUDGE is a neuro-symbolic RL method that combines neural networks and differentiable logic to create interpretable and explainable policies, performing better than pure neural methods in various environments.
Yashaswini Murthy, Mehrdad Moharrami, R. Srikant
https://openreview.net/forum?id=PaSpImjKm2
Keywords: Average Reward MDPs, Reinforcement Learning Theory, Approximate Policy Iteration, Policy Based Methods, Performance Bounds
Compressor summary: The authors propose a solution to the open problem of obtaining performance bounds for approximate policy iteration and reinforcement learning algorithms in the average-reward setting, by deriving non-trivial finite time error bounds that converge to zero as policy evaluation and improvement errors decrease.
Mingyu Xu, Zheng Lian, Lei Feng, Bin Liu, Jianhua Tao
https://openreview.net/forum?id=PYSfn5xXEe
Keywords: Partial label learning; Noisy label learning
Compressor summary: The paragraph introduces a new method called "Adjusting Label Importance Mechanism" (ALIM) for handling noisy partial label learning, which reduces the impact of detection errors by adjusting the candidate set and model outputs.
Jack Henry Good, Torin Kovach, Kyle Miller, Artur Dubrawski
https://openreview.net/forum?id=PYEgC56flW
Keywords: explainability, interpretability, decision tree, feature learning
Compressor summary: The text discusses a new system that combines sparse feature learning and differentiable decision tree construction to create small, interpretable trees with good performance.
Han Shao, Lee Cohen, Avrim Blum, Yishay Mansour, Aadirupa Saha, Matthew Walter
https://openreview.net/forum?id=PYASzxr2OP
Keywords: preference learning, algorithms, linear model, Markov decision processes, learning theory, multi-objective decision making, preference elicitation
Compressor summary: The paper presents a method to learn user preferences over multiple objectives and find near-optimal policies based on feedback from comparisons between policies or trajectories.
JunHoo Lee, Jayeon Yoo, Nojun Kwak
https://openreview.net/forum?id=PXsqbAjpQd
Keywords: meta learning, Hessian, Gradient-Based meta learning, Feature Reuse, Implicit Prior
Compressor summary: SHOT is a new algorithm that suppresses the Hessian in gradient-based meta-learning, improving performance on few-shot learning tasks.
Yatin Dandi, Ludovic Stephan, Florent Krzakala, Bruno Loureiro, Lenka Zdeborova
https://openreview.net/forum?id=PU3deePP2S
Keywords: theoretical analysis, high-dimensional statistics, Universality, weak convergence, mixture models, sampling, statistical physics
Compressor summary: The paragraph discusses results in high-dimensional statistics under the Gaussian mixture hypothesis and provides rigorous proofs for applying these results to a general class of datasets using generalized linear models and investigating their asymptotic joint statistics.
Depen Morwani, jatin batra, Prateek Jain, Praneeth Netrapalli
https://openreview.net/forum?id=PTvxck0QDE
Keywords: Simplicity Bias, Gradient Descent, Implicit Bias, Neural Networks
Compressor summary: The paper investigates simplicity bias in neural networks with one hidden layer, showing that they learn only the simplest features even when more complex ones exist, and propose an ensemble method to improve robustness.
Yuki Takezawa, Ryoma Sato, Han Bao, Kenta Niwa, Makoto Yamada
https://openreview.net/forum?id=PSngfm5B9q
Keywords: decentralized learning, distributed optimization, network topology, consensus rate
Compressor summary: The study proposes a new graph topology for decentralized learning that combines fast consensus rate and small maximum degree, improving accuracy and communication efficiency over existing topologies.
Jian Yao, Weiming Liu, Haobo Fu, Yaodong Yang, Stephen Marcus McAleer, QIANG FU, Yang Wei
https://openreview.net/forum?id=PRgvdEbhdH
Keywords: Policy Diversity, Policy-Space Response Oracles, Nash Equilibrium, Multi-agent Reinforcement Learning
Compressor summary: The paper proposes a new diversity metric for Policy Space Response Oracles (PSRO) that improves the approximation of Nash Equilibrium and presents PSD-PSRO, a variant that converges and performs better in experiments.
Jishnu Ray Chowdhury, Cornelia Caragea
https://openreview.net/forum?id=PR5znB6BZ2
Keywords: Recursive Models, Recursive Neural Networks, RvNNs, Length Generalization, Structured Encoding, Representation Learning
Compressor summary: The paper proposes strategies to reduce memory usage of BT-RvNN, a neural network for encoding sentences, and to use it as a token contextualizer.
Waverly Wei, Xinwei Ma, Jingshen Wang
https://openreview.net/forum?id=PMvudWa53L
Keywords: Adaptive Randomized Experiment; Adaptive Design; Causal Inference
Compressor summary: The paragraph describes a fair adaptive experiment strategy that balances data efficiency, equity, and participant welfare in randomized experiments without assuming parametric models on outcomes.
Jiang-Xin Shi, Tong Wei, Yuke Xiang, Yu-Feng Li
https://openreview.net/forum?id=PLzCXefcpE
Keywords: long-tail learning, class-imbalanced learning, re-sampling
Compressor summary: This paper investigates how re-sampling affects long-tail learning, finding that it can improve generalization if the training data has relevant contexts but may introduce spurious correlations otherwise; they propose a new context shift augmentation module to address this issue.
Michael Beukman, Devon Jarvis, Richard Klein, Steven James, Benjamin Rosman
https://openreview.net/forum?id=PJhjkSFlbG
Keywords: Deep Reinforment Learning, Contextual Markov Decision Process, Neural Network Architecture
Compressor summary: The paper proposes the Decision Adapter, a neural network architecture that incorporates context information into behaviour learning and improves generalisation performance and robustness.
Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Raghu Meka, Chiyuan Zhang
https://openreview.net/forum?id=PITeSdYQkv
Keywords: differential privacy, user-level privacy, PAC learning
Compressor summary: This paper presents algorithms for user-level differential privacy in the example-scarce regime, improving existing results and providing new bounds for various learning tasks.
Junfeng Guo, Yiming Li, Lixu Wang, Shu-Tao Xia, Heng Huang, Cong Liu, Bo Li
https://openreview.net/forum?id=PIDNxRRJ8w
Keywords: Ownership Verification, Dataset Protection, Copyright Protection, Backdoor Attack, AI Security
Compressor summary: The paper proposes a new approach for dataset ownership verification using domain watermarks that make watermarked DNNs classify hard samples correctly while maintaining stealthiness.
Xinyuan Cao, Santosh Vempala
https://openreview.net/forum?id=PHbqznMa1i
Keywords: Unsupervised Learning, Learning Halfspaces, Non-Gaussian Component analysis
Compressor summary: The paper presents a polynomial-time algorithm for learning high-dimensional halfspaces with margins from an unknown affine transformation and a symmetric one-dimensional logconcave distribution, using only the first two moments of suitable re-weightings and achieving TV distance guarantees.
Zhenbang Wu, Huaxiu Yao, David Liebovitz, Jimeng Sun
https://openreview.net/forum?id=PHKkBbuJWM
Keywords: healthcare, clinical predictive model, domain generalization
Compressor summary: SLGD is a self-learning framework that trains personalized classifiers for decoupled domains, improving deep learning models' domain generalization performance on EHR datasets.
Björn Deiseroth, Mayukh Deb, Samuel Weinbach, Manuel Brack, Patrick Schramowski, Kristian Kersting
https://openreview.net/forum?id=PBpEb86bj7
Keywords: explainability, attention manipulation, perturbation, large language model, multi-modality, generative decoder, efficiency, transformer
Compressor summary: AtMan explains generative transformer models efficiently by manipulating their attention mechanisms using cosine similarity neighborhood in the embedding space instead of backpropagation.
Zhouxing Shi, Nicholas Carlini, Ananth Balashankar, Ludwig Schmidt, Cho-Jui Hsieh, Alex Beutel, Yao Qin
https://openreview.net/forum?id=PAYXfIUKWY
Keywords: Effective robustness, natural distribution shifts, out-of-distribution robustness
Compressor summary: The paper introduces a new metric to evaluate and compare the effective robustness of models trained on different data distributions by controlling for accuracy on multiple in-distribution test sets.
Pha Nguyen, Kha Gia Quach, Kris M. Kitani, Khoa Luu
https://openreview.net/forum?id=PARMyW6xX0
Keywords: Grounded Object Tracking, Multiple Object Tracking, Vision Language
Compressor summary: The paper proposes a new method for tracking objects in videos using natural language captions and introduces a new dataset, GroOT, as well as evaluation protocols and metrics for this task. It also presents a fast and accurate model called MENDER based on tensor decomposition.
Xiangyu Sun, Oliver Schulte
https://openreview.net/forum?id=P9I2VQv1uC
Keywords: Causal Discovery, Cause-Effect Inference, Location-Scale Noise Models
Compressor summary: The paper discusses a problem of inferring cause-effect relationships between random variables and proposes a more robust method for selecting causal models when the noise distribution assumptions are incorrect.
Wenxiao Wang, Soheil Feizi
https://openreview.net/forum?id=P5vzRpoOj2
Keywords: Robustness, Data Poisoning, Security, Machine Learning, Backdoor, Adversarial
Compressor summary: The paper proposes a temporal threat model of data poisoning that considers how long an attack started and lasted, and evaluates a baseline defense called temporal aggregation.
Guy Bar-Shalom, Yonatan Geifman, Ran El-Yaniv
https://openreview.net/forum?id=P3n4wFJGs5
Keywords: Distribution shift detection, Window-based detection
Compressor summary: The authors propose a method to detect distribution deviations in deep neural networks, which is faster and more efficient than existing methods, and can handle large datasets.
Mohammad Mahdi Rahimi, Hasnain Irshad Bhatti, Younghyun Park, Humaira Kousar, Do-Yeon Kim, Jaekyun Moon
https://openreview.net/forum?id=P3Z59Okb5I
Keywords: evolutionary strategies, federated learning, gradient compression, distributed learning
Compressor summary: EvoFed is a novel approach for federated learning that reduces communication costs by using evolutionary strategies and fitness-based information sharing instead of exchanging model parameters.
Zhen Qin, Songlin Yang, Yiran Zhong
https://openreview.net/forum?id=P1TCHxJwLB
Keywords: RNN, Sequence Modeling, NLP
Compressor summary: The paper introduces HGRN, a gated linear RNN model that uses forget gates with a learnable lower bound to enable efficient long-term dependency modeling in different tasks.
Xinyu Tang, Ashwinee Panda, Vikash Sehwag, Prateek Mittal
https://openreview.net/forum?id=P0Avuii9iI
Keywords: Differential privacy, image classification, deep learning
Compressor summary: DP-RandP improves privacy-utility tradeoff of DP-SGD by learning priors from random image generation and transferring them to private data, achieving state-of-the-art accuracy on various datasets.
Xiaosen Wang, Kangheng Tong, Kun He
https://openreview.net/forum?id=OzpTd2EsH1
Keywords: Adversarial examples, Convolutional neural networks, Adversarial transferability, Backward propagation
Compressor summary: The paper proposes Backward Propagation Attack (BPA), a novel method that mitigates information loss in non-linear layers to improve the transferability of adversarial examples.
Yifan Yang, Peiyao Xiao, Kaiyi Ji
https://openreview.net/forum?id=OzjBohmLvE
Keywords: Stochastic bilevel optimization, Hessian-free algorithms, near-optimal complexity
Compressor summary: The paper introduces FdeHBO, a novel method for solving nonconvex-strongly-convex bilevel optimization without second-order derivative computation, achieving an $\mathcal{O}(\epsilon^{-1.5})$ sample complexity and iteration complexity.
Zhanpeng Zeng, Cole Hawkins, Mingyi Hong, Aston Zhang, Nikolaos Pappas, Vikas Singh, Shuai Zheng
https://openreview.net/forum?id=Ozc8XVzwd4
Keywords: efficient, transformer, roberta, T5, language modeling, question answering, summarization
Compressor summary: The paper proposes a method to improve the efficiency and performance of Transformers for ultra long sequences by compressing them based on VIP-tokens that are most relevant to the final prediction.
Jiaxin Lu, Yifan Sun, Qixing Huang
https://openreview.net/forum?id=OwpaO4w6K7
Keywords: Shape Matching; Reassembly; Shape Segmentation;
Compressor summary: Jigsaw is a novel framework that uses hierarchical features to assemble physically broken 3D objects from multiple pieces and outperforms existing methods.
Meghdad Kurmanji, Peter Triantafillou, Jamie Hayes, Eleni Triantafillou
https://openreview.net/forum?id=OveBaTtUAT
Keywords: machine unlearning, deep learning
Compressor summary: The paper introduces SCRUB, a novel unlearning algorithm for neural networks that performs well across various applications and metrics, including privacy protection.
Ye Yuan, Can Chen, Zixuan Liu, Willie Neiswanger, Xue Liu
https://openreview.net/forum?id=OvPnc5kVsb
Keywords: offline model-based optimization, co-teaching, meta-learning, sample reweighting
Compressor summary: ICT is a method that uses three symmetric proxies and pseudo-labeling to improve offline model-based optimization, addressing the out-of-distribution issue in gradient ascent.
Shicheng Liu, Minghui Zhu
https://openreview.net/forum?id=Ou1VRZ4j4y
Keywords: inverse reinforcement learning; distributed online bi-level optimization
Compressor summary: The paper presents MA-BIRDS, an algorithm that infers expert behaviors from sequential demonstrations using a distributed bi-level optimization framework and achieves consensus, low regret, and sub-linear constraint violation.
Qizhou Wang, Zhen Fang, Yonggang Zhang, Feng Liu, Yixuan Li, Bo Han
https://openreview.net/forum?id=OtU6VvXJue
Keywords: OOD Detection
Compressor summary: The text discusses a problem in open-world classification systems and proposes a solution called Distributional-Augmented OOD Learning (DAOL) that improves performance by reducing distribution discrepancy between auxiliary and unseen out-of-distribution data.
Kunhao Liu, Fangneng Zhan, Jiahui Zhang, MUYU XU, Yingchen Yu, Abdulmotaleb El Saddik, Christian Theobalt, Eric Xing, Shijian Lu
https://openreview.net/forum?id=Orp1K2dZvY
Keywords: 3D, open-vocabulary segmentation, neural radiance field
Compressor summary: The paper proposes a method to learn 3D open-vocabulary segmentation from 2D images and text using pre-trained models CLIP and DINO without any manual annotations.
Ilze Amanda Auzina, Cagatay Yildiz, Sara Magliacane, Matthias Bethge, Efstratios Gavves
https://openreview.net/forum?id=Op9z2QfXbC
Keywords: Neural ODEs, Modulator Variables, Dynamical Systems, Disentanglment
Compressor summary: The paper introduces MoNODEs, a novel framework for neural ordinary differential equations that improves generalization and far-horizon forecasting by learning time-invariant modulator variables from data.
JungWoo Chae, Hyunin Cho, Sooyeon Go, Kyungmook Choi, Youngjung Uh
https://openreview.net/forum?id=OoPLRGBKjM
Keywords: Generative model
Compressor summary: The paper presents a novel method for semantic image synthesis that uses a proxy mask derived from intermediate feature maps to guide a pretrained unconditional generator without heavy annotation.
Xinyu Zhou, Pinxue Guo, Lingyi Hong, Jinglun Li, Wei Zhang, Weifeng Ge, Wenqiang Zhang
https://openreview.net/forum?id=On0IDMYKw2
Keywords: object tracking;global representation memory;transformer
Compressor summary: The text introduces a new tracking paradigm that uses a relevance attention mechanism and a global representation memory to select and read the most relevant historical information for visual object tracking, improving performance and reducing redundancy.
Po-Yao Huang, Vasu Sharma, Hu Xu, Chaitanya Ryali, Haoqi Fan, Yanghao Li, Shang-Wen Li, Gargi Ghosh, Jitendra Malik, Christoph Feichtenhofer
https://openreview.net/forum?id=OmTMaTbjac
Keywords: self-supervised learning, audio representation learning, audio classification
Compressor summary: MAViL learns audio-visual representations using three forms of self-supervision, achieves state-of-the-art performance in audio-video classification and improves the quality of both modalities individually.
Haibo Yang, Zhuqing Liu, Jia Liu, Chaosheng Dong, Michinari Momma
https://openreview.net/forum?id=OlSTwlz96r
Keywords: Multi-Objective Learning, Federated Learning
Compressor summary: The text introduces a new federated multi-objective learning framework that allows multiple clients to collaboratively solve a multi-objective optimization problem while keeping their data private and proposes two new algorithms for this framework.
Lujun Li, Peijie Dong, Anggeng Li, Zimian Wei, Yang Ya
https://openreview.net/forum?id=OlMKa5YZ8e
Keywords: Knowledge distillation
Compressor summary: The paper introduces KD-Zero, a novel framework that uses evolutionary search to automatically discover optimal knowledge distillation designs for any teacher-student model pairs, achieving superior performance across various tasks and architectures.
Cyrus Cousins, Elita Lobo, Marek Petrik, Yair Zick
https://openreview.net/forum?id=OjlZqQzw51
Keywords: Reinforcement Learning, Bayesian Uncertainty, Robustness
Compressor summary: The paper proposes a new dynamic programming algorithm for reinforcement learning that optimizes the percentile criterion without explicitly creating uncertainty sets, leading to smaller sets and more efficient policies.
Runqi Lin, Chaojian Yu, Tongliang Liu
https://openreview.net/forum?id=Oj7Mrb4009
Keywords: adversarial training, catastrophic overfitting
Compressor summary: The paper proposes AAER, a method to prevent catastrophic overfitting in SSAT by regularizing the variation of abnormal adversarial examples, which are associated with classifier distortion.
Kaiyue Wen, Yuchen Li, Bingbin Liu, Andrej Risteski
https://openreview.net/forum?id=OitmaxSAUu
Keywords: Transformer, Self Attention, Dyck Language, Context Free Grammar, Formal Language, Theory, Interpretability
Compressor summary: This paper critically examines transformer interpretability methods that focus only on parts of the model instead of considering it holistically, using theoretical results and synthetic data experiments.
Liang Hou, Qi Cao, Yige Yuan, Songtao Zhao, Chongyang Ma, Siyuan Pan, Pengfei Wan, Zhongyuan Wang, Huawei Shen, Xueqi Cheng
https://openreview.net/forum?id=OiivS2mqQf
Keywords: generative adversarial networks, limited data, self-supervised learning
Compressor summary: The paper proposes a novel self-supervised discriminator that predicts the augmentation parameter and encourages the generator to produce realistic and augmentation-predictable data, improving data efficiency for training GANs with limited data.
Aaron Sidford, Chenyi Zhang
https://openreview.net/forum?id=OiatK9W6tR
Keywords: continuous optimization, quantum algorithms, stochastic optimization, gradient oracle
Compressor summary: The text describes new quantum methods for minimizing a Lipschitz convex function, achieving unachievable trade-offs and optimal rates in low dimensions, as well as quantum algorithms for solving smooth non-convex functions using quantum multivariate mean estimation and variance reduction techniques.
Shuhan Tan, Tushar Nagarajan, Kristen Grauman
https://openreview.net/forum?id=OfjVAKx44G
Keywords: Egocentric Video; IMU; Efficient Video Understanding
Compressor summary: EgoDistill is a distillation-based approach that improves efficiency of ego-centric video understanding by combining semantics from sparse video frames and head motion from IMU readings, achieving state-of-the-art results on Ego4D and EPIC-Kitchens datasets.
Aaditya K Singh, Stephanie C.Y. Chan, Ted Moskovitz, Erin Grant, Andrew M Saxe, Felix Hill
https://openreview.net/forum?id=Of0GBzow8P
Keywords: in-context learning, transformers, emergence, transience
Compressor summary: The paragraph discusses how in-context learning in transformer neural networks is often transient and not persistent, raising questions about overtraining and suggesting L2 regularization as a way to achieve more lasting ICL.
Antti Koskela, Tejas Kulkarni
https://openreview.net/forum?id=OeLInnFKUK
Keywords: differential privacy, hyperparameter tuning, Rényi differential privacy, computational efficiency, DP-SGD
Compressor summary: The paper proposes a method to lower the privacy and compute cost of differentially private hyperparameter tuning algorithms by using a random subset of sensitive data and extrapolating optimal values.
Stephen Casper, Tong Bu, Yuxiao Li, Jiawei Li, Kevin Zhang, Kaivalya Hariharan, Dylan Hadfield-Menell
https://openreview.net/forum?id=Od6CHhPM7I
Keywords: interpretability, benchmarking, auditing, diagnostics, debugging, adversarial attacks, feature synthesis
Compressor summary: The paper introduces a benchmark with human-interpretable trojans to test interpretability tools for model debugging, showing that current methods often fail to discover them.
Thomas Hartvigsen, Swami Sankaranarayanan, Hamid Palangi, Yoon Kim, Marzyeh Ghassemi
https://openreview.net/forum?id=Oc1SIKxwdV
Keywords: Model Editing, Continual Learning, Model Repair
Compressor summary: GRACE is a method that allows targeted edits on deployed language models without retraining or degrading performance, using streaming errors to update a latent codebook.
Vasilis Kontonis, Mingchen Ma, Christos Tzamos
https://openreview.net/forum?id=OaUT4hX40s
Keywords: Online Learning, Self-directed Learning, Hardness of Approximation
Compressor summary: The paper explores how to order datapoints in online linear regression to minimize regret and shows that efficient algorithms are hard for arbitrary datasets but can achieve a log(d)-approximation for structured ones.
Yuseung Lee, Kunho Kim, Hyunjin Kim, Minhyuk Sung
https://openreview.net/forum?id=OZEfMD7axv
Keywords: Diffusion model, Text-to-image generation, Panorama generation
Compressor summary: SyncDiffusion is a module that helps generate coherent image montages by synchronizing multiple diffusions using perceptual similarity loss, and it can be applied to various image generation tasks.
Roland S. Zimmermann, Thomas Klein, Wieland Brendel
https://openreview.net/forum?id=OZ7aImD4uQ
Keywords: feature visualization, interpretability, explainability, deep learning, neural networks, analysis, activation maximization, psychophysics
Compressor summary: The study finds that larger neural networks do not improve interpretability and suggests the need for more interpretable models and methods.
Stefano Massaroli, Michael Poli, Daniel Y Fu, Hermann Kumbong, Rom Nishijima Parnichkun, David W. Romero, Aman Timalsina, Quinn McIntyre, Beidi Chen, Atri Rudra, Ce Zhang, Christopher Re, Stefano Ermon, Yoshua Bengio
https://openreview.net/forum?id=OWELckerm6
Keywords: Long convolutions, recurrence, attention, language models, signal processing, throughput, auto-regressive generation
Compressor summary: The paper proposes methods to improve the efficiency of attention-free sequence models using linear state-space models and architectural improvements like Hyena, achieving better performance than Transformers and Hyena.
Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, Luke Zettlemoyer
https://openreview.net/forum?id=OUIFPHEgJU
Keywords: finetuning, llama, instructions, quantization
Compressor summary: QLora is a memory-efficient finetuning method that achieves near-ChatGPT performance with the Guanaco model family, while introducing new innovations to save memory without sacrificing performance.
Anh Viet Do, Aneta Neumann, Frank Neumann, Andrew M. Sutton
https://openreview.net/forum?id=ORmVvN94B9
Keywords: minimum weight base problem, multi-objective optimization, approximation, evolutionary algorithm
Compressor summary: The paper studies the multi-objective minimum weight base problem, proves properties of its convex hull, and analyzes the performance of the MOEA/D algorithm on this problem compared to another algorithm called GSEMO.
Nicholas Carlini, Milad Nasr, Christopher A. Choquette-Choo, Matthew Jagielski, Irena Gao, Pang Wei Koh, Daphne Ippolito, Florian Tramèr, Ludwig Schmidt
https://openreview.net/forum?id=OQQoD8Vc3B
Keywords: Adversarial examples, large language models, alignment
Compressor summary: The paper investigates how well large language models can resist harmful inputs from adversarial users, and finds that multimodal models are more vulnerable than text-only models.
Wenhan Yang, Jingdong Gao, Baharan Mirzasoleiman
https://openreview.net/forum?id=ONwL9ucoYG
Keywords: Contrastive Learning, Adversarial Learning, Model Robustness
Compressor summary: RoCLIP is a method for pre-training multimodal vision-language models that defends against targeted data poisoning and backdoor attacks by using random captions and image/text augmentations.
Zihang Shao, Xuanye Fang, Yaxin Li, Chaoran Feng, Jiangrong Shen, Qi Xu
https://openreview.net/forum?id=OMDgOjdqoZ
Keywords: spiking neural networks cycle learning spike encoding
Compressor summary: EICIL is a novel learning method for spiking neural networks that mimics biological neurons by integrating excitatory and inhibitory behaviors, improving bio-mimicry and adaptability, and expanding the representation space.
Marc Jourdan, Rémy Degenne, Emilie Kaufmann
https://openreview.net/forum?id=OLk3F64eSg
Keywords: multi-armed bandits, pure-exploration, epsilon best arm identification, Top Two algorithm, anytime
Compressor summary: EB-TCε is a new sampling method for finding the best arm in stochastic bandits with theoretical guarantees on its performance and simulation results showing its effectiveness.
Yujie Lu, Xianjun Yang, Xiujun Li, Xin Eric Wang, William Yang Wang
https://openreview.net/forum?id=OJ0c6um1An
Keywords: Text-to-Image Evaluation, Visio-linguistic Compositionality, Large Language Models
Compressor summary: LLMScore is a new framework that uses large language models to evaluate text-to-image synthesis by considering object-level compositionality, achieving high correlation with human judgments and outperforming existing metrics.
Chuning Zhu, Max Simchowitz, Siri Gadipudi, Abhishek Gupta
https://openreview.net/forum?id=OIJ3VXDy6s
Keywords: Model-Based Reinforcement Learning, Deep Reinforcement Learning
Compressor summary: The paper proposes a visual model-based RL method that learns a latent representation resilient to spurious variations and shows how it can adapt to different environments without relearning the dynamics and policy.
Ge Yuan, Xiaodong Cun, Yong Zhang, Maomao Li, Chenyang Qi, Xintao Wang, Ying Shan, Huicheng Zheng
https://openreview.net/forum?id=OGQWZ3p0Zn
Keywords: Text-to-Image Synthesis, Personalized Synthesis, Face Embedding
Compressor summary: The paper proposes a new personalization method for text-to-image models that uses one facial photo and 1024 learnable parameters to integrate a unique individual into the model, enabling generation of stunning images featuring the person in various scenarios.
Zifu Wang, Xuefei Ning, Matthew B. Blaschko
https://openreview.net/forum?id=OFMPrCAMKi
Keywords: Semantic Segmentation
Compressor summary: JMLs are compatible with soft labels and improve semantic segmentation performance in various scenarios, such as label smoothing, knowledge distillation, and semi-supervised learning.
Justin David Naggar Weltz, Tanner Fiez, Alexander Volfovsky, Eric Laber, Blake Mason, houssam nassif, Lalit K Jain
https://openreview.net/forum?id=OFDApY678F
Keywords: Heteroskedastic Variance, Linear Bandits, Experimental design
Compressor summary: The paragraph discusses a novel design for adaptive experimental design problems under heteroskedastic noise, which improves sample complexity and estimation accuracy by bounding the error of variance parameters.
Yury Demidovich, Grigory Malinovsky, Igor Sokolov, Peter Richtárik
https://openreview.net/forum?id=OCtv4NyahI
Keywords: Stochastic optimization, biased SGD, Non-convex analysis
Compressor summary: The paper studies SGD with biased gradient estimators, clarifies existing assumptions and relationships among them, introduces a new set of weaker assumptions, and shows advantages of biased estimators over unbiased ones in various settings using theory and experiments.
Donghao Ying, YUNKAI ZHANG, Yuhao Ding, Alec Koppel, Javad Lavaei
https://openreview.net/forum?id=O63qgtebjH
Keywords: Reinforcement Learning Theory, Safe reinforcement learning, Multi-agent reinforcement learning
Compressor summary: The paper presents a safe multi-agent reinforcement learning method with general utilities and safety constraints, and proves its convergence and sample efficiency. It also demonstrates its effectiveness in simulations.
Wayne WM Soo, Vishwa Goudar, Xiao-Jing Wang
https://openreview.net/forum?id=O453PHSthc
Keywords: neuroscience, recurrent neural network, neural circuits, cortical circuits, cognitive tasks, working memory
Compressor summary: The authors propose a method to train recurrent neural networks (RNNs) for cognitive tasks with long temporal dependencies by adding specialized skip-connections through time and revert to the original architecture, enabling RNNs to learn tasks that are difficult or impossible using conventional methods.
Sangwoo Mo, Minkyu Kim, Kyungmin Lee, Jinwoo Shin
https://openreview.net/forum?id=O1lYncfVOO
Keywords: vision-language model, semi-supervised learning, specialist domain
Compressor summary: S-CLIP is a semi-supervised learning method that uses unpaired images and pseudo-labeling strategies to improve vision-language models like CLIP in specialized domains with limited training data.
Alicia Curth, Alan Jeffares, Mihaela van der Schaar
https://openreview.net/forum?id=O0Lz8XZT2b
Keywords: Double Descent, Statistical Machine Learning, Interpolation Regime, Effective Parameters
Compressor summary: The paragraph discusses how recent work on the relationship between model complexity and prediction error has suggested a new phenomenon called double descent, but argues that this is not incompatible with conventional statistical wisdom if one considers multiple axes of complexity.
Nino Scherrer, Claudia Shi, Amir Feder, David Blei
https://openreview.net/forum?id=O06z2G18me
Keywords: Language Models, Moral Decision Making, Social Aspects of Machine Learning, Ethics
Compressor summary: The paper investigates how LLMs make moral choices and express uncertainty in ambiguous situations using a large survey with moral dilemmas.
Yue Lin, Wenhao Li, Hongyuan Zha, Baoxiang Wang
https://openreview.net/forum?id=NyQwBttTnG
Keywords: multi-agent reinforcement learning, multi-agent communication, information design, signaling gradient, obedience constraints
Compressor summary: The paragraph discusses how reinforcement learning agents can influence each other through information design, addressing challenges like non-stationarity and information obedience in a Markov signaling game framework.
Wonje Choi, Woo Kyung Kim, SeungHyun Kim, Honguk Woo
https://openreview.net/forum?id=Ny3GcHLyzj
Keywords: Prompt Learining, Domain Adaptation, Embodied AI
Compressor summary: ConPE is a novel framework that uses visual prompts and a pretrained vision-language model to enable efficient policy learning and adaptation for embodied agents in various environments.
Iulia Duta, Giulia Cassarà, Fabrizio Silvestri, Pietro Lio
https://openreview.net/forum?id=NvcVXzJvhX
Keywords: hypergraph neural networks, hypergraph, sheaf, higher-order
Compressor summary: The authors propose cellular sheaves for hypergraphs to enhance representation of complex interactions and develop two types of models that improve hypergraph node classification performance.
Ngoc-Bao Nguyen, Keshigeyan Chandrasegaran, Milad Abdollahzadeh, Ngai-man Cheung
https://openreview.net/forum?id=NuoIThPPag
Keywords: Model Inversion attacks, Generative models, Surrogate models, Knowledge transfer
Compressor summary: The authors propose LOKT, a method to perform label-only MI attacks using surrogate models and knowledge transfer from the target model, achieving significant improvements over existing methods.
Soumyabrata Pal, Arun Suggala, Karthikeyan Shanmugam, Prateek Jain
https://openreview.net/forum?id=Ntd6X7uWYF
Keywords: Blocked Bandits, Collaborative Filtering, Clustering
Compressor summary: B-LATTICE is a novel algorithm that efficiently solves the blocked collaborative bandit problem by clustering users with similar rewards, collaborating across clusters, and respecting the arm sampling budget.
Ido Ben-Shaul, Ravid Shwartz-Ziv, Tomer Galanti, Shai Dekel, Yann LeCun
https://openreview.net/forum?id=NsVEjx6YPd
Keywords: Self-Supervised Learning, Deep Learning, Representation Learning
Compressor summary: The paper analyzes self-supervised learning models to reveal that they naturally form semantic label-based clusters, which improve downstream tasks and information compression.
Achraf Azize, Marc Jourdan, Aymen Al Marjani, Debabrota Basu
https://openreview.net/forum?id=NsPbMwyxRl
Keywords: Differential Privacy, Multi-armed Bandits, Best Arm Identification, Fixed Confidence
Compressor summary: The authors study best arm identification problems under differential privacy and propose a new algorithm (AdaP-TT) with a good trade-off between privacy and utility.
Lam M. Nguyen, Trang H. Tran
https://openreview.net/forum?id=Nr1XSeDzpn
Keywords: stochastic gradient, shuffling type gradient method, global convergence
Compressor summary: Shuffling SGD, a scalable and efficient machine learning method, can converge to global solutions for certain non-convex functions even when over-parameterized.
Yeshu Li, Brian D Ziebart
https://openreview.net/forum?id=NpyZkaEEun
Keywords: structure learning, Bayesian network, robustness
Compressor summary: The paper proposes a method to learn the structure of discrete Bayesian networks from potentially corrupted data using distributionally robust optimization and regression, with guarantees for successful learning and efficient algorithms.
Zhenglin Huang, Xiaoan Bao, Na Zhang, Qingqi Zhang, Xiao mei Tu, Biao Wu, Xi Yang
https://openreview.net/forum?id=No52399wXA
Keywords: data augmentation, robustness, safety
Compressor summary: IPMix is a data augmentation technique that improves robustness of deep neural networks without sacrificing clean accuracy by applying image-level, patch-level, and pixel-level augmentations with structural complexity and random mixing.
Tracey Mills, Joshua B. Tenenbaum, Samuel J Cheyette
https://openreview.net/forum?id=NnXznLurw5
Keywords: pattern learning; probabilistic programs; program synthesis; gaussian process; human learning
Compressor summary: The paragraph discusses a study on how humans learn structured patterns from small amounts of data and suggests that people might "learn by programming" using probabilistic models, with the best fit being a structured "Language of Thought" model.
Diederik P Kingma, Ruiqi Gao
https://openreview.net/forum?id=NnMEadcdyD
Keywords: Diffusion Model, Evidence Lower Bound, Maximum Likelihood
Compressor summary: This paper shows that diffusion model objectives are closely related to the ELBO and explores new monotonic weightings for better performance.
Yuchao Gu, Xintao Wang, Jay Zhangjie Wu, Yujun Shi, Yunpeng Chen, Zihan Fan, WUYOU XIAO, Rui Zhao, Shuning Chang, Weijia Wu, Yixiao Ge, Ying Shan, Mike Zheng Shou
https://openreview.net/forum?id=NnIaEaBfXD
Keywords: Text-to-Image Diffusion Models, Concept Customization
Compressor summary: The paper proposes Mix-of-Show, a framework for decentralized multi-concept customization in large-scale text-to-image diffusion models, which uses embedding-decomposed LoRAs and regionally controllable sampling to preserve concept identity and handle attribute binding.
Nathan Rahn, Pierluca D'Oro, Harley Wiltzer, Pierre-Luc Bacon, Marc G Bellemare
https://openreview.net/forum?id=Nn0daSf6CW
Keywords: deep reinforcement learning, continuous control, return landscape, stability
Compressor summary: The authors study how deep reinforcement learning agents for continuous control behave over time and propose a method to improve their stability by navigating the return landscape.
Haonan Wang, Xiaomeng Li
https://openreview.net/forum?id=NibgkUin5n
Keywords: Volumetric Medical Image Segmentation, Semi-supervised Learning, Unsupervised Domain Adaptation, Semi-supervised Domain Generalization
Compressor summary: The paper proposes a new semi-supervised learning framework that can handle different settings, such as unsupervised domain adaptation and semi-supervised domain generalization, by addressing the issues of capturing distribution-invariant features and avoiding over-fitting to labeled data.
Qian Huang, Eric Zelikman, Sarah Li Chen, Yuhuai Wu, Gregory Valiant, Percy Liang
https://openreview.net/forum?id=NiQTy0NW1L
Keywords: Large Language Model, in-context learning, pretraining
Compressor summary: The paper explores lexinvariant language models that do not use fixed token embeddings but rely on context to determine token meanings, and shows their performance, properties, and potential applications.
Lukas Muttenthaler, Lorenz Linhardt, Jonas Dippel, Robert A. Vandermeulen, Katherine Hermann, Andrew Kyle Lampinen, Simon Kornblith
https://openreview.net/forum?id=Nh5dp6Uuvx
Keywords: representational alignment; human similarity judgments; neural networks; representation learning; few-shot learning; anomaly detection
Compressor summary: The authors explore how supervising the global structure of deep neural networks' representations with human similarity judgments can improve their performance on computer vision tasks such as few-shot learning and anomaly detection.
Jan Dubiński, Stanisław Pawlak, Franziska Boenisch, Tomasz Trzcinski, Adam Dziedzic
https://openreview.net/forum?id=NfpYgGZC3B
Keywords: model stealing, model defenses, self-supervised learning
Compressor summary: B4B is a novel defense for MLaaS APIs that protects against model stealing attacks by dynamically changing the quality and uniqueness of vector representations for each user.
Hanlin Chen, Chen Li, Mengqi Guo, Zhiwen Yan, Gim Hee Lee
https://openreview.net/forum?id=NemifGnD2E
Keywords: NeRF; Semantic Segmentation; 3D vision; Scene understanding; Generalizable
Compressor summary: The authors propose a generalizable 3D segmentation framework based on neural implicit representation that uses multi-view image features and semantic maps as inputs, and achieves comparable or better performance than existing methods with only 2D supervision.
Gen Li, Wenhao Zhan, Jason D. Lee, Yuejie Chi, Yuxin Chen
https://openreview.net/forum?id=Nd3FennRJZ
Keywords: reward-agnostic reinforcement learning, policy finetuning, offline reinforcement learning, online reinforcement learning
Compressor summary: The paper proposes a hybrid RL algorithm that combines tabular methods with reward-agnostic exploration and offline RL, achieving better sample complexity than pure offline or online RL methods without needing reward information during data collection.
Milad Sefidgaran, Abdellatif Zaidi, Piotr Krasnowski
https://openreview.net/forum?id=Ncb0MvVqRV
Keywords: Information Bottleneck, Representation Learning, Generalization Error, Minimum Description Length
Compressor summary: The paper proposes an information-theoretic framework to derive upper bounds on generalization error for representation learning algorithms, using the Minimum Description Length of labels or latent variables.
Semih Cayci, Atilla Eryilmaz
https://openreview.net/forum?id=NapL36HSBT
Keywords: temporal difference learning, natural actor-critic, reinforcement learning, policy evaluation, policy gradient, markov decision processes
Compressor summary: This paper proposes a dynamic gradient clipping mechanism for robustifying temporal difference and natural actor-critic methods against heavy-tailed rewards in reinforcement learning, and proves their sample complexities under certain assumptions.
Emile Mathieu, Vincent Dutordoir, Michael John Hutchinson, Valentin De Bortoli, Yee Whye Teh, Richard E Turner
https://openreview.net/forum?id=NaYAsbv2jF
Keywords: diffusion model, functional space, stochastic process, time-series, neural processes, Gaussian processes, random fields, invariance, equivariance, symmetries, stationarity
Compressor summary: The authors propose a geometric diffusion model that incorporates symmetries in infinite-dimensional modelling and show its applicability to weather data.
Wei Tang, Weijia Zhang, Min-Ling Zhang
https://openreview.net/forum?id=NYwbmCrrni
Keywords: Machine Learning, Multi-Instance Partial-Label Learning, Multi-Instance Learning, Partial-Label Learning
Compressor summary: DEMIPL is a novel algorithm that uses disambiguation attention to embed multi-instance bags into single vectors, improving performance in Multi-Instance Partial-Label Learning tasks for colorectal cancer classification.
Qi Qian, Yuanhong Xu, Juhua Hu
https://openreview.net/forum?id=NXLjaYdgaL
Keywords: zero-shot; clip; proxy learning
Compressor summary: The paper proposes InMaP, a method to learn vision proxies from unlabeled data using text proxies and refining pseudo labels for zero-shot transfer tasks.
Mingtian Zhang, Alex Hawkins-Hooker, Brooks Paige, David Barber
https://openreview.net/forum?id=NWrN6cMG2x
Keywords: denoising score-matching, gibbs sampling, diffusion model
Compressor summary: The authors propose a framework for efficient sampling from EBMs trained with DSM, which addresses the inconsistency issues of DSM and allows for scalability to high-dimensional data.
Samuel Goldman, John Bradshaw, Jiayi Xin, Connor W. Coley
https://openreview.net/forum?id=NWEbeI2HNQ
Keywords: molecules, prefix tree, mass spectra, mass spectrum prediction, metabolomics, GNNs, chemistry, biology
Compressor summary: The paper presents a new method for predicting mass spectra from molecules by treating them as sets of subformulae and using a prefix tree structure to decode them, improving on existing limitations in computational tools.
Rafail Fridman, Amit Abecasis, Yoni Kasten, Tali Dekel
https://openreview.net/forum?id=NU2kGsA4TT
Keywords: Computer Vision, Image & Video Editing, Video Generation, Perpetual View Generation, Texture Synthesis & Inpainting
Compressor summary: The paragraph describes a method for generating realistic long-term videos of various scenes based on text prompts and camera poses, using a combination of text-to-image and depth prediction models with online test-time training to ensure 3D consistency.
Julia Olkhovskaya, Jack Mayo, Tim van Erven, Gergely Neu, Chen-Yu Wei
https://openreview.net/forum?id=NTSbj2otOA
Keywords: contextual bandits, bandits, sequential learning, regret bounds
Compressor summary: The paper studies an online learning problem where the loss functions for each arm change over time and gives improved bounds on regret using a modified continuous exponential weights algorithm.
Haixiang Zhang, Ying Chen, Javad Lavaei
https://openreview.net/forum?id=NRnm5xO8Hz
Keywords: Low-rank matrix optimization, non-convex optimization
Compressor summary: This paper introduces a new condition called the $\Omega$-RIP condition for matrix sensing over graphs and proves that it leads to polynomial-time global convergence of saddle-avoiding methods.
Jinqiu Jin, Haoxuan Li, Fuli Feng, Sihao Ding, Peng Wu, Xiangnan He
https://openreview.net/forum?id=NP5xb00Y6a
Keywords: Recommender System, Fairness
Compressor summary: This paper proposes social attribute-aware item group fairness metrics for recommender systems, and develops a gradient-based optimization algorithm to balance direct and social utility in training such models.
Yuedong Yang, Hung-Yueh Chiang, Guihong Li, Diana Marculescu, Radu Marculescu
https://openreview.net/forum?id=NNtsO5L27J
Keywords: Low-rank backpropagation, model adaptation, transfer learning, vision transformer, Edge AI
Compressor summary: LBP-WHT is a new method that projects gradients into a low-rank space, reducing the computation needed for adapting large vision transformers (ViT) models and improving their accuracy on various datasets.
Ildus Sadrtdinov, Dmitrii Pozdeev, Dmitry P. Vetrov, Ekaterina Lobacheva
https://openreview.net/forum?id=NNooZoQpP4
Keywords: ensembles, transfer learning, loss landscape basins, model soups
Compressor summary: StarSSE is a modified Snapshot Ensembles method that improves the performance and diversity of neural network ensembles trained from a single pre-trained checkpoint by better exploring the pre-train basin without losing the benefits of transfer learning.
Felipe Pinto Coelho Nuti, Tim Franzmeyer, Joao F. Henriques
https://openreview.net/forum?id=NN60HKTur2
Keywords: Diffusion models, sequential decision making, inverse reinforcement learning
Compressor summary: The authors propose a method to extract a reward function from two diffusion models with different behavior, and show its effectiveness in navigation and image generation tasks.
Navdeep Kumar, Esther Derman, Matthieu Geist, Kfir Yehuda Levy, Shie Mannor
https://openreview.net/forum?id=NLpXRrjpa6
Keywords: robust Markov decision process, policy gradient
Compressor summary: The paper introduces robust policy gradient (RPG), a method to train reinforcement learning agents that account for transition uncertainty and is efficient in computation.
Jun Xia, Lecheng Zhang, Xiao Zhu, Yue Liu, Zhangyang Gao, Bozhen Hu, Cheng Tan, Jiangbin Zheng, Siyuan Li, Stan Z. Li
https://openreview.net/forum?id=NLFqlDeuzt
Keywords: Graph Neural Networks
Compressor summary: The study compares 12 models on MPP tasks, finding that deep models generally underperform non-deep ones, and proposes a feature mapping method to improve their performance.
Justin Lovelace, Varsha Kishore, Chao Wan, Eliot Seo Shekhtman, Kilian Q Weinberger
https://openreview.net/forum?id=NKdtztladR
Keywords: diffusion, language generation
Compressor summary: The paragraph discusses a new method for language generation using encoder-decoder language models and continuous diffusion models in the latent space, which improves over previous diffusion language models on various data sets.
Dayana Savostianova, Emanuele Zangrando, Gianluca Ceruti, Francesco Tudisco
https://openreview.net/forum?id=NJPSvv0u3R
Keywords: low-rank neural networks, Stiefel manifold, orthogonal neural networks, pruning, adversarial robustness, neural network condition number, neural network singular values
Compressor summary: The authors propose a new method to prune neural networks while maintaining accuracy, robustness, and efficiency in the face of adversarial attacks.
Thibault Christophe RANDRIANARISOA, Botond Szabo
https://openreview.net/forum?id=NJK3aSB0z4
Keywords: Linear inverse problems, Gaussian processes, Variational inference, Inducing variables, Asymptotics, Contraction rates
Compressor summary: The paper analyzes variational Bayesian methods for Gaussian process priors to solve linear inverse problems with different levels of ill-posedness and shows that inducing variable methods can attain the minimax estimation rate.
Ye Zhu, Yu Wu, Zhiwei Deng, Olga Russakovsky, Yan Yan
https://openreview.net/forum?id=NIrTSCiIZ7
Keywords: Diffusion probabilistic models, learning-free applications, high-dimensional semantic boundary, markov mixing
Compressor summary: BoundaryDiffusion is a learning-free method for efficient and effective image semantic editing using pre-trained generative denoising diffusion models without fine-tuning or extra networks.
Fivos Kalogiannis, Ioannis Panageas
https://openreview.net/forum?id=NGiq8qCQNk
Keywords: network games, Nash equilibrium, equilibrium, game theory, learning
Compressor summary: The paper proposes a class of zero-sum multi-agent Markov games with pairwise interactions and shows how to efficiently find an approximate Nash equilibrium using generalized techniques from previous work on coarse-correlated equilibria.
Michael Painter, Mohamed Baioumy, Nick Hawes, Bruno Lacerda
https://openreview.net/forum?id=NG4DaApavi
Keywords: Monte Carlo Tree Search, Planning, Entropy, Reinforcement Learning
Compressor summary: The paper introduces two improved MCTS methods (BTS and DENTS) that balance exploration and exploitation better than previous methods (MENTS and UCT) in automated planning problems like the game of Go.
DongHyeok Shin, Seungjae Shin, Il-chul Moon
https://openreview.net/forum?id=NEawU0TgKG
Keywords: Dataset distillation, Frequency domain, Dataset condensation
Compressor summary: FreD is a new method to create smaller synthetic datasets from large ones by optimizing frequency representations and selecting relevant dimensions, improving distillation performance and preserving original data information.
Volkan Cevher, Ashok Cutkosky, Ali Kavis, Georgios Piliouras, Stratis Skoulakis, Luca Viano
https://openreview.net/forum?id=NBMIsOS6B7
Keywords: Online Learning, Regret Minimization, Game Theory
Compressor summary: The paper proposes and analyzes two algorithms for alternating online linear optimization that achieve sub-linear regrets, depending on the problem domain and the number of actions.
Vasilis Kontonis, Fotis Iliopoulos, Khoa Trinh, Cenk Baykal, Gaurav Menghani, Erik Vee
https://openreview.net/forum?id=N7tw0QXx3z
Keywords: Distillation, teacher, student
Compressor summary: The paper proposes SLaM, a method to improve knowledge distillation with unlabeled data by mixing student and teacher labels, and provides theoretical guarantees and empirical results.
Pavel Kolev, Georg Martius, Michael Muehlebach
https://openreview.net/forum?id=N6YNe4KxDc
Keywords: online learning, online convex optimization, constrained optimization, adversarial nonlinear constraints, constraint violation oracle
Compressor summary: The CVV-Pro algorithm can learn from continuous non-stationary data streams with adversarial and time-varying constraints by using local sparse linear approximations of the feasible set.
Xiyang Liu, Prateek Jain, Weihao Kong, Sewoong Oh, Arun Suggala
https://openreview.net/forum?id=N6FhEMnxCU
Keywords: Differential Privacy; Private Estimation
Compressor summary: The paper proposes an efficient linear regression method under differential privacy that works well even when some data points are corrupted or in the no-corruption setting.
Yian Deng, Tingting Mu
https://openreview.net/forum?id=N5uUTWLz0E
Keywords: adversarial defense, ensemble diversity, robustness, curvature
Compressor summary: The paragraph describes a new error theory for ensemble adversarial defense, which improves robustness by selectively allocating challenging examples to base classifiers and addressing their weaknesses.
Peiyan Dong, LEI LU, Chao Wu, Cheng Lyu, Geng Yuan, Hao Tang, Yanzhi Wang
https://openreview.net/forum?id=N56hAiQvot
Keywords: Vision Transformers, Quantization, Real-time on mobile, Sub-8-bit
Compressor summary: The paper proposes PackQViT, a framework for efficiently and accurately accelerating ViTs on mobile devices using sub-8-bit quantization and activation-aware training techniques.
Jinbiao Chen, Zizhen Zhang, Zhiguang Cao, Yaoxin Wu, Yining Ma, Te Ye, Jiahai Wang
https://openreview.net/forum?id=N4JkStI1fe
Keywords: neural heuristic, diversity enhancement, deep reinforcement learning, multi-objective combinatorial optimization
Compressor summary: NHDE is a novel neural heuristic method that uses deep reinforcement learning and graph attention to generate diverse Pareto solutions for multi-objective combinatorial optimization problems, outperforming existing decomposition-based approaches.
Sanath Kumar Krishnamurthy, Ruohan Zhan, Susan Athey, Emma Brunskill
https://openreview.net/forum?id=N1feehMSG9
Keywords: Contextual Bandits; Adaptive Experimentation; Simple Regret; Reinforcement Learning
Compressor summary: The authors propose efficient bandit algorithms for learning optimal treatment policies in stochastic settings, using conformal arm sets to balance simple and cumulative regret guarantees.
Christian Lange, Elijah Cole, Grant Van Horn, Oisin Mac Aodha
https://openreview.net/forum?id=N0m9c0FqUV
Keywords: species range estimation, active learning, implicit networks
Compressor summary: The paper proposes an active learning method that uses transfer learned spatial representations to estimate the geographic range of unmapped species from limited observations and crowdsourced data.
Zihao Wang, Lei Wu
https://openreview.net/forum?id=N0KwVdaaaJ
Keywords: Convolutional neural network, Inductive bias, Universality, Sparse function, Equivariance group
Compressor summary: This paper analyzes CNNs' inductive biases, shows they can approximate any continuous function with depth $\mathcal{O}(\log d)$, and reveals the importance of weight sharing and locality for efficient learning.
Thomas FEL, Victor Boutin, Louis Béthune, Remi Cadene, Mazda Moayeri, Léo Andéol, Mathieu Chalvidal, Thomas Serre
https://openreview.net/forum?id=MziFFGjpkb
Keywords: Explainable AI, Concept-based explainability, Interpretability, Concept extraction, Concept importance, Attribution methods
Compressor summary: The paragraph introduces a unifying framework for concept-based explainability methods in ANNs, which connects concept extraction and importance estimation as special cases of dictionary learning and attribution methods, respectively, and provides tools to evaluate and visualize them.
Fanjie Kong, Shuai Yuan, Weituo Hao, Ricardo Henao
https://openreview.net/forum?id=Mxhb2lCOKL
Keywords: Vision-language, Fairness, Text-based Image Retrieval, Deep Learning, Application
Compressor summary: The paper proposes a post-processing technique, Post-hoc Bias Mitigation (PBM), to generate fair and unbiased image retrieval results from neutral textual queries while maintaining the performance of the vision-language model.
Mingxuan Ye, Yufei Kuang, Jie Wang, Rui Yang, Wengang Zhou, Houqiang Li, Feng Wu
https://openreview.net/forum?id=MvoMDD6emT
Keywords: Reinforcement learning, Representation learning, State sequences prediction, Fourier transform
Compressor summary: SPF uses Fourier transform of state sequences to learn efficient representations for data-efficient RL, improving sample efficiency and long-term decision-making.
Yifei Zhang, Hao Zhu, yankai Chen, Zixing Song, Piotr Koniusz, Irwin King
https://openreview.net/forum?id=MvCq52yt9Y
Keywords: Graph, Collaborative Filtering, Recommendation
Compressor summary: The paper proposes a decorrelation-enhanced graph-based collaborative filtering (GCF) method using non-Euclidean geometry to improve feature diversity and reduce popularity bias in personalized recommendation systems.
Ainesh Bakshi, Piotr Indyk, Rajesh Jayaram, Sandeep Silwal, Erik Waingarten
https://openreview.net/forum?id=Mv96iC6TMX
Keywords: chamfer distance, earth mover distance, high dimensional data analysis, nearest neighbor search, high dimensional data, high-dimensional geometry, sublinear algorithms, point clouds, theory
Compressor summary: The paper proposes an efficient algorithm for approximating the Chamfer distance between large point sets in high-dimensional spaces, which can enable new applications and analyses of such data.
Enneng Yang, Li Shen, Zhenyi Wang, Tongliang Liu, Guibing Guo
https://openreview.net/forum?id=Murj6wcjRw
Keywords: Data Condensation, Continual Learning, Few-shot Learning
Compressor summary: The authors propose a new dataset condensation method that uses low-rank matrices to represent raw images, improving performance and reducing privacy risks in training networks from synthetic data.
Bogdan Raonic, Roberto Molinaro, Tim De Ryck, Tobias Rohner, Francesca Bartolucci, Rima Alaifari, Siddhartha Mishra, Emmanuel de Bezenac
https://openreview.net/forum?id=MtekhXRP4h
Keywords: PDEs, Neural Operators, Scientific Machine Learning, Convolutional Neural Networks
Compressor summary: The authors propose convolutional neural operators (CNOs), which are adaptations of convolutional neural networks that can learn operators for partial differential equations (PDEs) while preserving their continuous nature, achieving better performance than existing methods.
Jianghui Wang, Yang Chen, Xingyu Xie, Cong Fang, Zhouchen Lin
https://openreview.net/forum?id=Mr4OpbZEiB
Keywords: Pre-training, Robustness, Multi-task learning
Compressor summary: The paper proposes a pre-training method that ensures good performance on downstream tasks by using a minimax loss and shows improved results on natural language processing and computer vision datasets.
Yifan Zhang, Haowei He, Zhiquan Tan, Yang Yuan
https://openreview.net/forum?id=MmCtXvW6GO
Keywords: AI interpretability, explainable AI, deep learning theory
Compressor summary: The paper shows that popular feature importance methods have a trade-off between interpretability, efficiency, and consistency, and proposes two new algorithms to minimize this trade-off using interpretation error as a metric.
Jiyoung Park, Ian Pelakh, Stephan Wojtowytsch
https://openreview.net/forum?id=MlrFYNo1yc
Keywords: Artificial neural network, interpolation, explicit regularization, implicit bias, weight decay, Barron class
Compressor summary: The paper explores how deep ReLU networks approach the minimal error solution when given enough data and relaxed regularization.
Hanyang Peng, Shuang Qin, Yue Yu, Jin Wang, Hui Wang, Ge Li
https://openreview.net/forum?id=Mlo2kM11ZB
Keywords: optimizer, 1-bit optimizer, distributed learning, optimization for deep networks, communication efficiency
Compressor summary: The paper proposes Birder, a 1-bit adaptive optimizer that offers fast distributed DNN training with similar convergence rate and inference performance as uncompressed SGDM/Adam, while reducing communication volume.
Bidipta Sarkar, Andy Shih, Dorsa Sadigh
https://openreview.net/forum?id=MljeRycu9s
Keywords: Multi-Agent RL, Multi-Agent Coordination, Human-AI Coordination
Compressor summary: The paper proposes a method for creating diverse and effective cooperative strategies in multi-agent games by combining self-play and cross-play optimization techniques.
Wenzhi Gao, Qi Deng
https://openreview.net/forum?id=MirclT6zpv
Keywords: Stochastic optimization, Distributed optimization, Prox-linear method, Stochastic gradient method
Compressor summary: The paper presents new delayed stochastic optimization algorithms with improved convergence rates and robustness against network delays, and shows their effectiveness in distributed weakly convex problems.
Leah Chrestien, Stefan Edelkamp, Antonin Komenda, Tomáš Pevný
https://openreview.net/forum?id=Mgy6sgslPY
Keywords: Learning heuristic functions, deep learning, Immitation learning, planning, A*, best first search
Compressor summary: The paper studies optimal heuristics for forward search algorithms and proposes ranking-based loss functions for imitation learning, with experiments showing good results.
Nate Gruver, Samuel Don Stanton, Nathan C. Frey, Tim G. J. Rudner, Isidro Hotzel, Julien Lafrance-Vanasse, Arvind Rajpal, Kyunghyun Cho, Andrew Gordon Wilson
https://openreview.net/forum?id=MfiK69Ga6p
Keywords: protein design, diffusion model, classifier guidance
Compressor summary: The authors propose NOS, a guidance method for discrete diffusion models that uses gradients from the denoising network to optimize protein design directly in sequence space, improving performance and addressing limitations of structure-based methods.
Eduardo Sany Laber, Lucas Murtinho
https://openreview.net/forum?id=MdJX5wwKwx
Keywords: Single Link, clustering, approximation algorithms, complexity, inter-group criterion
Compressor summary: The authors propose algorithms for optimizing inter-group criteria in clustering and show provable guarantees for two natural measures, with results for both unrestricted and constrained cases, and an empirical study on 10 real datasets.
Benjamin Hoover, Yuchen Liang, Bao Pham, Rameswar Panda, Hendrik Strobelt, Duen Horng Chau, Mohammed J Zaki, Dmitry Krotov
https://openreview.net/forum?id=MbwVNEx9KS
Keywords: Hopfield Network, Dense Associative Memory, Energy-based models, Attention Mechanism
Compressor summary: The Energy Transformer combines attention mechanism, energy-based models, and associative memory to improve machine learning tasks such as image completion, graph anomaly detection, and graph classification.
Shivani Bathla, Vinita Vasudevan
https://openreview.net/forum?id=MamHShmHiX
Keywords: Bayesian inference, posterior marginals, probabilistic graphical models
Compressor summary: The paper introduces a new algorithm for inferring marginals in probabilistic graphical models using a sequence of linked clique tree forests and a heuristic belief update method, which is faster and more accurate than existing methods.
Xiaojun Guo, Yifei Wang, Zeming Wei, Yisen Wang
https://openreview.net/forum?id=MYfqIVcQrp
Keywords: graph contrastive learning
Compressor summary: The paragraph discusses common phenomena in graph contrastive learning (GCL) methods that differ from visual contrastive learning (VCL), and suggests paying attention to the unique architecture of graph learning when designing GCL methods.
Davide Carbone, Mengjian Hua, Simon Coste, Eric Vanden-Eijnden
https://openreview.net/forum?id=MXxZ0Z5MNz
Keywords: Energy-Based Model, contrastive learning, generative models, Jarzynski identity, ULA
Compressor summary: The text describes a new method to train energy-based models more efficiently by using tools from nonequilibrium thermodynamics and sequential Monte-Carlo sampling, which improves performance over existing methods.
Soroush Ebadian, Aris Filos-Ratsikas, Mohamad Latifian, Nisarg Shah
https://openreview.net/forum?id=MWxsYPVmLS
Keywords: explainability, efficiency, voting, distortion, randomized decision-making
Compressor summary: The text discusses how adding simple randomization to deterministic voting rules can improve efficiency without losing explainability, and focuses on two families of such rules.
Yinan Liang, Ziwei Wang, Xiuwei Xu, Yansong Tang, Jie Zhou, Jiwen Lu
https://openreview.net/forum?id=MWp3SwoHmH
Keywords: Vision transformer, microcontroller, network architecture search
Compressor summary: The paper introduces MCUFormer, a method to deploy vision transformers on IoT devices with limited memory by jointly designing architecture and constructing an inference operator library for efficient memory usage.
Yige Hong, Qiaomin Xie, Yudong Chen, Weina Wang
https://openreview.net/forum?id=MWQjqtV1z4
Keywords: restless bandits, average reward MDP, simulation-based method, asymptotic optimality
Compressor summary: The paper proposes a simulation-based method to design near-optimal policies for multi-armed bandit problems without relying on complex assumptions.
Austin Tripp, Sergio Bacallado, Sukriti Singh, José Miguel Hernández-Lobato
https://openreview.net/forum?id=MV0INFAKGq
Keywords: Tanimoto, Kernel, MinMax, Gaussian process, molecule, chemistry, interpretability
Compressor summary: The paper introduces new random features that enable faster computation of the Tanimoto kernel for large molecular datasets, and extend it to real-valued vectors with theoretical and empirical analysis.
Russell Alan Hart, Linlin Yu, Yifei Lou, Feng Chen
https://openreview.net/forum?id=MUzdCW2hC6
Keywords: Uncertainty Quantification, Graph Posterior Network, Bayesian
Compressor summary: The paragraph discusses uncertainty quantification for interdependent node-level classification using graph posterior networks, highlighting the limitations of the uncertainty cross-entropy loss function and proposing a distance-based regularization to improve out-of-distribution and misclassification detection.
Gang Zhang, Chen Junnan, Guohuan Gao, Jianmin Li, Xiaolin Hu
https://openreview.net/forum?id=MUwr2YVJfN
Keywords: 3D object detection; encoder-decoder structure
Compressor summary: HEDNet is a hierarchical encoder-decoder network for 3D object detection in point clouds that captures long-range dependencies among features and outperforms previous methods on Waymo Open and nuScenes datasets.
Zhijian Zhou, Jie Ni, Jia-He Yao, Wei Gao
https://openreview.net/forum?id=MRiitgpcUy
Keywords: two-sample test, local significant difference, directional information
Compressor summary: The paper proposes ME$_\text{MaBiD}$, a new method for two-sample testing that uses multiple Mahalanobis kernels, partitions the embedding space into regions, and tests for local significant differences with a bi-directional approach.
Boya Zhang, Weijian Luo, Zhihua Zhang
https://openreview.net/forum?id=MOAHXRzHhm
Keywords: Adversarial Defense, Adversarial Attack, Score-based Models, Diffusion Models
Compressor summary: The paper proposes ScoreOpt, a novel defense against adversarial attacks for deep neural networks that optimizes adversarial samples towards clean data, achieving better robustness and faster inference than existing methods.
Weijian Luo, Tianyang Hu, Shifeng Zhang, Jiacheng Sun, Zhenguo Li, Zhihua Zhang
https://openreview.net/forum?id=MLIs5iRq4w
Keywords: diffusion model, data-free distillation, implicit generator, knowledge transfer
Compressor summary: Diff-Instruct is a framework for instructing the training of different generative models using pre-trained diffusion models, based on a novel divergence called Integral Kullback-Leibler (IKL) divergence.
Zhimin Chen, Longlong Jing, Yingwei Li, Bing Li
https://openreview.net/forum?id=MJbDy2155j
Keywords: 3D self-supervised learning, Multi-modal Representation Learning, Masked autoencoders, Knowledge distillation
Compressor summary: The authors propose Bridge3D, a method that uses features, semantic masks, and captions from foundation models to pre-train 3D models for better 3D scene representation learning and improve 3D object detection and semantic segmentation tasks.
Minyoung Hwang, Gunmin Lee, Hogun Kee, Chan Woo Kim, Kyungjae Lee, Songhwai Oh
https://openreview.net/forum?id=MIYBTjCVjR
Keywords: Reinforcement Learning; Reinforcement Learning from Human Feedback; Preference-based Reinforcement Learning; Human-Robot Interaction
Compressor summary: The proposed SeqRank framework uses sequential preference ranking to enhance feedback efficiency in reinforcement learning from human feedback, leading to higher average feedback efficiency and improved task performance.
Lie He, Shiva Kasiviswanathan
https://openreview.net/forum?id=MH7E7AME1r
Keywords: Optimization, Bilevel Optimization, Stochastic Optimization
Compressor summary: The paper proposes a general stochastic extrapolation technique to reduce bias and improve sample complexity for conditional stochastic optimization (CSO) problems, which have applications in various fields such as portfolio selection and reinforcement learning.
Chuanbo Hua, Federico Berto, Michael Poli, Stefano Massaroli, Jinkyoo Park
https://openreview.net/forum?id=MGPST5I9DO
Keywords: Graph, Spline Collocation Method, Graph Neural Networks, Simulation, Partial Differential Equations, PDEs, Physics, Scientific Computing, Surrogate Models, Weather Forecasting
Compressor summary: GraphSplineNets is a deep learning method that uses spline collocation to forecast physical systems faster by reducing grid size and iteration steps, with an adaptive strategy for sampling important regions.
Berfin Simsek, Amire Bendjeddou, Wulfram Gerstner, Johanni Brea
https://openreview.net/forum?id=MG0mYskXN2
Keywords: shallow neural networks, non-convex optimization, approximation error, loss landscape
Compressor summary: The paragraph discusses how an under-parameterized neural network (student) can be fitted to an over-parameterized one (teacher) and shows that "copy-average" configurations are critical points for shallow networks with specific activation functions.
Konstantin Makarychev, Liren Shan
https://openreview.net/forum?id=MFWgLCWgUB
Keywords: Clustering, k-medians, Decision Tree, Explainability
Compressor summary: The RandomCoordinateCut algorithm achieves the best possible performance for solving the explainable $k$-medians problem in the $\ell_1$ norm, with a competitive ratio of at most $2\ln k+2$.
Pini Zilber, Boaz Nadler
https://openreview.net/forum?id=MDxZYFR5Me
Keywords: Mixture regression model, Mixture of linear models, Iteratively reweighted least squares
Compressor summary: Mix-IRLS is a novel algorithm that efficiently solves mixed linear regression problems with imbalanced and challenging data settings, outperforming existing methods.
Ayush Sawarni, Soumyabrata Pal, Siddharth Barman
https://openreview.net/forum?id=MCkUS1P3Sh
Keywords: Sub-Poisson Distribution, Nash Social Welfare, Fairness Quantification, John Ellipsoid, Kiefer-Wolfowitz Optimal Design, Algorithmic Game Theory, Online Learning
Compressor summary: The paper presents a new notion of regret called Nash regret, which measures the performance of a bandit algorithm as collective welfare, and develops an algorithm that achieves tight upper bounds on it for stochastic linear bandits with sub-Poisson rewards.
Jinhang Zuo, Zhiyao Zhang, Zhiyong Wang, Shuai Li, Mohammad Hajiesmaili, Adam Wierman
https://openreview.net/forum?id=MCj7DLkYqS
Keywords: online learning to rank, adversarial attack, click model
Compressor summary: This paper investigates various attack strategies against online learning to rank algorithms and proposes a general attack strategy that works under different feedback models.
Abulhair Saparov, Richard Yuanzhe Pang, Vishakh Padmakumar, Nitish Joshi, Mehran Kazemi, Najoung Kim, He He
https://openreview.net/forum?id=MCVfX7HgPO
Keywords: large language models, reasoning, out-of-distribution generalization, chain-of-thought, in-context learning
Compressor summary: The paragraph discusses testing the general deductive reasoning ability of large language models on various deduction rules and complex proofs, using a new synthetic dataset.
Yueming Lyu
https://openreview.net/forum?id=M8CYKLHoEN
Keywords: Black-box Optimization, Derivate-free Optimization, Kernel methods
Compressor summary: RLTS is a novel technique that uses rank-1 lattices to improve query efficiency in high-dimensional optimization problems, enabling faster training and inference of Gaussian processes.
Johann Brehmer, Pim De Haan, Sönke Behrends, Taco Cohen
https://openreview.net/forum?id=M7r2CO4tJC
Keywords: Geometry, geometric algebra, equivariance, transformer
Compressor summary: The Geometric Algebra Transformer (GATr) is a new architecture for geometric data that leverages projective geometric algebra and symmetry to achieve better performance in various applications.
Alex Tamkin, Margalit Glasgow, Xiluo He, Noah Goodman
https://openreview.net/forum?id=M7hijAPA4B
Keywords: self-supervised learning, contrastive learning
Compressor summary: Augmentations that destroy labels can be beneficial in foundation model settings, where diverse, general-purpose representations are needed for multiple downstream tasks.
Jintong Gao, He Zhao, Zhuo Li, Dan dan Guo
https://openreview.net/forum?id=M7FQpIdo0X
Keywords: Long-tailed Classification, Optimal Transport, Image-mixing, Semantic Similarity
Compressor summary: The authors propose an adaptive image-mixing method based on optimal transport to generate semantically reasonable mixed images for minority classes in long-tailed classification tasks, improving performance and serving as a general data augmentation method.
Larry Han, Zhu Shen, Jose R Zubizarreta
https://openreview.net/forum?id=M6UccKMFGl
Keywords: Causal inference, Covariate mismatch, Federated learning, Multiple robustness, Transportation
Compressor summary: The paper proposes a new method to make valid causal inferences from multi-site studies while preserving privacy and accounting for differences between sites.
Geunwoo Kim, Pierre Baldi, Stephen Marcus McAleer
https://openreview.net/forum?id=M6OmjAZ4CX
Keywords: Large Language models, Web Navigation, Foundation Models, Decision Making
Compressor summary: The paper presents a new method called RCI, which uses a pre-trained language model to automate computer tasks via natural language commands with fewer demonstrations and without task-specific rewards, outperforming existing methods on benchmarks.
Guanqiang Zhou, Ping Xu, Yue Wang, Zhi Tian
https://openreview.net/forum?id=M4h1UAxI3b
Keywords: distributed learning, federated learning, fairness, robustness, Byzantine attack, norm-based screening, q-FFL, optimization, convergence analysis
Compressor summary: The paper proposes a framework called H-nobs that combines fairness and robustness in distributed learning systems, addressing challenges like data heterogeneity and trade-offs between the two properties.
Kwangjun Ahn, Xiang Cheng, Hadi Daneshmand, Suvrit Sra
https://openreview.net/forum?id=LziniAXEI9
Keywords: Transformers, In-context learning, adaptive gradient methods
Compressor summary: The authors explore if transformers can learn to implement algorithms like gradient descent by training over random problem instances and provide theoretical analysis of the loss landscape for linear transformers in this setting.
Vedant Nanda, Till Speicher, John P Dickerson, Krishna P. Gummadi, Soheil Feizi, Adrian Weller
https://openreview.net/forum?id=LyAuNoZkGP
Keywords: representation learning, redundancy, transfer learning, fairness
Compressor summary: This study analyzes how pre-trained neural network features encode diffuse redundancy, where a subset of neurons can perform similar tasks as the whole layer, depending on the downstream task and pre-training parameters.
Vladimir R Kostic, Karim Lounici, Pietro Novelli, massimiliano pontil
https://openreview.net/forum?id=Lt3jqxsbVO
Keywords: Statistical Learning Theory, Dynamical Systems
Compressor summary: The paper presents non-asymptotic learning bounds for Koopman eigenvalues and eigenfunctions in time-reversal-invariant stochastic dynamical systems, comparing two estimators (EDMD and RRR) and discussing their bias and variance trade-off.
Siqiao Xue, Yan Wang, Zhixuan Chu, Xiaoming Shi, Caigao JIANG, Hongyan Hao, Gangwei Jiang, Xiaoyun Feng, James Y. Zhang, JUN ZHOU
https://openreview.net/forum?id=LswqtKU9op
Keywords: prompt, point process, event sequence, continual learning.
Compressor summary: The text introduces PromptTPP, a framework that uses Continual Learning to monitor Neural Temporal Point Processes and learn event sequences without forgetting or buffering past data.
Cassidy Laidlaw, Stuart Russell, Anca Dragan
https://openreview.net/forum?id=Lr2swAfwff
Keywords: reinforcement learning, RL theory, theory of reinforcement learning, instance-dependent bounds, empirical validation of theory
Compressor summary: The authors introduce BRIDGE, a new dataset for comparing deep RL algorithms and bounds, and propose the effective horizon as a better measure of MDP complexity that correlates with empirical performance and exploration techniques.
Huizong Yang, Yuxin Sun, Ganesh Sundaramoorthi, Anthony Yezzi
https://openreview.net/forum?id=Lqv7VS1iBF
Keywords: Implicit Neural Representation, Surface Reconstruction
Compressor summary: The authors propose a novel approach to learning implicit neural representations of shapes, which addresses the instability issues in existing methods and enables capturing finer shape details.
Thaddäus Wiedemer, Prasanna Mayilvahanan, Matthias Bethge, Wieland Brendel
https://openreview.net/forum?id=LqOQ1uJmSx
Keywords: compositional generalization, compositionality, generalization, combinatorial generalization, out-of-distribution, out-of-domain, identifiability, disentanglement, object-centric learning, DSprites
Compressor summary: The text discusses an approach to improve machine learning's compositional generalization by considering it as a property of the data-generating process, rather than the data itself, and provides conditions and empirical validation for this approach.
Yue Song, T. Anderson Keller, Nicu Sebe, Max Welling
https://openreview.net/forum?id=LnySNEJAQt
Keywords: Generative modelling, latent disentanglement, variational autoencoders
Compressor summary: The authors propose Flow Factorized Representation Learning, a novel approach to structured representation learning that uses latent flow paths generated by learned potentials to achieve more efficient and better structured representations than existing methods.
Guy Kornowski, Gilad Yehudai, Ohad Shamir
https://openreview.net/forum?id=LnZuxp3Tx7
Keywords: benign overfitting, implicit bias, interpolating predictors, neural networks, theory
Compressor summary: The text discusses how the type of overfitting in neural networks depends on input dimension, sample size, architecture, and training algorithm, and provides theoretical and empirical evidence for this claim.
Stav Belogolovsky, Ido Greenberg, Danny Eytan, Shie Mannor
https://openreview.net/forum?id=Lmxo0RVNx2
Keywords: personalized medicine, dosing dynamics, sequential prediction, stochastic differential equations, Kalman filter, recurrent neural networks, medical drug control
Compressor summary: NESDE is a novel algorithm that uses neural differential equations to model biological dynamics in individual patients, generalize to new treatment policies, and handle different noise levels while providing fast and continuous predictions.
Nicolas Keriven, Samuel Vaiter
https://openreview.net/forum?id=LmmjiTwYm0
Keywords: graph neural network; random graph; positional encoding
Compressor summary: This paper investigates the expressive power of equivariant Graph Neural Networks (GNNs) for node-tasks on large graphs, considering input features and positional encodings, and provides theoretical and empirical insights.
Shangtong Gui, Chenze Shao, Zhengrui Ma, Xishan Zhang, Yunji Chen, Yang Feng
https://openreview.net/forum?id=LloZFVwWvj
Keywords: Machine translation, Non-autoregressive generation, Probabilistic Context-free Grammar
Compressor summary: PCFG-NAT improves non-autoregressive neural machine translation by using a probabilistic context-free grammar to capture complex dependencies among output tokens and enhance explainability.
Erin George, Michael Murray, William Joseph Swartworth, Deanna Needell
https://openreview.net/forum?id=LlERoXEKjh
Keywords: benign overfitting, neural networks, relu, hinge loss
Compressor summary: The paper investigates how two-layer ReLU networks with gradient descent and hinge loss handle linearly separable data with some noisy labels, and identifies three possible outcomes depending on the margin conditions of the clean data.
Yexiong Lin, Yu Yao, Xiaolong Shi, Mingming Gong, Xu Shen, Dong Xu, Tongliang Liu
https://openreview.net/forum?id=Lkc0KjsDFv
Keywords: learning with label errors
Compressor summary: The paper proposes a method to infer content factors from noisy labels by using data augmentation and regularization, improving the learning of hard examples for image classification.
Joe Suk, Arpit Agarwal
https://openreview.net/forum?id=LjWJLkSpjh
Keywords: non-stationary, multi-armed bandits, dueling bandits, preference-based learning
Compressor summary: The paper investigates how preferences can change over time in dueling bandits problems and explores the feasibility of adaptive algorithms with low dynamic regret under different preference distributions.
Daiki E. Matsunaga, Jongmin Lee, Jaeseok Yoon, Stefanos Leonardos, Pieter Abbeel, Kee-Eung Kim
https://openreview.net/forum?id=LhVJdq4cZm
Keywords: Offline Reinforcement Learning, Multi-Agent Reinforcement Learning
Compressor summary: AlberDICE is an offline multi-agent reinforcement learning algorithm that trains agents individually based on stationary distribution optimization and avoids out-of-distribution joint actions, achieving better performance than baselines.
Jin Xu, Emilien Dupont, Kaspar Märtens, Tom Rainforth, Yee Whye Teh
https://openreview.net/forum?id=Lg1ODJGGiI
Keywords: Neural Processes, Bayesian Nonparammetric Models
Compressor summary: Markov Neural Processes combine neural networks with stochastic processes, preserving their properties and improving performance on various tasks.
Hanlin Yang, Chao Yu, peng sun, Siji Chen
https://openreview.net/forum?id=LftAvFt54C
Keywords: reinforcement learning, sparse reward, exploration, learning from demonstrations
Compressor summary: The paper proposes a novel RL algorithm called HYPO that uses imperfect demonstrations to train an offline guider policy to instruct an online agent policy for efficient exploration in challenging tasks with sparse rewards.
Jia Lin Hau, Erick Delage, Mohammad Ghavamzadeh, Marek Petrik
https://openreview.net/forum?id=LelK6Mfoey
Keywords: reinforcement learning, markov decision processes, monetary risk measures
Compressor summary: The paragraph discusses the limitations of common risk-averse optimization methods in Markov decision processes and proposes a new decomposition for Value-at-Risk that avoids suboptimality.
Xuanle Zhao, Duzhen Zhang, Liyuan Han, Tielin Zhang, Bo XU
https://openreview.net/forum?id=LdvVd0bNyO
Keywords: neural ode, POMDPs, reinforcement learning
Compressor summary: The paragraph discusses a novel ODE-based recurrent model combined with model-free RL for solving POMDPs and demonstrates its effectiveness in various tasks, including robustness against irregular observations.
Chenran Li, Chen Tang, Haruki Nishimura, Jean Mercat, Masayoshi Tomizuka, Wei Zhan
https://openreview.net/forum?id=LaNeRwDrTk
Keywords: reinforcement learning, imitation learning
Compressor summary: The paragraph discusses imitation learning and introduces a new problem setting called policy customization, which involves training a policy that inherits the characteristics of a prior policy while satisfying additional requirements from a downstream task, using a novel framework called Residual Q-learning.
Yang Liu, Feng Wang, Naiyan Wang, Zhaoxiang Zhang
https://openreview.net/forum?id=LZzsn51DPr
Keywords: 4D Radar; Transformer; Multi-modality
Compressor summary: The paper introduces EchoFusion, a novel radar signal processing method that fuses raw radar data with other sensors to improve autonomous driving performance and approach LiDAR accuracy.
Junpei Komiyama, Masaaki Imaizumi
https://openreview.net/forum?id=LZ4WgwmrUJ
Keywords: multi-armed bandits, linear bandits, contextual bandits, overparameterized models, high-dimensional models, online learning
Compressor summary: The paper studies how to solve high-dimensional linear contextual bandit problems using explore-then-commit and adaptive explore-then-commit algorithms, without imposing sparsity on regression coefficients.
Ethan Brooks, Logan A Walls, Richard Lewis, Satinder Singh
https://openreview.net/forum?id=LWxjWoBTsr
Keywords: Reinforcement Learning, In-Context Learning, Foundation Models
Compressor summary: The authors propose a method to use a large language model for policy iteration without expert demonstrations or gradients, by iteratively updating the prompt and interacting with an RL environment.
Zebin You, Yong Zhong, Fan Bao, Jiacheng Sun, Chongxuan Li, Jun Zhu
https://openreview.net/forum?id=LVHEcVgEGm
Keywords: diffusion models, semi-supervised generation, semi-supervised diffusion models, semi-supervised classification, image generation.
Compressor summary: The authors propose dual pseudo training, a method that uses partial labels to train a classifier, then a generative model, and finally retrains the classifier with real and generated images, achieving state-of-the-art results in semi-supervised image generation and classification.
Zaid Khan, Vijay Kumar b g, Samuel Schulter, Manmohan Chandraker, Yun Fu
https://openreview.net/forum?id=LV83JEihHu
Keywords: visual question answering, in-context learning, vision-language
Compressor summary: The authors propose a question decomposition strategy for visual question answering, which improves performance by selectively decomposing questions and using human-written decompositions to guide large vision-language models.
Lars Holdijk, Yuanqi Du, Ferry Hooft, Priyank Jaini, Bernd Ensing, Max Welling
https://openreview.net/forum?id=LUVqEs90mq
Keywords: Transition Path Sampling, Stochastic Optimal Control
Compressor summary: The paragraph describes a problem of finding transition paths between metastable states in molecular systems, proposes a machine learning method called PIPS that does not rely on collective variables, and demonstrates its effectiveness on small proteins.
Jinheng Xie, Kai Ye, Yudong Li, Yuexiang Li, Kevin Qinghong Lin, Yefeng Zheng, Linlin Shen, Mike Zheng Shou
https://openreview.net/forum?id=LUT4b9gOtS
Keywords: Visual Prior, Generative Pre-Training, Conditional Image Synthesis
Compressor summary: The paper proposes VisorGPT, a method to learn and customize visual prior using generative pre-training and prompt engineering for conditional image synthesis tasks.
Nikhil Vyas, Alexander Atanasov, Blake Bordelon, Depen Morwani, Sabarish Sainathan, Cengiz Pehlevan
https://openreview.net/forum?id=LTdfYIvbHc
Keywords: mean-field, muP, feature learning, infinite width, deep ensembles
Compressor summary: The dynamics of feature-learning neural networks differ across various architectures and datasets, but their behavior can be captured by infinite-width limits for simple tasks, while deviations grow systematically for harder tasks due to initialization-dependent output variance scaling and the bias of narrower width.
Honghao Wei, Xin Liu, Weina Wang, Lei Ying
https://openreview.net/forum?id=LTbIUkN95h
Keywords: Reinforcement Learning, Mixed Systems, Queueing Network, Sample Efficient
Compressor summary: The paper proposes a sample-efficient reinforcement learning method for mixed systems with stochastic and pseudo-stochastic states, which improves learning by generating augmented data samples and reducing the sample complexity.
Jannik Kossen, Mark Collier, Basil Mustafa, Xiao Wang, Xiaohua Zhai, Lucas Beyer, Andreas Peter Steiner, Jesse Berent, Rodolphe Jenatton, Effrosyni Kokiopoulou
https://openreview.net/forum?id=LSYQB4CwD3
Keywords: three towers, contrastive learning, transformers, vision transformers, pretrained models, representation learning, finetuning, CLIP, ALIGN, classification, zero-shot, few-shot, retrieval
Compressor summary: Three Towers (3T) is a method that enhances contrastive learning of vision-language models by combining pretrained image classifiers with contrastive training, improving performance on retrieval tasks and classification for some pretraining datasets.
Xiao-Yue Gong, Mark Sellke
https://openreview.net/forum?id=LROEcjVkv5
Keywords: pure exploration, multi-armed bandits, Fisher information
Compressor summary: The paper studies how to efficiently select one high quality arm in a setting with infinitely many arms, considering both fixed confidence and fixed budget scenarios.
Xianhang Li, Zeyu Wang, Cihang Xie
https://openreview.net/forum?id=LMU2RNwdh2
Keywords: CLIP, inverse scaling, efficient training
Compressor summary: The paper finds an inverse scaling law for CLIP training, reducing its computational cost and enabling faster training with limited resources.
Kyle Hsu, Will Dorrell, James C. R. Whittington, Jiajun Wu, Chelsea Finn
https://openreview.net/forum?id=LLETO26Ga2
Keywords: disentanglement, unsupervised learning, quantization
Compressor summary: The authors propose an inductive bias for latent space encoding and decoding using quantization and regularization, which improves disentangled representation learning in various generative models, and introduce new metrics called InfoMEC to evaluate disentanglement reliably.
Sangwoong Yoon, Frank C. Park, Gunsu S YUN, Iljung Kim, Yung-Kyun Noh
https://openreview.net/forum?id=LIsJHQHi4z
Keywords: Kernel Density Estimation, KL-divergence, Density Ratio
Compressor summary: Kernel density estimation (KDE) improves predictions and information-theoretic measures by using an optimal weight function derived from multidimensional calculus of variations for density ratios.
Zheng Wang, Shikai Fang, Shibo Li, Shandian Zhe
https://openreview.net/forum?id=LGqIAn2OaZ
Keywords: Tensor Decomposition, Representation Learning
Compressor summary: DEMOTE is a method that uses neural networks to capture temporal structure in sparse multiway data, by estimating dynamic embeddings for entities in each mode of a tensor.
Sara Pieri, Jose Renato Restom, Samuel Horváth, Hisham Cholakkal
https://openreview.net/forum?id=LGKxz9clGG
Keywords: Computer Vision, Federated Learning, Image Classification, Neural Network Architectures, Transformer, CNN, Data Hetereogenity, non-IID
Compressor summary: The paper reviews federated learning for visual recognition, highlighting the crucial role of architecture choices to achieve optimal performance and narrow the gap with centralized learning.
Mahyar Fazlyab, Taha Entesari, Aniket Roy, Rama Chellappa
https://openreview.net/forum?id=LDhhi8HBO3
Keywords: Deep Learning, Adversarial Robustness, Certified Radius, Lipschitz Constants
Compressor summary: The paper proposes a differentiable regularizer that uses Lipschitz bounds on neural networks to improve robustness against adversarial perturbations, and shows better results than existing methods on image classification tasks.
Tamas Sarlos, Xingyou Song, David Woodruff, Qiuyi Zhang
https://openreview.net/forum?id=LCwToX315b
Keywords: Low rank approximation, kernel methods, fine-grained complexity
Compressor summary: The authors study low rank approximation of a function involving two matrices, and show that certain conditions are necessary for fast algorithms, while also giving new hardness results and efficient algorithms for some cases.
Sébastien Herbreteau, Emmanuel Moebel, Charles Kervrann
https://openreview.net/forum?id=LCnjG1IEfm
Keywords: equivariance, normalization, image denoising, activation functions, ReLU, interpretability, robustness, deep learning, analysis of neural networks
Compressor summary: The text proposes a method to adapt neural networks for normalization-equivariance, which is beneficial for many applications, by replacing existing layers and functions with new ones.
Dairui Wang, Junyu Cao, Yan Zhang, Wei Qi
https://openreview.net/forum?id=LClyG4vZmS
Keywords: delayed feedback, recommender system, frequency control
Compressor summary: This paper proposes a new bandit problem in dynamic recommender systems and presents several algorithms to optimize recommendation frequency and user engagement based on delayed feedback.
Paul-Edouard Sarlin, Eduard Trulls, Marc Pollefeys, Jan Hosang, Simon Lynen
https://openreview.net/forum?id=LCHmP68Gtj
Keywords: neural maps, visual positioning, semantic mapping
Compressor summary: SNAP is a deep network that learns 2D neural maps from images and can create better, more detailed, and semantically richer maps than traditional methods.
Christos Sourmpis, Carl C. H. Petersen, Wulfram Gerstner, Guillaume Bellec
https://openreview.net/forum?id=LAbxkhkjbD
Keywords: neuroscience, spiking networks, data-constrained modeling, electrophysiological recordings, optimal transport, trial variability, RNN, interpretable machine learning
Compressor summary: The authors use a neural network model fitted to mouse behavioral and neural data to simulate sensory-motor processes and identify different modes of variability in the data.
Siu Lun Chau, Krikamol Muandet, Dino Sejdinovic
https://openreview.net/forum?id=LAGxc2ybuH
Keywords: Gaussian Processes, Shapley values, Uncertainty Modelling
Compressor summary: The authors introduce a method to explain Gaussian processes using Shapley values from stochastic cooperative games, which allows for uncertainty quantification and predictive explanations.
Sima Behpour, Thang Doan, Xin Li, Wenbin He, Liang Gou, Liu Ren
https://openreview.net/forum?id=L9nTuSbAws
Keywords: out-of-distribution detection, OOD, uncertainty estimation, gradient projection
Compressor summary: GradOrth is a novel approach for detecting out-of-distribution data that uses important features from lower-rank subspaces of in-distribution data, achieving better performance than existing methods.
Miltiadis Kofinas, Erik J Bekkers, Naveen Shankar Nagaraja, Efstratios Gavves
https://openreview.net/forum?id=L9ZTvJ5jVx
Keywords: Graph Neural Networks, Neural Fields, Field Discovery, Equivariance, Interacting Dynamical Systems, Geometric Graphs
Compressor summary: The authors propose a method to infer hidden force fields from observed dynamics using equivariant graph networks and neural fields, which can improve prediction of system evolution.
Osama Hanna, Lin Yang, Christina Fragouli
https://openreview.net/forum?id=L7Whl9pXd0
Keywords: efficient bandit algorithms, contextual linear bandits
Compressor summary: The paper presents an efficient batched algorithm for contextual linear bandits with large action spaces using a novel soft elimination approach and achieving low regret bounds.
Vishal Asnani, Abhinav Kumar, Suya You, Xiaoming Liu
https://openreview.net/forum?id=L74NTrzH1O
Keywords: Object detection, proactive, Camouflage, 2D
Compressor summary: PrObeD is a proactive scheme that enhances 2D object detection performance by learning an image-dependent template that acts as a mask for the input images, improving detection for generic and camouflaged objects.
Md Ashiqur Rahman, Raymond A. Yeh
https://openreview.net/forum?id=L0QwnevT0F
Keywords: Scale Equivariance, Fourier Neural Network
Compressor summary: The paper presents a new scale-equivariant convolutional neural network architecture that considers anti-aliasing in down-scaling and performs well on image classification tasks.
Zihao Yue, Anwen Hu, Liang Zhang, Qin Jin
https://openreview.net/forum?id=Kvaa3DhvlZ
Keywords: Image Captioning, Learning Objective, Natural Language Processing
Compressor summary: SMILE is a method for image captioning that encourages richer and more detailed descriptions by preventing conciseness optimization.
Ajay Subramanian, Elena Sizikova, Najib J. Majaj, Denis G. Pelli
https://openreview.net/forum?id=KvPwXVcslY
Keywords: object recognition, critical band masking, spatial-frequency channels, shape bias, adversarial robustness
Compressor summary: Critical band masking shows humans and neural networks use different frequency filters for object recognition, with networks having broader filters that make them less robust to noise.
Aravind Gollakota, Adam Klivans, Konstantinos Stavropoulos, Arsen Vasilyan
https://openreview.net/forum?id=Kv8GJkV19S
Keywords: testable learning, pac learning, agnostic learning, Massart label noise, adversarial label noise, distribution testing
Compressor summary: The paragraph describes a universal tester-learner for halfspaces that works on many structured distributions, has low error, and uses sum-of-squares programs to check hypercontractivity of unknown distributions.
Zihao Wang, Shaofei Cai, Guanzhou Chen, Anji Liu, Xiaojian Ma, Yitao Liang
https://openreview.net/forum?id=KtvPdGb31Z
Keywords: open-ended learning, multi task, large language models, zero-shot planning
Compressor summary: The paper presents DESPs, a novel interactive planning method for Minecraft agents using LLMs, which enables a zero-shot multi-task agent to achieve better results in various domains.
Zhenchao Jin, Xiaowei Hu, Lingting Zhu, Luchuan Song, Li Yuan, Lequan Yu
https://openreview.net/forum?id=KtHquQuyA5
Keywords: semantic segmentation, relation modeling, object detection
Compressor summary: IDRNet is a novel context modeling paradigm that uses deletion diagnostics to improve dense prediction tasks by enhancing pixel-level representations with semantic-level interactions.
Jialv Zou, Xinggang Wang, JiaHao Guo, Wenyu Liu, Qian Zhang, Chang Huang
https://openreview.net/forum?id=KsICioDlYs
Keywords: EDA, Circuit Design, Congestion prediction, DRC violation prediction
Compressor summary: The authors propose a new method for circuit design using Transformer-based point cloud perception to extract features directly from raw data, achieving high performance in various tasks.
Futoshi Futami, Masahiro Fujisawa
https://openreview.net/forum?id=Ks0RSFNxPO
Keywords: SGLD, Langevin dynamics, Generalization, Information theoretic analysis
Compressor summary: The paper proposes new generalization bounds for SGLD that are time-independent, decay with sample size, and account for stability of datasets, improving upon previous work.
Kihyuk Sohn, Lu Jiang, Jarred Barber, Kimin Lee, Nataniel Ruiz, Dilip Krishnan, Huiwen Chang, Yuanzhen Li, Irfan Essa, Michael Rubinstein, Yuan Hao, Glenn Entis, Irina Blok, Daniel Castro Chin
https://openreview.net/forum?id=KoaFh16uOc
Keywords: text-to-image synthesis, fine-tuning, stylization
Compressor summary: StyleDrop is a versatile method that allows synthesizing images with specific styles using text-to-image models, even with just one image as input, and outperforms other methods in style tuning.
Yuyang Deng, Mohammad Mahdi Kamani, Pouria Mahdavinia, Mehrdad Mahdavi
https://openreview.net/forum?id=KoQgA0coZ9
Keywords: distributed learning, heterogeneous data, heterogeneous system, convergence analysis
Compressor summary: The paper introduces PERM, a new learning method for heterogeneous data that adapts models to each client's needs and constraints, using a distributed algorithm that optimizes PERM objectives and allows different model architectures.
TIANCHI LIU, Kong Aik Lee, Qiongqiong Wang, Haizhou Li
https://openreview.net/forum?id=KoFYzuwjCA
Keywords: speaker recognition, disentanglement learning, self-supervision
Compressor summary: The paper introduces a framework that uses three Gaussian inference layers to model speaker traits and content variability in speech, disentangling them without labels other than speaker identities, and shows its effectiveness on VoxCeleb and SITW datasets.
Tongtong Fang, Nan Lu, Gang Niu, Masashi Sugiyama
https://openreview.net/forum?id=KmdlUP23qh
Keywords: importance weighting, distribution shift, deep learning
Compressor summary: The paper proposes a generalized importance weighting method to handle distribution shift problems in machine learning, addressing the limitations of existing methods.
Ravid Shwartz-Ziv, Randall Balestriero, Kenji Kawaguchi, Tim G. J. Rudner, Yann LeCun
https://openreview.net/forum?id=KipjqOPaZ0
Keywords: Self-Supervised Learning, Generalization Bounds, Information-Theory, Deep Neural Networks
Compressor summary: The paper explores the mechanisms behind VICReg, an SSL method, and introduces new SSL methods based on information-theoretic principles that perform better than existing ones.
Klim Kireev, Maksym Andriushchenko, Carmela Troncoso, Nicolas Flammarion
https://openreview.net/forum?id=Kig2YJVYfq
Keywords: Tabular data, Categorical data, Robust ML, Adversarial Robustness
Compressor summary: The paper proposes a method to train robust deep networks and transfer robustness to other classifiers using categorical feature embeddings for tabular data, which improves performance under adversarial attacks compared to existing methods.
Dheeraj Baby, Saurabh Garg, Tzu-Ching Yen, Sivaraman Balakrishnan, Zachary Chase Lipton, Yu-Xiang Wang
https://openreview.net/forum?id=Ki6DqBXss4
Keywords: online learning, label shift, distribution shift, unsupervised domain adaptation
Compressor summary: The paper presents algorithms for adapting to online label shift in both supervised and unsupervised settings, using bootstrapped online regression oracles to track drifting class marginals and achieve optimal dynamic regret with high accuracy.
Bowen Li, Jiashun Wang, Yaoyu Hu, Chen Wang, Sebastian Scherer
https://openreview.net/forum?id=KgqucdSwIe
Keywords: Unseen object detection, instance perception, voxel representation
Compressor summary: The text introduces VoxDet, a 3D geometry-aware framework that uses voxel representation and matching techniques to improve object detection in challenging scenarios involving occlusion and pose variations.
Xinyi Wu, Amir Ajorlou, Zihui Wu, Ali Jadbabaie
https://openreview.net/forum?id=Kg65qieiuB
Keywords: graph neural networks, attention mechanisms, oversmoothing, dynamical systems, theory
Compressor summary: The paper analyzes the effect of graph attention on oversmoothing in different types of Graph Neural Networks (GNNs) using mathematical tools and shows that it does not prevent oversmoothing and loses expressive power exponentially.
Rajat Vadiraj Dwaraknath, Ishani Karmarkar, Aaron Sidford
https://openreview.net/forum?id=KffE8iXAw7
Keywords: effective resistances, spectral sketch, fine-grained complexity, triangle detection, numerical linear algebra
Compressor summary: The paper presents new algorithms for estimating effective resistances in undirected expander graphs with faster runtimes than previous methods, as well as a conditional lower bound showing that these algorithms are optimal up to a factor of $\epsilon$.
Xingjian Bai, Guangyi He, Yifan Jiang, Jan Obloj
https://openreview.net/forum?id=KfOUAlraMP
Keywords: adversarial attack, adversarial robustness of DNN, adversarial training, Wasserstein distance, distributionally robust optimization, sensitivity analysis, asymptotic bounds
Compressor summary: The authors propose a method for evaluating the robustness of deep neural networks against adversarial attacks using Wasserstein distributionally robust optimization and show its effectiveness on various image recognition tasks.
Muhammed Fatih Balin, Umit Catalyurek
https://openreview.net/forum?id=Kd5W4JRsfV
Keywords: Graph Neural Networks, Graph Sampling, GNN, Layer Sampling, Minibatch Training
Compressor summary: LABOR is a new sampling algorithm for Graph Neural Networks that samples fewer vertices with better performance and convergence than existing methods.
Hao Liu, Pieter Abbeel
https://openreview.net/forum?id=KbqQMoqfLQ
Keywords: Language Model, Long Context Modeling, Reinforcement Learning
Compressor summary: The Blockwise Parallel Transformer (BPT) is a new method that reduces memory costs and enables longer input sequences for natural language processing models by processing them in blocks, outperforming previous methods on language modeling and reinforcement learning tasks.
Nishanth Anand, Doina Precup
https://openreview.net/forum?id=KakzVASqul
Keywords: reinforcement learning, continual reinforcement learning, lifelong learning, never-ending learning, prediction, control, multi-task learning, complementary learning systems
Compressor summary: The paper proposes a value function decomposition into permanent and transient components for continual reinforcement learning, showing theoretical and empirical benefits.
Jia Guo, shuai lu, LIze JIa, Weihang Zhang, Huiqi Li
https://openreview.net/forum?id=KYxD9YCQBH
Keywords: Unsupervised Anomaly Detection, Contrastive Learning, Medical Anomaly Detection, Transfer Learning
Compressor summary: The paper introduces ReContrast, a new unsupervised anomaly detection method that adapts to different domains by optimizing the entire network for feature reconstruction and contrastive learning.
Gergely Flamich, Stratis Markou, José Miguel Hernández-Lobato
https://openreview.net/forum?id=KXbAgvLi2l
Keywords: Compression, Learnt Compression, Relative Entropy Coding, Information Theory
Compressor summary: The paper introduces Greedy Rejection Coding (GRC), a new algorithm for relative entropy coding that improves runtime and codelength for continuous distributions, and evaluates it on a compression pipeline with variational autoencoders.
Mathieu Molina, Nicolas Gast, Patrick Loiseau, Vianney Perchet
https://openreview.net/forum?id=KUBFYAPdqN
Keywords: Fairness, Online allocation, Bandits algorithms
Compressor summary: The paper proposes an algorithm that balances fairness and efficiency in online allocation problems by choosing data sources and allocating resources based on multi-armed bandit theory.
Steven Morad, Ryan Kortvelesy, Stephan Liwicki, Amanda Prorok
https://openreview.net/forum?id=KTfAtro6vP
Keywords: reinforcement learning, partially observable, POMDP, memory, rnn, transformer
Compressor summary: Fast and Forgetful Memory is a memory model for Reinforcement Learning that improves reward and training speed compared to recurrent neural networks, without changing hyperparameters or adding more resources.
Andrea Coletta, Sriram Gopalakrishnan, Daniel Borrajo, Svitlana Vyetrenko
https://openreview.net/forum?id=KTZttLZekH
Keywords: time-series, generative models, constrained optimization, machine learning
Compressor summary: The paper introduces novel methods for generating realistic constrained time series using a constrained optimization framework and a guided diffusion model, achieving better performance and efficiency than existing approaches.
Andreas Auer, Martin Gauch, Daniel Klotz, Sepp Hochreiter
https://openreview.net/forum?id=KTRwpWCMsC
Keywords: time series, uncertainty, prediction interval, conformal prediction, modern hopfield networks
Compressor summary: HopCPT is a novel conformal prediction method for time series that handles temporal dependencies and performs better than existing methods on various datasets.
Giorgio Giannone, Akash Srivastava, Ole Winther, Faez Ahmed
https://openreview.net/forum?id=KTR33hMnMX
Keywords: diffusion models, engineering design, generative optimization, trajectory matching
Compressor summary: Diffusion Optimization Models (DOM) align diffusion model sampling with physics-based optimization for efficient, feasible, and high-performance designs in constrained settings with limited data.
Rui Sun, Huayu Mai, Tianzhu Zhang, Feng Wu
https://openreview.net/forum?id=KRlG7NJUCD
Keywords: semi-supervised semantic segmentation
Compressor summary: The paper analyzes a trade-off in semi-supervised semantic segmentation methods when handling pseudo-labels and proposes Distribution-Aware Weighting (DAW) to improve the model's generalization performance.
Zhuoyan Luo, Yicheng Xiao, Yong Liu, Shuyan Li, Yitong Wang, Yansong Tang, Xiu Li, Yujiu Yang
https://openreview.net/forum?id=KQyXyIAfK8
Keywords: Referring Video Object Segmentation, Video-Level Multi-Modal Understanding, Object Cluster, Visual-Linguistic Contrastive Learning
Compressor summary: The paper proposes a new approach, Semantic-assisted Object Cluster (SOC), that improves referring video object segmentation by aggregating video content and textual guidance for unified temporal modeling and cross-modal alignment.
Thanh-Dat Truong, Hoang-Quan Nguyen, Bhiksha Raj, Khoa Luu
https://openreview.net/forum?id=KQ25VgEvOJ
Keywords: Fairness Continual Learning; Semantic Segmentation; Contrastive Clustering;
Compressor summary: The paper proposes a new fairness continual learning framework for semantic segmentation that addresses challenges like catastrophic forgetting and background shift using novel losses and achieves state-of-the-art results on three benchmarks.
Josh Alman, Zhao Song
https://openreview.net/forum?id=KOVWXcrFIK
Keywords: fast attention computation, algorithm, hardness
Compressor summary: The paper investigates faster algorithms for inner product attention computation using implicit use of the attention matrix A, and shows a sharp transition at matrix dimensions B = Θ(√log n).
Zichang Liu, Zhaozhuo Xu, Benjamin Coleman, Anshumali Shrivastava
https://openreview.net/forum?id=KMxRQO7P98
Keywords: Distribution sketch, federated learning
Compressor summary: The paper proposes a one-pass distribution sketch that measures data heterogeneity in federated learning, improving client selection and handling cold start problems.
Aleksandr Beznosikov, Sergey Samsonov, Marina Sheshukova, Alexander Gasnikov, Alexey Naumov, Eric Moulines
https://openreview.net/forum?id=KMeFZopsqP
Keywords: convex optimization, stochastic optimization, Markovian noise, acceleration, variational inequalities, lower bounds
Compressor summary: The paper proposes a unified approach for analyzing first-order gradient methods for stochastic optimization and variational inequalions with Markovian noise, eliminating some limiting assumptions and achieving optimal dependence on mixing time.
Laurence Illing Midgley, Vincent Stimper, Javier Antoran, Emile Mathieu, Bernhard Schölkopf, José Miguel Hernández-Lobato
https://openreview.net/forum?id=KKxO6wwx8p
Keywords: Boltzmann generator, normalizing flow
Compressor summary: Coupling normalizing flows with SE(3) and permutation equivariance are proposed for fast probabilistic modeling of physical systems, enabling efficient sampling and estimations.
Manuel Brack, Felix Friedrich, Dominik Hintersdorf, Lukas Struppek, Patrick Schramowski, Kristian Kersting
https://openreview.net/forum?id=KIPAIy329j
Keywords: diffusion, text-to-image, generation, semantics
Compressor summary: The paper introduces a method to interact with text-to-image diffusion models for semantic guidance, enabling users to control the image generation process more effectively.
Tangyu Jiang, Haodi Wang, Rongfang Bie
https://openreview.net/forum?id=KFm2lZiI7n
Keywords: Neural Architecture Search, Zero-Cost Proxy, Evaluation Strategy, Feature Map
Compressor summary: The paper proposes a novel zero-cost proxy called $\mathsf{MeCo}$ for neural architecture search, which uses the Pearson correlation matrix of feature maps and requires only one random data for a single forward pass.
Lucas Nunes Alegre, Ana L. C. Bazzan, Ann Nowe, Bruno Castro da Silva
https://openreview.net/forum?id=KFj0Q1EXvU
Keywords: generalized policy improvement, successor features, transfer learning, model-based reinforcement learning
Compressor summary: The paper presents a new method, h-GPI, for zero-shot transfer in reinforcement learning that combines model-free and model-based approaches and improves performance over existing methods.
Ta Duy Nguyen, Alina Ene, Huy Nguyen
https://openreview.net/forum?id=KF4LCXz8Np
Keywords: high probability, generalization, convex optimization, nonconvex optimization
Compressor summary: The authors propose a new method to analyze the generalization error of stochastic mirror descent for quadratically bounded losses and obtain improved bounds in various settings.
Sitan Chen, Sinho Chewi, Holden Lee, Yuanzhi Li, Jianfeng Lu, Adil Salim
https://openreview.net/forum?id=KD6MFeWSAd
Keywords: DDIM, deterministic samplers, diffusion models, predictor-corrector, probability flow ODE, score-based generative modeling
Compressor summary: The text describes a probabilistic flow ODE method for generative modeling that has polynomial-time convergence guarantees and outperforms DDPM in terms of dimension dependence.
Yichen Xie, Mingyu Ding, Masayoshi Tomizuka, Wei Zhan
https://openreview.net/forum?id=KBXcDAaZE7
Keywords: data selection, unsupervised learning
Compressor summary: The paper proposes FreeSel, an efficient data selection method that uses existing models to select data without training or supervision, achieving significant speedups compared to current active learning methods.
Chunting Zhou, Pengfei Liu, Puxin Xu, Srini Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, LILI YU, Susan Zhang, Gargi Ghosh, Mike Lewis, Luke Zettlemoyer, Omer Levy
https://openreview.net/forum?id=KBMOKmX2he
Keywords: large language models, supervised instruction fine-tuning, chat assistant
Compressor summary: The paragraph discusses LIMA, a language model fine-tuned with supervised loss on a small number of prompts and responses, which shows strong performance and generalization, suggesting that pretraining is the most important stage for large language models.
Vaishnav Potlapalli, Syed Waqas Zamir, Salman Khan, Fahad Khan
https://openreview.net/forum?id=KAlSIL4tXU
Keywords: Image Restoration
Compressor summary: The paper introduces PromptIR, a method that uses prompts to guide a restoration network and achieve high-quality image restoration from various degradation types and levels without prior knowledge of the corruptions.
Niv Giladi, Shahar Gottlieb, Moran Shkolnik, Asaf Karnieli, Ron Banner, Elad Hoffer, Kfir Yehuda Levy, Daniel Soudry
https://openreview.net/forum?id=KAWaeKOEkx
Keywords: distributed optimization, large-scale parallel SGD, synchronous training
Compressor summary: The paper analyzes how worker variability affects the scalability and robustness of synchronous distributed DNN training, and proposes a decentralized method to reduce this variation.
Haotao Wang, Ziyu Jiang, Yuning You, Yan Han, Gaowen Liu, Jayanth Srinivasa, Ramana Rao Kompella, Zhangyang Wang
https://openreview.net/forum?id=K9xHDD6mic
Keywords: graph neural networks
Compressor summary: The GMoE model improves GNNs by allowing nodes to select expert aggregators that capture different graph structures and information distances, leading to better performance on various graph tasks.
Junfeng Zuo, Xiao Liu, Ying Nian Wu, Si Wu, Wenhao Zhang
https://openreview.net/forum?id=K9dmkfZcMu
Keywords: temporal-scaling group, equivariant representation, disentangled representation, motor timing, continuous attractor networks
Compressor summary: The study proposes a recurrent circuit model that explains how the brain generates temporal sequences at different speeds, using control inputs and Lie group operators.
Chenlu Ye, Rui Yang, Quanquan Gu, Tong Zhang
https://openreview.net/forum?id=K9M7XNS9BX
Keywords: offline RL; adversarial corruption; general function approximation
Compressor summary: The paper proposes a new algorithm to solve offline RL problems under corruption and provides a suboptimality bound that depends on the corruption level and other factors, showing tightness for linear MDPs.
Antonio H. Ribeiro, Dave Zachariah, Francis Bach, Thomas B. Schön
https://openreview.net/forum?id=K8gLHZIgVW
Keywords: adversarial training; regularization; linear models
Compressor summary: Adversarial training for linear models can lead to minimum-norm interpolators, equivalent parameter shrinking methods, or independent of noise variance in some cases.
Jonathan Zedaka, Elisha Halperin, Evgeny Blaichman, Amit Berman
https://openreview.net/forum?id=K7u3RkoBP9
Keywords: WGAN, GAN, Autoencoder, Unsupervised Learning, Generative models, Flash Memory, NAND, Modulation, Reliability, Flash, Communication system
Compressor summary: The authors propose a machine learning approach that reduces errors and data degradation in NAND flash memory by using a neural modulator that adapts programming operations for each memory cell.
Mengyue Yang, Yonggang Zhang, Zhen Fang, Yali Du, Furui Liu, Jean-Francois Ton, Jianhong Wang, Jun Wang
https://openreview.net/forum?id=K5e5tFZuur
Keywords: OOD Generalization, Invariant Representation Learning
Compressor summary: The paper proposes a method to improve out-of-distribution generalization by using probability of sufficiency and necessary causes (PNS) to learn representations with high PNS value, which is theoretically analyzed and experimentally validated.
Lecheng Kong, Jiarui Feng, Hao Liu, Dacheng Tao, Yixin Chen, Muhan Zhang
https://openreview.net/forum?id=K4FK7I8Jnl
Keywords: Graph Neural Network; Reinforcement Learning
Compressor summary: MAG-GNN is a reinforcement learning boosted graph neural network that uses combinatorial optimization to find the optimal subset of subgraphs for efficient and expressive graph learning tasks.
Hilal Asi, Vitaly Feldman, Jelani Nelson, Huy Nguyen, Kunal Talwar
https://openreview.net/forum?id=K3JgUvDSYX
Keywords: Differential Privacy, mean estimation, private federated learning, communication complexity
Compressor summary: The paper proposes ProjUnit, a simple and efficient algorithmic framework for private mean estimation of high-dimensional vectors with low communication and run-time complexity.
Zeju Qiu, Weiyang Liu, Haiwen Feng, Yuxuan Xue, Yao Feng, Zhen Liu, Dan Zhang, Adrian Weller, Bernhard Schölkopf
https://openreview.net/forum?id=K30wTdIIYc
Keywords: Text-to-image, diffusion models, finetuning, generative models, orthogonality
Compressor summary: The Orthogonal Finetuning (OFT) method helps adapt text-to-image diffusion models for different tasks by preserving their semantic generation ability and improving finetuning stability.
Siyu Jiao, Yunchao Wei, Yaowei Wang, Yao Zhao, Humphrey Shi
https://openreview.net/forum?id=K1Uzj8tuwd
Keywords: Zero-Shot Segmentation, Open-Vocabulary Segmentation, Fine-tuning
Compressor summary: The paper proposes Mask-aware Fine-tuning (MAFT), a method to improve zero-shot segmentation by making CLIP more responsive to different mask proposals using a new encoder and two types of loss functions.
Jui-Nan Yen, Sai Surya Duvvuri, Inderjit S Dhillon, Cho-Jui Hsieh
https://openreview.net/forum?id=JzQlGqBm8d
Keywords: Second Order Optimization, Optimization for deep networks
Compressor summary: The paper proposes a low-rank approximation method for preconditioning in second-order optimization that reduces time and memory complexity while maintaining performance on deep neural networks.
Dibyadip Chatterjee, Fadime Sener, Shugao Ma, Angela Yao
https://openreview.net/forum?id=JzQ7QClAdF
Keywords: video understanding, egocentric videos, open vocabulary
Compressor summary: The paper introduces a new action recognition task that can handle verbs and objects not seen during training using an object-agnostic verb encoder and a prompt-based object encoder based on CLIP representations.
Xiaoyan Hu, Ho-fung Leung
https://openreview.net/forum?id=JwNXeBdkeo
Keywords: Learning with Options, Offline RL, Provably Efficient RL
Compressor summary: The paper studies how using options in offline reinforcement learning can improve sample efficiency and provide a novel information-theoretic lower bound for this scenario. It also proposes an algorithm (PEVIO) and analyzes two data-collection methods.
Chaoqi Chen, Luyao Tang, Yue Huang, Xiaoguang Han, Yizhou Yu
https://openreview.net/forum?id=Jw0KRTjsGA
Keywords: Domain generalization, domain shift, open class, source compaction, target disambiguation
Compressor summary: CODA is a novel framework for learning compact representations and adapting to open classes in machine learning, addressing domain shift and unseen categories by introducing virtual unknown classes and optimizing a new training objective.
Qingkai Fang, Yan Zhou, Yang Feng
https://openreview.net/forum?id=JvYSSPtQyk
Keywords: speech-to-speech translation, non-autoregressive translation, speech translation, directed acyclic transformer
Compressor summary: DASpeech is a non-autoregressive direct speech-to-speech translation model that uses a two-pass architecture, achieving fast and high-quality translations while preserving the source speaker's voice.
Ido Greenberg, Shie Mannor, Gal Chechik, Eli Meirom
https://openreview.net/forum?id=JvOZ4IIjwP
Keywords: meta reinforcement learning, robust reinforcement learning, safe reinforcement learning, risk sensitive reinforcement learning
Compressor summary: The paper proposes Robust Meta RL (RoML), a meta-algorithm that improves the performance of standard Meta RL methods in high-risk or difficult tasks by over-sampling them, thus increasing system reliability.
Etienne Boursier, Nicolas Flammarion
https://openreview.net/forum?id=JtIqG47DAQ
Keywords: Neural networks, Min norm interpolators, Sparsity, Representational cost
Compressor summary: This paper investigates how controlling the norm of neural network parameters affects generalization and shows that using a specific weighting factor ensures unique and sparse solutions for one hidden ReLU layer networks with unidimensional data.
Yang Li, Jinpei Guo, Runzhong Wang, Junchi Yan
https://openreview.net/forum?id=JtF0ugNMv2
Keywords: Machine Learning, Combinatorial Optimization, Generative Modeling, Diffusion Model
Compressor summary: The paragraph describes a new framework called T2TCO that improves Combinatorial Optimization problem solving by using generative modeling and gradient-based search, achieving significant performance gains on TSP and MIS problems.
Xiaolei Ru, Xin-Ya Zhang, Zijia Liu, Jack Murdoch Moore, Gang Yan
https://openreview.net/forum?id=JpU5YmMKx7
Keywords: Directed coupled network reconstruction; Neuronal dynamics; Mutual information estimator; Attention mechanism; Transfer entropy.
Compressor summary: The text describes a novel method to reconstruct coupled biological neural networks from time series data using attention mechanism and Attentive Transfer Entropy, which can identify critical regions of strong coupling-drive.
Kexun Zhang, Danqing Wang, Jingtao Xia, William Yang Wang, Lei Li
https://openreview.net/forum?id=JolrEmMim6
Keywords: Large Language Models, Code Generation, Code Intelligence, Automatic Verification
Compressor summary: The proposed ALGO framework synthesizes algorithms with LLM-generated oracles to guide code generation and verify correctness, achieving significantly improved one-submission pass rates on algorithmic problems.
Aaron Zweig, Loucas Pillaud-Vivien, Joan Bruna
https://openreview.net/forum?id=JkmvrheMe7
Keywords: gradient descent, shallow neural networks
Compressor summary: This paper studies how gradient descent methods can learn non-linear features from data using single-index models, and shows that it works well for non-Gaussian data as well.
Youzhi Luo, Chengkai Liu, Shuiwang Ji
https://openreview.net/forum?id=Jkc74vn1aZ
Keywords: material generation, symmetries, variational auto-encoder, score-based diffusion model
Compressor summary: SyMat is a new method for generating periodic materials with different symmetries using deep models, which can create atom types, lattices, and coordinates that are invariant to various symmetry transformations.
Chen Liang, Jiahui Yu, Ming-Hsuan Yang, Matthew Brown, Yin Cui, Tuo Zhao, Boqing Gong, Tianyi Zhou
https://openreview.net/forum?id=JhQP33aMx2
Keywords: Multimodality foundation models, knowledge distillation
Compressor summary: The proposed method uses a modified-Thompson sampling algorithm to selectively distill different modules of a multimodal foundation model based on their contributions to reduce its size and maintain performance.
Shivam Gupta, Jasper C.H. Lee, Eric Price, Paul Valiant
https://openreview.net/forum?id=JeKXmYb4kd
Keywords: location estimation, minimax estimation
Compressor summary: The paper proposes two optimal location estimators for parametric statistics with different criteria, one minimizing estimation error and success probability, and another minimizing expected squared interval width.
Allan Zhou, Kaien Yang, Yiding Jiang, Kaylee Burns, Winnie Xu, Samuel Sokota, J Zico Kolter, Chelsea Finn
https://openreview.net/forum?id=JdhyIa0azI
Keywords: equivariance, permutation, implicit neural representation, generalization, transformers, attention
Compressor summary: The paper proposes neural functional Transformers (NFTs) that use attention mechanisms to process neural networks' weights and achieve better performance in weight-space methods, enabling improved classification of implicit neural representations (INRs).
Zichang Liu, Aditya Desai, Fangshuo Liao, Weitao Wang, Victor Xie, Zhaozhuo Xu, Anastasios Kyrillidis, Anshumali Shrivastava
https://openreview.net/forum?id=JZfg6wGi6g
Keywords: Large language model; KV Cache Compression
Compressor summary: Scissorhands is a system that reduces memory usage and inference time for large language models by storing pivotal tokens more frequently.
Duo Peng, Li Xu, Qiuhong Ke, Ping Hu, Jun Liu
https://openreview.net/forum?id=JYUN0vYjh9
Keywords: Privacy Preservation, Action Recognition, Meta-Learning
Compressor summary: The paper proposes MPPAR, a meta-learning framework for privacy-preserving action recognition that can adapt to novel privacy attributes and attack models.
Jaemin Na, Jung-Woo Ha, Hyung Jin Chang, Dongyoon Han, Wonjun Hwang
https://openreview.net/forum?id=JXvszuOqY3
Keywords: Semi-supervised Learning, Semantic Segmentation
Compressor summary: The paper proposes Dual Teacher, a method to improve semi-supervised semantic segmentation by using two temporary teachers that shift roles and prevent the student model from losing its unique features.
Suyoung Lee, Myungsik Cho, Youngchul Sung
https://openreview.net/forum?id=JX6UloWrmE
Keywords: Deep reinforcement learning, Meta-reinforcement learning, Subtask decomposition
Compressor summary: SDVT is a novel meta-RL approach that decomposes non-parametric tasks into subtasks, uses a Gaussian mixture VAE to meta-learn the decomposition, and employs virtual training for better generalization.
Jiazheng Xu, Xiao Liu, Yuchen Wu, Yuxuan Tong, Qinkai Li, Ming Ding, Jie Tang, Yuxiao Dong
https://openreview.net/forum?id=JVzeOYEx6d
Keywords: Generative Models, Text-to-Image, Learning from Human Feedback, Multimodality, Evaluation
Compressor summary: The authors introduce ImageReward, a new model that encodes human preferences for text-to-image synthesis, and ReFL, an algorithm to optimize diffusion models using ImageReward.
Zhiqing Sun, Yiming Yang
https://openreview.net/forum?id=JV8Ff0lgVV
Keywords: neural-symbolic reasoning, combinatorial optimization, diffusion models
Compressor summary: The paper introduces DIFUSCO, a graph-based diffusion framework that generates high-quality solutions for NP-complete problems using neural networks and outperforms previous state-of-the-art methods.
LILI YU, Daniel Simig, Colin Flaherty, Armen Aghajanyan, Luke Zettlemoyer, Mike Lewis
https://openreview.net/forum?id=JTmO2V9Xpz
Keywords: byte level language model, model architecture, efficient pretraining
Compressor summary: Megabyte is a multi-scale decoder architecture that enables efficient and large-scale tokenization-free autoregressive sequence modeling for various domains such as language, image, and audio.
Jules Berman, Benjamin Peherstorfer
https://openreview.net/forum?id=JTKd7zYROf
Keywords: numerical methods, deep networks, evolution equations, scientific computing, partial differential equations, model reduction
Compressor summary: The paper proposes Neural Galerkin schemes that use randomized sparse updates to train neural networks for time-dependent partial differential equations, improving accuracy and speed over dense updates.
Alberto Silvio Chiappa, Alessandro Marin Vargas, Ann Huang, Alexander Mathis
https://openreview.net/forum?id=JSVXZKqfLU
Keywords: Reinforcement learning, efficient exploration, curse of dimensionality, motor control, musculoskeletal control
Compressor summary: Lattice introduces temporally-correlated noise into policy networks to improve exploration and control in overactuated systems, achieving state of the art results on locomotion and musculoskeletal tasks.
Felix Biggs, Antonin Schrab, Arthur Gretton
https://openreview.net/forum?id=JOkgEY9os2
Keywords: Testing, MMD, Kernel Methods, Two-sample testing
Compressor summary: The paper introduces new statistics that improve the ability to test if two samples come from the same distribution using a modified version of the Maximum Mean Discrepancy (MMD) test, and shows their effectiveness on both synthetic and real data.
Phil Pope, David Jacobs
https://openreview.net/forum?id=JOHp5SmckS
Keywords: graph neural networks, equivariance, materials science, chemistry, density functional theory, combinatorial generalization, catalysts
Compressor summary: The authors propose using pointwise learning of the Kohn-Sham charge-density to model catalysts, which shows improved convergence and combinatorial generalization compared to energy prediction.
Ammar Fayad, Majd Ibrahim
https://openreview.net/forum?id=JMuKfZx2xU
Keywords: Mutual information, Information Theory
Compressor summary: The paper proposes a framework to find the optimal distribution of slices for measuring mutual information between high-dimensional variables, improving on existing methods that use uniform distributions and discard informative features.
Ryan Kortvelesy, Steven Morad, Amanda Prorok
https://openreview.net/forum?id=JMrIeKjTAe
Keywords: Aggregation, Graph Neural Networks
Compressor summary: The paper introduces GenAgg, a generalised aggregation operator for graph neural networks that improves performance by parametrising a function space including all standard aggregators.
Alessandro Conti, Enrico Fini, Massimiliano Mancini, Paolo Rota, Yiming Wang, Elisa Ricci
https://openreview.net/forum?id=JKhyQHpx7B
Keywords: language and vision, zero-shot classification, image classification
Compressor summary: The paragraph introduces a new task called Vocabulary-free Image Classification (VIC), which aims to classify images without knowing the categories beforehand, and proposes CaSED, a method that uses a pre-trained vision-language model and an external database to achieve this goal.
Jingzhou Hu, Kejun Huang
https://openreview.net/forum?id=JK2oPrP8B3
Keywords: dictionary learning, matrix volume, nonconvex optimization
Compressor summary: The authors propose a novel dictionary learning method that ensures global identifiability of the groundtruth matrices by minimizing the volume of the dictionary matrix while maintaining unit $\ell_1$ norm for each row of the sparse coefficient matrix, without requiring mutual incoherence of dictionary atoms.
Moses Charikar, Monika Henzinger, Lunjia Hu, Maximilian Vötsch, Erik Waingarten
https://openreview.net/forum?id=JIYdbHDonF
Keywords: clustering, k-means, random projection, massive datasets
Compressor summary: A simple randomized clustering algorithm with improved running time and cluster quality compared to popular methods is introduced for data analysis tasks such as k-means clustering and coreset construction.
Weijia Wu, Yuzhong Zhao, Hao Chen, Yuchao Gu, Rui Zhao, Yefei He, Hong Zhou, Mike Zheng Shou, Chunhua Shen
https://openreview.net/forum?id=JIKM2vS8XU
Keywords: Diffusion Model; Text-guided dataset generation
Compressor summary: DatasetDM is a model that generates synthetic images and perception annotations using a diffusion model, enabling efficient training of various perception models on downstream tasks with minimal effort and cost.
Jun-Yi Hang, Min-Ling Zhang
https://openreview.net/forum?id=JDw50IX4TY
Keywords: Machine learning, multi-label learning, partial multi-label learning, label disambiguation
Compressor summary: The paper proposes a probabilistic graphical model for partial multi-label learning, which uses stochastic gradient variational Bayes to infer ground-truth labeling information from inaccurate annotations and performs better than existing methods.
Ambar Pal, Jeremias Sulam, Rene Vidal
https://openreview.net/forum?id=JDoA6admhv
Keywords: Adversarial Robustness, Geometry in Data, Low Dimensional Modeling
Compressor summary: The authors explore whether adversarial examples are unavoidable by analyzing how data distribution concentration affects the existence of robust machine learning classifiers.
Alon Albalak, Colin Raffel, William Yang Wang
https://openreview.net/forum?id=JDnLXc4NOn
Keywords: Few-shot learning, natural language processing, few shot learning, NLP, multi-armed bandit, multi armed bandit
Compressor summary: FLAD is a method to improve few-shot learning with auxiliary data, and this paper introduces two algorithms that explore and exploit different datasets, achieving better results than previous methods and enabling larger language models.
Yue Li, Yueyi Zhang, Juntian Ye, Feihu Xu, Zhiwei Xiong
https://openreview.net/forum?id=JCN9YsZiwB
Keywords: Non-line-of-sight imaging, Transient Recovery, Volume Reconstruction
Compressor summary: The paragraph describes a new deep learning approach to non-line-of-sight imaging that uses under-scanning measurements to reconstruct hidden volumes with high resolution and robustness, outperforming traditional methods.
Wenlong Huang, Fei Xia, Dhruv Shah, Danny Driess, Andy Zeng, Yao Lu, Pete Florence, Igor Mordatch, Sergey Levine, Karol Hausman, brian ichter
https://openreview.net/forum?id=JCCi58IUsh
Keywords: robotics, language models, embodied agents
Compressor summary: The paragraph discusses the challenge of combining language models' semantic knowledge with grounded robots' real-world understanding and proposes a solution that decodes action sequences using both types of models.
Baohao Liao, Shaomu Tan, Christof Monz
https://openreview.net/forum?id=J8McuwS3zY
Keywords: large language model, parameter-efficient learning, memory-efficient learning, reversible neural network
Compressor summary: MEFT is a memory-efficient fine-tuning method that uses adapters to insert reversibility into PLMs, preserving their starting point and achieving comparable performance to full fine-tuning with much less activation memory.
Xinyin Ma, Gongfan Fang, Xinchao Wang
https://openreview.net/forum?id=J8Ajf9WfXP
Keywords: model compression, structural pruning, large language model
Compressor summary: The paper introduces LLM-pruner, a method to compress large language models without relying on their original training data, by selectively removing non-critical structures based on gradient information.
Thomas FEL, Thibaut Boissin, Victor Boutin, Agustin Martin Picard, Paul Novello, Julien Colin, Drew Linsley, Tom ROUSSEAU, Remi Cadene, Lore Goetschalckx, Laurent Gardes, Thomas Serre
https://openreview.net/forum?id=J7VoDuzuKs
Keywords: explainable AI, feature visualization, interpretability, optimization
Compressor summary: MACO is a method for improving feature visualization in neural networks by optimizing image phase spectrum and introducing three metrics for evaluation.
Ziyi Bai, Ruiping Wang, Xilin CHEN
https://openreview.net/forum?id=J6Niv3yrMq
Keywords: Video Question Answering; Multi-Event Reasoning; Spatial-Temporal Reasoning
Compressor summary: The Glance- Focus model uses dynamic event memories to improve Video Question Answering by mimicking human reasoning strategies, achieving state-of-the-art results on four benchmarks.
Ethan Nicholas Epperly, Elvira Moreno Ferreira
https://openreview.net/forum?id=J66ptjMkAG
Keywords: kernel quadrature, Nyström approximation, reproducing kernel Hilbert space, randomly pivoted Cholesky
Compressor summary: The paper introduces a fast and accurate kernel quadrature method using randomly pivoted Choleskey nodes that works well with complex geometries and kernels.
Taira Tsuchiya, Shinji Ito, Junya Honda
https://openreview.net/forum?id=J3taqrzyyA
Keywords: follow-the-regularized-leader, adaptive learning rate, multi-armed bandits, partial monitoring, data-dependent bound, sparsity, game-dependency, best-of-both-worlds
Compressor summary: The paper proposes a new adaptive learning rate for FTRL called SPA, which enables algorithms with three types of adaptivity in sequential decision-making problems, including sparsity, game-dependency, and best-of-both-worlds.
Liunian Harold Li, Zi-Yi Dou, Nanyun Peng, Kai-Wei Chang
https://openreview.net/forum?id=J2Cso0wWZX
Keywords: Vision language, fine-grained recognition, object detection
Compressor summary: The text describes a new vision-language model for object detection that uses rich language descriptions and context-sensitive queries to improve accuracy and adaptability.
Paul Pu Liang, Yun Cheng, Xiang Fan, Chun Kai Ling, Suzanne Nie, Richard J. Chen, Zihao Deng, Nicholas Allen, Randy Auerbach, Faisal Mahmood, Ruslan Salakhutdinov, Louis-Philippe Morency
https://openreview.net/forum?id=J1gBijopla
Keywords: multimodal learning, feature interactions, partial information decomposition, information theory, quantification, model selection
Compressor summary: The paragraph discusses an information-theoretic approach called PID to measure interactions between input modalities and output tasks in multimodal applications, and its usefulness in various aspects of multimodal modeling.
Ron Levie
https://openreview.net/forum?id=J0RD92Tmfc
Keywords: graph neural network, graphon, generalization, stability, sampling, Szemerédi regularity lemma
Compressor summary: The paper proposes a new similarity measure for graph-signals in message passing graph neural networks (MPNNs) and shows how it leads to Lipschitz continuity and generalization bounds for MPNNs.
Jingyang Xiang, Siqi Li, Jun Chen, Guang Dai, Shipeng Bai, Yukai Ma, Yong Liu
https://openreview.net/forum?id=J0Pvvxspmz
Keywords: Soft Uniform Block Pruning, Block Angular Redundancy, Hardware Acceleration
Compressor summary: The paper proposes a new method to train sparse neural networks that reduces storage, improves performance, and balances workload across threads.
Adam Block, Max Simchowitz, Russ Tedrake
https://openreview.net/forum?id=Izt7rDD7jN
Keywords: Smoothed Online Learning, Piecewise Affine Prediction, Learning Dynamics
Compressor summary: The paper presents polynomial-regret algorithms for prediction and simulation in piecewise affine systems using smoothed online learning and introduces new technical tools.
Parikshit Gopalan, Michael P. Kim, Omer Reingold
https://openreview.net/forum?id=IzlRh5qwmG
Keywords: Agnostic Learning, Omniprediction, Multicalibration
Compressor summary: Swap Agnostic Learning is a game where a predictor selects a hypothesis and an adversary tries to minimize the loss, and the paper shows how it relates to other concepts like Omniprediction and Multicalibration.
Tao Lei, Junwen Bai, Siddhartha Brahma, Joshua Ainslie, Kenton Lee, Yanqi Zhou, Nan Du, Vincent Y Zhao, Yuexin Wu, Bo Li, Yu Zhang, Ming-Wei Chang
https://openreview.net/forum?id=IyYyKov0Aj
Keywords: conditional computation, inference efficiency, parameter efficiency, large models
Compressor summary: CoDA is a transfer learning method that boosts inference efficiency while maintaining accuracy and parameter efficiency on various tasks.
Weijie Gan, Shirin Shoushtari, Yuyang Hu, Jiaming Liu, Hongyu An, Ulugbek Kamilov
https://openreview.net/forum?id=IyWpP2e0bF
Keywords: inverse problems, plug-and-play priors, computational imaging, nonconvex optimization
Compressor summary: BC-PnP is a new method that uses learned denoisers to solve blind imaging inverse problems by combining physical measurements and image estimation.
Francesco Montagna, Atalanti A. Mastakouri, Elias Eulig, Nicoletta Noceti, Lorenzo Rosasco, Dominik Janzing, Bryon Aragam, Francesco Locatello
https://openreview.net/forum?id=IyTArtpuCK
Keywords: Causal discovery; empirical study; robust inference; benchmark
Compressor summary: The paper benchmarks various causal discovery algorithms on observational data with different conditions, finding that score matching-based methods perform well and providing theoretical explanations.
Vinay Shukla, Zhe Zeng, Kareem Ahmed, Guy Van den Broeck
https://openreview.net/forum?id=IyAHCbMq3a
Keywords: weakly supervised learning, constraint, label proportion, learning from positive and unlabeled data, multiple instance learning
Compressor summary: The paper proposes a method called count-based weakly-supervised learning that learns from data with inferred weak labels by computing and penalizing deviations from expected label counts.
Austin Xu, Andrew McRae, Jingyan Wang, Mark A. Davenport, Ashwin Pananjady
https://openreview.net/forum?id=IwyymRXfzL
Keywords: human querying, high dimensional low rank matrix estimation, metric learning
Compressor summary: The paper introduces PAQ, a query mechanism that combines cardinal and ordinal feedback to learn an unknown Mahalanobis distance using a two-stage estimator with sample complexity guarantees.
Ahmed Alaa, Zaid Ahmad, Mark van der Laan
https://openreview.net/forum?id=IwnINorSZ5
Keywords: Heterogeneous treatment effects, conformal prediction
Compressor summary: The paper proposes a method to estimate uncertainty in individual treatment effects using machine learning and conformal prediction techniques, showing that it is valid under certain conditions and performs well in simulations.
Qiyang Li, Jason Zhang, Dibya Ghosh, Amy Zhang, Sergey Levine
https://openreview.net/forum?id=Itorzn4Kwf
Keywords: Reinforcement Learning, Exploration
Compressor summary: The paper proposes a method to improve exploration in sparse reward tasks by using prior data without reward labels to guide an agent, leading to better performance in challenging domains.
Jimmy Z. Di, Jack Douglas, Jayadev Acharya, Gautam Kamath, Ayush Sekhari
https://openreview.net/forum?id=Isy7gl1Hqc
Keywords: Machine unlearning, new attack vector, Camouflaging poisoning attacks
Compressor summary: The paper introduces a new data poisoning attack that uses camouflaged points to affect model predictions when some points are removed during model retraining or unlearning.
Daniil Tiapkin, Denis Belomestny, Daniele Calandriello, Eric Moulines, Remi Munos, Alexey Naumov, pierre perrault, Michal Valko, Pierre MENARD
https://openreview.net/forum?id=IrjXmIKFyx
Keywords: reinforcement learning, exploration, q-learning
Compressor summary: Randomized Q-learning (RandQL) is a new algorithm for minimizing regret in episodic MDPs that uses random learning rates and achieves better performance than existing methods.
Chirag Raman, Alec Nonnemaker, Amelia Villegas-Morcillo, Hayley Hung, Marco Loog
https://openreview.net/forum?id=IrEYkhuxup
Keywords: Probabilistic Forecasting, Saliency, Explainability, XAI, Probabilistic Regression
Compressor summary: The authors propose a method to explain probabilistic time-series forecasts by identifying salient timesteps based on human visual cognition and test it on synthetic and real data.
Walter Gerych, Kevin Hickey, Luke Buquicchio, Kavin Chandrasekaran, Abdulaziz Alajaji, Elke Rundensteiner, Emmanuel Agu
https://openreview.net/forum?id=Iq7v0sZw2H
Keywords: generative models, generative modeling, bias, GANs, debiasing
Compressor summary: The distribution mapping module proposes a way to debias generative models without retraining them by using fair noise distribution, and it works better than current methods.
Stella Biderman, USVSN Sai Prashanth, Lintang Sutawika, Hailey Schoelkopf, Quentin Gregory Anthony, Shivanshu Purohit, Edward Raff
https://openreview.net/forum?id=Iq0DvhB4Kf
Keywords: large language model, emergent properties, memorization
Compressor summary: The paper proposes a method to predict memorization in language models using lower-compute trials, and provides insights on memorization behavior and recommendations for reliability.
Rachel Emily Redberg, Antti Koskela, Yu-Xiang Wang
https://openreview.net/forum?id=IpUJd3KG3c
Keywords: differential privacy, empirical risk minimization, objective perturbation
Compressor summary: The paper improves the objective perturbation mechanism for privacy-preserving machine learning, making it more efficient and competitive with differentially private stochastic gradient descent.
Mayee F Chen, Nicholas Roberts, Kush Bhatia, Jue WANG, Ce Zhang, Frederic Sala, Christopher Re
https://openreview.net/forum?id=IoizwO1NLf
Keywords: language models, data selection
Compressor summary: The authors propose a framework for selecting and ordering training data based on language model skills, which improves performance on downstream tasks with fewer tokens.
Songtao Lu
https://openreview.net/forum?id=IobxuwPnWt
Keywords: Functional constrained optimization, bilevel optimization, primal dual method, Lagrangian method
Compressor summary: The paper proposes a smoothed first-order Lagrangian method for solving nonconvex functional constrained optimization problems with nonconvex constraints, and shows its effectiveness through theoretical convergence guarantees and numerical results.
Devon R. Graham, Kevin Leyton-Brown, Tim Roughgarden
https://openreview.net/forum?id=InB9Loet1u
Keywords: algorithm configuration, algorithm selection, data-driven algorithm design, utility of runtime
Compressor summary: The authors propose a new method for configuring heuristic algorithms that maximizes utility and offers theoretical guarantees about performance, unlike existing methods that only minimize expected runtime.
Meena Jagadeesan, Michael Jordan, Jacob Steinhardt, Nika Haghtalab
https://openreview.net/forum?id=IltQ87ZdT6
Keywords: competition, equilibria, inverse scaling, digital marketplaces
Compressor summary: Increasing the scale of machine learning models can hurt overall predictive accuracy when multiple providers compete for users, as improving data representation quality may reduce social welfare.
Junqi Gao, Biqing Qi, Yao Li, Zhichang Guo, Dong Li, Yuming Xing, Dazhi Zhang
https://openreview.net/forum?id=IkD1EWFF8c
Keywords: Adversarial Attacks; Generative Attack; Transferable Targeted Attack
Compressor summary: The paper proposes ESMA, a generative targeted attack strategy that improves transferability by adding perturbations to easy samples in the target class, reducing storage space and computation time compared to current methods.
Jungwuk Park, Dong-Jun Han, Jinho Kim, Shiqiang Wang, Christopher Brinton, Jaekyun Moon
https://openreview.net/forum?id=IjZa2fQ8tL
Keywords: Federated Learning, Domain Generalization
Compressor summary: The paper proposes StableFDG, a learning strategy for federated domain generalization, which uses style and attention-based methods to improve domain diversity and learn domain-invariant characteristics in data-poor scenarios.
Xin Zou, Weiwei Liu
https://openreview.net/forum?id=IiwTFcGGTq
Keywords: Adversarial Robustness, Out-of-distribution Generalization
Compressor summary: The text discusses out-of-distribution generalization, its vulnerability to attacks, and proposes methods to improve it.
Charles Thomas Marx, Sofian Zalouk, Stefano Ermon
https://openreview.net/forum?id=IhxD94i5ra
Keywords: Uncertainty Quantification, Calibration, Decision Making, Probabilistic Forecasting
Compressor summary: The authors propose kernel-based calibration metrics for both classification and regression that improve predictive uncertainty, sharpness, and decision-making by incorporating them into empirical risk minimization.
Shu Tew, Mario Boley, Daniel F. Schmidt
https://openreview.net/forum?id=Ih2yL7o2Gq
Keywords: Ridge Regression, Cross validation, Expectation Maximisation, Bayesian methods
Compressor summary: A faster and better method for tuning ridge regression hyper-parameter $\lambda$ using Bayesian formulation and iterative EM procedure is proposed.
Zakaria Mhammedi, Adam Block, Dylan J Foster, Alexander Rakhlin
https://openreview.net/forum?id=IgDa5Ynm9l
Keywords: Reinforcement learning, Representation Learning, Low-rank MDPs, Model-Free Learning
Compressor summary: The paper proposes SpanRL, a sample-efficient exploration algorithm for high-dimensional reinforcement learning that works without structural assumptions and uses barycentric spanners for efficient exploration.
Felipe Maia Polo, Yuekai Sun, Moulinath Banerjee
https://openreview.net/forum?id=Ifq8GMdqJK
Keywords: conditional independence, hypothesis testing, misspecification
Compressor summary: The paper studies how regression-based tests for conditional independence can fail due to misspecified models or algorithms, and proposes a new test (RBPT) that is robust against such failures.
Zihan Zhu, Ethan X Fang, Zhuoran Yang
https://openreview.net/forum?id=IdF7VT6eEs
Keywords: Performative Prediction, Nash Equilibrium, Reproducing Kernel Hilbert Space, Online Learning, Stochastic Gradient Methods
Compressor summary: The paper proposes a novel online algorithm (OPGD) to find the Nash equilibrium of decision-dependent games in the bandit feedback setting, where agents' actions affect population data and traditional gradient-based methods are infefficient without a gradient oracle.
Lijun Zhang, Peng Zhao, Zhenhua Zhuang, Tianbao Yang, Zhi-Hua Zhou
https://openreview.net/forum?id=IcIQbCWoFj
Keywords: Group distributionally robust optimization, Stochastic mirror descent, Non-oblivious online learning, Sample complexity, Stochastic mirror-prox algorithm, Mini-batch
Compressor summary: The paper studies group distributionally robust optimization (GDRO) and proposes methods to reduce sample complexity using techniques from online learning and non-uniform sampling.
Swarnadeep Saha, Peter Hase, Mohit Bansal
https://openreview.net/forum?id=IacxcFpvWQ
Keywords: Language Models, Reasoning, Explanations
Compressor summary: The study explores how to make LLMs good teachers for weaker agents by using natural language explanations and a communication budget. It shows that teacher LLMs can improve student performance by intervening at the right time, personalizing explanations, and avoiding misinformation.
Alfredo De Goyeneche, Shreya Ramachandran, Ke Wang, Ekin Karasan, Joseph Yitan Cheng, Stella X. Yu, Michael Lustig
https://openreview.net/forum?id=Ia4dmqst0Z
Keywords: Inverse problem, MRI, Medical Imaging, Computational Imaging, Deep Learning, Off-Resonance
Compressor summary: The paper proposes a physics-informed deep learning framework for correcting off-resonance artifacts in MRI using synthetic data, which could enable faster and more accurate scans with non-Cartesian sampling trajectories.
Zeshuai Deng, Zhuokun Chen, Shuaicheng Niu, Thomas H. Li, Bohan Zhuang, Mingkui Tan
https://openreview.net/forum?id=IZRlMABK4l
Keywords: Image Super-resolution, Test-time Adaptation, Self-supervised Learning, Second-Order Degradation
Compressor summary: The authors propose a test-time adaptation framework for image super-resolution, called SRTTA, that can handle different degradation types by predicting the degradation type and adapting the model using feature-level reconstruction learning.
Arash Ahmadian, Saurabh Dash, Hongyu Chen, Bharat Venkitesh, Zhen Stephen Gou, Phil Blunsom, Ahmet Üstün, Sara Hooker
https://openreview.net/forum?id=IYe8j7Gy8f
Keywords: Quantization, optimization, language modelling, efficiency
Compressor summary: This paper investigates if quantization cliffs, which cause performance drops in large models, are solely a factor of scale and proposes a method to optimize for efficient quantization by suppressing outlier dimensions.
Ido Greenberg, Netanel Yannay, Shie Mannor
https://openreview.net/forum?id=IXWaWPkGke
Keywords: non-linear filtering, Kalman filter, noise estimation, optimization, Cholesky parameterization
Compressor summary: The Optimized Kalman Filter (OKF) improves the performance of the standard Kalman Filter by optimizing both its architecture and parameters, making it competitive with non-linear models like neural networks.
Weizhe Lin, Jinghong Chen, Jingbiao Mei, Alexandru Coca, Bill Byrne
https://openreview.net/forum?id=IWWWulAX7g
Keywords: knowledge-based visual question answering, knowledge retrieval, multi-modality, vision-and-language
Compressor summary: FLMR improves knowledge retrieval in RA-VQA by using finer-grained image and question embeddings and complementary image representations from a vision model.
Rylan Schaeffer, Brando Miranda, Sanmi Koyejo
https://openreview.net/forum?id=ITw9edRDlD
Keywords: large language models, foundation models, natural language processing, language modeling, emergent abilities
Compressor summary: The paper suggests that emergent abilities in large language models are due to the choice of metric rather than scale, and provides evidence for this claim through various analyses and experiments.
Jingfeng Wu, Vladimir Braverman, Jason D. Lee
https://openreview.net/forum?id=IT9mWLYNpQ
Keywords: gd; implicit bias; edge of stability
Compressor summary: This paper studies how gradient descent with constant stepsize behaves on linearly separable data with non-monotonic losses, proving its ability to minimize logistic loss and revealing its divergence under exponential loss in the edge of stability regime.
A. Feder Cooper, Wentao Guo, Khiem Pham, Tiancheng Yuan, Charlie F. Ruan, Yucheng Lu, Christopher De Sa
https://openreview.net/forum?id=ISRyILhAyS
Keywords: permuted example ordering, distributed training, scalable training, herding
Compressor summary: CD-GraB is a new method that improves the speed of distributed machine learning by using example ordering based on stale gradients, leading to faster convergence than existing methods.
Jiangxing Wang, Deheng Ye, Zongqing Lu
https://openreview.net/forum?id=IQRc3FrYOG
Keywords: Multi-Agent Reinforcement Learning
Compressor summary: The text proposes using mutual information as an augmented reward to help multi-agent reinforcement learning algorithms adapt to changing team compositions and perform well in complex cooperative tasks.
Hedi Hadiji, Sarah Sachs, Tim van Erven, Wouter M Koolen
https://openreview.net/forum?id=IPNg84RF1k
Keywords: game theory, minimax optimization, lower bounds
Compressor summary: The paragraph discusses the query complexity of learning equilibria in zero-sum matrix games, presenting new lower bounds and upper bounds for different values of K and epsilon, and proposing future research directions.
Han Liu, Zhi Xu, Xiaotong Zhang, Feng Zhang, Fenglong Ma, Hongyang Chen, Hong Yu, Xianchao Zhang
https://openreview.net/forum?id=IOuuLBrGJR
Keywords: High-quality adversarial example, Black-box hard-label textual adversarial attack
Compressor summary: HQA-Attack is a simple framework for generating high quality adversarial examples on text data under black-box hard-label attack scenarios, by substituting original words and using synonym sets to optimize the example while minimizing perturbation rate and query number.
Daiwen Sun, He Huang, Yao Li, Xinqi Gong, Qiwei Ye
https://openreview.net/forum?id=IOSaJ7ukgf
Keywords: Protein molecular dynamics, Protein surface representation, Implicit neural representation, Signed distance function, Continuous time modeling
Compressor summary: The authors propose a new neural network method for modeling protein dynamics using an implicit representation of protein surfaces based on signed distance functions, which can accurately capture and interpolate large-scale protein motions, and offer advantages over existing methods.
Hiren Madhu, Sundeep Prabhakar Chepuri
https://openreview.net/forum?id=INS3ltgjg7
Keywords: Simplicial representation learning, Self-supervised learning, Message passing simplicial networks
Compressor summary: $\texttt{TopoSRL}$ is a new self-supervised learning method for simplicial complexes that captures higher-order interactions and preserves topology, outperforming existing graph-based methods and supervised models.
Huining Yuan, Hongkun Dou, Xingyu Jiang, Yue Deng
https://openreview.net/forum?id=IN3hQx1BrC
Keywords: Model-based reinforcement learning, world model, generative model, meta-learning, bi-level optimization
Compressor summary: TEMPO is a bi-level model learning framework for reinforcement learning that combines maximum-likelihood and value-equivalent models using a meta weighter network to improve task awareness and semantic information.
Shuang Li, Ke Li, Wei Li
https://openreview.net/forum?id=IMiGRqltQQ
Keywords: Bayesian Optimization, Termination Criterion, Looking Backward
Compressor summary: The paper presents a novel, theoretically grounded method to terminate Bayesian Optimization early when it is in a convex region, saving computation while maintaining good performance.
Yao Mu, Qinglong Zhang, Mengkang Hu, Wenhai Wang, Mingyu Ding, Jun Jin, Bin Wang, Jifeng Dai, Yu Qiao, Ping Luo
https://openreview.net/forum?id=IL5zJqfxAa
Keywords: Embodied AI, Multi-modal Foundation Model, Embodied Control
Compressor summary: EmbodiedGPT is an end-to-end multi-modal model for embodied AI that uses a large language model to generate plans and extract features for effective embodied tasks.
Jiayuan Ye, Zhenyu Zhu, Fanghui Liu, Reza Shokri, Volkan Cevher
https://openreview.net/forum?id=IKvxmnHjkL
Keywords: overparameterized neural network, privacy
Compressor summary: This paper analyzes how over-parameterized models in machine learning algorithms can leak information about their training data and explores how initialization distribution affects this privacy loss depending on the model's depth.
Haoyu Han, Xiaorui Liu, Feng Shi, MohamadAli Torkamani, Charu C. Aggarwal, Jiliang Tang
https://openreview.net/forum?id=IKjOMA8olL
Keywords: Graph Neural Networks, Label Position Bias, Graph Structure Learning
Compressor summary: The paper introduces label position bias as a new bias in graph neural networks and proposes an optimization framework to reduce it.
Di Luo, Jiayu Shen, Rumen Dangovski, Marin Soljacic
https://openreview.net/forum?id=IKQOS8rqwr
Keywords: Koopman operator, quantum optimization, machine learning
Compressor summary: QuACK is a novel framework that uses Koopman operator theory to significantly accelerate quantum optimization and machine learning by efficiently predicting gradient dynamics on quantum computers.
Zilai Zeng, Ce Zhang, Shijie Wang, Chen Sun
https://openreview.net/forum?id=IJblKO45YU
Keywords: reinforcement learning, offline RL, self-supervised learning
Compressor summary: This paper explores how sequence modeling can improve policy learning by condensing trajectories into useful representations using Goal-Conditioned Predictive Coding (GCPC).
Marco Fumero, Florian Wenzel, Luca Zancato, Alessandro Achille, Emanuele Rodolà, Stefano Soatto, Bernhard Schölkopf, Francesco Locatello
https://openreview.net/forum?id=IHR83ufYPy
Keywords: disentanglement, OOD generalization, multitask learning
Compressor summary: The authors propose a method to learn disentangled features from multiple supervised tasks without directly observing the factors of variation, and show its effectiveness on various real-world data.
Zehan Wang, Yang Zhao, Xize Cheng, Haifeng Huang, Jiageng Liu, Aoxiong Yin, Li Tang, Linjun Li, Yongqi Wang, Ziang Zhang, Zhou Zhao
https://openreview.net/forum?id=IGTbT9P1ti
Keywords: multi-modal, representation learning, contrastive learning
Compressor summary: The paper proposes a method called C-MCR to learn multi-modal representations without paired data by connecting existing MCRs trained on different modality pairs using overlapping modalities and enhancing semantic consistency.
Fabian Fumagalli, Maximilian Muschalik, Patrick Kolpaczki, Eyke Hüllermeier, Barbara Eva Hammer
https://openreview.net/forum?id=IEMLNF4gK4
Keywords: Explainable Artificial Intelligence, Feature Interaction, Shapley Interaction, Shapley Value
Compressor summary: The paper introduces SHAP-IQ, a method for computing Shapley interactions in XAI using a novel representation and sampling, with theoretical guarantees and improved calculations.
Yonghan Jung, Ivan Diaz, Jin Tian, Elias Bareinboim
https://openreview.net/forum?id=IEJzoOBM0z
Keywords: Causal Effect Estimation, Causal Effect Identification, Data Fusion, Double Machine Learning, Doubly Robust Estimator
Compressor summary: The paper proposes a new estimator for learning causal relations from observational and interventional data that is robust to bias and can handle multiple outcomes.
Guangyuan Jiang, Manjie Xu, Song-Chun Zhu, Wenjuan Han, Chi Zhang, Yixin Zhu
https://openreview.net/forum?id=I9xE1Jsjfx
Keywords: machine personality, machine behavior, personality trait theory, psychometric, large language models, prompt
Compressor summary: The authors propose using a Machine Personality Inventory (MPI) tool to study and induce specific personalities in large language models (LLMs), inspired by human psychometric tests and the Big Five Personality Factors theory.
Pengze Zhang, Hubery Yin, Chen Li, Xiaohua Xie
https://openreview.net/forum?id=I9GNrInbdf
Keywords: Discrete Probability Flow, Optimal Transport
Compressor summary: The paper introduces a new theory for discrete diffusion models, based on optimal transport principles, and shows how it leads to a better sampling method with more certain outcomes.
Zhenghai Xue, Qingpeng Cai, Shuchang Liu, Dong Zheng, Peng Jiang, Kun Gai, Bo An
https://openreview.net/forum?id=I8t9RKDnz2
Keywords: Reinforcement Learning, Dynamics Shift, Stationary State Distribution, Offline RL, Off-Policy RL
Compressor summary: The paper proposes a new Reinforcement Learning algorithm (SRPO) that uses stationary state distributions to regularize policies, enabling efficient data reuse and improving the performance of context-based algorithms in different environments with similar structures but different dynamics.
Somnath Basu Roy Chowdhury, Nicholas Monath, Kumar Avinava Dubey, Amr Ahmed, Snigdha Chaturvedi
https://openreview.net/forum?id=I6aOjhpcNQ
Keywords: Concept Erasure, Representation Learning, Rate distortion, Fairness, Debiasing
Compressor summary: The paper introduces KRaM, a new objective for concept erasure in distributed representations, which uses a rate-distortion function to transform representations while preserving other information. KRaM can handle various concepts and domains and shows good results in experiments.
Zhan Ling, Yunhao Fang, Xuanlin Li, Zhiao Huang, Mingu Lee, Roland Memisevic, Hao Su
https://openreview.net/forum?id=I5rsM4CY2z
Keywords: Chain-of-thought, Large language model, Reasoning
Compressor summary: The text discusses a proposed method to improve large language models' deductive reasoning through natural language-based reasoning verification in step-by-step subprocesses.
David Woodruff, Fred Zhang, Samson Zhou
https://openreview.net/forum?id=I5SM5y57k2
Keywords: online learning, memory efficiency, sub-linear algorithm, communication lower bound
Compressor summary: The paper studies robust algorithms for online learning with experts under memory constraints and shows a trade-off between space and regret, while also providing a lower bound on space usage for deterministic algorithms.
Suhas Kotha, Christopher Brix, J Zico Kolter, Krishnamurthy Dj Dvijotham, Huan Zhang
https://openreview.net/forum?id=I50HbChk3U
Keywords: Trustworthy ML, Formal Verification, Safe Control, OOD Detection
Compressor summary: The INVPROP algorithm efficiently over-approximates inputs that lead to specific outputs for neural networks and other systems, outperforming prior work in speed and precision.
Hannaneh Akrami, Kurt Mehlhorn, Masoud Seddighin, Golnoosh Shahkarami
https://openreview.net/forum?id=I3k2NHt1zu
Keywords: Algorithmic game theory, Fairness, Randomized, Allocation, Maximin-share, Fractionally Subadditive
Compressor summary: The paper studies how to fairly allocate indivisible items to agents with subadditive valuations, improving both deterministic and randomized allocation methods.
Vicente Vivanco Cepeda, Gaurav Kumar Nayak, Mubarak Shah
https://openreview.net/forum?id=I18BXotQ7j
Keywords: Geo-localization, Image-to-GPS retrieval, CLIP, Random Fourier Features
Compressor summary: GeoCLIP is a new method for pinpointing images' locations on Earth using GPS encoding and a hierarchical representation, overcoming limitations of previous approaches.
Masatoshi Uehara, Haruka Kiyohara, Andrew Bennett, Victor Chernozhukov, Nan Jiang, Nathan Kallus, Chengchun Shi, Wen Sun
https://openreview.net/forum?id=HwhRehMr4a
Keywords: Reinforcement learning theory, POMDP, PAC RL, Off-policy evaluation, Offilne reinforcement learning
Compressor summary: The paper proposes a new model-free off-policy evaluation method for partially observable MDPs using future-dependent value functions and conditional moment equations, with PAC result and Bellman completeness guarantees.
Qiuxia Lin, Kerui Gu, Linlin Yang, Angela Yao
https://openreview.net/forum?id=HvhagNdf5z
Keywords: pose estimation, domain adaptation
Compressor summary: The paper proposes a reconstruction-based strategy to adapt models trained on synthetic data to real-world domains using unlabelled data, improving keypoint localization accuracy for hand and human pose estimation tasks.
Pan Lu, Baolin Peng, Hao Cheng, Michel Galley, Kai-Wei Chang, Ying Nian Wu, Song-Chun Zhu, Jianfeng Gao
https://openreview.net/forum?id=HtqnVSCj3q
Keywords: large language models, compositional reasoning, tool use, multi-modal reasoning, mathematical reasoning
Compressor summary: Chameleon is an AI system that augments large language models with plug-and-play modules for compositional reasoning, enabling them to access up-to-date information and perform complex tasks.
Gowthami Somepalli, Vasu Singla, Micah Goldblum, Jonas Geiping, Tom Goldstein
https://openreview.net/forum?id=HtMXRGbUMt
Keywords: diffusion models, memorization, data replication, model safety
Compressor summary: The paper analyzes how text-to-image diffusion models can unintentionally copy images from their training data and proposes techniques to reduce this problem by modifying image captions.
Luke Taylor, Andrew J King, Nicol Spencer Harper
https://openreview.net/forum?id=Ht79ZTVMsn
Keywords: spiking neural network, surrogate gradient descent, adaptive leaky integrate and fire neuron, speed-accuracy trade-off, electrophysiological recordings
Compressor summary: The ALIF model is a brain simulation tool that can be made faster and more accurate by using parallel processing on GPUs and smaller time-steps.
Wendong Liang, Armin Kekić, Julius von Kügelgen, Simon Buchholz, Michel Besserve, Luigi Gresele, Bernhard Schölkopf
https://openreview.net/forum?id=HszLRiHyfO
Keywords: Causality, independent component analysis, causal inference, interventions, latent variable models, identifiability
Compressor summary: Causal Component Analysis (CauCA) is an extension of Independent Component Analysis (ICA) that incorporates causal relationships among latent variables, and can be seen as a special case of Causal Representation Learning (CRL).
Tianhe Wu, Shuwei Shi, Haoming Cai, Mingdeng Cao, Jing Xiao, Yinqiang Zheng, Yujiu Yang
https://openreview.net/forum?id=HrL1oblm1a
Keywords: Blind omnidirectional image quality assessment, Multi-sequence network, Viewport sequence
Compressor summary: The paper introduces Assessor360, a novel network for objectively assessing the quality of omnidirectional images (ODIs) used in virtual reality, by modeling the observer's browsing process and using a multi-sequence approach with recursive probability sampling, multi-scale feature aggregation, and temporal modeling.
Simon Schrodi, Danny Stoll, Binxin Ru, Rhea Sanjay Sukthanker, Thomas Brox, Frank Hutter
https://openreview.net/forum?id=Hpt1i5j6wh
Keywords: Neural Architecture Search, Hierarchical Search Space, Context-free Grammars, Bayesian Optimization
Compressor summary: The paper introduces a new framework for generating large hierarchical search spaces for Neural Architecture Search using context-free grammars and an efficient kernel design for Bayesian Optimization.
Yuan Jiang, Zhiguang Cao, Yaoxin Wu, Wen Song, Jie Zhang
https://openreview.net/forum?id=HoBbZ1vPAh
Keywords: Vehicle Routing Problem, Distribution shift, Deep Reinforcement Learning, Ensemble Learning
Compressor summary: The authors propose a diverse ensemble of sub-policies using deep reinforcement learning to improve vehicle routing performance across different instance distributions.
Jimmy Ba, Murat A Erdogdu, Taiji Suzuki, Zhichao Wang, Denny Wu
https://openreview.net/forum?id=HlIAoCHDWW
Keywords: random matrix theory, high-dimensional statistics, neural network, kernel method, feature learning
Compressor summary: The paper investigates how the strength of a low-dimensional component in high-dimensional data affects the learning performance of kernel methods and neural networks with gradient descent.
Zhekai Du, Jingjing Li
https://openreview.net/forum?id=HffQOS3mk8
Keywords: diffusion-based models, active learning, domain adaptation, source-free domain adaptation, uncertainty estimation
Compressor summary: The paragraph describes a probabilistic framework for active domain adaptation that uses variational inference to capture data-level and prediction-level uncertainties, enabling efficient sampling of all possible predictions and selective annotation of informative target samples based on p-values.
Damien Teney, LIN Yong, Seong Joon Oh, Ehsan Abbasnejad
https://openreview.net/forum?id=HZQZli6amV
Keywords: Generalisation; machine learning
Compressor summary: The paper reveals that there are cases where in-distribution and out-of-distribution performance trade-offs exist, contrary to previous findings, and suggests that relying solely on in-distribution performance for model selection can lead to suboptimal out-of-distribution generalization.
Jheng-Wei Su, Kuei-Yu Tung, Chi-Han Peng, Peter Wonka, Hung-Kuo Chu
https://openreview.net/forum?id=HYo2Ao3hP8
Keywords: deep learning, layout reconstruction
Compressor summary: The paper introduces SLIBO-Net, a novel transformer-based approach to reconstruct 2D floorplans from unstructured 3D point clouds with improved semantic quality, efficient representation, and local geometric details, achieving state-of-the-art results on the Structure3D dataset.
Ilia Sucholutsky, Thomas L. Griffiths
https://openreview.net/forum?id=HYGnmSLBCf
Keywords: representation learning, supervised learning, human alignment, few-shot learning
Compressor summary: AI systems with representations similar to humans perform better on few-shot learning tasks, are more robust to attacks and domain shifts, but having human alignment is not always necessary.
Alejandro Escontrela, Ademi Adeniji, Wilson Yan, Ajay Jain, Xue Bin Peng, Ken Goldberg, Youngwoon Lee, Danijar Hafner, Pieter Abbeel
https://openreview.net/forum?id=HWNl9PAYIP
Keywords: Reinforcement learning, generative modeling, learning from demonstrations, video prediction, unsupervised reinforcement learning
Compressor summary: VIPER is an algorithm that learns expert-level behaviors for reinforcement learning agents by extracting preferences from unlabeled videos using pretrained video prediction models.
Shuai Zhang, Hongkang Li, Meng Wang, Miao Liu, Pin-Yu Chen, Songtao Lu, Sijia Liu, Keerthiram Murugesan, Subhajit Chaudhury
https://openreview.net/forum?id=HWGWeaN76q
Keywords: Reinforcement learning, Deep Q Network, Convergence analysis, Sample complexity, Generalization analysis
Compressor summary: The paper analyzes the theoretical aspects of deep Q-Network with epsilon-greedy exploration in reinforcement learning and shows how its convergence and sample complexity depend on the epsilon value.
Gellért Weisz, András György, Csaba Szepesvari
https://openreview.net/forum?id=HV85SiyrsV
Keywords: Reinforcement learning, linear function approximation, online learning
Compressor summary: The paper proposes a novel online reinforcement learning algorithm for linearly $q^\pi$-realizable MDPs that can efficiently learn policies even in the presence of states with nearly equal action-values and misspecified features.
Rickard Karlsson, JH Krijthe
https://openreview.net/forum?id=HUuEMMM8Ik
Keywords: causal inference, hidden confounding, multiple environments, independent causal mechansisms, independence testing
Compressor summary: The paper proposes a method to detect hidden confounding factors in multiple observational datasets by testing conditional independencies and evaluates its performance using simulations and real-world data.
Gwen Legate, Nicolas Bernier, Lucas Caccia, Edouard Oyallon, Eugene Belilovsky
https://openreview.net/forum?id=HRGd5dcVfw
Keywords: federated learning; transfer learning; nearest mean classifier; continual learning;
Compressor summary: The text describes how using pre-trained models and efficient transfer learning methods can improve federated learning by reducing costs and improving performance.
Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher D Manning, Stefano Ermon, Chelsea Finn
https://openreview.net/forum?id=HPuSIXJaa9
Keywords: reinforcement learning from human feedback, language models, RLHF, preferences
Compressor summary: DPO is a simple and effective method for aligning large unsupervised language models with human preferences, outperforming existing methods like RLHF.
Zhenbo Song, XiangHui Ze, Jianfeng Lu, Yujiao Shi
https://openreview.net/forum?id=HPrd17Qvbp
Keywords: Optical flow, correspondence learning, cross-view, camera localization
Compressor summary: The paper proposes a novel method to estimate camera pose using pixel-wise flow fields learned from ground and satellite images, improving localization accuracy significantly.
Guillaume Huguet, Alexander Tong, Edward De Brouwer, Yanlei Zhang, Guy Wolf, Ian Adelstein, Smita Krishnaswamy
https://openreview.net/forum?id=HNd4qTJxkW
Keywords: manifold learning, heat diffusion, geodesic, metric preserving, dimensionality reduction, embedding
Compressor summary: The authors propose a new heat geodesic embedding method for manifold learning and denoising, which connects heat diffusion to manifold distances using Riemannian geometry and outperforms existing methods in preserving ground truth distances and cluster structure.
Cheems Wang, Yiqin Lv, Yanghe Feng, Zheng Xie, Jincai Huang
https://openreview.net/forum?id=HMqGYxnlpv
Keywords: meta learning, robust fast adaptation, model agnostic meta learning
Compressor summary: This paper proposes a distributionally robust optimization approach for meta learning, which improves its robustness to task distributions and reduces the worst fast adaptation risk.
Jean-Yves Franceschi, Mike Gartrell, Ludovic Dos Santos, Thibaut Issenhuth, Emmanuel de Bezenac, Mickael Chen, Alain Rakotomamonjy
https://openreview.net/forum?id=HMhEFKDQ6J
Keywords: deep learning, generative models, GANs, generative adversarial networks, diffusion, score-based, gradient flows
Compressor summary: The paper proposes a unified framework for particle and adversarial generative models, suggesting that a generator is an optional addition to any generative model, and empirically tests its applications.
Shashanka Venkataramanan, Ewa Kijak, laurent amsaleg, Yannis Avrithis
https://openreview.net/forum?id=HKueO74ZTB
Keywords: Interpolation based data augmentation, mixup, dense interpolation, robustness, representation learning
Compressor summary: The authors propose MultiMix, a data augmentation method that generates more interpolated examples in the embedding space than previous methods, leading to better performance on sequence data tasks.
Jiashuo WANG, Haozhao Wang, Shichao Sun, Wenjie Li
https://openreview.net/forum?id=HGFcM3UU50
Keywords: Aligned Models; Human-centric NLG
Compressor summary: The paper proposes a novel Bayesian framework called d-PM to account for the distribution of disagreements among human preferences in NLG systems and uses contrastive learning to train the model efficiently, achieving better results than previous methods.
Zitong Yang, Michal Lukasik, Vaishnavh Nagarajan, Zonglin Li, Ankit Singh Rawat, Manzil Zaheer, Aditya Krishna Menon, Sanjiv Kumar
https://openreview.net/forum?id=HFQFAyNucq
Keywords: deep learning, generalization, memorization, deep learning theory, boosting, nearest neighbor
Compressor summary: ResMem is a method to improve model generalization by explicitly memorizing training labels through fitting residuals with a nearest-neighbor based regressor.
Amira Abbas, Robbie King, Hsin-Yuan Huang, William J. Huggins, Ramis Movassagh, Dar Gilboa, Jarrod Ryan McClean
https://openreview.net/forum?id=HF6bnhfSqH
Keywords: Backpropagation, quantum computing, shadow tomography, gentle measurement
Compressor summary: The paper explores how to train quantum neural networks efficiently by using multiple copies of a state and shadow tomography, similar to backpropagation in classical deep learning.
Haocheng Xi, ChangHao Li, Jianfei Chen, Jun Zhu
https://openreview.net/forum?id=H9hWlfMT6O
Keywords: neural network quantization, transformer, matrix multiplication, randomized numerical linear algebra
Compressor summary: The paper proposes a new method to train transformer models with 4-bit precision, which is faster and can work on current GPUs, by using dedicated quantizers for activations, weights, and gradients.
Shogo Iwazaki, Shion Takeno, Tomohiko Tanabe, Mitsuru Irie
https://openreview.net/forum?id=H5pwAeYAun
Keywords: Gaussian process optimization, regret analysis, black-box optimization, Bayesian optimization
Compressor summary: F-GP-UCB is a black-box optimization method that handles observation failure with complex latent constraints and provides regret upper bounds and convergence guarantees.
Cai Zhou, Xiyuan Wang, Muhan Zhang
https://openreview.net/forum?id=H57w5EOj6O
Keywords: random walk on simplicials, Hodge Laplacian, graph neural networks, edge-level positional encoding
Compressor summary: The paper studies how random walk on different orders of simplicial complexes improves Graph Neural Networks' theoretical expressivity and proposes new positional encoding methods based on Hodge Laplacians.
Zeke Xie, Qian-Yuan Tang, Mingming Sun, Ping Li
https://openreview.net/forum?id=H4GsteoL0M
Keywords: Gradient Noise, SGD, Deep Learning
Compressor summary: This paper investigates the heavy-tail properties of stochastic gradients in deep neural networks and reveals new insights into their covariance structures.
Ryoma Yataka, Kazuki Hirashima, Masashi Shiraishi
https://openreview.net/forum?id=H2udtfMbl4
Keywords: Generative Models, Geometric Deep Learning, Normalizing Flows, Shape Analysis, Grassmann Manifold
Compressor summary: This paper proposes a novel method to generate stable shapes using Grassmann manifolds and continuous normalization flows in machine learning, outperforming existing methods.
Maxence Noble, Valentin De Bortoli, Arnaud Doucet, Alain Durmus
https://openreview.net/forum?id=H2SuXHbFIn
Keywords: Schrödinger bridge, optimal transport, diffusion model, Wasserstein barycenter
Compressor summary: The paper introduces TreeDSB, an extension of DSB for solving entropic multi-marginal optimal transport problems with tree-structured quadratic costs, which can be used for computing Wasserstein barycenters and applying to high-dimensional tasks like image interpolation and Bayesian fusion.
Xin Yuan, Pedro Henrique Pamplona Savarese, Michael Maire
https://openreview.net/forum?id=H1a7bVVnPK
Keywords: network growing, efficient network training
Compressor summary: The approach grows neural networks efficiently by designing parameterization and optimization strategies for the training dynamics, achieving comparable or better accuracy than fixed-size models with less computation budget and faster training speed.
Zhen Xiang, Zidi Xiong, Bo Li
https://openreview.net/forum?id=H1CQZqpgdQ
Keywords: backdoor, Trojan, certification, adversarial learning, deep neural network, conformal prediction
Compressor summary: The paper introduces a novel certified backdoor detector (CBD) that uses conformal prediction and local dominant probability statistics to detect and guarantee the detection of various backdoor attacks on deep neural networks.
Daniel Bertschinger, Christoph Hertrich, Paul Jungeblut, Tillmann Miltzow, Simon Weber
https://openreview.net/forum?id=H15KtcyHvn
Keywords: Neural Network Training, Computational Complexity, Existential Theory of the Reals, Algebraic Universality, Empirical Risk Minimization
Compressor summary: The paper shows that finding the optimal weights and biases for a two-layer neural network is as hard as solving real roots of multivariate polynomials with integer coefficients.
Dihong Jiang, Sun Sun, Yaoliang Yu
https://openreview.net/forum?id=GzlDKZlwie
Keywords: Renyi differential privacy, RKHS, MMD, Gaussian process, generative model
Compressor summary: The paper extends differential privacy to functional outputs and applies it to a generative model, improving the privacy-utility trade-off.
Seunghwan An, Jong-June Jeon
https://openreview.net/forum?id=GxL6PrmEUw
Keywords: Variational AutoEncoder, distributional learning, synthetic data generation, CRPS, asymmetric Laplace distribution
Compressor summary: The paper introduces a new VAE model with an infinite mixture of asymmetric Laplace distribution that can fit any continuous distribution and has better quantile estimation and privacy control properties.
Shaolei Zhang, Yang Feng
https://openreview.net/forum?id=GuErIOGLie
Keywords: Machine Translation, Speech Translation, Speech Recognition, Simultaneous Generation, Simultaneous Translation
Compressor summary: The paper introduces a framework called Seg2Seg that learns to generate target sequences adaptively and simultaneously from source sequences in different tasks, using latent segments as the bridge between them.
Amine Ouasfi, Adnane Boukhayma
https://openreview.net/forum?id=Gtse4R6iS4
Keywords: implicit neural representations, 3D reconstruction from unoriented point could, kernel ridge regression
Compressor summary: The authors propose a method to improve implicit shape reconstruction from point clouds by combining inter-shape and intra-shape priors, achieving better stability and efficiency.
Hong Chen, Xin Wang, Yuwei Zhou, Yijian Qin, Chaoyu Guan, Wenwu Zhu
https://openreview.net/forum?id=GtgFo5lmOB
Keywords: auxiliary learning, data-task joint generation
Compressor summary: The paper proposes a new method for auxiliary learning that generates data and tasks jointly to improve the primary task performance and avoid relying on domain knowledge for data collection.
gregoire pacreau, Karim Lounici
https://openreview.net/forum?id=GtYlxtwO74
Keywords: robust statistics, missing values, cell-wise contamination
Compressor summary: The paper proposes a new estimator for covariance that works well with missing data and cell-wise outliers, without needing to remove or impute any data points.
Suman Bhoi, Mong-Li Lee, Wynne Hsu, Ngiap Chuan Tan
https://openreview.net/forum?id=GsCTjmYe5v
Keywords: Fine-grained Medication recommendation, Drug Interaction Severity
Compressor summary: REFINE is a deep learning system that recommends fine-grained medications to patients with co-morbidities, considering their health conditions and drug interactions, and improves treatment outcomes and safety.
Sujin Jang, Dae Ung Jo, Sung Ju Hwang, Dongwook Lee, Daehyun Ji
https://openreview.net/forum?id=Grz2ijKrWI
Keywords: knowledge distillation, cross-modal learning, 3d object detection
Compressor summary: The paper proposes a novel framework called STXD to improve multi-view 3D object detection by transferring structural, temporal, and output knowledge from LiDAR-based teachers using cross-modal distillation.
Kazuto Fukuchi, Jun Sakuma
https://openreview.net/forum?id=GrFsx4mBWF
Keywords: demographic parity, regression, minimax optimal
Compressor summary: We explore the minimax optimal error associated with a demographic parity-constrained regression problem within the context of a linear model. Our proposed model encompasses a broader range of discriminatory bias sources compared to the model presented by Chzhen and Schreuder. Our analysis reveals that the minimax optimal error for the demographic parity-constrained regression problem under our model is characterized by $\Theta(\frac{dM}{n})$, where $n$ denotes the sample size, $d$ represents the dimensionality, and $M$ signifies the number of demographic groups arising from sensitive attributes. Moreover, we demonstrate that the minimax error increases in conjunction with a larger bias present in the model.
Sungbin Lim, Eunbi Yoon, Taehyun Byun, Taewon Kang, Seungwoo Kim, Kyungjae Lee, Sungjoon Choi
https://openreview.net/forum?id=GrElRvXnEj
Keywords: Generative Model, Score-based Method, Diffusion Model
Compressor summary: The paper introduces Hilbert Diffusion Model (HDM), which extends score-based generative models to function spaces by using stochastic evolution equations in Hilbert spaces and achieves better performance in sampling functions and synthesizing motion.
Simone Fioravanti, Michele Flammini, Bojana Kodric, Giovanna Varricchio
https://openreview.net/forum?id=GqtpYUCwnu
Keywords: Game Theory, Hedonic Games, Core stability, Coalition Formation, Social Choice, PAC learning
Compressor summary: The text introduces a new concept called $\varepsilon$-fractional core-stability to address the challenges of finding stable coalition structures in hedonic games, and presents efficient algorithms for two classes of games.
Zhaohan Xi, Tianyu Du, Changjiang Li, Ren Pang, Shouling Ji, Jinghui Chen, Fenglong Ma, Ting Wang
https://openreview.net/forum?id=GqXbfVmEPW
Keywords: few-shot learning, prompt learning, language model, backdoor defense
Compressor summary: The authors propose a defense method (MDP) for language models that are vulnerable to backdoor attacks in few-shot learning scenarios, which identifies poisoned samples by comparing their masking-sensitivity with clean samples.
Li Yang, Chunfeng Yuan, Ziqi Zhang, Zhongang Qi, Yan Xu, Wei Liu, Ying Shan, Bing Li, Weiping Yang, Peng Li, Yan Wang, Weiming Hu
https://openreview.net/forum?id=GlWzQhf2lV
Keywords: 3D Visual Grounding, Contextual Object, Contextual Relation
Compressor summary: The paper presents a new model, CORE-3DVG, that learns to identify objects in 3D scenes from natural language inputs by explicitly learning contextual information and achieves state-of-the-art performance on three datasets.
Pier Giuseppe Sessa, Pierre Laforgue, Nicolò Cesa-Bianchi, Andreas Krause
https://openreview.net/forum?id=GjJRbEZ1dc
Keywords: multitask learning, confidence intervals, online learning theory, active learning, regret
Compressor summary: The text discusses novel confidence intervals for multitask regression in the agnostic setting, regret guarantees, an online learning algorithm that adapts to task similarity, and a multitask active learning setup with no-regret algorithms.
Haitz Sáez de Ocáriz Borde, Alvaro Arroyo, Ismael Morales López, Ingmar Posner, Xiaowen Dong
https://openreview.net/forum?id=Gij638d76O
Keywords: Representation Learning, Product Manifolds, Bayesian Optimization, Gromov-Hausdorff Distance
Compressor summary: The paper proposes NLGS, a method to automatically find the best latent geometry for machine learning models using Bayesian optimization and metric geometry techniques.
Chanwoo Park, Kaiqing Zhang, Asuman E. Ozdaglar
https://openreview.net/forum?id=GiiOpKinGm
Keywords: Markov Games, Local Interaction, PPAD-Hardness, Fictitious Play
Compressor summary: The paragraph introduces a new class of Markov games called zero-sum NMGs, which model non-cooperative multi-agent sequential decision-making with networked separable interactions, and studies their equilibrium properties and learning dynamics.
Wenzhuo Zhou
https://openreview.net/forum?id=GiUe0ZFiVe
Keywords: Offline Reinforcement Learning, Sample Efficiency, Regret Bound, Data Coverage
Compressor summary: The paper proposes a bi-level structured policy optimization algorithm for offline reinforcement learning that addresses the distributional shift problem by constructing a confidence set of value estimates and maximizing a conservative value estimate at the upper level.
Wilka Carvalho, Andre Saraiva, Angelos Filos, Andrew Kyle Lampinen, Loic Matthey, Richard Lewis, Honglak Lee, Satinder Singh, Danilo Jimenez Rezende, Daniel Zoran
https://openreview.net/forum?id=GhNCFtLSsy
Keywords: deep reinforcement learning, successor features, transfer, generalization, feature-discovery
Compressor summary: The paper proposes a new method, Successor Features Keyboard (SFK), for transferring behavioral knowledge across tasks using discovered state-features and task encodings, and demonstrates its effectiveness in a 3D environment.
Zhihan Gao, Xingjian Shi, Boran Han, Hao Wang, Xiaoyong Jin, Danielle C. Maddix, Yi Zhu, Mu Li, Bernie Wang
https://openreview.net/forum?id=Gh67ZZ6zkS
Keywords: Machine Learning for Earth Science, Spatiotemporal Forecasting, Generative Models, Diffusion Models
Compressor summary: The authors propose a two-stage pipeline for probabilistic spatiotemporal forecasting using a conditional latent diffusion model (PreDiff) and a knowledge alignment mechanism to handle uncertainty and incorporate physical constraints.
Siddhartha Laghuvarapu, Zhen Lin, Jimeng Sun
https://openreview.net/forum?id=GgdFLb94Ld
Keywords: drug discovery, molecule property prediction, conformal prediction
Compressor summary: CoDrug is a method that uses an energy-based model and Kernel Density Estimation to create reliable uncertainty estimates for drug molecules, accounting for distribution shifts in drug discovery tasks.
Giannis Daras, Yuval Dagan, Alex Dimakis, Constantinos Costis Daskalakis
https://openreview.net/forum?id=GfZGdJHj27
Keywords: diffusion models, sampling drift, Fokker-Planck, invariances, Stochastic Differential Equations, Martingales
Compressor summary: The paper proposes a method to improve the quality of generated images by enforcing a Consistency property that ensures predictions on generated data are consistent across time, and shows empirical results on various image datasets.
Nina L. Corvelo Benz, Manuel Gomez Rodriguez
https://openreview.net/forum?id=GfITbjrIOd
Keywords: Calibration, Trustworthy Machine Learning, Human-Centric ML, Probabilistic Models and Methods
Compressor summary: The paper explores why decision makers struggle with confidence values provided by binary classifiers and proposes constructing more useful confidence values based on alignment with the decision maker's own confidence.
Rui Feng, Qi Zhu, Huan Tran, Binghong Chen, Aubrey Toland, Rampi Ramprasad, Chao Zhang
https://openreview.net/forum?id=Ge8Mhggq0z
Keywords: molecular pretraining, molecular representation learning
Compressor summary: The paper proposes a new force-centric pretraining model for 3D molecular conformations that covers both equilibrium and off-equilibrium data, improving forces accuracy, simulation performance, and inference speed.
Mitsuhiko Nakamoto, Yuexiang Zhai, Anikait Singh, Max Sobol Mark, Yi Ma, Chelsea Finn, Aviral Kumar, Sergey Levine
https://openreview.net/forum?id=GcEIvidYSw
Keywords: offline reinforcement learning, online fine-tuning
Compressor summary: The paper proposes calibrated Q-learning, a method that initializes offline RL policies with conservative value functions, enabling fast online fine-tuning and improving performance on benchmark tasks.
Marcello Massimo Negri, Fabricio Arend Torres, Volker Roth
https://openreview.net/forum?id=GYnbubCXhE
Keywords: normalizing flow, variational inference, graphical lasso, gaussian graphical model, bayesian inference
Compressor summary: The paper presents a new method to study conditional independence among many variables using Gaussian Graphical Models that unifies frequentist and Bayesian approaches and handles different sparsity-inducing norms.
Mingyuan Zhang, Huirong Li, Zhongang Cai, Jiawei Ren, Lei Yang, Ziwei Liu
https://openreview.net/forum?id=GYjV1M5s0D
Keywords: Motion Generation, Diffusion Model
Compressor summary: FineMoGen is a diffusion-based framework that generates and edits fine-grained motions using a novel spatio-temporal attention mechanism and a large dataset, enabling zero-shot motion editing with language models.
Wenqi Cui, Yan Jiang, Baosen Zhang, Yuanyuan Shi
https://openreview.net/forum?id=GWIRpKF6yU
Keywords: Control, Stability, Tracking, Passivity, Neural network-based controllers, Power systems
Compressor summary: The paper proposes neural PI controllers with provable stability and output tracking guarantees for multiple-input and multiple-output systems using equilibrium-independent passivity and strictly convex neural networks.
Zhiyuan Ren, Yiyang Su, Xiaoming Liu
https://openreview.net/forum?id=GTYaYNsFyv
Keywords: ChatGPT, Hierarchical Comparisons, Image Classification, Zero shot
Compressor summary: The authors propose a novel image classification framework using hierarchical comparisons of text embeddings to overcome CLIP's limitations and achieve better accuracy.
Jingkang Yang, Jun CEN, Wenxuan Peng, Shuai Liu, Fangzhou Hong, Xiangtai Li, Kaiyang Zhou, Qifeng Chen, Ziwei Liu
https://openreview.net/forum?id=GRHZiTbDDI
Keywords: Scene Graph Generation, 4D Understanding, 4D Perception.
Compressor summary: The paragraph introduces PSG-4D, a new 4D visual representation for AI, and presents a dataset, model, and application example for it.
Lukas Gosch, Simon Geisler, Daniel Sturm, Bertrand Charpentier, Daniel Zügner, Stephan Günnemann
https://openreview.net/forum?id=GPtroppvUM
Keywords: adversarial training, adversarial examples, robust graph learning, graph machine learning, graph neural networks, graphs
Compressor summary: The authors propose and evaluate a flexible graph neural network model with learnable graph diffusion that can defend against adversarial perturbations in the graph structure, while also introducing a new attack method for such perturbations.
Jialin Chen, Shirley Wu, Abhijit Gupta, Zhitao Ying
https://openreview.net/forum?id=GJtP1ZEzua
Keywords: Explainability, Graph Neural Network, Diffusion Model
Compressor summary: D4Explainer is a novel approach for generating reliable and diverse explanations for Graph Neural Networks by learning graph distributions that conform to the in-distribution property.
Giacomo Meanti, Antoine Chatalic, Vladimir R Kostic, Pietro Novelli, massimiliano pontil, Lorenzo Rosasco
https://openreview.net/forum?id=GItLpB1vhK
Keywords: dynamical systems, kernel methods, koopman operator, sketching, molecular dynamics, efficient machine learning
Compressor summary: The paper presents new methods to learn complex dynamics from data using Koopman operators and kernel methods, making them more efficient with random projections.
Anqi Mao, Christopher Mohri, Mehryar Mohri, Yutao Zhong
https://openreview.net/forum?id=GIlsH0T4b2
Keywords: learning to defer, learning theory
Compressor summary: The authors propose a new method to learn from multiple experts in two stages, using surrogate loss functions that ensure consistency and perform well in experiments.
Tom Coates, Alexander M. Kasprzyk, Sara Veneziale
https://openreview.net/forum?id=GI4Pp01prW
Keywords: mathematics, geometry, Fano varieties, terminal singularities, theorem discovery, neural network classifier, supervised learning
Compressor summary: The paper uses machine learning to classify eight-dimensional positively curved algebraic varieties with toric symmetry and Picard rank two, revealing the landscape of Q-Fano varieties and providing new evidence for machine learning's role in mathematical discovery.
Camilo Ruiz, Hongyu Ren, Kexin Huang, Jure Leskovec
https://openreview.net/forum?id=GGylthmehy
Keywords: Tabular Data, Deep Learning, Knowledge Graph, Regularization
Compressor summary: PLATO is a method that improves machine learning performance on tabular data with high dimensions and low samples by using a heterogeneous knowledge graph as an auxiliary source of information to regularize a multilayer perceptron.
Sungho Choi, Seungyul Han, Woojun Kim, Jongseong Chae, Whiyoung Jung, Youngchul Sung
https://openreview.net/forum?id=GGbBXSkX3r
Keywords: Reinforcement Learning, Deep Reinforcement Learning, Imitation Learning
Compressor summary: The paper presents a new method for teaching a robot to perform a task by observing other robots, even if they look different or are viewed from different angles.
Fatih Dinc, Adam Shai, Mark Schnitzer, Hidenori Tanaka
https://openreview.net/forum?id=GGIA1p9fDT
Keywords: brain-machine interfaces, recurrent neural networks, convex optimization, computational neuroscience
Compressor summary: CORNN is a new training method for recurrent neural networks that enables faster and more accurate modeling of large neural datasets, potentially enabling real-time analysis and control of animal behavior.
Tianyu Chen, Kevin Bello, Bryon Aragam, Pradeep Kumar Ravikumar
https://openreview.net/forum?id=GEtXhqKW6X
Keywords: distribution shifts, heterogeneous data, feature-shift, structural causal models, additive noise models, causality
Compressor summary: This paper proposes a method for identifying causal mechanism shifts in related datasets without estimating their full causal structure, assuming nonlinear additive noise models.
Jie Huang, Man Zhou, JingHao Zhang, Gang Yang, Mingde Yao, Chongyi Li, Zhiwei Xiong, Feng Zhao
https://openreview.net/forum?id=GEWzHeHpLr
Keywords: Image Enhancement, Normalization, Image Restoration
Compressor summary: The authors propose a novel normalization technique, Transition-Constant Normalization (TCN), for various image enhancement tasks that has several advantages and shows performance improvements in multiple experiments.
Xiaotong Yuan, Ping Li
https://openreview.net/forum?id=GEQZ52oqxa
Keywords: Uniform stability, Randomized learning algorithms, Confidence boosting, Generalization bounds, Stochastic gradient methods
Compressor summary: The paper introduces a new concept of uniform stability and proves strong exponential bounds on generalization error, leading to near-optimal results for randomized learning algorithms like SGD.
Ping Li, Xiaoyun Li
https://openreview.net/forum?id=GEMHw2sd9S
Keywords: Differential Privacy, Random Projection
Compressor summary: The text introduces new differential privacy algorithms using random projection and sign random projection methods, improving performance and utility for data protection in machine learning and search applications.
Fei Deng, Junyeong Park, Sungjin Ahn
https://openreview.net/forum?id=GDYuzX0rwj
Keywords: world models, structured state space sequence models, S4, long-term memory, model-based reinforcement learning
Compressor summary: This paper explores alternative world model backbones for improving long-term memory in model-based reinforcement learning, comparing Transformers, Structured State Space Sequence models, and recurrent neural networks across various tasks.
Dishank Bansal, Ricky T. Q. Chen, Mustafa Mukadam, Brandon Amos
https://openreview.net/forum?id=GCY9C43A4L
Keywords: task-based learning, decision-focused learning
Compressor summary: The proposed method improves deep learning models' performance on downstream tasks by using a task loss to learn a metric that emphasizes important information for the desired task without altering the original prediction model.
Thao Nguyen Truong, Balazs Gerofi, Edgar Josafat Martinez-Noriega, François Trahay, Mohamed Wahib
https://openreview.net/forum?id=GAsRl2ElHk
Keywords: Deep learning, Sampling
Compressor summary: The paper presents a method to hide less important samples in deep neural network training to save time without much accuracy loss.
Changyeon Kim, Younggyo Seo, Hao Liu, Lisa Lee, Jinwoo Shin, Honglak Lee, Kimin Lee
https://openreview.net/forum?id=G8nal7MpIQ
Keywords: Reinforcement Learning, Multimodal Representation, Imitation Learning
Compressor summary: The paper proposes ARP, an imitation learning framework that uses multimodal rewards based on natural language descriptions and visual observations to enhance generalization and adapt to unseen environments.
Liang Chen, Shuming Ma, Dongdong Zhang, Furu Wei, Baobao Chang
https://openreview.net/forum?id=G7sQlfTzmY
Keywords: Multilinugal Neural Machine Translation, Multitask Learning, Pareto Optimization
Compressor summary: The paper investigates how different translation directions perform in multilingual neural machine translation, proposes a method to predict and optimize their performance trade-offs, and shows that it improves over existing methods.
Anders Vestergaard Nørskov, Alexander Neergaard Zahid, Morten Mørup
https://openreview.net/forum?id=G7Y145tm2F
Keywords: zero-shot conversion, representations learning, contrastive learning, electroencephalography, autoencoder, subject variability, permutation invariant training
Compressor summary: The authors propose a new method (CSLP-AE) to convert EEG signals between tasks and subjects by extracting latent representations that account for both content and style, inspired by voice conversion technologies.
Yuval Kirstain, Adam Polyak, Uriel Singer, Shahbuland Matiana, Joe Penna, Omer Levy
https://openreview.net/forum?id=G5RwHpBUv0
Keywords: text-to-image, human-preferences, dataset
Compressor summary: The authors create a web app to collect user preferences for text-to-image generation and use it to build a large dataset (Pick-a-Pic) and a scoring function (PickScore) that predicts human preferences better than existing metrics and can improve existing models.
Yiwen Kou, Zixiang Chen, Quanquan Gu
https://openreview.net/forum?id=G560qr59Gi
Keywords: ReLU Neural Networks, Implicit Bias, Deep Learning Theory
Compressor summary: The paper investigates how gradient descent impacts the implicit bias of non-smooth neural networks, especially for leaky ReLU activation functions, and shows that it leads to stable rank and uniform normalized margin in some cases.
Justin Whitehouse, Aaditya Ramdas, Steven Wu
https://openreview.net/forum?id=G3aubF5Wnw
Keywords: Kernel Bandits, Online Learning, Self-Normalized Concentration, Online Regression
Compressor summary: The paper presents a new analysis of the popular Gaussian Process Upper Confidence Bound (GP-UCB) algorithm that shows it achieves nearly optimal regret and improves over previous results for the Matern kernel.
Shangshang Yang, Xiaoshan Yu, Ye Tian, Xueming Yan, Haiping Ma, Xingyi Zhang
https://openreview.net/forum?id=G14N38AjpU
Keywords: Knowledge tracing, intelligent education, neural architecture search, Transformer
Compressor summary: The paper proposes a new approach for knowledge tracing that uses convolution operations in Transformers, automates input feature selection, and applies evolutionary neural architecture search to balance local and global context modelling.
Minkyu Choi, Kuan Han, Xiaokai Wang, Yizhen Zhang, Zhongming Liu
https://openreview.net/forum?id=Fy1S3v4UAk
Keywords: brain-inspired AI, retina transformation, eye movements, deep neural networks
Compressor summary: The authors developed a dual-stream vision model inspired by human vision, which uses two CNN branches for spatial processing and object recognition, and found that it matches the dorsal and ventral pathways of the visual cortex in terms of functional alignment.
Kelsey Lieberman, James Diffenderfer, Charles Godfrey, Bhavya Kailkhura
https://openreview.net/forum?id=FxRfAIj4s2
Keywords: image compression, robustness, generalization
Compressor summary: The paper introduces a benchmark suite, inspection tools, and theoretical analysis to evaluate the out-of-distribution performance of neural image compression methods for real-world applications.
Suneel Belkhale, Yuchen Cui, Dorsa Sadigh
https://openreview.net/forum?id=FwmvbuDiMk
Keywords: Imitation Learning, Robotics, Data Quality
Compressor summary: The paper proposes two properties for defining high quality datasets in imitation learning: action divergence and transition diversity, which help the learned policy to stay in distribution at test time.
Badih Ghazi, Xiao Hu, Ravi Kumar, Pasin Manurangsi
https://openreview.net/forum?id=FviF8vuz5B
Keywords: privacy, sampling, Gaussian distribution, product distributions
Compressor summary: The paper proposes new differentially private algorithms for generating samples from multi-dimensional Gaussian distributions with different covariance assumptions and improves the performance in some settings.
Joseph T Costello, Hisham Temmar, Luis H Cubillos, Matthew J Mender, Dylan M Wallace, Matthew S Willsey, Parag G Patil, Cynthia Chestek
https://openreview.net/forum?id=FujJO3dsNj
Keywords: brain computer interface, brain machine interface, neural decoding, prosthetic control, recurrent neural network, RNN, transformer, real time, closed-loop, user interface
Compressor summary: The study evaluated recurrent neural networks (RNNs) for real-time brain-machine interface control, showing improved performance over other architectures in decoding finger movements from primate neural signals.
Shipra Agrawal, Yiding Feng, Wei Tang
https://openreview.net/forum?id=FtZ7lUwH99
Keywords: dynamic pricing, information design, regret minimization
Compressor summary: The paper proposes an algorithm that learns optimal pricing and advertising strategies for a seller in a dynamic setting with linear buyer valuations.
Fabian Zaiser, Andrzej S Murawski, Luke Ong
https://openreview.net/forum?id=FtNruwFEs3
Keywords: Bayesian statistics, probabliistic programming, exact inference, discrete models, probability generating functions
Compressor summary: The authors present an exact Bayesian inference method for discrete models using a probabilistic programming language and probability generating functions, which they implement in a tool called Genfer that outperforms existing tools on many problems.
Siqi Shen, Chennan Ma, Chao Li, Weiquan Liu, Yongquan Fu, Songzhu Mei, Xinwang Liu, Cheng Wang
https://openreview.net/forum?id=FskZtRvMJI
Keywords: multi-agent reinforcement learning, value factorization, individual global max, risk-sensitive
Compressor summary: RIGM principle introduces a new approach to coordinate risk-sensitive MARL policies, and RiskQ is a value factorization method that satisfies the RIGM principle for various risk metrics.
Tom Monnier, Jake Austin, Angjoo Kanazawa, Alexei A Efros, Mathieu Aubry
https://openreview.net/forum?id=FsQWxU5TOL
Keywords: 3D decomposition, 3D reconstruction, MVS, primitives, qualitative 3D
Compressor summary: The authors present a method to generate simple 3D world representations from images using textured superquadric meshes, which model transparency and handle varying numbers of primitives. The method operates on images through differentiable rendering and surpasses existing approaches on various scenes.
Dung Nguyen, Mahantesh M Halappanavar, Venkatesh Srinivasan, Anil Vullikanti
https://openreview.net/forum?id=Fqg9vGWy4k
Keywords: differential privacy, subgraph counting, smooth sensitivity, local sensitivity
Compressor summary: The paper proposes new algorithms for privately counting subgraphs in graph data, using approximated sensitivity metrics and reducing noise.
Haotian Xue, Antonio Torralba, Joshua B. Tenenbaum, Daniel LK Yamins, Yunzhu Li, Hsiao-Yu Tung
https://openreview.net/forum?id=Fp5uC6YHwe
Keywords: Intuitive Physics, Computer Vision
Compressor summary: The paper presents a framework to learn 3D visual intuitive physics models from videos of complex scenes with fluids using a conditional NeRF-style visual frontend and a point-based dynamics prediction backend, which can handle challenging scenarios without relying on dense point trajectory supervision.
Stefan Lionar, Xiangyu Xu, Min Lin, Gim Hee Lee
https://openreview.net/forum?id=FmpH0CYWiX
Keywords: single-view 3d reconstruction, neural fields, 3d reconstruction
Compressor summary: NU-MCC is a new approach for single-view 3D reconstruction that improves efficiency and detail recovery by using a neighborhood decoder and a repulsive unsigned distance function, achieving state-of-the-art results on the CO3D-v2 dataset.
Alexander Bukharin, Yan Li, Yue Yu, Qingru Zhang, Zhehui Chen, Simiao Zuo, Chao Zhang, Songan Zhang, Tuo Zhao
https://openreview.net/forum?id=FmZVRe0gn8
Keywords: Multi-Agent Reinforcement Learning, Theory of Robust Reinforcement Learning, Adversarial Regularization
Compressor summary: The paper introduces ERNIE, a robust multi-agent reinforcement learning framework that uses adversarial regularization to control policy's Lipschitz constant and improve stability against environmental changes.
Zebang Shen, Zhenfu Wang
https://openreview.net/forum?id=FkpMm9avyP
Keywords: Entropy-dissipation, McKean-Vlasov, Navier-Stokes, PDE, Coulomb, singular interaction
Compressor summary: The paper proposes a new neural network-based method to solve McKean-Vlasov equations involving singular interactions, using entropy dissipation as a control criterion.
Nouha Dziri, Ximing Lu, Melanie Sclar, Xiang Lorraine Li, Liwei Jiang, Bill Yuchen Lin, Sean Welleck, Peter West, Chandra Bhagavatula, Ronan Le Bras, Jena D. Hwang, Soumya Sanyal, Xiang Ren, Allyson Ettinger, Zaid Harchaoui, Yejin Choi
https://openreview.net/forum?id=Fkckkr3ya8
Keywords: Natural language processing, large language models, multi-step reasoning
Compressor summary: The authors examine transformer LLMs' ability to perform complex tasks requiring multi-step reasoning and find that these models rely on subgraph matching instead of developing problem-solving skills, which may limit their performance in more challenging tasks.
Gaon An, Junhyeok Lee, Xingdong Zuo, Norio Kosaka, Kyung-Min Kim, Hyun Oh Song
https://openreview.net/forum?id=FkAwlqBuyO
Keywords: Preference-based reinforcement learning, Contrastive learning, Offline reinforcement learning, RLHF
Compressor summary: The paper proposes a preference-based reinforcement learning (PbRL) method that uses contrastive learning to directly learn from human preferences without requiring a reward model.
Haizhou Shi, Hao Wang
https://openreview.net/forum?id=FiClXlUqA7
Keywords: Domain Incremental Learning, Continual Learning, Theory
Compressor summary: The paper proposes a unified framework for domain incremental learning that adapts to different domains with memory, improves generalization error bounds, and outperforms existing methods on various datasets.
Ting Wei Li, Qiaozhu Mei, Jiaqi Ma
https://openreview.net/forum?id=FgakGFpll1
Keywords: Graph Neural Networks, Metadata-Driven Analysis, Gini Coefficient of Degree Distribution
Compressor summary: This paper proposes a metadata-driven approach to analyze how graph data properties affect the performance of Graph Neural Networks (GNNs) using regression analysis, theoretical analysis, and experiments.
Alireza Mousavi-Hosseini, Denny Wu, Taiji Suzuki, Murat A Erdogdu
https://openreview.net/forum?id=Fe8PxP2F2p
Keywords: feature learning, neural networks, single-index model, gradient descent
Compressor summary: The paper studies how adding structure to the input data affects gradient-based learning of single index models and shows that appropriate normalization and exploiting the alignment between the input covariance and the target can improve performance.
Jieyu Zhang, Bohan Wang, Zhengyu Hu, Pang Wei Koh, Alexander Ratner
https://openreview.net/forum?id=Fe6fDq65aZ
Keywords: data-centric study, supervised pretraining, transfer learning
Compressor summary: The paper studies how the balance between intra-class and inter-class diversity in pre-training datasets affects downstream performance, and shows that the optimal class-to-sample ratio is independent of the pre-training dataset size.
Yihang Yao, Zuxin Liu, Zhepeng Cen, Jiacheng Zhu, Wenhao Yu, Tingnan Zhang, Ding Zhao
https://openreview.net/forum?id=FdtdjQpAwJ
Keywords: Safe Reinforcement Learning, Conditioned Reinforcement Learning, Multi-task Reinforcement Learning
Compressor summary: The paper proposes a new framework, CCPO, that learns safe policies in reinforcement learning that can adapt to changing safety constraints without retraining and achieve good performance.
Xiang Li, Chung-Ching Lin, Yinpeng Chen, Zicheng Liu, Jinglu Wang, Rita Singh, Bhiksha Raj
https://openreview.net/forum?id=FdsS51iif3
Keywords: Prompt-guided Segmentation, Generative models, Training-free
Compressor summary: PaintSeg is an unsupervised method that uses adversarial masked contrastive painting to create contrast between original and painted images, advancing the target segmentation mask towards ground truth without supervision or training.
Deepali Jain, Krzysztof Marcin Choromanski, Kumar Avinava Dubey, Sumeet Singh, Vikas Sindhwani, Tingnan Zhang, Jie Tan
https://openreview.net/forum?id=Fdfyga5i0A
Keywords: learnable optimizers, Transformers, efficient attention, spatio-temporal attention
Compressor summary: Mnemosyne is a new learnable optimizer that can train neural networks without task-specific tuning, performs well compared to popular LSTM optimizers, and has space complexity comparable to hand-designed optimizers.
Ivana Balazevic, David Steiner, Nikhil Parthasarathy, Relja Arandjelovic, Olivier J Henaff
https://openreview.net/forum?id=FasIQqsJhe
Keywords: transfer learning, adaptation, self-supervised learning, contrastive learning, scene understanding, representation learning, in-context learning, vision transformers
Compressor summary: The paragraph discusses a new approach called Hummingbird, which enables in-context learning for computer vision tasks such as semantic segmentation and depth estimation using nearest neighbor retrieval and attention-based pretraining.
Anand Bhattad, Daniel McKee, Derek Hoiem, David Forsyth
https://openreview.net/forum?id=FYqqvQdXhZ
Keywords: Generative models, StyleGAN, Depth, Normals, Segmentation, Intrinsic Images, Albedo, Shading
Compressor summary: This paper shows how StyleGAN can generate different types of intrinsic images easily and demonstrates its advantages over existing methods in terms of quality, robustness, and efficiency.
Jiayi Shen, Xiantong Zhen, Cheems Wang, Marcel Worring
https://openreview.net/forum?id=FXU4aR2uif
Keywords: data-insufficiency problem, episodic training, multi-task learning and neural processes
Compressor summary: The paper introduces HNPs, a method that uses meta-knowledge, heterogeneous information, and task-relatedness to adapt to novel tasks in episodic multi-task learning with limited data, and shows its superior performance over baselines.
David Simchi-Levi, Chonghuan Wang, Zeyu Zheng
https://openreview.net/forum?id=FV4ngfUlY0
Keywords: Adaptive Experimental Design, Non-stationary, Online Learning, Treatment Effect
Compressor summary: The paper proposes an efficient design for non-stationary experiments with linear trends, aiming to estimate the dynamic treatment effect and minimize welfare loss, while highlighting the challenge and trade-off between these objectives.
Minsoo Kim, Sihwa Lee, Janghwan Lee, Sukjin Hong, Du-Seong Chang, Wonyong Sung, Jungwook Choi
https://openreview.net/forum?id=FUnEkOkodU
Keywords: Generative Language Model, Quantization, QAT, Knowledge Distillation, Causal Attention, Language Modeling
Compressor summary: The paper proposes a new method for training large generative language models with less memory and maintaining accuracy in various tasks.
Khai Nguyen, Tongzheng Ren, Nhat Ho
https://openreview.net/forum?id=FT2q2B4cKZ
Keywords: Sliced Wasserstein, Generative Models, Optimal Transport
Compressor summary: The Markovian sliced Wasserstein (MSW) distance is a new family of sliced Wasserstein distances that imposes a first-order Markov structure on projecting directions, improving metricity, computational efficiency, and effectiveness over previous methods.
Syamantak Kumar, Purnamrita Sarkar
https://openreview.net/forum?id=FQGRkwmRzm
Keywords: Streaming PCA, Markov Chain, Mixing, Oja's algorithm
Compressor summary: The paper proposes a new method to estimate the top eigenvector of a covariance matrix using Oja's algorithm on data streams sampled from a Markov chain without downsampling.
Rongqing Li, Changsheng Li, Dongchun Ren, Guangyi Chen, Ye Yuan, Guoren Wang
https://openreview.net/forum?id=FOFJmR1oxt
Keywords: Trajectory prediction, instantaneous observation
Compressor summary: BCDiff is a novel framework for predicting pedestrian trajectories from instantaneous observations, using two coupled diffusion models that mutually guide each other to generate more accurate predictions.
Seunghyuk Cho, Juyong Lee, Dongwoo Kim
https://openreview.net/forum?id=FNn4zibGvw
Keywords: Hyperbolic space, VAE, Distribution on hyperbolic space, Hierarchical representation learning, Reinforcement Learning
Compressor summary: The GM-VAE is a new model that uses Gaussian distributions to represent data in a latent space shaped like a hyperbolic space, which improves density estimation and reinforcement learning tasks.
Sepideh Mahabadi, Stojan Trajanovski
https://openreview.net/forum?id=FM8thAWqiO
Keywords: Constrained Diversity Maximization, Fairness, Data Summarization, Core-sets, Approximation Algorithms
Compressor summary: The paragraph discusses algorithms that maximize diversity in partitioned sets under fairness constraints and shows their effectiveness in summarizing timed messages for a communication platform.
Royson Lee, Minyoung Kim, Da Li, Xinchi Qiu, Timothy Hospedales, Ferenc Huszár, Nicholas Donald Lane
https://openreview.net/forum?id=FM81CI68Iz
Keywords: federated learning; meta-learning; hyperparameter optimization
Compressor summary: The paper proposes a federated meta-learning approach to learn personalization strategies for different clients in federated learning problems, using meta-nets to adjust batch-norm and learning rate parameters based on local data statistics.
Aaron Lou, Minkai Xu, Adam Farris, Stefano Ermon
https://openreview.net/forum?id=FLTg8uA5xI
Keywords: Diffusion Models, Geometric Deep Learning, Manifolds, Numerical Algorithms
Compressor summary: The authors improve Riemannian diffusion models by leveraging symmetry and computationally efficient approximations, enabling their application to high dimensional tasks and reducing representation collapse.
Basile Confavreux, Poornima Ramesh, Pedro J. Goncalves, Jakob H. Macke, Tim P. Vogels
https://openreview.net/forum?id=FLFasCFJNo
Keywords: synaptic plasticity, spiking network, meta-learning, computational neuroscience
Compressor summary: The authors develop a simulation-based inference method (filter SBI) that allows them to infer complex and co-active plasticity rules in spiking networks, which can be used for deeper insights into brain function.
Theo Gruner, Boris Belousov, Fabio Muratore, Daniel Palenicek, Jan Peters
https://openreview.net/forum?id=FIv84qGPFT
Keywords: simulation-based inference, approximate Bayesian computation
Compressor summary: Pseudo-Likelihood Inference (PLI) is a new method that combines neural approximation with Approximate Bayesian Computation (ABC) to perform better on high-dimensional Bayesian system identification tasks, especially when more data is available.
Aditya Vardhan Varre, Maria-Luiza Vladarean, Loucas Pillaud-Vivien, Nicolas Flammarion
https://openreview.net/forum?id=FFdrXkm3Cz
Keywords: linear networks, spectral bias, low rank, singular values, mirror flow
Compressor summary: The paper analyzes how two-layer neural networks with linear activations and gradient flow training behave differently depending on initialization scale, revealing that low-rank structures arise in small scale regimes.
Xiaoyu Liu, Jiaxin Yuan, Bang An, Yuancheng Xu, Yifan Yang, Furong Huang
https://openreview.net/forum?id=FFOYWUpBca
Keywords: causal disentanglement, causal generative process, generative factors, confounder, inductive bias, disentanglement, causal inference
Compressor summary: The paper proposes C-Disentanglement, a framework that incorporates expert knowledge to identify causally disentangled factors in latent space, improving data generation and generalization.
Russell Tsuchida, Cheng Soon Ong, Dino Sejdinovic
https://openreview.net/forum?id=FDzQQTPqEJ
Keywords: probabilistic modelling; density estimation; exponential family;
Compressor summary: SNEFYs are a new type of probability distributions that can model complex data with flexible and tractable density models, and are applicable to many machine learning problems.
Jianyi Yang, Pengfei Li, Tongxin Li, Adam Wierman, Shaolei Ren
https://openreview.net/forum?id=FCwfZj1bQl
Keywords: Markov Decision Process, Constrained Reinforcement Learning, Anytime Competitive Constraints
Compressor summary: The paper introduces ACRL, a new algorithm for A-CMDP that guarantees cost bounds and optimizes expected reward, with experiments showing its effectiveness in carbon-intelligent computing.
Saptarshi Roy, Raymond K. W. Wong, Yang Ni
https://openreview.net/forum?id=FCwF5431IY
Keywords: Causal Embedding, Causal Discovery, Multivariate Functional Data, Directed Cyclic Graph, Causal Structure Learning, Bayesian Inference
Compressor summary: The article presents a functional linear structural equation model for learning causal structures from multivariate functional data, which preserves causal information in a low-dimensional space and can be inferred using a Bayesian framework.
Talia Konkle, George A. Alvarez
https://openreview.net/forum?id=FCIj5KMn2m
Keywords: convolutional neural networks, steerability, computer vision
Compressor summary: The paragraph discusses how adding long-range modulatory pathways to vision models can improve image recognition, robustness, and alignment, as well as allow for goal-directed visual encoding using output space vectors.
Michael Scherbela, Leon Gerard, Philipp Grohs
https://openreview.net/forum?id=FBNyccPfAu
Keywords: Computational Physics, Machine Learning for Science, Quantum Monte Carlo, Fermionic Neural Networks
Compressor summary: The authors propose a deep learning-based method for quantum chemistry that uses pre-training and fine-tuning to achieve high accuracy with less computation than conventional methods.
Jiachen T. Wang, Yuqing Zhu, Yu-Xiang Wang, Ruoxi Jia, Prateek Mittal
https://openreview.net/forum?id=FAZ3i0hvm0
Keywords: Data Valuation, Differential Privacy
Compressor summary: The paper introduces TKNN-Shapley, a privacy-friendly variant of KNN-Shapley, which is a method for data valuation in machine learning models, and shows that it offers better trade-offs between privacy and utility than the original method.
Haobo Wang, Yiwen Dong, Ruixuan Xiao, Fei Huang, Gang Chen, Junbo Zhao
https://openreview.net/forum?id=FAGY52HbyV
Keywords: Distant Supervision; Named Entity-Recognition; Biased Learning
Compressor summary: DesERT is a novel self-training framework that addresses two types of biases in distant supervision for named entity recognition and achieves state-of-the-art performance on five benchmark datasets.
Sijia Zhou, Yunwen Lei, Ata Kaban
https://openreview.net/forum?id=F6j16Qr6Vk
Keywords: PAC-Bayesian Bounds, Uniform Stability, Generalization Analysis
Compressor summary: The paper improves bounds for stable randomized algorithms using concentration of weakly dependent variables and introduces a sub-exponential stability parameter assumption, applicable to stochastic gradient descent and randomized coordinate descent.
Yuheng Ma, Han Zhang, Yuchao Cai, Hanfang Yang
https://openreview.net/forum?id=F5FVsfCxt8
Keywords: Local differential privacy, non-parametric regression, decision tree, public data
Compressor summary: The paper introduces a new algorithm called Locally differentially Private Decision Tree (LPDT) for enhancing private estimation using public data, which improves convergence rates and performance compared to existing methods.
Shibal Ibrahim, Gabriel Isaac Afriat, Kayhan Behdin, Rahul Mazumder
https://openreview.net/forum?id=F5DYsAc7Rt
Keywords: Generalized additive models, component selection, hierarchy, interpretability
Compressor summary: The paper introduces GRAND-SLAMIN, a framework for learning generalized additive models (GAMs) with sparse interactions and structural constraints using differentiable optimization and novel prediction bounds.
João B. S. Carvalho, Mengtao Zhang, Robin Geyer, Carlos Cotrini, Joachim M. Buhmann
https://openreview.net/forum?id=F1mv2L7Rkb
Keywords: anomaly detection, causal inference, distribution shifts
Compressor summary: The paper proposes a new regularization term for anomaly detection that improves its performance under distribution shifts, using tools from causal inference and ensuring invariant representations.
Josephine Lamp, Mark Derdzinski, Christopher Hannemann, Joost Van der Linden, Lu Feng, Tianhao Wang, David Evans
https://openreview.net/forum?id=Eysb8t3MJ5
Keywords: Synthetic Data, Time Series, Generative Adversarial Networks, Differential Privacy, Glucose, Diabetes
Compressor summary: The paper introduces GlucoSynth, a novel GAN framework that generates private and high-quality synthetic glucose traces by preserving the relationships among glucose events and temporal dynamics.
Runa Eschenhagen, Alexander Immer, Richard E Turner, Frank Schneider, Philipp Hennig
https://openreview.net/forum?id=Ex3oJEKS53
Keywords: deep learning, second-order, optimization, natural gradient, fisher, gauss-newton, k-fac, weight-sharing
Compressor summary: The paper proposes two flavors of K-FAC for linear weight-sharing layers in neural networks and shows that they can speed up training and reduce computational costs, especially for deep vision tasks.
Dayong Ren, Zhe Ma, Yuanpei Chen, Weihang Peng, Xiaode Liu, Yuhan Zhang, Yufei Guo
https://openreview.net/forum?id=Ev2XuqvJCy
Keywords: Spiking Neural Networks, Point Clouds
Compressor summary: The paper introduces Spiking PointNet, a spiking neural network model for efficient 3D point cloud recognition, overcoming the challenges of SNNs' optimization and PointNet's memory and computation cost.
Minghua Liu, Ruoxi Shi, Kaiming Kuang, Yinhao Zhu, Xuanlin Li, Shizhong Han, Hong Cai, Fatih Porikli, Hao Su
https://openreview.net/forum?id=Eu4Kkefq7p
Keywords: 3d, shape understanding, open-world understanding, zero-shot 3D classification, vision-language model
Compressor summary: OpenShape is a method that learns joint representations of text, image, and point clouds to enable open-world 3D shape understanding and achieve superior zero-shot classification performance on several benchmarks.
Alexia Atsidakou, Branislav Kveton, Sumeet Katariya, Constantine Caramanis, sujay sanghavi
https://openreview.net/forum?id=ErAP8kF4tG
Keywords: Bayesian bandits, logarithmic regret bounds, multi-armed bandits, linear bandits
Compressor summary: The paper presents new upper bounds for finite-time logarithmic Bayes regret in Bayesian bandits and applies simple and general techniques to linear bandits.
Yongqiang Chen, Yatao Bian, Kaiwen Zhou, Binghui Xie, Bo Han, James Cheng
https://openreview.net/forum?id=EqpR9Vtt13
Keywords: Graph Neural Networks, Out-of-Distribution Generalization, Invariant Learning
Compressor summary: The paper proposes a new framework, GALA, to learn invariant graph representations by incorporating an assistant model that detects environment changes and identifies maximally invariant subgraphs for out-of-distribution generalization.
Arnaud Robert, Ciara Pike-Burke, Aldo A. Faisal
https://openreview.net/forum?id=EqnZqrbFrc
Keywords: Hierarchical Reinforcement Learning, Sample Complexity
Compressor summary: The paper derives a lower bound on the sample complexity for goal-conditioned Hierarchical Reinforcement Learning (HRL) algorithms and proposes a new algorithm that leverages hierarchical decompositions, which is empirically validated on various tasks.
David Durfee
https://openreview.net/forum?id=Eq9AFZlAjt
Keywords: Differential privacy, Theory, Spars Vector Technique, Quantile
Compressor summary: The paper presents an efficient and accurate method for computing differential privacy using $\texttt{AboveThreshold}$ subroutine on unbounded data with improved privacy guarantees.
Yuchen Yan, Baoyu Jing, Lihui Liu, Ruijie Wang, Jinning Li, Tarek Abdelzaher, Hanghang Tong
https://openreview.net/forum?id=EoDpq18R30
Keywords: Network embedding
Compressor summary: The paper studies the trade-off between discrimination and monotonicity properties for node embeddings under different sampling strategies and proposes a new model (SENSEI) that improves network embedding performance.
John Isak Texas Falk, Luigi Bonati, Pietro Novelli, Michele Parrinello, massimiliano pontil
https://openreview.net/forum?id=Enzew8XujO
Keywords: GNN, Mean Embedding, Kernels, Atomistic Simulations, OCP, Transfer Learning, Molecular Dynamics, Kernel Ridge Regression, Neural Networks
Compressor summary: The authors propose a method to learn interatomic potentials using graph neural networks and kernel mean embeddings, which improves accuracy and interpretability for atomistic simulations of catalytic processes.
Francis Rhys Ward, Francesca Toni, Francesco Belardinelli, Tom Everitt
https://openreview.net/forum?id=EmxpDiPgRu
Keywords: Deception, Causality, Game Theory
Compressor summary: The paper introduces a formal definition of deception in structural causal games, based on philosophy literature, and provides graphical criteria for detecting and mitigating deception in AI systems like reinforcement learners and language models.
Mhairi Dunion, Trevor McInroe, Kevin Sebastian Luck, Josiah P. Hanna, Stefano V Albrecht
https://openreview.net/forum?id=EmYWJsyad4
Keywords: Reinforcement Learning, Representation Learning, Disentanglement
Compressor summary: The paper proposes a method to teach reinforcement learning agents to separate and represent correlated features in their latent representation, which improves their ability to generalize when the environment changes or when deployed in the real world.
Yasheng SUN, Yifan Yang, Houwen Peng, Yifei Shen, Yuqing Yang, Han Hu, Lili Qiu, Hideki Koike
https://openreview.net/forum?id=EmOIP3t9nk
Keywords: Image Manipulation, Visual Instruction
Compressor summary: The paper proposes ImageBrush, a novel image manipulation method that learns visual instructions from transformation images to accurately edit images without using language descriptions.
Peter Hase, Mohit Bansal, Been Kim, Asma Ghandeharioun
https://openreview.net/forum?id=EldbUlZtbd
Keywords: localization, model editing, mechanistic interpretability, language models
Compressor summary: The paper finds that editing facts in language models does not always follow the localization conclusions from representation denoising and suggests that a better understanding of model behavior may not be helpful for editing tasks.
Huanjing Yue, Yijia Cheng, Xin Liu, Jingyu Yang
https://openreview.net/forum?id=EkcO9tHm6S
Keywords: Raw image demoiréing, raw video demoiréing, video demoiréing dataset
Compressor summary: The authors propose a new network for removing moiré patterns from raw images and videos, which outperforms existing methods and introduce a new dataset with an alignment method.
Chaoran Cheng, Jian Peng
https://openreview.net/forum?id=EjiA3uWpnc
Keywords: Neural Operator Learning, Spectral Graph Theory, Graphon
Compressor summary: The paragraph describes a new model called InfGCN that uses a combination of continuous and discrete graph structures to learn mappings between 3D functions while preserving equivariance, and shows it performs better than existing architectures on electron density datasets.
Leon Klein, Andrew Y. K. Foong, Tor Erlend Fjelde, Bruno Kacper Mlodozeniec, Marc Brockschmidt, Sebastian Nowozin, Frank Noe, Ryota Tomioka
https://openreview.net/forum?id=EjMLpTgvKH
Keywords: Molecular Dynamics, Normalizing Flows, MCMC
Compressor summary: Timewarp is a method that uses a flow trained on MD trajectories to make large steps in time, enabling efficient simulation of processes like binding and folding over longer timescales for different molecular systems.
Matheus Aparecido Do Carmo Alves, Amokh Varma, Yehia Elkhatib, Leandro Soriano Marcolino
https://openreview.net/forum?id=EjG2G1PT2v
Keywords: Information-guided planning, Planning under uncertainty, Sequential decision making
Compressor summary: IB-POMCP is an online planning algorithm that uses entropy estimates of the world belief to guide a tree search process and improve decision-making in partially observable environments, outperforming existing methods in reward, time, and convergence.
Changyu Chen, Ramesha Karunasena, Thanh Hong Nguyen, Arunesh Sinha, Pradeep Varakantham
https://openreview.net/forum?id=Ehzj9F2Kmj
Keywords: Action constrained reinforcement learning, Normalizing flow, Generative modelling
Compressor summary: The authors propose a method for solving large discrete action space problems in Reinforcement Learning with validity constraints by using a conditional normalizing flow to represent a stochastic policy and an invalid action rejection technique to update the base policy.
Siddhant Agarwal, Ishan Durugkar, Peter Stone, Amy Zhang
https://openreview.net/forum?id=EhhPtGsVAv
Keywords: Goal Conditioned Reinforcement Learning, Shaping Rewards, Reward Design
Compressor summary: The paper proposes $f$-Policy Gradients, a method to encourage exploration in sparse reward RL problems by minimizing the f-divergence between the agent's state visitation distribution and the goal, which can lead to an optimal policy.
Xiaosong Ma, Jie ZHANG, Song Guo, Wenchao Xu
https://openreview.net/forum?id=EhdNQiOWgQ
Keywords: Test-Time Adaptation, Prompt Learning, Unsupervised Representation Learning
Compressor summary: SwapPrompt is a novel framework for test-time adaptation that leverages self-supervised contrastive learning to improve the performance of pre-trained vision-language models on unseen test domains, achieving state-of-the-art results and comparable performance with supervised methods.
Stephan Rabanser, Anvith Thudi, Abhradeep Guha Thakurta, Krishnamurthy Dj Dvijotham, Nicolas Papernot
https://openreview.net/forum?id=EgCjf1vjMB
Keywords: differential privacy, selective classification, selective prediction, abstain option, reject option, uncertainty quantification, misclassification detection
Compressor summary: The authors investigate selective classifiers under differential privacy and find that some methods are ineffective due to increased privacy leakage, while a recent approach works well; however, performance degrades with lower privacy levels.
Ramy Mounir, Sujal Vijayaraghavan, Sudeep Sarkar
https://openreview.net/forum?id=EfTMRQn00d
Keywords: predictive learning, hierarchical event segmentation, self-supervised learning, streaming processing, perceptual inputs, biologically-plausible.
Compressor summary: The paragraph describes STREAMER, a self-supervised architecture for hierarchical representation learning and segmentation of streaming inputs, which is evaluated on the egocentric EPIC-KITCHENS dataset for temporal event segmentation.
Sehoon Kim, Karttikeya Mangalam, Suhong Moon, Jitendra Malik, Michael W. Mahoney, Amir Gholami, Kurt Keutzer
https://openreview.net/forum?id=EfMyf9MC3t
Keywords: Transformer, efficient inference, efficient model, decoding
Compressor summary: BiLD improves text generation efficiency by using a small model for autoregressive tasks and a large model for refining predictions non-autoregressively, with two policies to coordinate them.
In Huh, changwook jeong, Jae Myung Choe, Young-Gu Kim, Dae Sin Kim
https://openreview.net/forum?id=EdgPb3ngR4
Keywords: representation learning, auto-encoders, geometry, symmetry
Compressor summary: The paper introduces IQVAEs, a novel auto-encoding framework that learns symmetry-preserving representations of data manifolds embedded in high-dimensional spaces using variational auto-encoders.
Hongyi Yuan, Zheng Yuan, Chuanqi Tan, Wei Wang, Songfang Huang, Fei Huang
https://openreview.net/forum?id=EdIGMCHk4l
Keywords: Large Language Model, Human Alignment
Compressor summary: RRHF is a novel learning paradigm for aligning large language models with human preferences using ranking loss and sampling responses from various sources, outperforming PPO in simplicity, efficiency, and performance.
Peiwen Yuan, Xinglin Wang, Jiayi Shi, Bin Sun, Yiwei Li, Kan Li
https://openreview.net/forum?id=Ecv1GMiXSk
Keywords: Natural language process, Automatic dialog evaluation
Compressor summary: BCR is a distribution-balanced self-supervised learning framework that improves the correlation and robustness of turn-level dialogue evaluation models using coherence-balanced training signals and a novel adaptive loss function.
Ziba Parsons, Fei Dou, Houyi Du, Zheng Song, Jin Lu
https://openreview.net/forum?id=EcmqyXekuP
Keywords: Mobilized Federated Networks, Personalized Federated Learning, Random Walk, Stochastic ADMM
Compressor summary: The paper proposes a novel FL approach for isolated nodes with data heterogeneity and wireless links, using Random Walk SADMM optimization algorithm that improves convergence, accuracy, and communication efficiency.
Ting Li, Chengchun Shi, Jianing Wang, Fan Zhou, Hongtu Zhu
https://openreview.net/forum?id=EcReRm7q9p
Keywords: Average treatment effect, Experimental design, Off-policy evaluation, Optimal treatment allocation
Compressor summary: The paper proposes three optimal allocation strategies for A/B testing in dynamic settings to minimize variance and improve accuracy of treatment effect estimators using off-policy evaluation methods.
Lu Mi, Trung Le, Tianxing He, Eli Shlizerman, Uygar Sümbül
https://openreview.net/forum?id=EcN3l6Xmnx
Keywords: population dynamics, neuronal representation, calcium imaging, cell types
Compressor summary: The authors propose a self-supervised learning method to assign stable representations to neurons based on population activity and improve predictions of gene expression and neuronal identity.
Tianxiao Li, Hongyu Guo, Filippo Grazioli, Mark Gerstein, Martin Renqiang Min
https://openreview.net/forum?id=Eb74zfBkWa
Keywords: protein engineering, disentangled representation, T cell receptor
Compressor summary: The proposed method uses a Wasserstein autoencoder to automatically identify functional residues in proteins and edit them without affecting the overall structure, improving efficiency and quality for protein engineering applications like T-cell receptor modification.
Chung-Wei Lee, Qinghua Liu, Yasin Abbasi-Yadkori, Chi Jin, Tor Lattimore, Csaba Szepesvari
https://openreview.net/forum?id=EY7Hpj8Ok6
Keywords: Contextual bandits, low-rank bandits, latent bandits, clustering bandits, stochastic bandit problems, context-lumpable bandits
Compressor summary: The paper proposes an algorithm for a contextual bandit problem where the learner groups the contexts by similarity and achieves near-optimal sample complexity and minimax regret.
Ryan Sullivan, Akarsh Kumar, Shengyi Huang, John P Dickerson, Joseph Suarez
https://openreview.net/forum?id=EY4OHikuBm
Keywords: Reinforcement Learning, Proximal Policy Optimization, Reward Normalization
Compressor summary: The paper applies DreamerV3's tricks to PPO and shows that they don't generally improve PPO, but have some specific cases where they work well.
Tackgeun You, Mijeong Kim, Jungtaek Kim, Bohyung Han
https://openreview.net/forum?id=EWNtYvepJh
Keywords: generative neural fields; implicit neural representation; model averaging
Compressor summary: The proposed method learns implicit neural representations and their coefficients to enlarge the capacity of generative neural fields, improving efficiency and effectiveness in generating diverse data.
Travis Dick, Jennifer Gillenwater, Matthew Joseph
https://openreview.net/forum?id=EUiIbwV379
Keywords: differential privacy, linear regression, sparse, feature selection, kendall
Compressor summary: The authors propose a differentially private feature selection method based on Kendall rank correlation to broaden the applicability of private linear regression algorithms in high-dimensional problems.
Ziyi Wu, Jingyu Hu, Wuyue Lu, Igor Gilitschenski, Animesh Garg
https://openreview.net/forum?id=ETk6cfS3vk
Keywords: Unsupervised object-centric learning, diffusion model, generative modeling
Compressor summary: SlotDiffusion is a new object-centric Latent Diffusion Model that improves visual generation and object manipulation in images and videos by enhancing slot-to-image decoding.
Chaofan Ma, Yuhuan Yang, Chen Ju, Fei Zhang, Ya Zhang, Yanfeng Wang
https://openreview.net/forum?id=ESEM1lNoeS
Keywords: Open-Vocabulary Semantic Segmentation, Attributes, Decomposition and Aggregation
Compressor summary: The paragraph describes a novel method for semantic segmentation that uses attribute decomposition and aggregation to handle ambiguous or incomplete textual category names, inspired by human cognition.
Zhaoxi Chen, Fangzhou Hong, Haiyi Mei, Guangcong Wang, Lei Yang, Ziwei Liu
https://openreview.net/forum?id=ESCafo3oD5
Keywords: neural rendering, 3D generative model, diffusion model, volumetric primitives, 3D human generation
Compressor summary: PrimDiffusion is a diffusion-based framework for generating 3D humans using volumetric primitives, which enables efficient, high-quality, and flexible rendering.
Yuzhong Wang, Xu Han, Weilin Zhao, Guoyang Zeng, Zhiyuan Liu, Maosong Sun
https://openreview.net/forum?id=ES32O8mBK3
Keywords: ML System, Parallelism Learning, Memory Optimization, Data Parallelism, Model Parallelism, Parameter Parallelism, ZeRO, Rematerialization, Checkpointing, Tensor Offloading, Dynamic Programming
Compressor summary: The paper proposes a framework called H3T that automatically integrates memory optimization and parallelism to improve the efficiency of training large Transformer-based models, achieving significant speedups and reduced memory overhead.
Nazarii Tupitsa, Abdulla Jasem Almansoori, Yanlin Wu, Martin Takáč, Karthik Nandakumar, Samuel Horváth, Eduard Gorbunov
https://openreview.net/forum?id=ER0bcYXvvo
Keywords: byzantine robustness, variational inequalities, min-max problems
Compressor summary: The authors present new methods for solving Byzantine-robust distributed variational inequalities, improving upon existing work and providing empirical evidence.
Yuancheng Wang, Zeqian Ju, Xu Tan, Lei He, Zhizheng Wu, Jiang Bian, sheng zhao
https://openreview.net/forum?id=EO1KuHoR0V
Keywords: audio editing, text-to-audio generation, diffusion models
Compressor summary: AUDIT is an instruction-guided audio editing model that uses diffusion and denoising processes, triplet training data, and only needs edit instructions as input to achieve state-of-the-art results in various audio editing tasks.
Kiarash Zahirnia, Yaochen Hu, Mark Coates, Oliver Schulte
https://openreview.net/forum?id=EI6BHFKA5p
Keywords: Graph Generation, Local Differential Privacy, Graph Statistics, Latent Adjacency Matrix
Compressor summary: The text describes a new method for learning a deep graph generative model from aggregate statistics without using the graph adjacency matrix, which can protect privacy and generate realistic graphs.
Junmin Zhong, Ruofan Wu, Jennie Si
https://openreview.net/forum?id=EGfYnTyEGv
Keywords: Deep reinforcement learning, Reward Estimation
Compressor summary: The paper introduces a new stage reward estimator, LNSS, for deep RL that mitigates high variance problem and improves performance in continuous control tasks using various baseline algorithms.
Jiawei Lin, Jiaqi Guo, Shizhao Sun, Zijiang James Yang, Jian-Guang Lou, Dongmei Zhang
https://openreview.net/forum?id=EF56cv8B3b
Keywords: graphic design, graphic layout, large language models, in-context learning
Compressor summary: The paper introduces LayoutPrompter, which uses large language models to generate layouts efficiently and versatilely without training or fine-tuning.
Guy Blanc, Jane Lange, Chirag Pabbaraju, Colin Sullivan, Li-Yang Tan, Mo Tiwari
https://openreview.net/forum?id=EEtJTfvNZx
Keywords: Decision Trees, Decision Tree Learning, Top-$k$, ID3, Greedy Algorithms
Compressor summary: Top-$k$ is a simple generalization of decision tree learning algorithms that considers multiple best attributes for splitting and shows improved accuracy, scalability, and applicability compared to existing methods.
Ben Chugg, Santiago Cortes-Gomez, Bryan Wilder, Aaditya Ramdas
https://openreview.net/forum?id=EEVpt3dJQj
Keywords: fairness, auditing, sequential analysis, martingales, testing by betting
Compressor summary: The authors propose efficient, sequential, and probabilistic methods for auditing the fairness of classification and regression models in real-world systems using inferential and game-theoretic frameworks.
Qianli Shen, Wai Hoh Tang, Zhun Deng, Apostolos Psaros, Kenji Kawaguchi
https://openreview.net/forum?id=EETqXXdqkI
Keywords: physics-informed learning, uncertainty quantification, deep learning
Compressor summary: The paper presents PICProp, a bi-level optimization method for estimating confidence intervals in physics-informed learning without strong assumptions.
Shinji Ito, Daisuke Hatano, Hanna Sumita, Kei Takemura, Takuro Fukunaga, Naonori Kakimura, Ken-Ichi Kawarabayashi
https://openreview.net/forum?id=EE1Uiu3Ryb
Keywords: bandit, combinatorial semi-bandits, bandits with badget
Compressor summary: The paper introduces a new bandit task assignment problem with processing time and combinatorial constraints, and proposes an UCB-based algorithm with phased updates that achieves near-optimal regret bounds.
Ying Wang, Tim G. J. Rudner, Andrew Gordon Wilson
https://openreview.net/forum?id=ECvtxmVP0x
Keywords: Interpretability, Attribution Maps, Information Bottleneck, Multi-Modal Learning, Vision-Language Pretrained Models
Compressor summary: M2IB is a method to improve interpretability of vision-language models like CLIP by compressing irrelevant information and preserving relevant features, without requiring ground truth labels.
Hao Li, Jingkuan Song, Lianli Gao, Xiaosu Zhu, Heng Tao Shen
https://openreview.net/forum?id=ECRgBK6sk1
Keywords: multimodal learning, cross-modal retrieval, robust learning, uncertainty
Compressor summary: The paper introduces a new framework to quantify and reduce uncertainty in cross-modal retrieval methods by using learnable prototypes and Subjective Logic Theory.
Wenjing YAN, Xuanyu Cao
https://openreview.net/forum?id=ECBK3TVmZl
Keywords: Performative prediction, decision-dependent distribution, inequality constraints, primal-dual algorithm.
Compressor summary: The paper studies how to optimize performative prediction problems with inequality constraints using a robust primal-dual framework and proposes an adaptive algorithm that achieves low regret and constraint violations.
Bohan Zhou, Ke Li, Jiechuan Jiang, Zongqing Lu
https://openreview.net/forum?id=E58gaxJN1d
Keywords: Learning from Observations, Offline Learning from Visual Observations, State-to-Go Transformer
Compressor summary: The paper proposes a two-stage framework that uses a pretrained State-to-Go Transformer to predict latent transitions from visual observation and provide intrinsic rewards for reinforcement learning tasks, achieving performance comparable to policy learned from environmental rewards in some cases.
Yicheng Li, Haobo Zhang, Qian Lin
https://openreview.net/forum?id=E4P5kVSKlT
Keywords: generalization, reproducing kernel Hilbert space, bias-variance trade-off
Compressor summary: The paper studies the learning curve of kernel ridge regression under realistic assumptions and shows how it relates to the choice of parameters, source condition, and noise, with implications for the benign overfitting phenomenon in over-parametrized neural networks.
Lauren E Conger, Franca Hoffman, Eric Mazumdar, Lillian J Ratliff
https://openreview.net/forum?id=E3ZUEaeFYS
Keywords: distribution shift, partial differential equations
Compressor summary: The paper proposes a new framework to analyze how learning algorithms and their environments influence each other over time, accounting for complex dynamics and proving its convergence in different settings.
Xiao Zhang, Ninglu Shao, Zihua Si, Jun Xu, Wenhan Wang, Hanjing Su, Ji-Rong Wen
https://openreview.net/forum?id=E2zoGTkTbW
Keywords: batched bandit, sketching, reward imputation, regret bound, ridge regression
Compressor summary: SPUIR is an efficient batched bandit algorithm that imputes unobserved rewards using sketching and achieves low regret bounds in both theoretical and practical settings.
Min Wu, Haoze Wu, Clark Barrett
https://openreview.net/forum?id=E2TJI6CKm0
Keywords: trustworthy machine learning, deep neural networks, explainability, interpretability, formal methods, automated verification
Compressor summary: VeriX is a system that creates optimal robust explanations and counterfactuals for machine learning models using constraint solving and feature-level sensitivity ranking.
Xidong Wu, Jianhui Sun, Zhengmian Hu, Junyi Li, Aidong Zhang, Heng Huang
https://openreview.net/forum?id=E0Gw1uz7lU
Keywords: Federated Learning, Conditional Stochastic Optimization, Nonconvex Optimization
Compressor summary: The paper introduces FCSG and FCSG-M, two new federated learning algorithms for nonconvex conditional stochastic optimization, and Acc-FCSG-M, an accelerated version with lower complexity, which are tested and validated on various tasks.
Michael Bereket, Theofanis Karaletsos
https://openreview.net/forum?id=DzaCE00jGV
Keywords: Disentagled representation learning, VAE, generative models, sparse mechanism shift, perturbation modeling, cellular modeling
Compressor summary: SAMS-VAE is a novel generative model that combines compositionality, disentanglement, and interpretability for perturbation models in drug discovery, and shows strong performance and correlations to known biological mechanisms.
Yipeng Li, Xinchen Lyu
https://openreview.net/forum?id=Dxhv8Oja2V
Keywords: Federated Learning, Convergence analysis
Compressor summary: The paper analyzes the convergence theory of sequential federated learning (SFL) for non-convex objectives on heterogeneous data and shows that SFL performs better than parallel federated learning (PFL) in cross-device settings.
Shai Ben-David, Alex Bie, Gautam Kamath, Tosca Lechner
https://openreview.net/forum?id=Dx99y3okbL
Keywords: robustness, distribution learning
Compressor summary: The paper investigates how different types of contamination affect the relationship between learnability and robust learnability in distribution learning, and discusses implications for other settings like compression and privacy.
Joshua Southern, Jeremy Wayland, Michael M. Bronstein, Bastian Rieck
https://openreview.net/forum?id=Dt71xKyabn
Keywords: Curvature, topology, persistent homology, graph learning, generative model, machine learning, geometric deep learning
Compressor summary: The authors propose using graph curvature and topological data analysis to create effective ways of assessing graph generative models.
Connor Toups, Rishi Bommasani, Kathleen Creel, Sarah H Bana, Dan Jurafsky, Percy Liang
https://openreview.net/forum?id=Ds8iLujo3p
Keywords: homogenous outcomes, societal impact of ML, deployed ML, systemic failure
Compressor summary: The authors propose a new approach called ecosystem-level analysis to study the societal impact of machine learning models by considering their collective effects on users, and they apply it to three modalities (text, images, speech) and find that systemic failure is common and not reduced by individual model improvements.
Leyla Biabani, Annika Hennes, Morteza Monemizadeh, Melanie Schmidt
https://openreview.net/forum?id=Ds7Vd83HlC
Keywords: $k$-center clustering, outliers, dynamic algorithms
Compressor summary: The paragraph describes a problem of finding the optimal set of centers for a point cloud with some outliers, and presents an efficient data structure that can handle dynamic updates and provide approximate solutions in sublinear time.
Amirhossein Kazemnejad, Inkit Padhi, Karthikeyan Natesan, Payel Das, Siva Reddy
https://openreview.net/forum?id=Drrl2gcjzl
Keywords: Transformers, Positional Encoding, Length Generalization
Compressor summary: The paper compares five positional encoding methods for decoder-only Transformers and finds that no explicit encoding is better than others for length generalization in downstream tasks.
Yangdi Jiang, Xiaotian Chang, Yi Liu, Lei Ding, Linglong Kong, Bei Jiang
https://openreview.net/forum?id=DrIZZwEZtM
Keywords: Differential Privacy, Gaussian Differential Privacy, Differential Geometry, Riemannian Manifold, Homogeneous Riemannian Manifold, Frechet Mean
Compressor summary: The authors develop a new method for preserving privacy in data analysis on curved spaces using a special type of probability distribution that accounts for the geometry of the space.
Carsten Tim Lüth, Till J. Bungert, Lukas Klein, Paul F Jaeger
https://openreview.net/forum?id=Dqn715Txgl
Keywords: Active Learning, Evaluation, Study
Compressor summary: The paper discusses the challenges of evaluating active learning methods, proposes a framework to address these issues, and presents an empirical study to demonstrate its usefulness.
Mengfan Xu, Diego Klabjan
https://openreview.net/forum?id=DqfdhM64LI
Keywords: decentralized multi-agent MAB, heterogeneous light-tailed and heavy-tailed rewards, time dependent random graphs
Compressor summary: The paper proposes a novel algorithm for decentralized multi-agent bandit problems with time-varying graphs and varying rewards, achieving near-optimal regret bounds with high probability.
Matthias Gerstgrasser, Tom Danino, Sarah Keren
https://openreview.net/forum?id=DpuphOgJqh
Keywords: multi-agent reinforcement learning, reinforcement learning, deep q learning, cooperative ai
Compressor summary: Selective Multi-Agent Prioritized Experience Relay is a new method for multi-agent reinforcement learning where agents share a small number of relevant experiences to improve learning efficiency and performance.
Ahmad Chamma, Denis Engemann, Bertrand Thirion
https://openreview.net/forum?id=DoE3esTIEM
Keywords: Interpretability, Variable Importance, Machine Learning, Deep Learning, Statistical Inference
Compressor summary: The authors propose CPI, a new method to assess variable importance in machine learning models, which overcomes the limitations of standard permutation importance by providing accurate type-I error control and performs well on benchmarks and real-world data analysis.
Ziyu Chen, Wei Zhu
https://openreview.net/forum?id=DnVjDRLwVu
Keywords: implicit bias, equivariant steerable networks, data augmentation, margin, generalization bound
Compressor summary: The text discusses how gradient flow helps linear equivariant steerable networks find group-invariant classifiers with a maximum margin, and shows that they are equivalent to data augmentation under certain conditions.
Mircea Petrache, Shubhendu Trivedi
https://openreview.net/forum?id=DnO6LTQ77U
Keywords: Equivariance, Invariance, Generalization, Equivariant Neural Networks, Approximation Error
Compressor summary: The paper investigates how incorporating task-specific symmetries in machine learning models improves generalization and how to handle approximate or partial symmetries between models and data distributions.
Gang Liu, Eric Inae, Tong Zhao, Jiaxin Xu, Tengfei Luo, Meng Jiang
https://openreview.net/forum?id=DmakwvCJ7l
Keywords: Graph property prediction, Molecular property prediction, Diffusion model, Unlabeled data, Data augmentation, Transfer learning
Compressor summary: The paper proposes a method to use unlabeled graphs for property prediction tasks by generating task-specific graph examples and labels with a diffusion model.
Kaiyue Wen, Zhiyuan Li, Tengyu Ma
https://openreview.net/forum?id=Dkmpa6wCIx
Keywords: Sharpness, Flatness, Generalization, Generalization Bound, SAM
Compressor summary: The paper investigates how flatness (sharpness) of two-layer ReLU neural networks relates to generalization and finds that it depends on data distribution and architecture, suggesting that sharpness minimization alone does not guarantee better generalization.
Dan Busbridge, Jason Ramapuram, Pierre Ablin, Tatiana Likhomanenko, Eeshan Gunesh Dhekane, Xavier Suau, Russell Webb
https://openreview.net/forum?id=DkeeXVdQyu
Keywords: Optimization, scaling rules, EMA, exponential moving average, self-supervised learning, pseudo-labelling, semi-supervised learning, BYOL, distillation, speech, vision
Compressor summary: The paper proposes a scaling rule for optimization with model EMA that preserves training dynamics across batch sizes and improves the performance of various machine learning methods, including pseudo-labeling and SSL.
Y. Jennifer Sun, Stephen Newman, Elad Hazan
https://openreview.net/forum?id=DkKHSsmVuA
Keywords: Bandit control, online learning
Compressor summary: The paper proposes an algorithm for optimal control under adversarial conditions that achieves sublinear regret and introduces a novel memory-based technique for bandit convex optimization.
Yun Yi, Haokui Zhang, Rong Xiao, Nannan Wang, Xiaoyu Wang
https://openreview.net/forum?id=DjX2Nr15kY
Keywords: Transformer, Graph Neural Network, neural network encoding, representation learning, neural architecture search, neural network deployment
Compressor summary: The paper proposes a modified Transformer-based model that can learn efficient representations from both cell-structured and entire networks by incorporating inductive representation learning capability of GNNs and enhancing Transformer's ability to work with graph structures, achieving better performance in predicting latency and accuracy.
Zihao Zhou, Rose Yu
https://openreview.net/forum?id=Deb1yP1zMN
Keywords: spatiotemporal modeling, neural point processes, integration method
Compressor summary: The paper introduces `Auto-STPP`, a novel method for efficiently integrating flexible spatiotemporal neural point processes using a decomposable parametrization and ProdNet, which outperforms existing methods in recovering complex intensity functions from irregular events.
Vladimir Feinberg, Xinyi Chen, Y. Jennifer Sun, Rohan Anil, Elad Hazan
https://openreview.net/forum?id=DeZst6dKyi
Keywords: online convex optimization, deep learning, matrix sketching, frequent directions
Compressor summary: The paper proposes a low-rank sketching approach using Frequent Directions (FD) to reduce memory and compute requirements of matrix preconditioners in deep learning training tasks, achieving similar performance to Shampoo and Adam with less memory.
Fei Zhang, Tianfei Zhou, Boyang Li, Hao He, Chaofan Ma, Tianjiao Zhang, Jiangchao Yao, Ya Zhang, Yanfeng Wang
https://openreview.net/forum?id=DdViWdxCTs
Keywords: Weakly (Text-based) Open-Vocabulary Semantic Segmentation, Vision-Language Pretraining, Prototypical Knowledge
Compressor summary: This paper introduces a new method for semantic segmentation of arbitrary classes using image-text pairs, which improves the performance by incorporating explicit supervision and multi-modal regularization for group tokens.
Dingkang Yang, Kun Yang, Yuzheng Wang, Jing Liu, Zhi Xu, Rongbin Yin, Peng Zhai, Lihua Zhang
https://openreview.net/forum?id=Dbaxm9ujq6
Keywords: Collaborative perception, multi-agent communication
Compressor summary: How2comm is a collaborative perception framework for driving scenarios that balances performance and communication bandwidth by using mutual information-aware communication, spatial-channel filtering, flow-guided delay compensation, and a collaboration transformer to handle various noises.
Jin Li, Yaoming Wang, XIAOPENG ZHANG, Bowen Shi, Dongsheng Jiang, Chenglin Li, Wenrui Dai, Hongkai Xiong, Qi Tian
https://openreview.net/forum?id=DVm0xxaEq1
Keywords: vision transformer, dense prediction
Compressor summary: The paper proposes an adaptive resolution method for vision transformers to speed up dense prediction tasks like semantic segmentation or object detection by clustering tokens based on importance and merging adjacent ones.
Sumedh Anand Sontakke, Jesse Zhang, Séb Arnold, Karl Pertsch, Erdem Biyik, Dorsa Sadigh, Chelsea Finn, Laurent Itti
https://openreview.net/forum?id=DVlawv2rSI
Keywords: Reinforcement Learning, Vision and Language Models
Compressor summary: RoboCLIP is an online imitation learning method that uses VLMs to generate rewards from a single video or text demonstration, enabling high zero-shot performance in robot manipulation tasks without manual reward design or domain matching.
Zhimeng Jiang, Xiaotian Han, Hongye Jin, Guanchu Wang, Rui Chen, Na Zou, Xia Hu
https://openreview.net/forum?id=DVjyq5eCAD
Keywords: Model Weight Perturbation, fairness, distribution shift
Compressor summary: The paper proposes a method called robust fairness regularization (RFR) that improves algorithmic fairness under distribution shifts by considering the worst case weight perturbation for each sensitive attribute group.
Zhilin Zhao, Longbing Cao
https://openreview.net/forum?id=DVWIA9v9Jm
Keywords: non-IID, Distribution Discrepancy, Data Divergence, Two-sample Test
Compressor summary: R-divergence is a method to measure how similar or different two datasets' probability distributions are by learning a minimum hypothesis on mixed data and comparing their expected risks.
Raanan Yehezkel Rohekar, Yaniv Gurwicz, Shami Nisimov
https://openreview.net/forum?id=DS4rKySlYC
Keywords: Self-Attention, Causal Discovery, Reasoning, Explainability, Zero-shot, Transformer
Compressor summary: The authors propose a way to understand self-attention in Transformers as a causal structure over input symbols, allowing them to use pre-trained models for zero-shot causal discovery and provide explanations for the outcomes of two tasks.
Yutong Kou, Jin Gao, Bing Li, Gang Wang, Weiming Hu, Yizheng Wang, Liang Li
https://openreview.net/forum?id=DQgTewaKzt
Keywords: Visual tracking, non-uniform resizing, HVS-inspired processing
Compressor summary: The paper proposes a method to improve the speed and performance of trackers by non-uniformly resizing the input image, allowing for better target detection despite smaller input size.
Shunya Minami, Kenji Fukumizu, Yoshihiro Hayashi, Ryo Yoshida
https://openreview.net/forum?id=DPeBX79eNz
Keywords: Machine learning, Transfer learning
Compressor summary: The paper introduces a new transfer learning regression method called affine model transfer that has theoretical properties and outperforms existing methods in scenarios with scarce data.
Shafi Goldwasser, David Gruber, Adam Tauman Kalai, Orr Paradise
https://openreview.net/forum?id=DP2lioYIYl
Keywords: Theory, Unsupervised Machine Translation
Compressor summary: The paragraph discusses a theoretical framework for analyzing unsupervised machine translation (UMT) between two languages with different structures, and how it could potentially help understand animal communication systems.
Qitong Gao, Ge Gao, Juncheng Dong, Vahid Tarokh, Min Chi, Miroslav Pajic
https://openreview.net/forum?id=DOdaV0Hqdy
Keywords: Off-policy evaluation (OPE), Variational latent model for trajectory representation learning, Reinforcement learning and OPE for adaptive neurostimulation
Compressor summary: The paper proposes an OPEHF framework for estimating human feedback signals in RL using offline data, which outperforms existing OPE methods in three real-world and simulation experiments.
Ibrahim El Shar, Daniel R. Jiang
https://openreview.net/forum?id=DNubFPV5Dy
Keywords: Reinforcement learning, Deep Reinforcement Learning, Weakly Coupled MDPs
Compressor summary: The paper introduces WCDQN, a reinforcement learning algorithm that improves performance in WCMDPs by training multiple subagents and combining their solutions to guide the main agent towards optimality.
Rui Jiao, Wenbing Huang, Peijia Lin, Jiaqi Han, Pin Chen, Yutong Lu, Yang Liu
https://openreview.net/forum?id=DNdN26m2Jk
Keywords: crystal structure prediction, equivariant graph neural networks, diffusion generative models
Compressor summary: DiffCSP is a novel diffusion model that leverages fractional coordinates and equivariant denoising to efficiently predict crystal structures with high accuracy.
Qi CHEN, Changjian Shui, Ligong Han, Mario Marchand
https://openreview.net/forum?id=DNHGKeOhLl
Keywords: continual meta-learning; transfer learning; stability-plasticity dilemma;
Compressor summary: The paragraph describes Continual Meta-Learning (CML), which aims to balance stability and plasticity in learning from non-i.i.d. tasks, and proposes a novel algorithm that adapts meta-parameter and learning rate to environment change.
Emmeran Johnson, Ciara Pike-Burke, Patrick Rebeschini
https://openreview.net/forum?id=DKHEkP7Idx
Keywords: Reinforcement Learning Theory, Policy Mirror Descent, Policy Gradient
Compressor summary: The paper introduces and analyzes Policy Mirror Descent (PMD), a general family of algorithms in reinforcement learning that achieves the optimal convergence rate and step-size when using exact policy evaluation.
Duc N.M Hoang, Souvik Kundu, Shiwei Liu, Zhangyang Wang
https://openreview.net/forum?id=DIBcdjWV7k
Keywords: Pruning at Initialization, Pruning at Training, LTH, DST, Ramanujan, graph
Compressor summary: This paper explores how the connectivity of a deep network's architecture as a graph affects its performance, and proposes a new pruning method that maximizes the distance between sparse subnetworks and their original initialization.
Ilias Diakonikolas, Sushrut Karmalkar, Jongho Park, Christos Tzamos
https://openreview.net/forum?id=DI6KQhgqUr
Keywords: Oblivious noise, Robust Statistics, Heavy-tailed Stochastic Optimization, Approximate Gradients, Inexact Gradients
Compressor summary: The study explores stochastic optimization with oblivious noise and presents an efficient list-decodable learner that finds a close candidate or a single solution depending on the noise level.
Sam Adam-Day, Theodor-Mihai Iliant, Ismail Ilkan Ceylan
https://openreview.net/forum?id=DGmxTUCHYs
Keywords: graph neural networks, graph convolutional networks, zero-one law, expressivity, asymptotic behavior
Compressor summary: The paper analyzes how graph neural networks (GNNs) represent and extrapolate information on large graphs, finding theoretical and empirical limitations to their performance.
Shankar Padmanabhan, Yasumasa Onoe, Michael JQ Zhang, Greg Durrett, Eunsol Choi
https://openreview.net/forum?id=DFaGf3O7jf
Keywords: Knowledge editing, NLP, Distillation, deep learning, fine-tuning
Compressor summary: The authors propose a method for updating language models with new entity definitions and show that their approach enables better inferences based on these definitions without sacrificing performance in other contexts.
Stephanie Fu, Netanel Yakir Tamir, Shobhita Sundaram, Lucy Chai, Richard Zhang, Tali Dekel, Phillip Isola
https://openreview.net/forum?id=DEiNSfh1k7
Keywords: perceptual similarity, foundation model, perception, computer vision, image metric
Compressor summary: The paper introduces DreamSim, a new perceptual similarity metric that assesses images holistically by capturing mid-level similarities in layout, object pose, and semantic content, and shows its effectiveness on retrieval and reconstruction tasks using a new dataset of human similarity judgments.
Hao Zheng, Hongming Li, Yong Fan
https://openreview.net/forum?id=DEC7NxDJLh
Keywords: Brain MRIs, cortical surface reconstruction, deep learning
Compressor summary: The authors propose a new deep learning framework that reconstructs cortical surfaces from brain MRIs by jointly estimating the midthickness surface and diffeomorphic flows to optimize and deform it into inner, outer, and midthickness surfaces while accounting for topological correctness and estimating cortical thickness.
Yossi Azar, Debmalya Panigrahi, Noam Touitou
https://openreview.net/forum?id=DDmH3H78iJ
Keywords: Online, Predictions, Learning-augmented, Facility Location, Set Cover
Compressor summary: The paper proposes discrete-smooth algorithms for online covering problems, such as facility location and set cover, which balance consistency, robustness, and smoothness in machine-learning predictions.
Daoze Zhang, Zhizhang Yuan, Yang Yang, Junru Chen, Jingjing Wang, Yafeng Li
https://openreview.net/forum?id=DDkl9vaJyE
Keywords: Foundation model, Brain signal, Pretraining, Medicine
Compressor summary: Brant is a pre-trained foundation model that learns powerful representations of intracranial neural signals, achieving state-of-the-art performance on different brain signal tasks.
Lu Yan, ZHUO ZHANG, Guanhong Tao, Kaiyuan Zhang, Xuan Chen, Guangyu Shen, Xiangyu Zhang
https://openreview.net/forum?id=DD0QJvPbTD
Keywords: NLP, backdoor attack, fuzzing
Compressor summary: The paper proposes a method to detect stealthy backdoor attacks in NLP models by paraphrasing inputs to remove triggers and comparing model predictions before and after paraphrasing.
Kyriakos Lotidis, Panayotis Mertikopoulos, Nicholas Bambos, Jose Blanchet
https://openreview.net/forum?id=DCIsNIUCV7
Keywords: quantum games, Matrix Multiplicative Weights, zero-sum games, Nash equilibrium
Compressor summary: The paper presents new methods for learning in quantum games and other semidefinite games with minimal information feedback, using zeroth-order gradient samplers and achieving different convergence rates depending on the information framework.
Aniket Das, Dheeraj Mysore Nagaraj
https://openreview.net/forum?id=DBz9E5aZey
Keywords: Stein Variational Gradient Descent, Variational Inference, Sampling
Compressor summary: The paper introduces virtual particles to improve the efficiency and convergence rate of Stein Variational Gradient Descent (SVGD), a particle-based variational inference algorithm, in the finite-particle regime.
Tahereh Toosi, Elias Issa
https://openreview.net/forum?id=DBlkX8Nczr
Keywords: Visual inference, Bio-plausible learning algorithm, Feedback connections, Visual imagery, Occlusions, Noise
Compressor summary: The paper proposes a learning algorithm, Feedback-Feedforward Alignment (FFA), that aligns feedforward and feedback pathways in neural networks to co-optimize tasks and enable emergent visual inference functions similar to those in natural vision.
Hao Wang, Jiajun Fan, Zhichao Chen, Haoxuan Li, Weiming Liu, Tianqiao Liu, Quanyu Dai, Yichao Wang, Zhenhua Dong, Ruiming Tang
https://openreview.net/forum?id=DAdfU1ASLb
Keywords: treatment effect estimation, optimal transport, wasserstein, causal inference, counterfactual
Compressor summary: The paper proposes a new method called Entire Space CounterFactual Regression (ESCFR) to better estimate individual treatment effects from observational data by addressing two common issues in existing methods.
Joel Daniel Andersson, Rasmus Pagh
https://openreview.net/forum?id=DAKAkMhjSR
Keywords: differential privacy, continual observation, binary mechanism
Compressor summary: The paper proposes an efficient alternative method for releasing differentially private estimates based on a dataset that evolves over time, improving on previous methods in terms of noise distribution and computation time.
Songhua Liu, Xinchao Wang
https://openreview.net/forum?id=D9CMRR5Lof
Keywords: Dataset Distillation, Dataset Condensation, Efficient Learning, Conditional Generation, Meta Learning
Compressor summary: The paper introduces a faster and more flexible dataset distillation method using a generative approach and meta-learning, achieving significant speedup and improvement over existing techniques.
Yixuan Even Xu, Steven Jecmen, Zimeng Song, Fei Fang
https://openreview.net/forum?id=D94QKZA7UP
Keywords: peer review, randomized paper assignment, mitigating malicious behavior, convex optimization
Compressor summary: The paper proposes a practical method for randomized paper assignment in peer review processes that balances various considerations like expertise, robustness, diversity, and anonymity.
Huan Ma, Changqing Zhang, Yatao Bian, Lemao Liu, Zhirui Zhang, Peilin Zhao, Shu Zhang, Huazhu Fu, Qinghua Hu, Bingzhe Wu
https://openreview.net/forum?id=D8oHQ2qSTj
Keywords: large language models, prompts, classification
Compressor summary: The paper introduces a metric to measure predictive bias of prompts for in-context learning and proposes a search strategy to find near-optimal prompts, improving the performance on various downstream tasks.
Laura Fee Nern, Harsh Raj, Maurice Georgi, Yash Sharma
https://openreview.net/forum?id=D8nAMRRCLS
Keywords: Machine Learning Theory, Transfer Learning, Adversarial Robustness
Compressor summary: The study shows how to measure and improve the robustness of linear predictors on downstream tasks by using representations with desired robustness properties.
Van-Anh Nguyen, Trung Le, Anh Tuan Bui, Thanh-Toan Do, Dinh Phung
https://openreview.net/forum?id=D7LdL2SCCi
Keywords: Distributional Robustness, Sharpness-aware, SAM
Compressor summary: The paper proposes an optimal transport-based distributional robustness framework for training deep learning models that can be applied to single, ensemble, or Bayesian models and shows improved performance in various experiments.
Jonas Bernhard Wildberger, Maximilian Dax, Simon Buchholz, Stephen R Green, Jakob H. Macke, Bernhard Schölkopf
https://openreview.net/forum?id=D2cS6SoYlP
Keywords: simulation-based inference, likelihood-free inference, machine learning for physical sciences
Compressor summary: FMPE is a continuous normalizing flows technique for simulation-based inference that offers flexibility, exact density evaluation, fast training, and scalability, and outperforms discrete flows methods in gravitational-wave inference.
Junhyuk So, Jungwon Lee, Daehyun Ahn, Hyungjun Kim, Eunhyeok Park
https://openreview.net/forum?id=D1sECc9fiG
Keywords: deep learning optimization, quantization, diffusion model, generative model
Compressor summary: The paper introduces a novel quantization method for diffusion models that adapts to time steps, improving performance without increasing computation cost.
Yihua Zhang, Yimeng Zhang, Aochuan Chen, Jinghan Jia, Jiancheng Liu, Gaowen Liu, Mingyi Hong, Shiyu Chang, Sijia Liu
https://openreview.net/forum?id=D0MII7rP3R
Keywords: Dataset pruning, transfer learning
Compressor summary: The authors propose new dataset pruning methods for transfer learning that reduce data usage by up to 80% without compromising performance on downstream tasks, leading to significant speed-ups during pretraining.
Jon Schneider, Julian Zimmert
https://openreview.net/forum?id=CzkOzKWpMa
Keywords: bandits, first-price auction, sleeping bandits, contextual bandits
Compressor summary: The paper proposes an efficient algorithm for designing contextual bandit algorithms with a nearly tight regret bound, which removes correlations between estimation of the unknown context distribution and the actions played by the algorithm.
Albert Tseng, Tao Yu, Toni J.B. Liu, Christopher De Sa
https://openreview.net/forum?id=CzAFnfwbGd
Keywords: Hyperbolic Entailment Cones, Hyperbolic Space, Entailment Cones, Attention, Dot Product, Hierarchy, Transformers
Compressor summary: Cone attention is a new attention mechanism for neural networks that uses hyperbolic cones to model the hierarchy between data points, improving performance over dot product attention and requiring fewer parameters.
Chudi Zhong, Zhi Chen, Jiachang Liu, Margo Seltzer, Cynthia Rudin
https://openreview.net/forum?id=CzAAbKOHQW
Keywords: Interpretability, human-model interaction, generalized additive model, Rashomon set
Compressor summary: The text discusses how approximating the Rashomon set of models allows for better interaction between machine learning models and domain experts, and presents algorithms to efficiently approximate this set for sparse, generalized additive models with fixed support sets.
Matthew Wallingford, Vivek Ramanujan, Alex Fang, Aditya Kusupati, Roozbeh Mottaghi, Aniruddha Kembhavi, Ludwig Schmidt, Ali Farhadi
https://openreview.net/forum?id=Cyn1PvuZsB
Keywords: Transfer Learning, Distribution Shift, Test-Time Training
Compressor summary: Neural Priming is a technique that adapts pretrained models to new tasks and distributions using recalled data from previous training, improving accuracy in various benchmarks.
Cathy Yuanchen Li, Emily Wenger, Zeyuan Allen-Zhu, Francois Charton, Kristin E. Lauter
https://openreview.net/forum?id=CxzCoFDeQf
Keywords: machine learning, cryptography, cryptanalysis
Compressor summary: VERDE is an improved machine learning attack on the Learning with Errors problem, which can recover sparse secrets and work with larger dimensions and smaller moduli, while using less time and power.
Jacob A Zavatone-Veth, Paul Masset, William Lingxiao Tong, Joseph Zak, Venkatesh N Murthy, Cengiz Pehlevan
https://openreview.net/forum?id=Cxn1FpnNvG
Keywords: Olfaction, Bayesian inference, neural circuits, normative models, population geometry
Compressor summary: The proposed rate-based Poisson compressed sensing circuit model for the olfactory bulb accurately detects tens of odors within a single sniff and performs uncertainty estimation by matching receptor properties with the geometry of the neural code.
Marc Jourdan, Rémy Degenne
https://openreview.net/forum?id=CxjmYRP9Ji
Keywords: multi-armed bandits, best-arm identification, Gaussian bandits, Top Two algorithm, fixed confidence, finite confidence
Compressor summary: The paper derives the first non-asymptotic upper bound on the expected sample complexity of a Top Two bandit algorithm with any error level, and proposes a UCB-based algorithm that has both theoretical guarantees and good empirical performance.
Hyosoon Jang, Seonghyun Park, Sangwoo Mo, Sungsoo Ahn
https://openreview.net/forum?id=CxUuCydMDU
Keywords: diffusion model, graph neural network, structured prediction, node classification
Compressor summary: The paper proposes a new framework for structured node classification on graphs that considers label dependencies and uses a diffusion probabilistic model to learn from partially labeled data.
Jameel Hassan Abdul Samadh, Hanan Gani, Noor Hazim Hussein, Muhammad Uzair Khattak, Muzammal Naseer, Fahad Khan, Salman Khan
https://openreview.net/forum?id=CusNOTRkQw
Keywords: Vision-Language models, Prompt Learning, Test-Time Adaptation
Compressor summary: The paper proposes a method to improve zero-shot generalization of vision-language models using prompt tuning to align test sample statistics with source data, achieving better results than existing techniques on domain generalization benchmarks.
Moran Barenboim, Vadim Indelman
https://openreview.net/forum?id=Cupr2yTFSx
Keywords: POMDPs, Planning under uncertainty, Robotics
Compressor summary: The authors propose a method to find a simplified solution for planning under uncertainty in POMDPs that has deterministic guarantees and can be integrated with existing solvers.
Theo Gnassounou, Rémi Flamary, Alexandre Gramfort
https://openreview.net/forum?id=CuHymkHRus
Keywords: Neuroscience, Domain adaptation, Optimal Transport
Compressor summary: In many machine learning applications on signals and biomedical data, especially electroencephalogram (EEG), one major challenge is the variability of the data across subjects, sessions, and hardware devices. In this work, we propose a new method called Convolutional Monge Mapping Normalization ($\texttt{CMMN}$), which consists in filtering the signals in order to adapt their power spectrum density (PSD) to a Wasserstein barycenter estimated on training data. $\texttt{CMMN}$ relies on novel closed-form solutions for optimal transport mappings and barycenters and provides individual test time adaptation to new data without needing to retrain a prediction model. Numerical experiments on sleep EEG data show that $\texttt{CMMN}$ leads to significant and consistent performance gains independent from the neural network architecture when adapting between subjects, sessions, and even datasets collected with different hardware. Notably our performance gain is on par with much more numerically intensive Domain Adaptation (DA) methods and can be used in conjunction with those for even better performances.
Senzhang Wang, Jun Yin, Chaozhuo Li, Xing Xie, Jianxin Wang
https://openreview.net/forum?id=CtXXOaxDw7
Keywords: Explainable AI, Graph Neural Networks, Machine Learning
Compressor summary: V-InfoR is a robust graph neural network (GNN) explainer that uses variational inference to extract graph representations and graph information bottleneck optimization to generate explanations for corrupted graphs.
Jinpeng Chen, Runmin Cong, Yuxuan LUO, Horace Ip, Sam Kwong
https://openreview.net/forum?id=Ct0zPIe3xs
Keywords: Continual learning, Class-incremental semantic segmentation, Prototype replay
Compressor summary: STAR is a method for semantic segmentation that addresses class imbalance by storing compact prototypes and repeating background pixels, achieving better performance with less storage than previous methods.
Xin-Qiang Cai, Yu-Jie Zhang, Chao-Kai Chiang, Masashi Sugiyama
https://openreview.net/forum?id=CswEebv5Hn
Keywords: imitation learning, vague feedback, risk rewriting, mixture propotion estimation
Compressor summary: The paragraph discusses imitation learning from human feedback, which aims to train agents using relative comparisons of demonstrations, and proposes a method for handling vague feedback in this setting.
Peter Dixon, A. Pavan, Jason Vander Woude, N V Vinodchandran
https://openreview.net/forum?id=Cs9ea2Gbgx
Keywords: Replicability, learning algorithms, sample complexity, PAC Learning
Compressor summary: The authors study replicable learning algorithms and design optimal list and certificate complexities for estimating coin biases and learning hypothesis classes with nonadaptive statistical queries, while minimizing sample complexity.
Wenhan Xian, Heng Huang
https://openreview.net/forum?id=CruxS0C0LS
Keywords: second-order optimality, decentralized optimization
Compressor summary: The paper introduces PEDESTAL, a new decentralized stochastic algorithm for nonconvex optimization that achieves second-order optimality with theoretical guarantees and matches state-of-the-art results in centralized methods.
Yanbang Wang, Jon Kleinberg
https://openreview.net/forum?id=CrpL8mGa0Q
Keywords: social networks, spectral analysis, link recommendation, polarization and conflict
Compressor summary: The paper analyzes how link recommendations in online social networks affect relevance and conflict, finding that some algorithms reduce conflict more effectively than others.
Ho Man Kwan, Ge Gao, Fan Zhang, Andy Gower, David Bull
https://openreview.net/forum?id=CpoS56pYnU
Keywords: Video compression, Implicit neural representations
Compressor summary: The paper introduces HiNeRV, a new Implicit Neural Representation for video compression that combines light weight layers with hierarchical positional encodings and achieves significant improvement over existing methods.
Lingbing Guo, Weiqing Wang, Zhuo Chen, Ningyu Zhang, Zequn Sun, Yixuan Lai, Qiang Zhang, Huajun Chen
https://openreview.net/forum?id=CnvZ7FIyAD
Keywords: Equivariant Graph Neural Networks, Molecular Dynamics, N-body System, Human Motion
Compressor summary: The paper introduces a new method for predicting the future state of system dynamics using velocity estimations and Newton–Cotes formulas, which improves upon existing GNNs-based approaches.
Trang Nguyen, Amin Mansouri, Kanika Madan, Nguyen Duy Khuong, Kartik Ahuja, Dianbo Liu, Yoshua Bengio
https://openreview.net/forum?id=CniUitfEY3
Keywords: Out-of-Distribution Generalization, Slotwise, Visual Reasoning, Video Prediction, Reusable Mechanism, Dynamics modeling
Compressor summary: The paper proposes a framework called Reusable Slotwise Mechanisms (RSM) that models object dynamics using communication among slots and a modular architecture, achieving improved robustness, generalization, and performance in various tasks.
Xinhong Ma, Yiming Wang, Hao Liu, Tianyu Guo, Yunhe Wang
https://openreview.net/forum?id=ChGGbmTNgE
Keywords: unsupervised domain adaptation, semantic segmentation, visual prompt tuning
Compressor summary: The text introduces a new method (Uni-UVPT) for adapting a semantic segmentation model to an unlabeled target domain without using the source data, by using a lightweight prompt adapter and a pseudo-label correction strategy.
Yanshu Zhang, Shichong Peng, Seyed Alireza Moazenipourasil, Ke Li
https://openreview.net/forum?id=CgJJvuLjec
Keywords: point cloud learning, point cloud rendering
Compressor summary: PAPR is a novel point-based scene representation and differentiable renderer that learns accurate scene geometry and texture using a parsimonious set of points and demonstrates practical applications.
Sergey Pozdnyakov, Michele Ceriotti
https://openreview.net/forum?id=CdSRFn1fVe
Keywords: geometric deep learning, point clouds, equivariance, machine learning potentials, GNN, transformer, atomic-scale modeling
Compressor summary: The authors propose a method to add rotational symmetry to point-cloud models for chemical and materials modeling, and introduce the Point Edge Transformer (PET) architecture that achieves state-of-the-art performance on molecular and solid datasets.
Cian Eastwood, Shashank Singh, Andrei Liviu Nicolicioiu, Marin Vlastelica, Julius von Kügelgen, Bernhard Schölkopf
https://openreview.net/forum?id=Cc2fjBBlBD
Keywords: invariant prediction, spurious correlations, out-of-distribution generalization, domain generalization, domain adaptation, test-time domain adaptation
Compressor summary: The paper proposes a method to improve performance in test domain by using pseudo-labels based on stable features, which are conditionally independent from unstable ones given the label.
Leonard Salewski, Stephan Alaniz, Isabel Rio-Torto, Eric Schulz, Zeynep Akata
https://openreview.net/forum?id=CbsJ53LdKc
Keywords: large language models, impersonation, vision language models, reasoning
Compressor summary: The text explores how LLMs can take on different roles by assuming personas based on social identity or domain expertise, and how this affects their performance in various tasks, revealing both strengths and biases.
Fabien Casenave, Brian Staber, Xavier Roynard
https://openreview.net/forum?id=Ca78M3awPw
Keywords: Gaussian process, mesh morphing, mesh parametrization, finite element interpolation, simulation, physics, predictive uncertainties, nonparametrized geometries
Compressor summary: The authors propose a machine learning method that uses mesh morphing, dimensionality reduction, and Gaussian processes to model physical phenomena in industrial designs without relying on graph neural networks, providing predictive uncertainties and handling large meshes efficiently.
Hideaki Kim
https://openreview.net/forum?id=CYCzfXn6cZ
Keywords: survival analysis, temporal point process, Bayesian estimation, permanental process, representer theorem, kernel method
Compressor summary: The paper proposes a non-parametric Bayesian survival model for analyzing nonlinear dependence on time-varying covariates using a permanental process, which is fast and accurate.
XiMing Xing, Chuang Wang, Haitao Zhou, Jing Zhang, Qian Yu, Dong Xu
https://openreview.net/forum?id=CY1xatvEQj
Keywords: Sketch; Vector Sketch; Sketch Generation; Diffusion Models
Compressor summary: The paper introduces DiffSketcher, an algorithm that creates vectorized free-hand sketches from natural language input using a pre-trained text-to-image diffusion model and optimizing Bézier curves with a modified loss function.
Jihui Jin, Etienne Ollivier, Richard Touret, Matthew McKinley, Karim Sabra, Justin Romberg
https://openreview.net/forum?id=CXrRMfs5eY
Keywords: Inverse Problems, Neural Adjoint, Hybrid Machine Learning, Physics
Compressor summary: The paragraph discusses a new method for recovering an underlying signal using a learned weighted average model that incorporates physics-based linearizations, which improves accuracy and provides better gradient information for the inverse problem of ocean acoustic tomography.
Lijun Yu, Yong Cheng, Zhiruo Wang, Vivek Kumar, Wolfgang Macherey, Yanping Huang, David A Ross, Irfan Essa, Yonatan Bisk, Ming-Hsuan Yang, Kevin Patrick Murphy, Alexander G Hauptmann, Lu Jiang
https://openreview.net/forum?id=CXPUg86A1D
Keywords: multimodal, generation, large language model
Compressor summary: The paper presents SPAE, a technique that allows frozen large language models (LLMs) to perform image understanding and generation tasks by converting raw pixels to lexical tokens.
Zeqiu Wu, Yushi Hu, Weijia Shi, Nouha Dziri, Alane Suhr, Prithviraj Ammanabrolu, Noah A. Smith, Mari Ostendorf, Hannaneh Hajishirzi
https://openreview.net/forum?id=CSbGXyCswu
Keywords: Language Model, Reinforcement Learning with Human Feedback, Long-Form Text Generation
Compressor summary: The paper proposes Fine-Grained RLHF, a framework that uses reward functions based on fine-grained human feedback to train language models for better text generation.
Zhiyong Wang, Jize Xie, Xutong Liu, Shuai Li, John C.S. Lui
https://openreview.net/forum?id=CQuRzAgjg9
Keywords: online clustering of bandits
Compressor summary: This paper introduces two robust clustering of bandits algorithms that can handle misspecified user models, achieving near-optimal regret bounds and outperforming existing methods in simulations.
Harry Bendekgey, Gabriel Hope, Erik B. Sudderth
https://openreview.net/forum?id=CQqBt46FUD
Keywords: Generative Models, Graphical Models, Variational Inference, Amortized Inference
Compressor summary: The authors propose a new generative model that combines graphical models and deep learning, handle multimodal uncertainty with latent variables, and improve optimization techniques to learn interpretable discrete data representations.
Palak Jain, Iden Kalemaj, Sofya Raskhodnikova, Satchit Sivakumar, Adam Smith
https://openreview.net/forum?id=CQ38aC92WY
Keywords: distinct elements, differential privacy, continual release, turnstile streams
Compressor summary: The paper discusses privacy challenges and error bounds for systems that learn from data streams with inserts and deletions, and proposes a new item-level differentially private mechanism with low additive error for streams with low maximum flippancy.
Tomer Galanti, Mengjia Xu, Liane Galanti, Tomaso Poggio
https://openreview.net/forum?id=COPzNA10hZ
Keywords: generalization bounds, convolution, rademacher, generalization, sparsity
Compressor summary: The paper develops norm-based generalization bounds for sparse ReLU neural networks that account for their sparse structure and convolutional filters, leading to tighter bounds than previous methods and better estimation of generalization.
Jaskirat Singh, Liang Zheng
https://openreview.net/forum?id=CLjBBd8u2j
Keywords: Text-to-Image Generation
Compressor summary: The paper proposes a decompositional approach to evaluate and improve text-to-image alignment, using a novel score that measures how well each assertion in a complex caption aligns with the generated image.
Yanbo Chen, Weiwei Liu
https://openreview.net/forum?id=CJY7NEXVwC
Keywords: Learning Theory
Compressor summary: This paper proposes a theoretical model that explains the transferability of adversarial examples in black-box attacks and shows that it depends on the curvature of the data manifold.
Thomas M. Sutter, Alain Ryser, Joram Liebeskind, Julia E Vogt
https://openreview.net/forum?id=CJWQGDwa6u
Keywords: random partition model, continuous relaxation, reparameterization, generative models, vae, representation learning, weak supervision, variational clustering, deep learning
Compressor summary: The paper proposes a novel method for inferring partitions in machine learning problems that allows gradient-based optimization and shows its versatility on three different challenging experiments.
Yifan Zang, Jinmin He, Kai Li, Haobo Fu, QIANG FU, Junliang Xing, Jian Cheng
https://openreview.net/forum?id=CGj72TyGJy
Keywords: MARL, Cooperative Multi-Agent Reinforcement Learning, Coordination and Cooperation, Automatic Grouping, Group-Wise Learning
Compressor summary: The paper introduces GoMARL, a novel multi-agent reinforcement learning approach that learns automatic grouping for efficient cooperation in natural systems.
Yuxin Guo, Shijie Ma, Hu Su, Zhiqing Wang, Yuhao Zhao, Wei Zou, Siyang Sun, Yun Zheng
https://openreview.net/forum?id=CFhpBJ8eZ5
Keywords: audio-visual learning, audio-visual source localization, semi-supervised learning
Compressor summary: The paper proposes a semi-supervised learning framework called Dual Mean-Teacher for audio-visual source localization, which improves localization accuracy by using two teachers to generate high-quality pseudo-labels and outperforms state-of-the-art methods with limited labeled data.
Tosca Lechner, Vinayak Pathak, Ruth Urner
https://openreview.net/forum?id=CFQBcz7k8n
Keywords: adversarially robust learning
Compressor summary: The paper studies a new setting of learning with imperfect knowledge of adversarial perturbation sets and proposes methods to improve robustness against them.
Anh Tuan Nguyen, Nikos Karampatziakis, Weizhu Chen
https://openreview.net/forum?id=CEk6JK71Mb
Keywords: language modeling, pre-training, deep learning, NLP
Compressor summary: The paper introduces a new language model pre-training method called Meet in the Middle (MIM) that trains models in both left-to-right and right-to-left directions, improving data efficiency and performance on generation and infilling tasks.
Christos Boutsikas, Petros Drineas, Marios Mertzanidis, Alexandros Psomas, Paritosh Verma
https://openreview.net/forum?id=CDTifMbUNc
Keywords: mechanism design, revenue maximization, randomized linear algebra, active regression
Compressor summary: This paper proposes a method to design revenue-maximizing mechanisms for sellers with many items and strategic bidders, using topic models and active learning techniques from randomized linear algebra.
Gaurav Shrivastava, Ser-Nam Lim, Abhinav Shrivastava
https://openreview.net/forum?id=CCq73CGMyV
Keywords: Computational Photography, Deep Internal Learning, low-level vision, video denoising, video super-resolution, video frame interpolation, video inpainting
Compressor summary: The paper proposes a robust framework for low-level vision tasks that learns from corrupted test sequences and uses a novel spatial pyramid loss to handle noise and degradation in videos, achieving state-of-the-art results on several tasks.
Koen Minartz, Yoeri Poels, Simon Martinus Koop, Vlado Menkovski
https://openreview.net/forum?id=CCVsGbhFdj
Keywords: stochastic simulation, equivariance, dynamical systems, probabilistic simulation, generative models
Compressor summary: Equivariant Probabilistic Neural Simulation (EPNS) is a framework for simulating stochastic systems with improved accuracy, efficiency, and stability by incorporating domain symmetries into probabilistic neural networks.
Joel Ye, Jennifer L Collinger, Leila Wehbe, Robert Gaunt
https://openreview.net/forum?id=CBBtMnlTGq
Keywords: Pretraining, Scaling Laws, Neuroscience, Brain-computer interfaces
Compressor summary: NDT2 is a spatiotemporal Transformer that uses unsupervised pretraining to learn rich representations from neural spiking data across sessions, subjects, and tasks for iBCI control.
Aditya Chattopadhyay, Ryan Pilgrim, Rene Vidal
https://openreview.net/forum?id=CAF4CnUblx
Keywords: Information Maximization, Sparse Coding, Orthogonal Matching Pursuit, Explainable AI, Information Pursuit
Compressor summary: The paper explores using Orthogonal Matching Pursuit (OMP) as an alternative to Information Pursuit (IP) for greedily selecting queries and shows how IP with random projections can almost reduce to OMP, leading to a simple explainable AI algorithm that encodes images as sparse combinations of semantically meaningful dictionary atoms.
Luke Travis, Kolyan Ray
https://openreview.net/forum?id=CA8tMQiscx
Keywords: Gaussian process, sparse variational Bayes, uncertainty quantification, theoretical guarantees
Compressor summary: The paper investigates how well a specific type of Gaussian process prior performs in estimating and quantifying uncertainty, and under what conditions it is accurate or misleading.
Siddharth Prasad, Nina Balcan, Tuomas Sandholm
https://openreview.net/forum?id=C9wlNF1Ooj
Keywords: mechanism design, revenue maximization, welfare maximization, side information, weakest competitors, algorithms with predictions, learning-augmented algorithms
Compressor summary: The paper presents a new method for multidimensional mechanism design using side information, which can generate high social welfare and revenue while accounting for errors and varying quality of the input data.
Adam Li, Amin Jaber, Elias Bareinboim
https://openreview.net/forum?id=C9wTM5xyw2
Keywords: causal inference, causal discovery, transportability, multi-domain learning
Compressor summary: The paper proposes an approach to learn causal structures in non-Markovian systems using data from multiple domains, by connecting interventional distributions with graphical criteria and introducing a new causal discovery algorithm called S-FCI.
Jian-Feng CAI, José Vinícius De Miranda Cardoso, Daniel P. Palomar, Jiaxi Ying
https://openreview.net/forum?id=C9cgwmJ8Pt
Keywords: MTP2, Total Positivity, Generalized graph Laplacian, Precision matrix estimation, Nonnegative partial correlations
Compressor summary: The paper proposes a new algorithm for estimating precision matrices in Gaussian distributions with specific properties, which is faster and theoretically proven to converge.
Chengjie Wu, Pingzhong Tang, Jun Yang, Yujing Hu, Tangjie Lv, Changjie Fan, Chongjie Zhang
https://openreview.net/forum?id=C8pvL8Qbfa
Keywords: reinforcement learning, opponent exploitation, multi-agent
Compressor summary: The paper introduces a novel approach for offline policy adaptation in multi-agent games using conservative reinforcement learning that can exploit weaknesses or enable cooperation without online interaction.
Runnan Chen, Youquan Liu, Lingdong Kong, Nenglun Chen, Xinge ZHU, Yuexin Ma, Tongliang Liu, Wenping Wang
https://openreview.net/forum?id=C8JdyM7B8I
Keywords: label-free, scene understanding
Compressor summary: The paper explores using CLIP and SAM for label-free scene understanding and proposes a novel method to supervise 2D and 3D networks with noisy labels from these models, improving performance on several datasets.
Lingxiao Li, Samuel Hurault, Justin Solomon
https://openreview.net/forum?id=C6fvJ2RfsL
Keywords: JKO, mass-conservation, PDE, Fokker-Planck, scalable, discretization-free, neural ODE
Compressor summary: The authors propose a scalable framework for solving mass-conserving PDEs using an iterative formulation with a biased gradient estimator that overcomes computational challenges and outperforms existing methods.
Lu Yin, Gen Li, Meng Fang, Li Shen, Tianjin Huang, Zhangyang Wang, Vlado Menkovski, Xiaolong Ma, Mykola Pechenizkiy, Shiwei Liu
https://openreview.net/forum?id=C6IIwFHWkF
Keywords: dynamic sparsity, dynamic sparse training, sparse training
Compressor summary: Chase is a new dynamic sparse training method that converts unstructured sparsity to GPU-friendly channel-level sparsity, improving inference speedup without sacrificing accuracy.
Henry Tsang, Thomas Dybdahl Ahle
https://openreview.net/forum?id=C4rRqkXFyC
Keywords: Embedding table compression, Clustering and sketching, Memory-efficient training
Compressor summary: CCE combines clustering-based compression with dynamic methods to efficiently represent large embedding tables for recommendation systems, and has theoretical convergence guarantees.
Mathieu Serrurier, Franck Mamalet, Thomas FEL, Louis Béthune, Thibaut Boissin
https://openreview.net/forum?id=ByDy2mlkig
Keywords: 1-Lipschitz neural network, explicability
Compressor summary: The paper shows that 1-Lipschitz neural networks, trained with optimal transportation problem's dual loss, produce saliency maps with low noise and high XAI properties, aligning well with human explanations and allowing counterfactual explanations that reveal the direction of transportation plans.
Feng Zhang, Ming Tian, Zhiqiang Li, Bin Xu, Qingbo Lu, Changxin Gao, Nong Sang
https://openreview.net/forum?id=BxqPN7KuQS
Keywords: tone mapping; learnable local laplacian filter; laplacian pyramid; 3D lookup table
Compressor summary: The paper proposes a novel strategy for converting HDR to LDR images using image-adaptive 3D LUTs and local Laplacian filters learned from annotated data, achieving better tone mapping and edge preservation than existing methods.
Yuxi Xie, Kenji Kawaguchi, Yiran Zhao, Xu Zhao, Min-Yen Kan, Junxian He, Qizhe Xie
https://openreview.net/forum?id=Bw82hwg5Q3
Keywords: Large Language Models, Multistep Reasoning, Stochastic Beam Search, LLM Self-Evaluation
Compressor summary: The paper proposes a decoding algorithm that integrates self-evaluation guidance to improve the reasoning quality of large language models and outperforms baselines on various benchmarks.
Le Yang, Siyang Gao, Chin Pang Ho
https://openreview.net/forum?id=BvslVXlUvF
Keywords: best arm identification, knowledge gradient, asymptotic optimality, convergence rate
Compressor summary: The improved knowledge gradient (iKG) algorithm is a better version of the knowledge gradient (KG) algorithm for identifying the best arm in multiple-arm bandit problems, as it is asymptotically optimal and can handle more variants of the problem.
Yuanzhi Wang, Yong Li, Zhen Cui
https://openreview.net/forum?id=BuGFwUS9B3
Keywords: Multimodal emotion recognition, Incomplete multimodalities
Compressor summary: IMDer is a method to improve multimodal emotion recognition by recovering missing modalities using a score-based diffusion model and embedding available modalities as conditions for semantic alignment.
Allan Raventos, Mansheej Paul, Feng Chen, Surya Ganguli
https://openreview.net/forum?id=BtAz4a5xDg
Keywords: in-context learning, Bayesian inference, transformers, task diversity, emergence
Compressor summary: The study investigates how task diversity during pretraining affects transformers' ability to learn new linear regression tasks without updating weights, finding a threshold beyond which they can optimally solve unseen tasks.
Yaru Hao, Zewen Chi, Li Dong, Furu Wei
https://openreview.net/forum?id=BsZNWXD3a1
Keywords: prompt adaptation, automatic prompt engineering, text-to-image generation
Compressor summary: Prompt adaptation is a framework that automatically improves text-to-image models by adapting user input to model preferences using supervised fine-tuning and reinforcement learning, resulting in better images that match user intentions.
Weizhi Wang, Li Dong, Hao Cheng, Xiaodong Liu, Xifeng Yan, Jianfeng Gao, Furu Wei
https://openreview.net/forum?id=BryMFPQ4L6
Keywords: large language models, long-term memory, long-text modeling and understanding, residual side-network, in-context learning
Compressor summary: The LongMem framework allows large language models to store and use long past contexts, improving performance on long-context modeling tasks and memory-augmented in-context learning.
Daehyun Kim, Sungyong Baik, Tae Hyun Kim
https://openreview.net/forum?id=BqZ70BEtuW
Keywords: Anomaly Detection, Visual Anomaly Detection, Computer Vision, Normalizing Flow, Anomaly Localization
Compressor summary: The authors propose a method to improve anomaly detection using normalizing flow to map different features of an image to different distributions, enhancing discriminability between normal and abnormal data.
Xing Han, Tongzheng Ren, Tan Minh Nguyen, Khai Nguyen, Joydeep Ghosh, Nhat Ho
https://openreview.net/forum?id=BqTv1Mtuhu
Keywords: Transformers, Kernel Density Estimation, Robustness
Compressor summary: The paper proposes new transformer architectures that are more robust to adversarial attacks and data contamination by adapting classical kernel density estimation methods to the self-attention mechanism.
Fnu Devvrit, Sai Surya Duvvuri, Rohan Anil, Vineet Gupta, Cho-Jui Hsieh, Inderjit S Dhillon
https://openreview.net/forum?id=BopG5dhH7L
Keywords: Optimization, Second order methods; Deep Learning
Compressor summary: The paper introduces Sparsified Online Newton (SONew), a memory-efficient second-order algorithm that improves convergence and performance for deep neural network training using structured sparsity patterns.
Ruitu Xu, Yifei Min, Tianhao Wang
https://openreview.net/forum?id=BnV2M2WFaY
Keywords: Linear Contextual Bandit, Thompson Sampling, Noise-Adaptive
Compressor summary: The paper proposes a noise-adaptive Thompson sampling-style algorithm for linear contextual bandits with heteroscedastic noise that achieves better regret bounds in both constant-variance and deterministic scenarios.
Petar Bevanda, Max Beier, Armin Lederer, Stefan Georg Sosnowski, Eyke Hüllermeier, Sandra Hirche
https://openreview.net/forum?id=BmIW6U0rz8
Keywords: Kernel Methods, Regression, Statistical Learning Theory, Koopman Operator, Mode Decomposition, Dynamical Systems, Supervised Learning
Compressor summary: The paragraph discusses a new method called Koopman Kernel Regression (KKR) that uses linear time-invariant equations to simplify complex forecasts, providing better learning guarantees and performance than existing approaches.
Hyunseung Kim, Byungkun Lee, Hojoon Lee, Dongyoon Hwang, Sejik Park, Kyushik Min, Jaegul Choo
https://openreview.net/forum?id=Bkrmr9LjeI
Keywords: Unsupervised skill discovery, Reinforcement Learning
Compressor summary: The paper proposes a new unsupervised skill discovery algorithm that improves exploration by selecting a guide skill and guiding other skills to follow it while maximizing their discriminability in unexplored states, and shows its effectiveness in challenging environments.
Vivien Cabannes, Stefano Vigogna
https://openreview.net/forum?id=BklIgOO76D
Keywords: Statistical learning, breaking the curse, kernel methods
Compressor summary: This paper studies how smoothness helps overcome the curse of dimensionality in statistical learning and explores the impact of constants and transient periods on generalization error.
Sergey A. Shuvaev, Evgeny M Amelchenko, Dmitry Smagin, Natalia Kudryavtseva, Grigori Enikolopov, Alexei A. Koulakov
https://openreview.net/forum?id=BkQM8huiIc
Keywords: neuroscience, decision-making, normative modeling, game theory, Bayesian methods, POMDP, inverse rational control, belief, theory of mind
Compressor summary: The study uses neural data from mice to develop a Theory of Mind model explaining how they form beliefs about strength and behavior in social conflict situations.
Hailey James, Chirag Nagpal, Katherine A Heller, Berk Ustun
https://openreview.net/forum?id=Bj1QSgiBPP
Keywords: healthcare, algorithmic fairness, data privacy, classification, interpretability
Compressor summary: The authors propose participatory systems, which allow individuals to opt into personalization of prediction models using their group attributes, improving both performance and privacy.
Hamed Nilforoshan, Michael Moor, Yusuf H Roohani, Yining Chen, Anja Šurina, Michihiro Yasunaga, Sara Oblak, Jure Leskovec
https://openreview.net/forum?id=BfQJrIiOZC
Keywords: causal inference, CATE, CATE estimation, causal machine learning, causal ML, heterogenous treatment effects, causality, potential outcomes, treatment effect
Compressor summary: CaML is a framework that predicts personalized effects of novel interventions using meta-learning and individual features, outperforming existing methods in medical and cell-line datasets.
Yu Zhang, Yepeng Liu, Duoqian Miao, Qi Zhang, Yiwei Shi, Liang Hu
https://openreview.net/forum?id=Bf6WFWNCUP
Keywords: Efficient AI, Vision Transformer, Image Classification, Multi-Granularity, Three-Way Decisions
Compressor summary: The paper introduces a multi-granularity strategy for compressing Vision Transformers that reduces computational cost without sacrificing performance on classification, detection, and segmentation tasks.
Ryan Thompson, Amir Dezfouli, Robert Kohn
https://openreview.net/forum?id=BdvCo8RVlx
Keywords: feature selection, sparsity, sparse regression, varying coefficients, deep learning
Compressor summary: The contextual lasso is a new method for creating interpretable sparse linear models that adapt to different contexts using a deep neural network and a novel lasso regularizer.
Viraj Uday Prabhu, Sriram Yenamandra, Prithvijit Chattopadhyay, Judy Hoffman
https://openreview.net/forum?id=BbIxB4xnbq
Keywords: image classification, robustness, guided diffusion models, counterfactuals
Compressor summary: The paper introduces an algorithm to create realistic test images based on language input to stress-test visual models and reveal hidden biases.
Anirudhan Badrinath, Yannis Flet-Berliac, Allen Nie, Emma Brunskill
https://openreview.net/forum?id=BYywOFbRFz
Keywords: offline reinforcement learning, reinforcement learning via supervised learning, behavioral cloning
Compressor summary: The Waypoint Transformer is a new reinforcement learning method that improves offline learning by using automatically-generated waypoints to guide the agent's trajectory.
Xiangsen Wang, Haoran Xu, Yinan Zheng, Xianyuan Zhan
https://openreview.net/forum?id=BXQtgwA2n0
Keywords: Offline reinforcement learning; multi-agent reinforcement learning; multi-agent cooperation
Compressor summary: OMIGA is a new offline multi-agent RL algorithm with implicit global-to-local value regularization that outperforms existing methods on various MuJoCo and StarCraft II tasks.
Seungtae Nam, Daniel Rho, Jong Hwan Ko, Eunbyung Park
https://openreview.net/forum?id=BW6nZf7TnK
Keywords: Novel view synthesis, Neural radiance fields
Compressor summary: mip-Grid is a novel approach that improves grid-based neural radiance field methods by integrating anti-aliasing techniques, achieving fast training speed and high rendering quality.
Weiwen Xu, Xin Li, Wenxuan Zhang, Meng Zhou, Wai Lam, Luo Si, Lidong Bing
https://openreview.net/forum?id=BVN9Kgvwzv
Keywords: Machine Reading Comprehension, Pre-training, Natural Language Understanding
Compressor summary: PMR is a novel method that improves machine reading comprehension models by retrofitting pre-trained masked language models using Wikipedia data and a Wiki Anchor Extraction task, leading to better performance, explainability, and versatility in various extraction and classification tasks.
Nived Rajaraman, Fnu Devvrit, Aryan Mokhtari, Kannan Ramchandran
https://openreview.net/forum?id=BTRcVP7ZJn
Keywords: Greedy Pruning; Matrix Sensing; Lasso regularization
Compressor summary: The paper investigates the theory behind pruning and fine-tuning methods for reducing model complexity in overparameterized matrix sensing problems, and shows that pruning columns with low norms leads to good generalization.
Darko Drakulic, Sofia Michel, Florian Mai, Arnaud Sors, Jean-Marc Andreoli
https://openreview.net/forum?id=BRqlkTDvvm
Keywords: Combinatorial Optimization, Markov Decision Processes, Bisimulation, Policy Learning, Out-of-Distribution Generalization, Routing Problems, TSP, CVRP, KP.
Compressor summary: The paper proposes a method for solving combinatorial optimization problems using Markov decision processes and bisimulation quotienting to exploit symmetries, achieving state-of-the-art results and better out-of-distribution generalization.
Andrew Kyle Lampinen, Stephanie C.Y. Chan, Ishita Dasgupta, Andrew Joo Hun Nam, Jane X Wang
https://openreview.net/forum?id=BRpi8YAfac
Keywords: passive; causal; offline; agency; language models
Compressor summary: The paragraph discusses how passively-trained agents can learn generalizable strategies for determining and using causal structures by intervening at test time, and how natural language explanations can further enhance their performance in complex environments.
Georgios Kaissis, Alexander Ziller, Stefan Kolek, Anneliese Riess, Daniel Rueckert
https://openreview.net/forum?id=BRSgVw85Mc
Keywords: Differential Privacy, Membership Inference Attack, Hypothesis Testing, Data Reconstruction Attack, Security
Compressor summary: The paper studies how private machine learning models can be attacked by less powerful adversaries with partial data, and suggests ways to set noise levels based on this risk.
Qi Zhang, Yifei Wang, Yisen Wang
https://openreview.net/forum?id=BQA7wR2KBF
Keywords: Self-supervised Learning, Contrastive Learning, Identifiability, Representation Learning
Compressor summary: The paper introduces triCL, a contrastive learning method that uses a 3-factor contrast to learn more interpretable and identifiable features by using a learnable diagonal matrix $S$ that captures the importance of each feature.
Yan Liu, Xiaokang Chen, Yan Gao, Zhe Su, Fengji Zhang, Daoguang Zan, Jian-Guang Lou, Pin-Yu Chen, Tsung-Yi Ho
https://openreview.net/forum?id=BOP5McdqGy
Keywords: Social bias, code fairness
Compressor summary: This paper investigates the social bias problem in automatic code generation tools like Copilot and proposes a method to construct code prompts that reveal and measure these biases.
Ruizhe Chen, Jianfei Yang, Huimin Xiong, Jianhong Bai, Tianxiang Hu, Jin Hao, YANG FENG, Joey Tianyi Zhou, Jian Wu, Zuozhu Liu
https://openreview.net/forum?id=BL9Pc7xsdX
Keywords: Model Debias, Bias Mitigation, Machine Unlearning, Counterfactual Fairness
Compressor summary: The paper proposes FMD, an efficient method for identifying, evaluating, and removing biases in deep neural networks using counterfactuals and machine unlearning.
Lemin Kong, Jiajin Li, Jianheng Tang, Anthony Man-Cho So
https://openreview.net/forum?id=BKAFLUcpBS
Keywords: Gromov Wasserstein, Robust Optimization, Nonconvex Optimization
Compressor summary: The paper introduces a new robust version of Gromov-Wasserstein distance called RGW that minimizes the impact of outliers in graph learning tasks using efficient algorithms and validates its effectiveness on real data.
Long-Kai Huang, Peilin Zhao, Junzhou Huang, Sinno Pan
https://openreview.net/forum?id=BJ1vOqh3hJ
Keywords: Model Repair; Fine-tuning
Compressor summary: The paper proposes a new approach to repair deep learning models by identifying and separating harmful data from beneficial data using a clean set and enhancing the model's performance.
Shashank Rajput, Nikhil Mehta, Anima Singh, Raghunandan Hulikal Keshavan, Trung Vu, Lukasz Heldt, Lichan Hong, Yi Tay, Vinh Q. Tran, Jonah Samost, Maciej Kula, Ed H. Chi, Maheswaran Sathiamoorthy
https://openreview.net/forum?id=BJ0fQUU32w
Keywords: Recommender Systems, Generative Retrieval, Vector Quantization
Compressor summary: The paper proposes a generative retrieval approach using semantically meaningful Semantic IDs for items, which improves recommender system performance and generalization.
Kunjal Panchal, Sunav Choudhary, Nisarg Parikh, Lijun Zhang, Hui Guan
https://openreview.net/forum?id=BI031mw7iS
Keywords: federated learning, personalization, statistical heterogeneity, dynamic routing
Compressor summary: Flow is a per-instance personalization FL algorithm that adapts dynamic models to each client's data distributions and instances, improving clients' accuracy.
Jiachang Liu, Sam Rosen, Chudi Zhong, Cynthia Rudin
https://openreview.net/forum?id=BHxsP5fSHv
Keywords: Sparse Ridge Regression, Dynamical Systems
Compressor summary: The authors propose OKRidge, a fast algorithm for identifying sparse governing equations in nonlinear dynamical systems using a novel lower bound calculation and beam search warm-starting.
Kiarash Banihashem, MohammadTaghi Hajiaghayi, Suho Shin, Max Springer
https://openreview.net/forum?id=BHZsJ2sTkG
Keywords: contextual bandits, adversarial bandits, oracle-efficient online learning
Compressor summary: The paper presents a new algorithm for adversarial contextual bandits that improves existing regret bounds and makes fewer calls to an offline optimization oracle per round.
Shibo Hao, Tianyang Liu, Zhen Wang, Zhiting Hu
https://openreview.net/forum?id=BHXsb69bSx
Keywords: large language model, tool learning
Compressor summary: The authors propose ToolkenGPT, which enables large language models to master tools by using vector embeddings of tools that are plugged into the LLM head and executed during text generation.
Cansu Sancaktar, Justus Piater, Georg Martius
https://openreview.net/forum?id=BHHrX3CRE1
Keywords: Intrinsic Motivation, Reinforcement Learning, Model-based Planning, Regularity, Manipulation, Zero-shot Generalization, Unsupervised Exploration
Compressor summary: We propose regularity as a novel reward signal for intrinsically-motivated reinforcement learning. Taking inspiration from child development, we postulate that striving for structure and order helps guide exploration towards a subspace of tasks that are not favored by naive uncertainty-based intrinsic rewards. Our generalized formulation of Regularity as Intrinsic Reward (RaIR) allows us to operationalize it within model-based reinforcement learning. In a synthetic environment, we showcase the plethora of structured patterns that can emerge from pursuing this regularity objective. We also demonstrate the strength of our method in a multi-object robotic manipulation environment. We incorporate RaIR into free play and use it to complement the model’s epistemic uncertainty as an intrinsic reward. Doing so, we witness the autonomous construction of towers and other regular structures during free play, which leads to a substantial improvement in zero-shot downstream task performance on assembly tasks.
Xinyi Wang, Wanrong Zhu, Michael Saxon, Mark Steyvers, William Yang Wang
https://openreview.net/forum?id=BGvkwZEGt7
Keywords: Large language models, Bayesian explanation, in-context learning
Compressor summary: The study examines in-context learning using Bayesian methods and proposes an algorithm to select optimal demonstrations for few-shot text classification, improving performance on real-world datasets.
Shiwei Liu, Tian Zhu, Milong Ren, Yu Chungong, Dongbo Bu, Haicang Zhang
https://openreview.net/forum?id=BGP5Vjt93A
Keywords: Riemannian Diffusion Probabilistic Model, Mutation, Protein-protein binding
Compressor summary: SidechainDiff is a new method that uses unlabelled experimental data to predict how mutations affect protein-protein binding by learning side-chain conformations and structural context representations.
Tejas Jayashankar, Gary C.F. Lee, Alejandro Lancho, Amir Weiss, Yury Polyanskiy, Gregory Wornell
https://openreview.net/forum?id=BFGQQKicuu
Keywords: Diffusion models, score-based models, source separation, digital communications, maximum a posteriori (MAP) estimation, alpha-posterior, Gaussian smoothing, score distillation sampling, radio frequency systems, interference mitigation
Compressor summary: The authors propose a new method for separating mixed signals using diffusion-based generative models that improves bit error rate in radio-frequency systems and has applications beyond conditional sampling.
Bipasha Sen, Gaurav Singh, Aditya Agarwal, Rohith Agaram, Madhava Krishna, Srinath Sridhar
https://openreview.net/forum?id=BExDjNDYkN
Keywords: neural radiance field, hypernetwork, multi-hash encoding, NeRF
Compressor summary: HyP-NeRF is a method for learning generalizable NeRF priors using hypernetworks that improves quality and enables multiple downstream tasks.
Guillem Simeon, Gianni De Fabritiis
https://openreview.net/forum?id=BEHlPdBZ2e
Keywords: Neural network interatomic potentials, Equivariant graph neural network, Message passing neural network
Compressor summary: TensorNet is a novel neural network architecture that uses Cartesian tensor representations to efficiently model molecular systems with reduced computational cost and improved performance.
Alex Foo, Wynne Hsu, Mong-Li Lee
https://openreview.net/forum?id=BDno5qWEFh
Keywords: Object-Centric Learning, Multi-Object Representation Learning
Compressor summary: The paragraph describes a new method for discovering object-centered representations from images that improves robustness, sample efficiency, and interpretability, and outperforms existing methods on various types of real-world images.
Alvin Heng, Harold Soh
https://openreview.net/forum?id=BC1IJdsuYB
Keywords: generative models, forgetting
Compressor summary: Selective Amnesia is a technique that allows controlling the forgetting of specific concepts in deep generative models to prevent misuse for harmful or inappropriate content.
Zijian Li, Ruichu Cai, Guangyi Chen, Boyang Sun, Zhifeng Hao, Kun Zhang
https://openreview.net/forum?id=BACQLWQW8u
Keywords: Domain Adaptation, Identification
Compressor summary: The paper proposes a new method for transferring knowledge between multiple labeled source domains to an unlabeled target domain using subspace identification and variational inference, which works better than existing methods.
Oren Mangoubi, Nisheeth K Vishnoi
https://openreview.net/forum?id=BA7NHAzbpO
Keywords: Logconcave sampling, Dikin walk, Markov chain Monte Carlo, interior point methods
Compressor summary: The paper presents a new Dikin walk algorithm to sample from a log-concave distribution defined by a Lipschitz or smooth convex function on a bounded polytope, with improved efficiency and acceptance ratio compared to previous works.
Phoenix Neale Williams, Ke Li
https://openreview.net/forum?id=B94G0MXWQX
Keywords: Evolutionary Strategy, Adversarial Attack, Adversarial Patches, Computational Art, Computer Vision
Compressor summary: Deep neural networks (DNNs) have demonstrated vulnerabilities to adversarial examples, which raises concerns about their reliability in safety-critical applications. While the majority of existing methods generate adversarial examples by making small modifications to the entire image, recent research has proposed a practical alternative known as adversarial patches. Adversarial patches have shown to be highly effective in causing DNNs to misclassify by distorting a localized area (patch) of the image. However, existing methods often produce clearly visible distortions since they do not consider the visibility of the patch. To address this, we propose a novel method for constructing adversarial patches that approximates the appearance of the area it covers. We achieve this by using a set of semi-transparent, RGB-valued circles, drawing inspiration from the computational art community. We utilize an evolutionary strategy to optimize the properties of each shape, and employ a simulated annealing approach to optimize the patch's location. Our approach achieves better or comparable performance to state-of-the-art methods on ImageNet DNN classifiers while achieving a lower $l_2$ distance from the original image. By minimizing the visibility of the patch, this work further highlights the vulnerabilities of DNNs to adversarial patches.
Zhengmian Hu, Heng Huang
https://openreview.net/forum?id=B7QkdEnjL9
Keywords: Bayesian learning, Generalization
Compressor summary: The paper introduces TansBL, which combines ERM and Bayesian learning for neural networks, balancing generalization and computational complexity.
Yudong Luo, Guiliang Liu, Pascal Poupart, Yangchen Pan
https://openreview.net/forum?id=B7QRV4XXiK
Keywords: risk-averse RL, mean-variance RL
Compressor summary: The paper proposes a new risk measure for Reinforcement Learning, called Gini deviation, which addresses the limitations of variance-based methods and improves policy learning.
Yongbo Chen, Hanna Kurniawati
https://openreview.net/forum?id=B6qZdrGRpm
Keywords: robotics, Partially Observable Markov Decision Process (POMDP), object search
Compressor summary: The paper proposes a method for mobile robots to search for objects in complex environments using a POMDP formulation, a perception module, and a planning algorithm that improves performance over baseline approaches.
Jing Dong, Yuichi Yoshida
https://openreview.net/forum?id=B6HSIgvyJ3
Keywords: online learning, random model setting
Compressor summary: The framework transforms offline approximation algorithms into online ones with low $\epsilon$-approximate regret and applies to various problems like clustering, matrix approximation, and regression.
Taoli Zheng, Linglingzhi Zhu, Anthony Man-Cho So, Jose Blanchet, Jiajin Li
https://openreview.net/forum?id=B6FihisDBl
Keywords: Nonconvex-Nonconcave Minimax Optimization, Limit Cycle
Compressor summary: The paper proposes a new optimization algorithm, DS-GDA, that works for various nonconvex-nonconcave problems and achieves convergence with better complexity than existing methods.
Junyi Li, Feihu Huang, Heng Huang
https://openreview.net/forum?id=B5XwENgy0T
Keywords: Federated Learning, Bilevel Optimization
Compressor summary: The paper proposes FedBiOAcc, a communication-efficient algorithm for Federated Bilevel Optimization problems with good theoretical and empirical performance.
Udaya Ghai, Arushi Gupta, Wenhan Xia, Karan Singh, Elad Hazan
https://openreview.net/forum?id=B5LpWAaBVA
Keywords: Control, Reinforcement Learning, Online Learning, Regret Minimization, Bandit Linear Control
Compressor summary: The authors propose a novel policy class based on disturbance signals and introduce efficient algorithms for optimizing them, improving robustness in model-free reinforcement learning agents facing adversarial or dynamic environments.
Alexander Tyurin, Peter Richtárik
https://openreview.net/forum?id=B4xF1wfQnF
Keywords: nonconvex optimization, convex optimization, parallel methods, asynchronous methods, lower bounds
Compressor summary: The paper introduces a new protocol for parallel optimization methods with an unbiased stochastic gradient oracle, establishes minimax complexities, and reveals implications for asynchronous optimization methods.
Bailin Wang, Zi Wang, Xuezhi Wang, Yuan Cao, Rif A. Saurous, Yoon Kim
https://openreview.net/forum?id=B4tkwuzeiY
Keywords: semantic parsing, large language models, PDDL, AI planning, molecule generation, data efficiency, grammar-based learning
Compressor summary: Grammar prompting helps large language models generate outputs for structured languages by using external knowledge and domain-specific constraints from a grammar in BNF during in-context learning.
Peter Macgregor, He Sun
https://openreview.net/forum?id=B4G87Bq5wA
Keywords: similarity graphs, spectral clustering
Compressor summary: The paper proposes a new algorithm to efficiently construct a sparse similarity graph that preserves cluster structure for data points in high dimensions using kernel density estimation.
YIZHOU CHEN, Anxiang Zeng, Qingtao Yu, Kerui Zhang, Cao Yuanpeng, Kangle Wu, Guangda Huzhang, Han Yu, Zhiming Zhou
https://openreview.net/forum?id=B3UDx1rNOy
Keywords: temporal graph, temporal graph network, temporal graph model expressiveness, continuous-time dynamic graph
Compressor summary: The paragraph discusses a novel framework for temporal neighbor aggregation that uses recurrent neural networks to integrate information from all historical neighbors, improving accuracy and expressiveness in graph networks.
Yo Joong Choe, Aditya Gangrade, Aaditya Ramdas
https://openreview.net/forum?id=B2DEcj4a7i
Keywords: abstaining classifiers, black-box model evaluation, causal inference, missing data
Compressor summary: The paper proposes a method to evaluate abstaining classifiers by treating their abstentions as missing data and estimating their counterfactual performance.
Yongrui Chen, Shenyu Zhang, Guilin Qi, Xinnan Guo
https://openreview.net/forum?id=B01uiWhjpc
Keywords: semantic parsing, continual learning, few-shot learning
Compressor summary: This paper introduces a novel method combining parameter-efficient fine-tuning and in-context tuning for training a continual table semantic parser that overcomes challenges like catastrophic forgetting and limited supervision.
Valentyn Melnychuk, Dennis Frauen, Stefan Feuerriegel
https://openreview.net/forum?id=AygwZzdCM0
Keywords: causal inference, counterfactual inference, partial identification, sensitivity model, normalizing flows, causal machine learning
Compressor summary: The paper proposes a novel sensitivity model and deep generative model to relax assumptions and achieve partial counterfactual identification of continuous outcomes in structural causal models.
Dylan J Foster, Noah Golowich, Jian Qian, Alexander Rakhlin, Ayush Sekhari
https://openreview.net/forum?id=Ay3WvSrtpO
Keywords: Decision making, learning theory, bandits, reinforcement learning theory, online learning, decision-estimation coefficient
Compressor summary: The paper proposes a method to improve interactive decision making by combining a meta-algorithm with a specialized estimation technique, achieving better guarantees and accommodating more lenient notions of estimation error.
Pascal Notin, Ruben Weitzman, Debora Susan Marks, Yarin Gal
https://openreview.net/forum?id=AwzbQVuDBk
Keywords: Non-parametric transformers, protein design, protein property prediction, fitness prediction, Bayesian optimization, ProteinGym
Compressor summary: ProteinNPT is a non-parametric transformer variant for protein sequence optimization that outperforms existing methods and enables iterative protein design in multi-task settings with label scarcity.
Yule Wang, Zijing Wu, Chengrui Li, Anqi Wu
https://openreview.net/forum?id=AuXd54odxm
Keywords: Neural distribution alignment, Diffusion model, Neuroscience, Neural decoding
Compressor summary: The proposed ERDiff method aligns neural signals across domains by using a diffusion model to preserve spatio-temporal structures in latent dynamics, resulting in better alignment quality and neural decoding performance.
Yeyuan Chen, Dingmin Wang
https://openreview.net/forum?id=AtHJ7TLheF
Keywords: Knowledge Graphs, First-Order Logic, Temporal Knowledge Graph, Graph Neural Networks
Compressor summary: The paper analyzes the expressiveness of Graph Neural Networks (GNNs) for Boolean node classification over multi-relational graphs, proposes a graph transformation technique to improve GNNs' capabilities, and tests their approach empirically.
Dipam Goswami, Yuyang Liu, Bartłomiej Twardowski, Joost van de Weijer
https://openreview.net/forum?id=Asx5eDqFZl
Keywords: Continual Learning, Class-Incremental Learning
Compressor summary: The paper proposes a prototypical network approach for class-incremental learning that uses anisotropic Mahalanobis distance and feature covariance modeling, achieving state-of-the-art results without updating the backbone network.
Junlin Wu, Andrew Clark, Yiannis Kantaros, Yevgeniy Vorobeychik
https://openreview.net/forum?id=ArRycLMoUg
Keywords: nonlinear systems, Lyapunov stability, neural Lyapunov control
Compressor summary: The authors propose a novel method for learning neural Lyapunov control in discrete-time systems using mixed-integer linear programming, sublevel sets computation, and gradient-based counterexample finding, achieving faster and better results on four benchmarks than existing approaches.
Kavosh Asadi, Rasool Fakoor, Shoham Sabach
https://openreview.net/forum?id=AnFUgNC3Yc
Keywords: Deep Reinforcement Learning, Rainbow Algorithm, Atari benchmark, Adam Optimizer
Compressor summary: The text discusses how resetting internal parameters of optimization algorithms, such as Adam, can improve the performance of deep reinforcement learning on Atari games.
Jacy Reese Anthis, Victor Veitch
https://openreview.net/forum?id=AmwgBjXqc3
Keywords: causal graphs, causality, counterfactual fairness, domain generalization, fairness, robustness, machine learning, artificial intelligence
Compressor summary: The paper proposes a causal context approach to relate counterfactual fairness, robust prediction, and group fairness in data-generating processes and shows how counterfactual fairness is equivalent to different group fairness metrics in various scenarios.
Yang He, Lingao Xiao, Joey Tianyi Zhou
https://openreview.net/forum?id=AlTyimRsLf
Keywords: Dataset Condensation, Dataset Pruning
Compressor summary: YOCO is a method to adjust the size of a condensed dataset for efficient on-device training using simple rules and improving accuracy.
Dayal Singh Kalra, Maissam Barkeshli
https://openreview.net/forum?id=Al9yglQGKj
Keywords: Optimization dynamics, Phase diagrams, learning rate transition, Catapult effect
Compressor summary: The paper studies how the learning rate, depth, and width of deep neural networks affect their optimization dynamics and the sharpness of the loss landscape during training with stochastic gradient descent.
Metod Jazbec, James Urquhart Allingham, Dan Zhang, Eric Nalisnick
https://openreview.net/forum?id=Akslsk891N
Keywords: anytime algorithms, early-exit neural networks, conditional monotonicity, anytime uncertainty
Compressor summary: The paper proposes a modification to early-exit neural networks that allows them to provide better predictions as computation time increases, enabling anytime predictive modeling with these architectures.
Nikita Tsoy, Nikola Konstantinov
https://openreview.net/forum?id=AkK3S2spZs
Keywords: Incentives, collaborative learning, federated learning, game theory, competition, oligopolistic markets, strategic behavior, Nash equilibrium
Compressor summary: The paper presents a framework to analyze how data sharing affects machine learning model quality and profitability in collaborative learning settings, and finds that market conditions and task difficulty influence collaboration incentives.
Zijie Geng, Xijun Li, Jie Wang, Xiao Li, Yongdong Zhang, Feng Wu
https://openreview.net/forum?id=AiEipk1X0c
Keywords: Learning to Optimize, Machine Learning for Combinatorial Optimization, Mixed-Integer Linear Programming, Graph Generation
Compressor summary: G2MILP is a deep generative framework that creates realistic and novel mixed-integer linear programs (MILPs) from bipartite graphs without prior expert knowledge, improving MILP solver performance under limited data availability.
Ayoub El Hanchi, Murat A Erdogdu
https://openreview.net/forum?id=Ah2Q8mLH96
Keywords: Excess risk bounds, Linear regression, Lp-norm, Fast rates
Compressor summary: The paper investigates how many samples are needed for different p-norms in linear regression and proves bounds on the performance of empirical risk minimization.
Subhojyoti Mukherjee, Qiaomin Xie, Josiah P. Hanna, Robert D Nowak
https://openreview.net/forum?id=AfC8PVQZ9z
Keywords: Linear Bandits, Experimental design, Pure Exploration, Representation Learning
Compressor summary: The paper proposes GOBLIN, a multi-task representation learning algorithm for bilinear bandits that reduces exploration time and improves sample complexity by sharing a common low-dimensional linear representation across tasks.
Fuzhao Xue, Yao Fu, Wangchunshu Zhou, Zangwei Zheng, Yang You
https://openreview.net/forum?id=Af5GvIj3T5
Keywords: Large Language Model, Transformer Scaling, Foundation Model Pre-training
Compressor summary: The study investigates how repeating pre-training data affects large language models, causing overfitting and multi-epoch degradation, and explores factors contributing to it, as well as regularization techniques and hyper-parameter tuning methods.
Omid Sadeghi, Maryam Fazel
https://openreview.net/forum?id=AesN5bYnJr
Keywords: incentive-compatible, online prediction with expert advice, forecasting
Compressor summary: The authors study a generalization of online binary prediction with expert advice where experts act strategically and aim to maximize their influence, and they design algorithms that encourage truthful reporting and achieve sublinear regret.
Ganyu Wang, Bin Gu, Qingsong Zhang, Xiang Li, Boyu Wang, Charles Ling
https://openreview.net/forum?id=AYiRHZirD2
Keywords: Vertical Federated Learning, Zeroth Order Optimization, Communication Efficiency, Privacy
Compressor summary: The paper proposes a cascaded hybrid optimization approach for Vertical Federated Learning that improves convergence and privacy protection by using a zeroth-order gradient on the output layer and a first-order gradient elsewhere, with experimental results showing its advantages over existing methods.
Jiahui Li, Kun Kuang, Baoxiang Wang, Xingchen Li, Fei Wu, Jun Xiao, Long Chen
https://openreview.net/forum?id=AYLlZMmUbo
Keywords: multi-agent reinforcement learning, influence-based exploration
Compressor summary: COIN is an exploration method for cooperative multi-agent reinforcement learning that combines curiosity-based and influence-based intrinsic rewards to effectively explore large and complex state spaces.
Hanhan Zhou, Tian Lan, Guru Prasadh Venkataramani, Wenbo Ding
https://openreview.net/forum?id=AWpWaub6nf
Keywords: Federated Learning, Optimization, Deep Learning
Compressor summary: The paper proposes a framework for heterogeneous FL algorithms with online model extraction and proves their convergence under certain conditions for IID and non-IID data, while introducing the concept of minimum coverage index to improve efficiency.
Xin Liu, Zheng Li, Yifan Gao, Jingfeng Yang, Tianyu Cao, Zhengyang Wang, Bing Yin, Yangqiu Song
https://openreview.net/forum?id=AV3iZlDrzF
Keywords: session-based recommendation, representation learning, pattern mining
Compressor summary: FAPAT is a new method for session-based recommendation in E-commerce that uses attribute transition graphs and frequent attribute patterns to capture user intents and outperforms existing methods by 4.5% on average.
Yunchang Yang, Han Zhong, Tianhao Wu, Bin Liu, Liwei Wang, Simon Shaolei Du
https://openreview.net/forum?id=AT6NaLPwy0
Keywords: sequential decision making, delay, reinforcement learning
Compressor summary: The paper presents a framework to improve sample efficiency in sequential decision making with stochastic delays using different multi-batched algorithms.
Sivan Doveh, Assaf Arbelle, Sivan Harary, Roei Herzig, Donghyun Kim, Paola Cascante-Bonilla, Amit Alfassy, Rameswar Panda, Raja Giryes, Rogerio Feris, Shimon Ullman, Leonid Karlinsky
https://openreview.net/forum?id=ARrwf7Ev2T
Keywords: computer vision, deep learning, vision and language models
Compressor summary: The paper identifies two factors that limit VL models' compositional reasoning performance and proposes a fine-tuning approach to improve it on a standard paired VL dataset, leading to significant performance increase.
Lennert De Smet, Emanuele Sansone, Pedro Zuidberg Dos Martires
https://openreview.net/forum?id=AQyqxXctsN
Keywords: gradient estimation, categorical random variables, probability theory, discrete distributions
Compressor summary: The CatLog-Derivative trick improves upon the Log-Derivative trick by accounting for the discrete nature of categorical distributions, leading to a more efficient and unbiased gradient estimator called IndeCateR.
Tianwei Ni, Michel Ma, Benjamin Eysenbach, Pierre-Luc Bacon
https://openreview.net/forum?id=APGXBNkt6h
Keywords: Memory-based RL, Transformers, Credit Assignment, Online RL, Model-free RL
Compressor summary: The paper investigates why Transformer architecture works well for reinforcement learning problems involving long-term dependencies, finding that it enhances memory capability but not credit assignment.
Royi Rassin, Eran Hirsch, Daniel Glickman, Shauli Ravfogel, Yoav Goldberg, Gal Chechik
https://openreview.net/forum?id=AOKU4nRw1W
Keywords: syntax, diffusion, stable diffusion, attribute, attention
Compressor summary: SynGen improves text-to-image generation by using syntax to guide cross-attention maps and maintain correct associations between entities and their visual attributes.
Xingrui Wang, Wufei Ma, Zhuowan Li, Adam Kortylewski, Alan Yuille
https://openreview.net/forum?id=AMIJEupsNq
Keywords: VQA, reasoning, 3D scene understanding, analysis-by-synthesis, neural modular network, neuro-symbolic reasoning
Compressor summary: The paragraph introduces the task of 3D-aware VQA, which involves answering questions about the 3D structure of visual scenes, and presents a new dataset (Super-CLEVR-3D) and a novel model (PO3D-VQA) to address this challenge.
Zhijin Ge, Xiaosen Wang, Hongying Liu, Fanhua Shang, Yuanyuan Liu
https://openreview.net/forum?id=AKAMNDe2Sw
Keywords: Adversarial attack, Adversarial transferability, Black-box Attack
Compressor summary: The authors propose a method to generate adversarial examples that transfer better by optimizing the gradient norm and using random sampling and averaging of gradients, achieving improved transferability on both normal and adversarial models.
Hoki Kim, Jinseong Park, Yujin Choi, Jaewook Lee
https://openreview.net/forum?id=AGVBqJuL0T
Keywords: Adversarial Robustness, Generalization, Measures
Compressor summary: The study analyzes how different measures of robust generalization relate to adversarial examples in various settings using over 1,300 models trained on CIFAR-10 and more than 100 models from RobustBench.
James Cheshire, Vincent Laurent, Stephan Clémençon
https://openreview.net/forum?id=AGMVzMGcGP
Keywords: bipartite ranking, multi armed bandits, active learning
Compressor summary: The paper presents an active learning framework for bipartite ranking and proposes a new algorithm called active-rank that minimizes the distance between the ROC curve of the ranking function and the optimal one.
Xunhan Hu, Jian Zhao, Wengang Zhou, Ruili Feng, Houqiang Li
https://openreview.net/forum?id=AG9A7Ae9r3
Keywords: Experience Replay; Reinforcement Learning; Multi-Agent System
Compressor summary: The paper proposes DIFFER, a framework to decompose individual rewards in cooperative multi-agent reinforcement learning for more efficient and fair learning.
Chen-Hao Chao, Wei-Fang Sun, Yen-Chang Hsu, Zsolt Kira, Chun-Yi Lee
https://openreview.net/forum?id=AALLvnv95q
Keywords: flow-based models, score-matching methods
Compressor summary: The paper introduces EBFlow, a flow-based modeling method that connects flow- and energy-based models, optimizes with score-matching for faster training and better performance than existing methods.
Yu-Hu Yan, Peng Zhao, Zhi-Hua Zhou
https://openreview.net/forum?id=AA1xrgAP5z
Keywords: online learning
Compressor summary: The paper proposes an online convex optimization method that adapts to different function types and can achieve problem-dependent guarantees, while using only one gradient query per round.
Dieterich Lawson, Michael Y. Li, Scott Linderman
https://openreview.net/forum?id=A9mHph8GJk
Keywords: sequence models, probabilistic inference, reweighted wake-sleep, sequential monte carlo, smoothing, mechanistic models
Compressor summary: NAS-X is a method for approximating inferences and model learning in complex sequential latent variable models using neural adaptive smoothing and twisting, which improves performance over previous methods in various applications, including neuronal dynamics.
Yun Yue, Zhiling Ye, Jiadi Jiang, Yongchao Liu, Ke Zhang
https://openreview.net/forum?id=A954O4tDmU
Keywords: adaptive optimizer, gradient difference, auto switch, AGD
Compressor summary: AGD is a new optimizer that combines a dynamic preconditioning matrix and an auto-switching function between SGD and adaptive optimizers, improving generalization performance on NLP, CV, and RecSys datasets.
RENCHUNZI XIE, Hongxin Wei, Lei Feng, Yuzhou Cao, Bo An
https://openreview.net/forum?id=A86JTXllHa
Keywords: Machine Learning, Uncertainty Estimation
Compressor summary: The paper proposes a dataset-level score based on feature dispersion to estimate test accuracy under distribution shift, showing that inter-class dispersion is strongly correlated with model accuracy on out-of-distribution data.
Michael Tschannen, Manoj Kumar, Andreas Peter Steiner, Xiaohua Zhai, Neil Houlsby, Lucas Beyer
https://openreview.net/forum?id=A7feCufBhL
Keywords: contrastive learning, CLIP, CapPa, Cap, vision-language, image captioning, visual representation learning, weakly supervised learning, VLM, multimodal learning, VQA, image classification
Compressor summary: The paper compares image captioning and contrastive pretraining for vision backbones, finding that captioning alone produces competitive or better results on various tasks.
Caleb Dahlke, Jason Pacheco
https://openreview.net/forum?id=A7ESFTMJWs
Keywords: entropy, Gaussian mixture model, uncertainty quantification, approximate inference
Compressor summary: The paper proposes a new Taylor series for Gaussian mixture model entropy that converges to the true value and improves accuracy over previous methods.
Minghua Liu, Chao Xu, Haian Jin, Linghao Chen, Mukund Varma T, Zexiang Xu, Hao Su
https://openreview.net/forum?id=A6X9y8n4sT
Keywords: single image reconstruction, 3d generation, mesh reconstruction, diffusion models
Compressor summary: The authors propose a novel method that generates a 360-degree 3D textured mesh from a single image using a view-conditioned 2D diffusion model and an SDF-based neural surface reconstruction method, achieving better geometry, faster runtime, and supporting the text-to-3D task.
Jiachen Liang, RuiBing Hou, Hong Chang, Bingpeng Ma, Shiguang Shan, Xilin CHEN
https://openreview.net/forum?id=A6PRwRjI8V
Keywords: semi-supervised learning, self-supervised learning
Compressor summary: The paper proposes a new semi-supervised learning setting where unlabeled data has a different feature distribution than labeled data, and introduces SSFA, a framework that adapts the model features to improve pseudo-label quality.
Chao Li, Chen GONG, Qiang He, Xinwen Hou
https://openreview.net/forum?id=A6JDQDv7Nt
Keywords: Reinforcement Learning, Ensemble Exploration, Control Tasks
Compressor summary: TEEN is a new ensemble RL algorithm that enhances diversity and performance of sub-policies in complex sequential decision-making problems.
Alexander Immer, Emanuele Palumbo, Alexander Marx, Julia E Vogt
https://openreview.net/forum?id=A6EquH0enk
Keywords: Heteroscedastic Regression, Marginal Likelihood, Bayesian Neural Networks, Uncertainty Estimaton, Model Selection, Laplace Approximation
Compressor summary: The authors propose an efficient method to handle uncertainties in complex regression problems using neural networks and Gaussian processes, which improves generalization and requires no hyperparameter tuning.
Giorgos Chionas, Dariusz Rafal Kowalski, Piotr Krysta
https://openreview.net/forum?id=A5yMv7XPuA
Keywords: Combinatorial Group Testing, Adversarial Equilibrium, Contention Resolution, selfish agents, learning time, adaptive learning algorithms
Compressor summary: The paper presents a game-theoretic approach to Combinatorial Group Testing with selfish agents and shows how adaptive strategies can achieve near-optimal learning time depending on whether the number of active agents is known or not.
Zhihan Liu, Miao Lu, Wei Xiong, Han Zhong, Hao Hu, Shenao Zhang, Sirui Zheng, Zhuoran Yang, Zhaoran Wang
https://openreview.net/forum?id=A57UMlUJdc
Keywords: reinforcement learning; online learning; game
Compressor summary: MEX is an easy-to-implement RL framework that balances exploration and exploitation automatically and outperforms baselines in various MuJoCo environments.
Yong Liu, Chenyu Li, Jianmin Wang, Mingsheng Long
https://openreview.net/forum?id=A4zzxu82a7
Keywords: Time series forecasting, Deep learning
Compressor summary: Koopa is a novel Koopman forecaster that uses modern theory to handle non-stationary time series with complex dynamics, achieving competitive performance while reducing training time and memory.
Yankun Huang, Qihang Lin
https://openreview.net/forum?id=A383wMho4h
Keywords: Constrained optimization, first-order method, non-smooth optimization, non-convex optimization
Compressor summary: The paper analyzes the oracle complexity of the switching subgradient method for solving non-convex constrained optimization problems with weakly convex or convex constraint functions, and shows that it outperforms double-loop methods in terms of simplicity and efficiency.
Dongsheng Wang, Miaoge Li, Xinyang Liu, MingSheng Xu, Bo Chen, Hanwang Zhang
https://openreview.net/forum?id=A253n2EXCd
Keywords: Multi-modal prompt learning; Optimal transport
Compressor summary: The text describes a new method for improving vision-language models by tuning multiple prompts across modalities using optimal transportation, which enhances semantic alignment and leads to better image recognition performance.
Filip Ekström Kelvinius, Dimitar Georgiev, Artur Toshev, Johannes Gasteiger
https://openreview.net/forum?id=A18PgVSUgf
Keywords: GNN, graph neural networks, knowledge distillation, molecules, molecular simulations
Compressor summary: The paper explores using knowledge distillation to accelerate molecular graph neural networks without sacrificing predictive accuracy in energy and force prediction tasks.
Chenglin Fan, Ping Li, Xiaoyun Li
https://openreview.net/forum?id=9zV2OXCrVF
Keywords: privacy, clustering
Compressor summary: The paper introduces a new initialization method for k-median problem in metric spaces called HST, which produces better and more efficient initial centers than existing methods, and can be extended to generate private initial centers with differential privacy.
Valentino delle Rose, Alexander Kozachinskiy, Cristobal Rojas, Mircea Petrache, Pablo Barcelo
https://openreview.net/forum?id=9yhYcjsdab
Keywords: euclidean graphs, point clouds, WL test, graph neural networks
Compressor summary: The Weisfeiler-Lehman test can distinguish point clouds in Euclidean space using a limited number of iterations and is complete for three dimensions or higher.
Marco Bellagente, Manuel Brack, Hannah Benita Teufel, Felix Friedrich, Björn Deiseroth, Constantin Eichenberg, Andrew Dai, Robert John Nicholas Baldock, Souradeep Nanda, Koen Oostermeijer, Andres Felipe Cruz-Salinas, Patrick Schramowski, Kristian Kersting, Samuel Weinbach
https://openreview.net/forum?id=9ych3krqP0
Keywords: diffusion, image generation, multimodal
Compressor summary: MultiFusion is a system that enables complex and nuanced image generation from multiple modalities and languages by fusing pre-trained models without extensive training.
Jean Tarbouriech, Tor Lattimore, Brendan O'Donoghue
https://openreview.net/forum?id=9yQ2aaArDn
Keywords: Reinforcement learning, Bayesian inference, Exploration
Compressor summary: The paper proposes VAPOR, a new variational Bayesian approach for Reinforcement learning that efficiently approximates the posterior probability of state-action optimality and leads to better exploration policies.
Lisa Dunlap, Alyssa Umino, Han Zhang, Jiezhi Yang, Joseph E. Gonzalez, Trevor Darrell
https://openreview.net/forum?id=9wrYfqdrwk
Keywords: data augmentation, diffusion, vision and language
Compressor summary: ALIA uses natural language descriptions and large vision models to generate diverse and visually consistent image augmentations for fine-grained classification tasks, improving performance on domain generalization and contextual bias.
Yutong He, Xinmeng Huang, Kun Yuan
https://openreview.net/forum?id=9v6gpFTfCM
Keywords: Communication Compression, Distributed Optimization, Unbiased Compression, Optimal Complexity
Compressor summary: The paper investigates when unbiased communication compression reduces total communication cost in distributed optimization, showing that it can achieve this under certain conditions, and provides theoretical and empirical evidence for its effectiveness.
Long-Fei Li, Peng Zhao, Zhi-Hua Zhou
https://openreview.net/forum?id=9tUjsRLjf2
Keywords: dynamic regret, adversarial MDPs, linear mixture MDPs, policy optimization
Compressor summary: The paper proposes a new algorithm for reinforcement learning in adversarial MDPs with unknown transition kernels, achieving near-optimal dynamic regret.
Matthew A Fisher, Chris J. Oates
https://openreview.net/forum?id=9rmwPAjk9O
Keywords: Bayesian, discrepancy, kernel, sampling, Stein's method
Compressor summary: The paper proposes non-canonical Stein discrepancies, which don't need derivatives, for posterior approximation in complex models, with convergence guarantees and applications in sampling and variational inference.
Guanlin Liu, Lifeng Lai
https://openreview.net/forum?id=9qlJGjO7bA
Keywords: adversarial attacks; multi agent reinforcement learning;
Compressor summary: The paper investigates how adversarial attacks affect multi-agent reinforcement learning and proposes a mixed attack strategy with both action poisoning and reward poisoning that can efficiently target MARL agents.
Beier Zhu, Kaihua Tang, Qianru Sun, Hanwang Zhang
https://openreview.net/forum?id=9qG6cMGUWk
Keywords: Foundation Model, Class Bias, Generalized Logit Adjustment
Compressor summary: The paper proposes a method called Generalized Logit Adjustment (GLA) to reduce biases in foundation models like CLIP and improve their performance on various tasks.
Yuanze Wang, Yichao Yan, Dianxi Shi, Wenhan Zhu, Jianqiang Xia, Tan Jeff, Songchang Jin, KE GAO, XIAOBO LI, Xiaokang Yang
https://openreview.net/forum?id=9pLaDXX8m3
Keywords: NeRF, Image-Based Visual Servoing (IBVS), visual localization, visual navigation
Compressor summary: The paper proposes a novel visual localization method that uses few posed images with coarse pseudo-3D labels from NeRF to train a coordinate regression network, estimate a coarse pose with PNP, optimize the pose with image-based visual servo, and provide navigation prior for navigation without custom markers or depth sensors.
Yusong Wang, Shaoning Li, Tong Wang, Bin Shao, Nanning Zheng, Tie-Yan Liu
https://openreview.net/forum?id=9o6KQrklrE
Keywords: Geometric Deep Learning, Molecular Modeling, Positional Encoding
Compressor summary: The paper proposes Geoformer, a novel geometric Transformer that uses Interatomic Positional Encoding to effectively model molecular structures for various property prediction and outperforms existing algorithms on benchmark datasets.
Qingyao Sun, Kevin Patrick Murphy, Sayna Ebrahimi, Alexander D'Amour
https://openreview.net/forum?id=9mJXDcr17V
Keywords: Distribution shift, Spurious correlation, Group robustness
Compressor summary: The paper proposes a method (TTLSA) to correct for label shifts in test data that includes additional meta-data labels, and shows improved performance on various datasets.
Jia-Xing Zhong, Ta-Ying Cheng, Yuhang He, Kai Lu, Kaichen Zhou, Andrew Markham, Niki Trigoni
https://openreview.net/forum?id=9lygTqLdWn
Keywords: Dynamic Point Cloud Analytics, Multi-body Motion
Compressor summary: The authors propose a novel method for segmenting and estimating motion in 3D scenes without category information, using an unsupervised training strategy and lightweight architecture.
Yeqi BAI, Ben Fei, Youquan Liu, Tao MA, Yuenan Hou, Botian Shi, Yikang LI
https://openreview.net/forum?id=9kFQEJSyCM
Keywords: 3D Detection, Autonomous Driving
Compressor summary: RangePerception is an efficient and accurate RV-based 3D object detection framework that addresses two challenges in existing methods, achieving higher performance and speed than BEV-based methods.
Xiaoyuan Zhang, Xi Lin, Bo Xue, Yifan Chen, Qingfu Zhang
https://openreview.net/forum?id=9ieV1hnuva
Keywords: multiobjective optimization;multitask learning;hypervolume maximization;Pareto set learning
Compressor summary: The paper proposes a new neural network-based method for modeling the entire Pareto set in multiobjective optimization and shows its effectiveness on several problems.
Yun Xing, Jian Kang, Aoran Xiao, Jiahao Nie, Ling Shao, Shijian Lu
https://openreview.net/forum?id=9iafshF7s3
Keywords: language-supervised semantic segmentation, vision-language pre-training
Compressor summary: CoCu is a method that uses CLIP to find and incorporate missing visual concepts from image-text pairs into pre-training, improving zero-shot recognition and language-supervised segmentation performance.
Elan Rosenfeld, Saurabh Garg
https://openreview.net/forum?id=9i8MD9btc8
Keywords: accuracy estimation, error bounds, distribution shift, unsupervised domain adaptation
Compressor summary: The paper proposes a new bound on the error of deep neural networks under distribution shift using unlabeled test data, which is simpler, more accurate, and easier to evaluate than previous methods.
Guangyu Shen, Siyuan Cheng, Guanhong Tao, Kaiyuan Zhang, Yingqi Liu, Shengwei An, Shiqing Ma, Xiangyu Zhang
https://openreview.net/forum?id=9fb975Au9G
Keywords: backdoor detection; object detection;
Compressor summary: The paper proposes Django, a backdoor detection framework for object detection models that uses dynamic Gaussian weighting to prioritize vulnerable boxes and improve trigger inversion efficiency.
Kaiwen Zheng, Cheng Lu, Jianfei Chen, Jun Zhu
https://openreview.net/forum?id=9fWKExmKa0
Keywords: diffusion models, fast sampling, ODE solver
Compressor summary: The paper proposes a new fast ODE solver for DPMs that improves sampling efficiency and sample quality, especially with fewer function evaluations or large guidance scales.
Wei Chen, Zichen Miao, Qiang Qiu
https://openreview.net/forum?id=9eneYFIGKq
Keywords: Neural Network Similarity, Filter Subspace
Compressor summary: The paper proposes a new method to quickly compare neural network models by simplifying their representation to filter subspace distance, which is efficient and robust.
Dongjie Wang, Meng Xiao, Min Wu, pengfei wang, Yuanchun Zhou, Yanjie Fu
https://openreview.net/forum?id=9dp35y5C0p
Keywords: Feature Transformation, Autoregressive Generation, Continuous Space Optimization
Compressor summary: The authors propose a new method to optimize feature transformation as a continuous space optimization task, and demonstrate its effectiveness and robustness through experiments and case studies.
Mingzhen He, FAN He, Ruikai Yang, Xiaolin Huang
https://openreview.net/forum?id=9cQzO3rXgR
Keywords: Asymmetric kernels, diffusion maps, magnetic transform, dimension reduction
Compressor summary: The paper introduces MagDM, a new diffusion map technique that can handle asymmetric data using the magnetic transform, and shows its effectiveness on synthetic and real-world datasets.
Kwangjun Ahn, Sebastien Bubeck, Sinho Chewi, Yin Tat Lee, Felipe Suarez, Yi Zhang
https://openreview.net/forum?id=9cQ6kToLnJ
Keywords: Gradient descent, edge of stability, generalization
Compressor summary: The paper analyzes how large learning rates affect two-layer neural networks and shows that using a high learning rate can improve generalization by enabling the training of threshold-like neurons.
Valerii Iakovlev, Markus Heinonen, Harri Lähdesmäki
https://openreview.net/forum?id=9cF6RUwMe7
Keywords: neural, PDEs, neural PDEs, partial observations, space time continuous
Compressor summary: The paper presents a new neural network model that learns partial differential equations from noisy and partial observations without relying on regular grids, achieving high performance and data efficiency.
Liyuan Wang, Jingyi Xie, Xingxing Zhang, Mingyi Huang, Hang Su, Jun Zhu
https://openreview.net/forum?id=9XieH21Tlf
Keywords: Continual Learning, Catastrophic Forgetting, Pre-training, Prompt Tuning
Compressor summary: The paper proposes Hierarchical Decomposition (HiDe-)Prompt, an approach that improves the performance of continual learning with self-supervised pre-training by optimizing hierarchical components with task-specific prompts and representation statistics.
Noah Hollmann, Samuel Müller, Frank Hutter
https://openreview.net/forum?id=9WSxQZ9mG7
Keywords: AutoML, AutoDS, Automated Feature Engineering, LLM Code Generation, Tabular Data, Feature Engineering, Automated Data Science, Automated Machine Learning
Compressor summary: The authors present CAAFE, a feature engineering method that uses large language models to generate semantically meaningful features for tabular datasets based on their descriptions, improving performance and interpretability in automated machine learning systems.
Andrew Luo, Margaret Marie Henderson, Leila Wehbe, Michael J. Tarr
https://openreview.net/forum?id=9VqMaSjf7U
Keywords: neuroscience, brain, fmri, generative models, diffusion models, image synthesis, visual cortex
Compressor summary: The authors present BrainDiVE, a data-driven approach to explore the fine-grained functional organization of the brain's higher visual cortex by synthesizing images that activate specific brain regions, revealing novel subdivisions and functional differences within these regions.
Jiajun Tang, Haofeng Zhong, Shuchen Weng, Boxin Shi
https://openreview.net/forum?id=9UxUTGCteW
Keywords: Illumination, Image Generation, Conditional Image Repainting
Compressor summary: The paper introduces a new task (LuminAIRe) and a dataset (Car-LuminAIRe) to improve conditional image repainting methods by using estimated 3D geometry and physically-based illumination rendering.
Mineui Hong, Minjae Kang, Songhwai Oh
https://openreview.net/forum?id=9Tx2znbyTm
Keywords: Multi-task decision-making, Offline reinforcement learning, Planning, Diffusion model
Compressor summary: The paper proposes a diffusion-based generative sequence model to plan milestones in a latent space for long-term planning, vision-based control, and multi-task decision-making, achieving superior results on various tasks.
Taiji Suzuki, Denny Wu, Atsushi Nitanda
https://openreview.net/forum?id=9STYRIVx6u
Keywords: mean-field regime, interacting particle system, propagation of chaos, Neural network optimization, MMD minimization, kernel stein discrepancy
Compressor summary: The paper proposes a framework to analyze the MFLD method for training two-layer neural networks and shows its convergence rate and applicability to various learning problems and gradient estimators.
Gecia Bravo-Hermsdorff, David Watson, Jialin Yu, Jakob Zeitler, Ricardo Silva
https://openreview.net/forum?id=9S8oVumknA
Keywords: Causality, experimental design
Compressor summary: The paper proposes a method to generalize causal inference from past experiments to novel conditions using factor graph models that abstract away unmeasured confounding and feedback mechanisms.
Juliette Bertrand, Giorgos Kordopatis-Zilos, Yannis Kalantidis, Giorgos Tolias
https://openreview.net/forum?id=9QsdPQlWiE
Keywords: VOS, video object segmentation, test-time training, test-time adaptation
Compressor summary: The paper proposes test-time training strategies for video object segmentation tasks under distribution shifts, such as video corruptions and style transfers, and introduces a new test set with extreme distribution shifts.
Yihe Deng, Yu Yang, Baharan Mirzasoleiman, Quanquan Gu
https://openreview.net/forum?id=9QEVJ9qm46
Keywords: spurious correlation, robustness, robust learning
Compressor summary: The paper analyzes how two-layer convolutional neural networks can be influenced by non-generalizable features and proposes a new algorithm called PDE that improves robustness and performance on various tasks.
Aniket Murhekar, Zhuowen Yuan, Bhaskar Ray Chaudhury, Bo Li, Ruta Mehta
https://openreview.net/forum?id=9OqezkNxnX
Keywords: Federated learning, Nash equilibrium, Mechanism design, Welfare maximization
Compressor summary: The paper proposes a budget-balanced mechanism to improve the welfare of agents participating in collaborative federated learning by ensuring optimal trade-offs between learning payoff and data sharing costs.
Yan-Shuo Liang, Wu-Jun Li
https://openreview.net/forum?id=9Oi3YxIBSa
Keywords: Continual Learning, stability, plasticity
Compressor summary: LODE is a method for task-agnostic continual learning that separates the objectives of distinguishing new classes from old and between different new classes, achieving better stability and plasticity than existing replay-based methods.
Haoyi Duan, Yan Xia, Mingze Zhou, Li Tang, Jieming Zhu, Zhou Zhao
https://openreview.net/forum?id=9MwidIH4ea
Keywords: audio-visual, multi-modal prompt, clip, cross-modal attention
Compressor summary: The paper proposes a novel DG-SCT attention mechanism that uses audio and visual modalities as soft prompts to dynamically adjust the parameters of pre-trained models for multi-modal tasks, achieving state-of-the-art results and few-shot/zero-shot performance.
Xiang Gu, Liwei Yang, Jian Sun, Zongben Xu
https://openreview.net/forum?id=9Muli2zoFn
Keywords: optimal transport, diffusion probabilistic model, conditional score-based model, unpaired super-resolution, image-to-image translation
Compressor summary: OTCS is a novel model for conditional generation of target data using unpaired or partially paired data based on optimal transport, and it achieves effective data transport and image translation.
Kaiwen Wu, Kyurae Kim, Roman Garnett, Jacob R. Gardner
https://openreview.net/forum?id=9KtX12YmA7
Keywords: Bayesian optimization, convergence rates
Compressor summary: The text discusses the benefits of using local optimization strategies in Bayesian optimization for high-dimensional problems, and provides a rigorous analysis of a recent algorithm's performance and convergence rates.
Emmanuel Abbe, Elisabetta Cornacchia, Aryo Lotfi
https://openreview.net/forum?id=9Ihu0VBOTq
Keywords: curriculum learning, parities, time complexity, sample complexity, neural networks, SGD
Compressor summary: The text explains that curriculum learning, starting with simpler tasks before more complex ones, can help neural networks learn parities faster than traditional methods, especially when the data is a mix of sparse and dense inputs.
Man Yao, JiaKui Hu, Zhaokun Zhou, Li Yuan, Yonghong Tian, Bo XU, Guoqi Li
https://openreview.net/forum?id=9FmolyOHi5
Keywords: Spiking Neural Networks; Transformer; Neuromorphic Computing; Event-driven; Linear Attention
Compressor summary: The paper proposes a spike-driven Transformer that uses event-driven and binary spike communication to achieve energy efficiency and state-of-the-art performance in the spiking neural network domain.
Shikai Qiu, Tim G. J. Rudner, Sanyam Kapoor, Andrew Gordon Wilson
https://openreview.net/forum?id=9EndFTDiqh
Keywords: Function-Space Modeling, Maximum A Posteriori Estimation, Generalization
Compressor summary: The paper explores how estimating the most likely function implied by a model and data can improve predictive performance, but warns of potential pathological solutions when using neural networks.
Zhen Liu, Peitian Ma, Dongliang Chen, Wenbin Pei, Qianli Ma
https://openreview.net/forum?id=9D0fELXbrg
Keywords: time series classification, deep neural networks, noisy labels
Compressor summary: The paper proposes a deep learning approach called Scale-teaching to handle noisy labels in time series data by using multiple DNNs, cross-scale fusion, and label propagation.
An Thai Le, Georgia Chalvatzaki, Armin Biess, Jan Peters
https://openreview.net/forum?id=9B9J8X23LK
Keywords: Motion Planning, Trajectory Optimization, Optimal Transport
Compressor summary: MPOT is a gradient-free method that optimizes smooth trajectories over nonlinear costs using the Sinkhorn Step and regular polytopes, outperforming other motion planners in various problems.
Shaokai Ye, Jessy Lauer, Mu Zhou, Alexander Mathis, Mackenzie W Mathis
https://openreview.net/forum?id=9AcG3Tsyoq
Keywords: ChatGPT, GPT3.5, GPT4, behavioral analysis, LLMs, human-AI interaction, behavioral neuroscience
Compressor summary: The paragraph introduces AmadeusGPT, a natural language interface that converts descriptions of animal behaviors into machine code using large-language models and a dual-memory mechanism, improving the efficiency and accuracy of behavior analysis.
Ethan Pronovost, Meghana Reddy Ganesina, Noureldin Hendy, Zeyu Wang, Andres Morales, Kai Wang, Nicholas Roy
https://openreview.net/forum?id=99MHSB98yZ
Keywords: Deep Learning, (Other) Applications, (Other) Machine Learning Topics
Compressor summary: Scenario Diffusion generates diverse and controllable synthetic traffic scenarios for autonomous vehicles using latent diffusion, object detection, and trajectory regression.
Pierre Marion
https://openreview.net/forum?id=992vogTP1L
Keywords: residual neural networks, neural ODEs, generalization bound
Compressor summary: The paper develops a generalization bound for continuous-in-time parameterized ODEs, which includes time-dependent neural ODEs and deep residual networks, by using a Lipschitz-based argument that relates the magnitude of weight matrix differences to generalization.
Dongyoung Kim, Jinwoo Shin, Pieter Abbeel, Younggyo Seo
https://openreview.net/forum?id=97E3YXvcFM
Keywords: Reinforcement Learning, State Entropy, Exploration
Compressor summary: The paper proposes a new exploration technique for reinforcement learning that considers the value of states when maximizing state entropy to prevent bias and improve performance in various tasks.
Jing Xu, Jiaye Teng, Yang Yuan, Andrew C Yao
https://openreview.net/forum?id=966yOmwk6d
Keywords: data-algorithm dependent generalization analysis, overparameterized linear regression
Compressor summary: The paper proposes a new concept called data-algorithm compatibility to analyze generalization in overparameterized models by considering the entire training trajectory instead of just the final model.
Xueting Li, Shalini De Mello, Sifei Liu, Koki Nagano, Umar Iqbal, Jan Kautz
https://openreview.net/forum?id=95q46MpBGZ
Keywords: Neural Radiance Field, Portrait Reconstruction and Animation
Compressor summary: The authors present a method that can reconstruct and animate realistic 3D head avatars from single-view images, generalizing to different identities and capturing details beyond the face. The method uses three branches with tri-planes and volumetric rendering for high fidelity results.
Junru Zhou, Jiarui Feng, Xiyuan Wang, Muhan Zhang
https://openreview.net/forum?id=94rKFkcm56
Keywords: Cycle counting, graph neural networks
Compressor summary: The paper introduces $d$-Distance-Restricted FWL(2) GNNs, a novel class of graph neural networks that can efficiently count certain graph substructures like cycles by exploiting graph sparsity and avoiding expensive preprocessing steps.
Yuyuan Li, Chaochao Chen, Yizhao Zhang, Weiming Liu, Lingjuan Lyu, Xiaolin Zheng, Dan Meng, Jun Wang
https://openreview.net/forum?id=93NLxUojvc
Keywords: recommendation unlearning, machine unlearning, recommender systems, ensemble learning
Compressor summary: UltraRE is a new framework that enhances RecEraser's recommendation unlearning by addressing redundancy, relevance, and combination losses.
Görkay Aydemir, Weidi Xie, Fatma Guney
https://openreview.net/forum?id=919tWtJPXe
Keywords: Unsupervised Object Discovery, Unsupervised Video Object Segmentation, Object-Centric Learning, Unsupervised Video Multi Object Segmentation
Compressor summary: The paper presents an unsupervised method for segmenting multiple objects in real-world video sequences using object-centric learning and a masking strategy.
Zhun Deng, Thomas P Zollo, Jake Snell, Toniann Pitassi, Richard Zemel
https://openreview.net/forum?id=917crxqJdA
Keywords: societal dispersion, distribution-free uncertainty quantification
Compressor summary: The paragraph discusses the importance of controlling the dispersion of loss distribution in machine learning models for high-stakes applications and proposes a new framework to handle various statistical functionals.
Hao Zheng, Hui Lin, Rong Zhao
https://openreview.net/forum?id=90O5cvFZkZ
Keywords: neuronal coherence, combinatorial generalization, perceptual grouping, unsupervised learning
Compressor summary: GUST is an iterative neural network architecture that mimics neuronal coherence in the human brain to learn how to group sensory information effectively and create symbolic representations of scenes.
Ruoxi Jiang, Peter Y. Lu, Elena Orlova, Rebecca Willett
https://openreview.net/forum?id=8xx0pyMOW1
Keywords: Neural operators, contrastive learning, optimal transport, chaotic attractors, invariant measures
Compressor summary: The paper proposes new methods for training neural networks to forecast chaotic systems over long time horizons while preserving their statistical properties using optimal transport or contrastive learning.
rongkun Zheng, Lu Qi, Xi Chen, Yi Wang, Kun Wang, Yu Qiao, Hengshuang Zhao
https://openreview.net/forum?id=8xTOtxinMH
Keywords: taxonomy-aware, multiple-datasets, video instance segementation
Compressor summary: The paper proposes TMT-VIS, a model that uses taxonomy information to improve video instance segmentation performance on multiple datasets.
Ayça Takmaz, Elisabetta Fedele, Robert Sumner, Marc Pollefeys, Federico Tombari, Francis Engelmann
https://openreview.net/forum?id=8vuDHCxrmy
Keywords: open-world, open-vocabulary, 3D vision, point cloud, instance segmentation, 3D instance segmentation
Compressor summary: OpenMask3D is a method for open-vocabulary 3D instance segmentation that uses multi-view fusion of CLIP-based image embeddings to separate multiple objects in scenes.
Eduard Tulchinskii, Kristian Kuznetsov, Kushnareva Laida, Daniil Cherniavskii, Sergey Nikolenko, Evgeny Burnaev, Serguei Barannikov, Irina Piontkovskaya
https://openreview.net/forum?id=8uOZ0kNji6
Keywords: generated texts detection, intrinsic dimension, TDA, Persistent Homology, ChatGPT
Compressor summary: The authors propose a new method to distinguish between natural and AI-generated texts based on the intrinsic dimensionality of text embeddings, which is lower for AI-generated texts and stable across different languages and domains.
Anders Aamand, Justin Y. Chen, Allen Liu, Sandeep Silwal, Pattara Sukprasert, Ali Vakilian, Fred Zhang
https://openreview.net/forum?id=8rDbUoYc0p
Keywords: clustering, fairness, approximation algorithms
Compressor summary: The paper introduces IP stability as a new clustering objective, shows that an O(1)-IP stable clustering always exists for general metrics, and provides efficient algorithms for various generalizations of IP stability.
Zhiwei Hao, Jianyuan Guo, Kai Han, Yehui Tang, Han Hu, Yunhe Wang, Chang Xu
https://openreview.net/forum?id=8qePPvL1VY
Keywords: knowledge distillation, feature distillation, heterogeneous architectures
Compressor summary: The paper proposes OFA-KD, a framework for distilling knowledge between heterogeneous models that improves performance by aligning features and adaptively enhancing the target.
Qing Su, Anton Netchaev, Hai Li, Shihao Ji
https://openreview.net/forum?id=8pOBo5NgTQ
Keywords: Transformer, ViT, Dense Prediction, Self-supervised Learning, Mean Shift, Self-attention, Representation learning
Compressor summary: The paper proposes a new SSL method, FLSL, that uses transformer for joint embedding and clustering, which improves dense prediction tasks like object detection and segmentation.
Rylan Schaeffer, Mikail Khona, Tzuhsuan Ma, Cristobal Eyzaguirre, Sanmi Koyejo, Ila R Fiete
https://openreview.net/forum?id=8ox2vrQiTF
Keywords: self-supervised learning, unsupervised learning, grid cells, neuroscience, systems neuroscience, representation learning
Compressor summary: The authors propose a new self-supervised learning framework to synthesize multi-periodic grid cells in deep recurrent neural networks, drawing from dynamical systems, coding theory, function optimization and supervised deep learning approaches.
Mingxuan Zhang, Yan Sun, Faming Liang
https://openreview.net/forum?id=8niGwlkLAX
Keywords: Sparse Deep Learning, Uncertainty Quantification, Model Compression, Variable Selection, Dependent Data
Compressor summary: The paper studies sparse deep learning with dependent data, showing that sparse recurrent neural networks can be consistently estimated and outperform existing methods in uncertainty quantification and model compression for time series data.
Rui Min, Zeyu Qin, Li Shen, Minhao Cheng
https://openreview.net/forum?id=8muKbaAgsh
Keywords: Backdoor Defense, Model-tuning
Compressor summary: The paper proposes Feature Shift Tuning (FST), a method to disentangle backdoor and clean features for tuning-based backdoor purification, which performs well against diverse attack scenarios with low tuning costs.
Georg Bökman, Fredrik Kahl
https://openreview.net/forum?id=8lbFwpebeu
Keywords: loss landscape, network merging, linear mode connectivity, equivariance, group convolutional neural network, permutation, group, symmetry, invariance, weight space ensembling
Compressor summary: The authors study how neural networks' equivariance properties affect their layers and propose a conjecture related to the recent permutation conjecture, while providing experiments to support their findings.
Jarosław Błasiok, Parikshit Gopalan, Lunjia Hu, Preetum Nakkiran
https://openreview.net/forum?id=8kyIChWsAG
Keywords: calibration, deep learning, theory, optimization
Compressor summary: This work studies under what conditions optimizing a proper loss function over a restricted family of predictors yields calibrated models, and provides a rigorous answer by introducing the concept of local optimality.
Ruo-Chun Tzeng, Po-An Wang, Alexandre Proutiere, Chi-Jen Lu
https://openreview.net/forum?id=8jg8z3ASiw
Keywords: best-arm identification; combinatorial semi-bandit; no-regret learning;
Compressor summary: Perturbed Frank-Wolfe Sampling (P-FWS) is an algorithm that efficiently identifies the best arm in combinatorial semi-bandits by solving an optimization problem with a single iteration of the Frank-Wolfe algorithm and leveraging structural properties.
Bo Jiang, Ya-Feng Liu
https://openreview.net/forum?id=8hKCNVqrlf
Keywords: Barzilai-Borwein method, exponential augmented Lagrangian, inexact gradient, Stiefel manifold, Sinkhorn iteration, Wasserstein distance
Compressor summary: The paper proposes a new method (REALM) to compute the projection robust Wasserstein distance more efficiently and accurately than existing methods, by using an inexact Riemannian Barzilai-Borwein method with Sinkhorn iteration.
Lingjiong Zhu, Mert Gurbuzbalaban, Anant Raj, Umut Simsekli
https://openreview.net/forum?id=8fLatmFQgF
Keywords: Algorithmic stability, SGD, Wasserstein distance
Compressor summary: The paper introduces a unified proof technique for deriving Wasserstein stability bounds for stochastic optimization algorithms like SGD, showing that ergodicity is crucial for time-uniform bounds and extending previous results to more general cases.
Shizhe Ding, Boyang Xia, Dongbo Bu
https://openreview.net/forum?id=8d9wVXri89
Keywords: Interpolation algorithm, scattered data, deep learning, residual learning
Compressor summary: HINT is a novel neural network-based interpolation method that uses residuals on observed points to guide target function estimation and hierarchical local constraints in correlation modeling to improve interpolation accuracy.
Roberto Cipollone, Anders Jonsson, Alessandro Ronca, Mohammad Sadegh Talebi
https://openreview.net/forum?id=8bQc7oRnjm
Keywords: Reinforcement Learning, Offline Reinforcement Learning, Regular Decision Processes, Sample complexity, Automata
Compressor summary: The paper introduces RegORL, an algorithm that combines automata learning and offline RL to learn near-optimal policies in unknown episodic RDPs using pre-collected data without further exploration.
Haochuan Li, Jian Qian, Yi Tian, Alexander Rakhlin, Ali Jadbabaie
https://openreview.net/forum?id=8aunGrXdkl
Keywords: Optimization, Convergence, Generalized smoothness
Compressor summary: The paper proposes a generalized non-uniform smoothness condition for optimization problems, which enables convergence rates for various methods in both convex and non-convex settings without gradient clipping or heavy-tailed noise restrictions.
Zixing Song, Yifei Zhang, Irwin King
https://openreview.net/forum?id=8aDG51pxFc
Keywords: Graph Neural Networks, Expected Model Change Maximization
Compressor summary: The paper proposes an active learning method for graph neural networks that uses the Expected Model Change Maximization principle to improve prediction performance on unlabeled data.
Yongtao Wu, Fanghui Liu, Grigorios Chrysos, Volkan Cevher
https://openreview.net/forum?id=8ZveVHfmIE
Keywords: Transformer, convergence, scaling, initialization, over-parameterization
Compressor summary: The paper develops a global convergence theory for shallow Transformers with encoder-only architecture under realistic settings and analyzes the impact of architectures, initialization, scaling, and softmax on their performance.
Kansei Ushiyama, Shun Sato, Takayasu Matsuo
https://openreview.net/forum?id=8YN62t19AW
Keywords: Convex optimization, Numerical analysis, Ordinary differential equations, Convergence estimate
Compressor summary: This paper introduces a concept called weak discrete gradient (wDG) that allows transitioning from continuous to discrete optimization methods in the differential equation approach for convex optimization, simplifying analysis and enabling faster convergence rates.
Gaurav Bhatt, Deepayan Das, Leonid Sigal, Vineeth N. Balasubramanian
https://openreview.net/forum?id=8Xn3D9OtqI
Keywords: part-based learning, interpretability, few-shot learning, vision transformers
Compressor summary: The study proposes two regularization methods to improve interpretability and generalization of part-based representations in intelligent systems by reducing incidental background correlations.
David Woodruff, Peilin Zhong, Samson Zhou
https://openreview.net/forum?id=8XRMbNAP6Z
Keywords: clustering, streaming algorithms, sliding window model
Compressor summary: The paper presents the first algorithm that achieves near-optimal clustering in the sliding window model and develops an online coreset data structure for this purpose.
Yutao Cui, Tianhui Song, Gangshan Wu, Limin Wang
https://openreview.net/forum?id=8WvYAycmDJ
Keywords: Efficient Tracking, Fully Transformer, Distillation, Model Pruning
Compressor summary: The paper proposes MixFormerV2, a transformer-based tracker that uses four special prediction tokens and a new distillation-based model reduction paradigm to achieve high accuracy and efficiency on both GPU and CPU platforms.
Fraser Mince, Dzung Dinh, Jonas Kgomo, Neil Thompson, Sara Hooker
https://openreview.net/forum?id=8VTbfVfAfI
Keywords: hardware, software, meta study, portability
Compressor summary: The lack of portability in popular ML frameworks across different hardware types hinders exploratory research and innovation in machine learning due to significant function loss and performance slowdown.
Konwoo Kim, Gokul Swamy, Zuxin Liu, Ding Zhao, Sanjiban Choudhury, Steven Wu
https://openreview.net/forum?id=8U31BCquNF
Keywords: constraints, inverse reinforcement learning, safe reinforcement learning
Compressor summary: The paper proposes a method to learn safety constraints from expert demonstrations for robotic tasks, using inverse reinforcement learning and diverse demonstrations to avoid overly conservative constraints.
Yuchen BAI, Jean-Baptiste Durand, Grégoire Laurent Vincent, Florence Forbes
https://openreview.net/forum?id=8SUtvEZCF2
Keywords: UAV, Deep Learning, Semantic Segmentation, Lidar, Class Imbalance, Point Cloud
Compressor summary: Lidar technology helps monitor forests and track climate change, but identifying leaf points from wood points in UAV data is challenging; a new neural network model based on Pointnet ++ architecture uses point geometry only to overcome these issues.
Wenlong Zhang, Xiaohui Li, Guangyuan SHI, Xiangyu Chen, Yu Qiao, Xiaoyun Zhang, Xiao-Ming Wu, Chao Dong
https://openreview.net/forum?id=8SCz56sUGP
Keywords: Image super-resolution
Compressor summary: The paper proposes a multi-task learning approach to improve real-world image super-resolution by grouping similar degradation tasks together and training them separately, which reduces task competition and enhances performance.
Shahriar Talebi, Amirhossein Taghvaei, Mehran Mesbahi
https://openreview.net/forum?id=8S9Fbee743
Keywords: Optimal filtering, data-driven control, stochastic optimization, learning
Compressor summary: The paper studies how to learn the best filtering policy for a linear system with unknown noise matrices using noisy data, and proposes a method that minimizes prediction error, has good convergence properties, and scales well with problem size.
Nina Balcan, Anh Tuan Nguyen, Dravyansh Sharma
https://openreview.net/forum?id=8QGukmdAbh
Keywords: Elastic Net, logistic regression, data-driven algorithm design, learning theory, regularization
Compressor summary: This paper studies how many samples are needed to tune regularization parameters in linear and logistic regressions with different constraints, and provides new upper and lower bounds for their validation loss functions.
Michael Arbel, Romain Menegaux, Pierre Wolinski
https://openreview.net/forum?id=8Oukmqfek2
Keywords: implicit bias, gauss newton
Compressor summary: The paper analyzes how Gauss Newton's method optimizes over-parameterized networks, finding that it converges faster than gradient descent but has an implicit bias that affects generalization.
Ying Fan, Olivia Watkins, Yuqing Du, Hao Liu, Moonkyung Ryu, Craig Boutilier, Pieter Abbeel, Mohammad Ghavamzadeh, Kangwook Lee, Kimin Lee
https://openreview.net/forum?id=8OTPepXzeh
Keywords: Diffusion models, RLHF
Compressor summary: The authors propose using online reinforcement learning to fine-tune text-to-image diffusion models, achieving better alignment and image quality than supervised fine-tuning methods.
Qi-Wei Wang, Da-Wei Zhou, Yi-Kai Zhang, De-Chuan Zhan, Han-Jia Ye
https://openreview.net/forum?id=8NAxGDdf7H
Keywords: Few-Shot Class-Incremental Learning, Continual Learning, Class-Incremental Learning
Compressor summary: The paper introduces a new strategy (TEEN) to improve few-shot class-incremental learning by enhancing the discriminability of new classes using weighted prototypes, and shows its effectiveness on standard benchmarks and few-shot learning tasks.
Yiren Jian, Chongyang Gao, Soroush Vosoughi
https://openreview.net/forum?id=8Kch0ILfQH
Keywords: vision-language pretraining, multi-modal learning, uni-modal auxiliary learning
Compressor summary: The Prompt-Transformer (P-Former) is a model that predicts optimal prompts for language models to align with visual features, improving performance and reducing the need for image-text pairs in vision-language pre-training.
Mohamed El Amine Boudjoghra, Salwa K. Al Khatib, Jean Lahoud, Hisham Cholakkal, Rao Muhammad Anwer, Salman Khan, Fahad Khan
https://openreview.net/forum?id=8JsbdJjRvY
Keywords: open-world, 3d instance segmentation
Compressor summary: The paper presents a new method for 3D indoor instance segmentation that can handle unknown classes and improve its performance using pseudo-labels and realistic open-world scenarios.
Zhengdong Hu, Yifan Sun, Jingdong Wang, Yi Yang
https://openreview.net/forum?id=8JMexYVcXB
Keywords: deep learning, computer vision, object detection, transformer
Compressor summary: The paper presents a method called Divide-And-Conquer DETR (DAC-DETR) that improves object detection by separating cross-attention and self-attention layers in the decoder, leading to better performance on popular benchmarks.
Taylor Whittington Webb, Shanka Subhra Mondal, Jonathan Cohen
https://openreview.net/forum?id=8JCZe7QrPy
Keywords: relational reasoning, object-centric representations, abstract rule learning, relational inductive biases, systematic generalization
Compressor summary: The Object-Centric Relational Abraction (OCRA) model combines object-centric and relational abstraction approaches to achieve strong systematic generalization in complex visual tasks.
Yaoyu Zhu, Wei Fang, Xiaodong Xie, Tiejun Huang, Zhaofei Yu
https://openreview.net/forum?id=8IvW2k5VeA
Keywords: Spiking neural networks, Spike encoding, Time-based training
Compressor summary: The paper proposes an enhanced counting loss for training spiking neural networks (SNNs) with time-based schemes, which better utilizes temporal information and improves performance on most datasets.
Zhenxing Ge, Zheng Xu, Tianyu Ding, Wenbin Li, Yang Gao
https://openreview.net/forum?id=8HzOyg1ngp
Keywords: Subgame solving, extensive-form game, imperfect information
Compressor summary: The GS2 framework uses a generation function to identify important subgame nodes, reducing their size and improving performance in large imperfect information games.
Jialong Wu, Haoyu Ma, Chaoyi Deng, Mingsheng Long
https://openreview.net/forum?id=8GuEVzAUQS
Keywords: Model-based reinforcement learning, world model, pre-training
Compressor summary: The paper proposes Contextualized World Models to pre-train world models with in-the-wild videos for efficient learning of visual control tasks in reinforcement learning.
Liting Chen, Jie Yan, Zhengdao Shao, Lu Wang, Qingwei Lin, Saravan Rajmohan, Thomas Moscibroda, Dongmei Zhang
https://openreview.net/forum?id=8GSCaoFot9
Keywords: Offline Reinforcement Learning
Compressor summary: The paper proposes a new approach called Conservative State Value Estimation (CSVE) for offline reinforcement learning, which improves state value estimation by penalizing out-of-distribution states and performs well in classic continual control tasks of D4RL.
Jack Jewson, Sahra Ghalebikesabi, Christopher C. Holmes
https://openreview.net/forum?id=8FbuHeVU7D
Keywords: differential privacy, beta-divergence, posterior sampling, generalised Bayesian inference
Compressor summary: $\beta$D-Bayes is a new method that allows more accurate and private estimation for various types of statistical models without altering the original data or model.
Aaron J Havens, Alexandre Araujo, Siddharth Garg, Farshad Khorrami, Bin Hu
https://openreview.net/forum?id=8F3Lutda7R
Keywords: Deep equilibrium models, Lipschitz networks, certifiable robustness
Compressor summary: The paper explores how different deep equilibrium models can be reparameterized as Lipschitz-bounded equilibrium networks, which could improve their certified robustness.
Yilan Chen, Wei Huang, Hao Wang, Charlotte Loh, Akash Srivastava, Lam M. Nguyen, Tsui-Wei Weng
https://openreview.net/forum?id=8Ba7VJ7xiM
Keywords: generalization, deep learning theory, neural tangent kernel, neural architecture search
Compressor summary: This paper introduces a new kernel called loss path kernel, which measures data similarity using loss gradient agreement along gradient flow paths, and shows its application in deriving tight generalization bounds for neural networks and guiding neural architecture search.
Wentong Li, Yuqian Yuan, Song Wang, Wenyu Liu, Dongqi Tang, Jian liu, Jianke Zhu, Lei Zhang
https://openreview.net/forum?id=8BPzLxF9p5
Keywords: Computer Vision, Segmentation, Weakly-supervised Learning
Compressor summary: The paper proposes an affinity propagation method that uses local and global pairwise affinity terms to generate accurate soft pseudo labels for weakly-supervised segmentation tasks, reducing the need for pixel-wise labeling and improving performance on three types of tasks.
Arthur Conmy, Augustine N. Mavor-Parker, Aengus Lynch, Stefan Heimersheim, Adrià Garriga-Alonso
https://openreview.net/forum?id=89ia77nZ8u
Keywords: Mechanistic Interpretability, Pruning, Science of Deep Learning, AI Safety
Compressor summary: The paper presents a systematic method to interpret transformer models and automates one step of the process using novel algorithms, demonstrating their effectiveness on GPT-2 Small.
Shengjie Zhu, Abhinav Kumar, Masa Hu, Xiaoming Liu
https://openreview.net/forum?id=898RcRYWCg
Keywords: Monocular Camera Calibration; Camera Pose Estimation; Image Editing
Compressor summary: The paper proposes a method to calibrate intrinsic camera parameters using monocular 3D priors from depthmaps and surface normals, which can improve 3D sensing tasks and applications.
Haotian Zheng, Qizhou Wang, Zhen Fang, Xiaobo Xia, Feng Liu, Tongliang Liu, Bo Han
https://openreview.net/forum?id=87Qnneer8l
Keywords: out-of-distribution detection
Compressor summary: The paper proposes ATOL, a data generation-based learning method for OOD detection that uses an auxiliary task to relieve mistaken OOD generation and improve reliability.
Zuheng Xu, Trevor Campbell
https://openreview.net/forum?id=87Nu9SagB7
Keywords: variational flow, numerical instability, shadowing property
Compressor summary: The paper examines how numerical instability affects variational flows' reliability in sampling, density evaluation, and ELBO estimation, and proposes a diagnostic procedure to validate results from unstable flows.
Zongyi Li, Nikola Borislavov Kovachki, Chris Choy, Boyi Li, Jean Kossaifi, Shourya Prakash Otta, Mohammad Amin Nabian, Maximilian Stadler, Christian Hundt, Kamyar Azizzadenesheli, Anima Anandkumar
https://openreview.net/forum?id=86dXbqT5Ua
Keywords: partial differential equation, computational fluid dynamics, neural operator
Compressor summary: GINO is an efficient method to learn solutions for large-scale PDEs with varying geometries, achieving significant speed-up and accuracy improvements over existing methods on 3D fluid simulation.
Stephen Pasteris, Chris Hicks, Vasilios Mavroudis
https://openreview.net/forum?id=86ADcKOHAw
Keywords: Nearest Neighbours, Contextual Bandits
Compressor summary: The paper presents an efficient algorithm for contextual bandits that adapts the nearest neighbour rule and handles adversarial settings with minimal assumptions, providing regret bounds and application to online classification.
Qihang Yu, Ju He, Xueqing Deng, Xiaohui Shen, Liang-Chieh Chen
https://openreview.net/forum?id=83LJRUzXWj
Keywords: open-vocabulary panoptic segmentation, panoptic segmentation, vision and language
Compressor summary: The authors propose a single-stage framework for open-vocabulary segmentation using a frozen convolutional CLIP backbone that simplifies the existing two-stage pipeline, achieves better accuracy-cost trade-off, and outperforms previous methods on various benchmarks.
Yifan Zhang, Daquan Zhou, Bryan Hooi, Kai Wang, Jiashi Feng
https://openreview.net/forum?id=82HeVCqsfh
Keywords: Dataset Expansion, Guided Imagination
Compressor summary: The paper proposes a Guided Imagination Framework that uses generative models to create new labeled data for dataset expansion, improving model accuracy on various image and medical datasets.
Jerome Bolte, Edouard Pauwels, Samuel Vaiter
https://openreview.net/forum?id=81snFfE3vR
Keywords: automatic differentiation, implicit differentiation, super-linear algorithms, bilevel optimization.
Compressor summary: One-step differentiation is an easy and efficient method for derivative estimation that works well for fast optimization algorithms and has theoretical and practical implications in bilevel optimization.
Bryan Andrews, Joseph Ramsey, Ruben Sanchez Romero, Jazmin Camchong, Erich Kummerfeld
https://openreview.net/forum?id=80g3Yqlo1a
Keywords: Causal Discovery, Directed Acyclic Graphs, DAGs, fMRI, Graphical Models, High Dimension, Densely Connected
Compressor summary: BOSS and GSTs are efficient algorithms for learning directed acyclic graphs (DAGs) from highly connected variables, achieving high accuracy and fast execution time in various conditions, and applied to fMRI data analysis.
Kai Yan, Alex Schwing, Yu-Xiong Wang
https://openreview.net/forum?id=805CW5w2CY
Keywords: offline Imitation learning, learning from observations, positive-unlabeled learning
Compressor summary: TAILO is a novel imitation learning method that uses a discriminator to scale weights for weighted behavior cloning based on expert states, improving robustness and performance, especially when learning from incomplete trajectories.
Ningyi Liao, Siqiang Luo, Xiang Li, Jieming Shi
https://openreview.net/forum?id=7zkFc9TGKz
Keywords: Graph neural networks, Scalability, Heterophilous Graphs, Non-Homophily
Compressor summary: LD2 is a scalable heterophilous GNN that simplifies the learning process by decoupling graph propagation and generating expressive embeddings prior to training, achieving optimal time complexity and memory footprint, and improving speed and efficiency on large-scale graphs.
Edith Cohen, Xin Lyu
https://openreview.net/forum?id=7yjsYrajlt
Keywords: Differential Privacy; Adaptive Composition; Sparse Vector Technique
Compressor summary: TCT is a framework that improves privacy in interactive settings by charging a small overhead for computations that hit their targets, while keeping most others essentially free.
Zhou Lu
https://openreview.net/forum?id=7xlrdSOm3g
Keywords: Multimodal Learning
Compressor summary: The paper presents a theory explaining how multimodal machine learning can generalize better than unimodal learning, especially when there is diversity and connection among different sensory inputs.
Pei Chen, Soumajyoti Sarkar, Leonard Lausen, Balasubramaniam Srinivasan, Sheng Zha, Ruihong Huang, George Karypis
https://openreview.net/forum?id=7vqlzODS28
Keywords: Tabular Language Model, Tabular Representation Learning, Pretraining, Tabular Data, Table, Hypergraph
Compressor summary: HyTrel is a tabular language model that uses hypergraphs to capture structural properties of tabular data and improve performance on downstream tasks.
Xuming An, Li Shen, Han Hu, Yong Luo
https://openreview.net/forum?id=7uPnuoYqac
Keywords: federated learning; manifold regularization; update reaggregation
Compressor summary: FedMRUR is a novel federated learning framework that uses hyperbolic graph manifolds and a new optimizer to reduce model inconsistency and improve convergence speed, achieving state-of-the-art results.
Wenxuan Bao, Francesco Pittaluga, Vijay Kumar b g, Vincent Bindschaedler
https://openreview.net/forum?id=7rm3OcASkg
Keywords: differential privacy, deep learning, data augmentation
Compressor summary: The paper proposes two new data augmentation techniques for differentially private learning that improve performance and are compatible with privacy constraints.
Tung Nguyen, Sudhanshu Agrawal, Aditya Grover
https://openreview.net/forum?id=7qfkImn0dL
Keywords: experimental design, few-shot, black-box optimization, synthetic pretraining, in-context learning, transformer
Compressor summary: The paper introduces Experiment Pretrained Transformers (ExPT), a foundation model for few-shot experimental design that uses synthetic pretraining and in-context learning to generate optimal input designs with limited labeled data.
Honghua Dong, Jiawei Xu, Yu Yang, Rui Zhao, Shiwen Wu, Chun Yuan, Xiu Li, Chris J. Maddison, Lei Han
https://openreview.net/forum?id=7p5YWe8GqG
Keywords: Long-Range Interactions, Hierachical Structure, Multi-Scale, Graph Pooling, Graph Neural Networks(GNNs)
Compressor summary: MeGraph is a model that combines local and hierarchical information in a multi-scale graph hierarchy to better capture long-range interactions and outperforms existing methods on various benchmarks.
Julian Rossbroich, Friedemann Zenke
https://openreview.net/forum?id=7otRtfrRqo
Keywords: Credit assignment, hebbian plasticity, inhibitory microcircuits, bio-plausible learning
Compressor summary: The paragraph discusses a microcircuit model that shows how error signals can be encoded in top-down synapses and influence Hebbian learning, resolving the discrepancy between functional models and experimental observations.
Eric Balkanski, Noemie Perivier, Clifford Stein, Hao-Ting Wei
https://openreview.net/forum?id=7ntySBR3Ey
Keywords: Scheduling, algorithms with predictions, speed scaling, energy minimization
Compressor summary: The paper proposes a learning-augmented algorithmic framework for energy-efficient scheduling that improves performance when predictions are accurate and maintains bounded competitive ratios regardless of prediction error, as shown by empirical results on real and synthetic datasets.
Jingyuan Li, Leo Scholl, Trung Le, Pavithra Rajeswaran, Amy L Orsborn, Eli Shlizerman
https://openreview.net/forum?id=7ntI4kcoqG
Keywords: Neuroscience and Cognitive Science, Neural Activity Forecasting, Graph Neural Network
Compressor summary: The text describes a model (AMAG) that uses deep learning and neural network interactions to forecast neural activity, capturing temporal causality and improving performance on synthetic and real neural data from macaques.
Ranran Shen, Pan Peng
https://openreview.net/forum?id=7nXaoclHed
Keywords: Sublinear-time algorithms, Spectral Clustering, Graph Clustering, Random Walks
Compressor summary: The paper proposes an efficient spectral clustering algorithm for graphs with strong clusterability, allowing sublinear-time preprocessing and queries, while handling some noise and deviation from ideal conditions.
Tianhang Cheng, Wei-Chiu Ma, Kaiyu Guan, Antonio Torralba, Shenlong Wang
https://openreview.net/forum?id=7irm2VJARb
Keywords: 3d reconstruction, inverse rendering, pose estimation, single view reconstruction, nerf, duplicates
Compressor summary: The paragraph introduces Structure from Duplicates (SfD), a novel inverse graphics framework that reconstructs geometry, material, and illumination from a single image containing multiple identical objects, using them as a robust prior for single-image inverse graphics and proposing an in-plane rotation-robust Structure from Motion formulation for joint 6-DoF object pose estimation.
Xingyue Huang, Miguel Romero Orth, Ismail Ilkan Ceylan, Pablo Barcelo
https://openreview.net/forum?id=7hLlZNrkt5
Keywords: graph neural networks, knowledge graphs, expressivity, logical characterization
Compressor summary: The authors analyze and compare different graph neural network models for link prediction on knowledge graphs and provide a logical characterization of their expressive power based on relational Weisfeiler-Leman algorithms.
Ruiyuan Kang, Tingting Mu, Panos Liatsis, Dimitrios Kyritsis
https://openreview.net/forum?id=7h1YaSGaHS
Keywords: Failure detection, Physical evaluation, Network-based optimization, Generative model, Hybrid surrogate model
Compressor summary: The authors propose GEESE, a method to detect and correct failed machine learning estimations in engineering problems using simulations and optimization, outperforming existing approaches.
Lujie Xia, Ziluo Ding, Rui Zhao, Jiyuan Zhang, Lei Ma, Zhaofei Yu, Tiejun Huang, Ruiqin Xiong
https://openreview.net/forum?id=7gbjsgcN5p
Keywords: Optical flow, unsupervised learning, spike camera
Compressor summary: The authors propose a dynamic timing representation for spike streams, an unsupervised learning method for optical flow estimation, and a synthetic validation dataset called SSES for evaluating their approach in autonomous driving scenarios.
Anh Do, Thanh Nguyen-Tang, Raman Arora
https://openreview.net/forum?id=7f6vH3mmhr
Keywords: Multi-agent, Bandits, Cooperative
Compressor summary: The paper proposes H-LINUCB, a novel distributed learning algorithm for linear contextual bandits, which optimally cooperates to minimize group regret when agents have knowledge of task similarity or dissimilarity.
Denis Blessing, Onur Celik, Xiaogang Jia, Moritz Reuss, Maximilian Xiling Li, Rudolf Lioutikov, Gerhard Neumann
https://openreview.net/forum?id=7eW6NzSE4g
Keywords: Imitation Learning, Verstile Skill Learning, Curriculum Learning
Compressor summary: The paper proposes a curriculum-based imitation learning method that uses weighted data and mixture of experts policy with maximum entropy objective to learn versatile policies from multimodal human demonstrations.
Amirkeivan Mohtashami, Martin Jaggi
https://openreview.net/forum?id=7eHn64wOVy
Keywords: large language models, memory, context length
Compressor summary: The paper proposes a new attention mechanism called landmark attention that allows Transformers to handle longer contexts without compromising flexibility or memory requirements.
Matthew Robert Wicker, Vihari Piratla, Adrian Weller
https://openreview.net/forum?id=7cnMLZvTy9
Keywords: Fairness, Individual Fairness, Deep Learning, Certification, Trustworthy ML
Compressor summary: The authors propose a novel method for certifying individual fairness in neural networks that is more computationally efficient and scalable than previous approaches, allowing them to analyze larger models and handle distributional shifts.
Mitchell Ostrow, Adam Joseph Eisen, Leo Kozachkov, Ila R Fiete
https://openreview.net/forum?id=7blSUMwe7R
Keywords: Computational Neuroscience, Neural Data Analysis, Statistical Shape Metrics, Representational Similarity Analysis, Recurrent Neural Networks, Dynamical Systems
Compressor summary: The paper introduces Dynamical Similarity Analysis (DSA), a novel method to compare recurrent neural networks based on their dynamics, which can distinguish different computations and learning rules that are not possible with standard geometric approaches.
L. Elisa Celis, Amit Kumar, Anay Mehrotra, Nisheeth K Vishnoi
https://openreview.net/forum?id=7b4oobeB4w
Keywords: bias, evaluation, maximum entropy, selection
Compressor summary: The paragraph discusses a model that explains how biases arise in evaluation processes, such as hiring or admissions, due to resource-information trade-offs and risk-averseness, and shows how this model can be used to study and reduce these biases.
Pratik Patil, Jin-Hong Du
https://openreview.net/forum?id=7aoVQkNmQ6
Keywords: subsampling, ridge regularization, asymptotic equivalences, proportional asymptotics
Compressor summary: The text shows that subsampling and ridge regularization have similar effects on ensemble ridge estimators, and provides methods to find the equivalent parameters for both.
Roi Livni
https://openreview.net/forum?id=7anW5TWbCJ
Keywords: Learning Theory
Compressor summary: The paper explores how mutual information relates to generalization in stochastic convex optimization and finds that existing information-theoretic bounds do not fully capture the performance of some learning algorithms.
Kiwan Maeng, Chuan Guo, Sanjay Kariyappa, G. Edward Suh
https://openreview.net/forum?id=7ZQiucQu2u
Keywords: privacy, instance encoding, split learning
Compressor summary: The paper introduces a new measure for evaluating the privacy of instance encoding based on Fisher information, which can bound its invertibility both theoretically and empirically.
SUBBA REDDY OOTA, Manish Gupta, Mariya Toneva
https://openreview.net/forum?id=7WeCyYy9TL
Keywords: Linguistic properties, fMRI, probing tasks, cognitive neuroscience, language models, NLP
Compressor summary: The correspondence between human brain processing and language models is affected by specific linguistic properties, particularly syntactic ones, which are crucial for maintaining alignment with fMRI recordings.
Guoxi Huang, Hongtao Fu, Adrian G. Bors
https://openreview.net/forum?id=7WTA298wts
Keywords: self-supervised learning, vision transformer, masked image modeling
Compressor summary: MIRL is a self-supervised learning framework that improves the training of deeper ViTs by teaching them to recover masked image residuals, leading to better accuracy and generalization on downstream tasks.
Jaemoo Choi, Jaewoong Choi, Myungjoo Kang
https://openreview.net/forum?id=7WQt1J13ex
Keywords: Optimal Transport, Generative modeling, Generative adversarial network
Compressor summary: The paper proposes a new generative model based on Unbalanced Optimal Transport, which improves robustness, stability, and convergence compared to traditional Optimal Transport methods, achieving lower FID scores on image generation tasks.
Ted Moskovitz, Samo Hromadka, Ahmed Touati, Diana L Borsa, Maneesh Sahani
https://openreview.net/forum?id=7Uix1eQZ8z
Keywords: reinforcement learning, successor features, successor representation, neuroscience
Compressor summary: The paper introduces a new state representation, the $\lambda$ representation (λR), that helps agents adapt to shifting reward priorities in multitask reinforcement learning and natural behaviors like foraging.
Vaishnavh Nagarajan, Aditya Krishna Menon, Srinadh Bhojanapalli, Hossein Mobahi, Sanjiv Kumar
https://openreview.net/forum?id=7UdVPRmpif
Keywords: knowledge distillation, regularization, understanding, underfitting, theory
Compressor summary: The paper investigates why students trained with knowledge distillation may overconfident but still outperform their teachers, and shows that this is due to an exaggerated bias of gradient descent that improves generalization.
Dami Choi, Derrick Xin, Hamid Dadkhahi, Justin Gilmer, Ankush Garg, Orhan Firat, Chih-Kuan Yeh, Andrew M. Dai, Behrooz Ghorbani
https://openreview.net/forum?id=7RMGI4slcb
Keywords: Multitask Optimization, Multilingual, Pre-training, Language Models, Language Sampling, Low Resource Languages, Overfitting
Compressor summary: This paper investigates how pre-training on high-resource tasks and fine-tuning on a mix of high/low-resource tasks improves multi-task learning with data imbalance, and demonstrates its effectiveness in neural machine translation and multi-lingual language modeling.
Hyunin Lee, Yuhao Ding, Jongmin Lee, Ming Jin, Javad Lavaei, Somayeh Sojoudi
https://openreview.net/forum?id=7R8noSP4vL
Keywords: Non-stationary RL, Reinforcement Learning
Compressor summary: The text proposes a method to synchronize time between an agent and a changing environment in reinforcement learning, improving policy performance.
Filippo Vannella, Alexandre Proutiere, Jaeseong Jeong
https://openreview.net/forum?id=7PJ6LaIOO4
Keywords: Multi-Agent Multi-Armed Bandits, Multi-Armed Bandits, Regret Minimization
Compressor summary: The paper presents a regret minimization algorithm for MAMABs with factor graph rewards, which approximates a lower bound using Mean Field techniques and shows its performance in radio communications networks.
Zhiwei Xu, Bin Zhang, Dapeng Li, Guangchong Zhou, Zeren Zhang, Guoliang Fan
https://openreview.net/forum?id=7LtzqnfuOs
Keywords: Multi-Agent Reinforcement Learning, Individual Global Max
Compressor summary: The paper introduces a new value decomposition method for multi-agent reinforcement learning that rejects the Individual Global Max principle and uses dual self-awareness to solve the credit assignment problem and improve exploration.
Jamie Hayes, Borja Balle, Saeed Mahloujifar
https://openreview.net/forum?id=7LZ4tZrYlx
Keywords: Differential privacy, reconstruction
Compressor summary: The text discusses how privacy levels in DP-SGD affect the risk of training data reconstruction and suggests that less noise may improve utility while emphasizing that the standard privacy guarantee may not reflect the actual risk.
Francesca Bartolucci, Emmanuel de Bezenac, Bogdan Raonic, Roberto Molinaro, Siddhartha Mishra, Rima Alaifari
https://openreview.net/forum?id=7LSEkvEGCM
Keywords: Operator Learning, Neural Operators, PDEs, Frame theory, Sampling theory
Compressor summary: The paper proposes ReNO, a framework to address issues in neural operator learning by measuring inconsistency between neural operators and their discrete representations using the concept of operator aliasing.
Zhuofan Ying, Peter Hase, Mohit Bansal
https://openreview.net/forum?id=7JuReDmGSL
Keywords: Computer vision, out-of-distribution generalization, representational geometry
Compressor summary: The paper investigates how biological vision systems adapt to context for out-of-distribution generalization and proposes new augmentation methods based on representational geometry analysis and causal intervention.
Tongxin Yin, Reilly Raab, Mingyan Liu, Yang Liu
https://openreview.net/forum?id=7INd5Yu9ET
Keywords: Long-term Fairness, Dynamics, Reinforcement Learning
Compressor summary: The paper proposes an online reinforcement learning approach for achieving long-term fairness in policy decisions affecting human populations, with algorithmic solutions that adapt to unknown dynamics and prove probabilistic bounds on loss and fairness violations.
Karolis Martinkus, Jan Ludwiczak, WEI-CHING LIANG, Julien Lafrance-Vanasse, Isidro Hotzel, Arvind Rajpal, Yan Wu, Kyunghyun Cho, Richard Bonneau, Vladimir Gligorijevic, Andreas Loukas
https://openreview.net/forum?id=7GyYpomkEa
Keywords: antibody generation, diffusion, equivariance
Compressor summary: AbDiffuser is a new diffusion model that generates antibody structures and sequences by using novel protein representation, physics-based constraints, and strong diffusion priors, achieving high in silico and in vitro performance.
Carlos Misael Madrid Padilla, Haotian Xu, Daren Wang, OSCAR HERNAN MADRID PADILLA, Yi Yu
https://openreview.net/forum?id=7Fb2lCwS76
Keywords: Multivariate; Nonparametric; Change point inference; short range dependence; Long-run variance; Confidence interval.
Compressor summary: The paper analyzes how to find changes in the underlying distributions of multivariate time series with smooth densities and unknown change points.
Berivan Isik, Wei-Ning Chen, Ayfer Ozgur, Tsachy Weissman, Albert No
https://openreview.net/forum?id=7ETbK9lQd7
Keywords: distributed mean estimation, privacy, compression, communication, federated analytics.
Compressor summary: The paper explores how to estimate a population mean with privacy guarantees and shared randomness, proposing an exact-optimal mechanism based on a rotationally symmetric codebook and $k$-closest encoding.
Ziniu Hu, Ahmet Iscen, Chen Sun, Kai-Wei Chang, Yizhou Sun, David A Ross, Cordelia Schmid, Alireza Fathi
https://openreview.net/forum?id=7EMphtUgCI
Keywords: large language model, visual question answering, dynamic decision making, Tool augmented LLM
Compressor summary: AVIS is a framework that uses an LLM to strategize the use of external tools and analyze their outputs to answer complex visual questions by following a user study-informed sequence of actions.
Zhenyi Wang, Li Shen, Tongliang Liu, Tiehang Duan, Yanjun Zhu, Donglin Zhan, David Doermann, Mingchen Gao
https://openreview.net/forum?id=7DZAVpOoAK
Keywords: Data-Free Model Extraction; Defense
Compressor summary: MeCo is a defense method against data-free model cloning that randomizes inputs to confuse attackers and protect the original model's utility.
Jiali Cui, Tian Han
https://openreview.net/forum?id=7962B4nXX7
Keywords: Energy-based model, MCMC, Joint-training, Generator model
Compressor summary: The paper proposes a joint learning framework for energy-based models that uses a complementary generator model to improve MCMC sampling and avoid bias in learning.
Aleksandar Petrov, Emanuele La Malfa, Philip Torr, Adel Bibi
https://openreview.net/forum?id=78yDLKi95p
Keywords: LLM, language model, tokenizer, multilingual, language, fairness
Compressor summary: The paper argues that current language models are not fair to all languages and suggests using multilingually fair subword tokenizers for better performance.
Haixin Wang, Hao Wu, Jinan Sun, Shikun Zhang, Chong Chen, Xian-Sheng Hua, Xiao Luo
https://openreview.net/forum?id=77i6itptQW
Keywords: domain adaption, binary descriptor, causal inference
Compressor summary: The paper proposes an IDEA model that learns to generate discriminative hash codes using causal and non-causal features, while minimizing the impact of non-causal effects on domain invariance for cross-domain retrieval.
Kha-Dinh Luong, Ambuj Singh
https://openreview.net/forum?id=77Nq1KjmLl
Keywords: Self-supervised Learning, Graph Neural Network, Molecule
Compressor summary: The authors propose GraphFP, a pretraining method for GNNs that leverages fragment-level information and improves performance on molecular property prediction tasks.
Annie Gray, Alexander Modell, Patrick Rubin-Delanchy, Nick Whiteley
https://openreview.net/forum?id=75v88kyyko
Keywords: agglomerative clustering, generative model, graphical model, hierarchical clustering, high-dimensional data
Compressor summary: The paper proposes a simple variant of agglomerative clustering that better recovers hierarchical structure in data using maximum average dot product and shows its advantages over other methods.
Mengzi Amy Guo, Donghao Ying, Javad Lavaei, Zuo-Jun Shen
https://openreview.net/forum?id=75Mxzfoeq7
Keywords: no-regret learning, price competition, reference effect, last-iterate convergence
Compressor summary: The paper proposes an algorithm for two firms to learn a stable equilibrium in a dynamic price competition with opaque information using online projected gradient ascent and stationary Nash equilibrium concepts.
Alex Damian, Eshaan Nichani, Rong Ge, Jason D. Lee
https://openreview.net/forum?id=73XPopmbXH
Keywords: statistical learning, learning theory, single index model, gradient descent, stochastic gradient descent
Compressor summary: The paper studies how many samples are needed to learn a single index model with respect to an isotropic Gaussian distribution using online SGD, and shows that $n \gtrsim d^{k^\star/2}$ samples suffice, closing the gap between upper and lower bounds.
Yang Yue, Rui Lu, Bingyi Kang, Shiji Song, Gao Huang
https://openreview.net/forum?id=71P7ugOGCV
Keywords: Offline RL, Theory
Compressor summary: The authors study the cause of Q-value estimation divergence in offline RL, propose a new metric to measure it, and introduce LayerNorm as a solution to improve extrapolation behavior and achieve better performance.
Ao Zhang, Hao Fei, Yuan Yao, Wei Ji, Li Li, Zhiyuan Liu, Tat-Seng Chua
https://openreview.net/forum?id=716PvHoDct
Keywords: Visual Prompt Generator, Efficient Transfer, Multimodality
Compressor summary: This paper explores transferring visual prompt generators (VPG) between multimodal large language models (MLLMs) to reduce training costs and presents a two-stage transfer framework called VPGTrans that achieves significant efficiency improvements.
Alexander Schlögl, Nora Hofer, Rainer Böhme
https://openreview.net/forum?id=6zyFgr1b8Q
Keywords: machine learning, security, reproducibility, forensics
Compressor summary: The study investigates numerical deviations in CNN inference results across different platforms due to hardware-specific optimizations, such as SIMD use on CPUs and convolution algorithms on GPUs.
Evangelia Gergatsouli, Christos Tzamos
https://openreview.net/forum?id=6wBkT2ndDu
Keywords: pandora's box, stochastic optimization, discrete optimization, learning from samples, algorithms under uncertainty
Compressor summary: The paper studies Pandora's Box problem with correlated value distributions and shows that Weitzman's rule works for both independent and correlated cases, improving approximation guarantees and simplifying the algorithm.
Xutao Wang, Hanting Chen, Tianyu Guo, Yunhe Wang
https://openreview.net/forum?id=6vtZIoxZoJ
Keywords: PU learning, causal inference, semi-supervised learning
Compressor summary: The paper proposes a PU learning enhancement algorithm using causal inference to improve accuracy on non-uniform label distribution datasets.
Ziyi Huang, Henry Lam, Haofeng Zhang
https://openreview.net/forum?id=6vnwhzRinw
Keywords: frequentist uncertainty, epistemic uncertainty, procedural variability, confidence intervals, batching, cheap bootstrap
Compressor summary: The paragraph discusses a new method to quantify and remove uncertainty in deep learning models using a single auxiliary network and light-computation resampling methods.
David Watson, Joshua O'Hara, Niek Tax, Richard Mudd, Ido Guy
https://openreview.net/forum?id=6rabAZhCRS
Keywords: Explainable AI, interpretable ML, feature attributions, information theory, Shapley values
Compressor summary: The authors propose a method to explain predictive uncertainty using Shapley values, which are related to information theory and conditional independence testing, and have applications in various tasks.
Arjun Majumdar, Karmesh Yadav, Sergio Arnaud, Yecheng Jason Ma, Claire Chen, Sneha Silwal, Aryan Jain, Vincent-Pierre Berges, Tingfan Wu, Jay Vakil, Pieter Abbeel, Jitendra Malik, Dhruv Batra, Yixin Lin, Oleksandr Maksymets, Aravind Rajeswaran, Franziska Meier
https://openreview.net/forum?id=6qLzQeFGio
Keywords: representation learning, pre-training, foundation models, embodied AI, reinforcement learning, imitation learning
Compressor summary: The paper evaluates various pre-trained visual representations for Embodied AI tasks using CortexBench, a new benchmark, and shows that adapting a large model to specific tasks improves performance.
Sebastian Shenghong Tay, Chuan-Sheng Foo, Daisuke Urano, Richalynn Leong, Bryan Kian Hsiang Low
https://openreview.net/forum?id=6oiux75UDj
Keywords: Bayesian optimization, Gaussian processes
Compressor summary: The paper proposes a novel algorithm for Bayesian optimization with cost-varying variable subsets, which balances informativeness and costs, and shows its superior performance over baselines.
Philippe Chatigny, Ivan Sergienko, Ryan Ferguson, Jordan Weir, Maxime Bergeron
https://openreview.net/forum?id=6lnoUqFd5R
Keywords: Efficient Frontier, Convex Optimization, Resource Allocation, Constrainted Optimization, Finance
Compressor summary: NeuralEF is a framework that uses neural networks to quickly and accurately predict the optimal portfolio for a given risk level in the efficient frontier problem, considering heterogeneous constraints and variable inputs.
Yuetian Weng, Mingfei Han, Haoyu He, Mingjie Li, Lina Yao, Xiaojun Chang, Bohan Zhuang
https://openreview.net/forum?id=6ljXBlojde
Keywords: Video Semantic Segmentation; Inference Efficiency
Compressor summary: The paper proposes MPVSS, an efficient mask propagation framework for video semantic segmentation that reduces computational costs by reusing predictions from key frames and warping them to non-key frames using learned queries.
Michael Noukhovitch, Samuel Lavoie, Florian Strub, Aaron Courville
https://openreview.net/forum?id=6lgugutkin
Keywords: reinforcement learning from human feedback (rlhf), language
Compressor summary: Elastic Reset is a new algorithm that reduces reward hacking and language drift in fine-tuning language models with reinforcement learning by periodically resetting the online model to an exponentially moving average of itself.
Tianyi Cheng, Dandan Shan, Ayda Sultan Hassen, Richard Ely Locke Higgins, David Fouhey
https://openreview.net/forum?id=6ldTxwhgtP
Keywords: human-object interaction; hand object detection; hand detection
Compressor summary: The new model produces detailed outputs about hand interactions, using large-scale annotations from multiple datasets to improve AI understanding of hands in contact with objects.
Mohammadamin Tavakoli, Pierre Baldi, Ann Marie Carlton, Yinting Chiu, Alexander Shmakov, David Van Vranken
https://openreview.net/forum?id=6kRQTPEVip
Keywords: Chemistry, Reactions, Contrastive, Radical, Graph
Compressor summary: The new reaction predictor system, RMechRP, uses contrastive learning and mechanistic pathways to provide accurate and interpretable predictions of radical reactions, overcoming limitations of existing deep learning-based methods that rely on US patents data.
Zhongzheng Xiong, Xiaoyi Zhu, Zengfeng Huang
https://openreview.net/forum?id=6kINNTYQcm
Keywords: Distributed Tracking, Adaptive Robustness, Differential Privacy, Generalization
Compressor summary: The paper studies how to track a function of items received by multiple sites communicating with a central server, considering adaptive adversaries and extending the differential privacy framework for optimal communication.
Cong Lu, Philip J. Ball, Yee Whye Teh, Jack Parker-Holder
https://openreview.net/forum?id=6jNQ1AY1Uf
Keywords: Reinforcement Learning, Diffusion Models, Synthetic Data, Sample-Efficient RL
Compressor summary: The authors propose Synthetic Experience Replay (SynthER), a method that uses generative models to upsample an agent's collected experience, improving the performance of deep reinforcement learning agents in both offline and online settings with limited data.
Behrooz Tahmasebi, Stefanie Jegelka
https://openreview.net/forum?id=6iouUxI45W
Keywords: invariances, manifolds, sample complexity
Compressor summary: The paper studies how encoding invariances into models reduces sample complexity for kernel ridge regression on compact manifolds with group action-invariant target functions, and presents a minimax optimal rate using differential geometry.
Hao-Kai Zhang, Chenghong Zhu, Mingrui Jing, Xin Wang
https://openreview.net/forum?id=6gcY0MGNhj
Keywords: quantum neural networks, quantum state learning, quantum computing, quantum machine learning, quantum optimization
Compressor summary: The paper develops a no-go theorem for learning an unknown quantum state with quantum neural networks (QNNs) and shows that local minima become more difficult to avoid as the qubit count increases, limiting the learnability and scalability of QNNs.
Zeyu Jia, Gene Li, Alexander Rakhlin, Ayush Sekhari, Nathan Srebro
https://openreview.net/forum?id=6gWpJ0IExE
Keywords: Agnostic Reinforcement Learning, Sample Complexity, Learning Theory, Complexity Measure
Compressor summary: The paper studies the number of interactions needed to learn an approximately optimal policy for unknown MDPs using a new complexity measure called spanning capacity, and introduces a new algorithm called POPLER that enables efficient online RL with certain conditions.
Angela Zhou
https://openreview.net/forum?id=6fuZs3ibGA
Keywords: causal inference, fairness in machine learning, algorithmic fairness, criminal justice, policy learning, off-policy evaluation
Compressor summary: The paragraph discusses how to optimize treatment recommendations for individuals who may not follow them, considering factors like who benefits most and fairness, using a new online learning algorithm.
Yu Gui, Rina Barber, Cong Ma
https://openreview.net/forum?id=6f320HfMeS
Keywords: matrix completion, conformal inference, uncertainty quantification
Compressor summary: The paper proposes a new method, conformalized matrix completion (cmc), which predicts missing entries in data matrices without requiring a low-rank assumption and provides valid prediction intervals using conformal prediction framework.
Xue Yan, Jiaxian Guo, Xingzhou Lou, Jun Wang, Haifeng Zhang, Yali Du
https://openreview.net/forum?id=6ePsuwXUwf
Keywords: Zero-Shot Coordination, Human-AI coordination, Training Efficiency, Partner Modeling
Compressor summary: E3T is an efficient and diverse agent for zero-shot human-AI coordination that uses a mixture of ego policy and random policy, partner modeling, and end-to-end training to improve collaboration efficiency and performance.
Yufan Li, Jialiang Mao, Iavor Bojinov
https://openreview.net/forum?id=6e86TccKyQ
Keywords: bandit algorithms, online learning, causality, Bayesian inference
Compressor summary: The paper proposes an algorithm that automatically determines the release percentage of new products or updates in phased releases, balancing risk and ramp-up speed.
Zexu Sun, Bowei He, Jinxin Liu, Xu Chen, Chen Ma, Shuai Zhang
https://openreview.net/forum?id=6d9Yxttb3w
Keywords: offline imitaion learning, counterfactual reasoning, data augmentation
Compressor summary: OILCA is a framework for offline imitation learning that uses counterfactual data augmentation to generate expert data and improve the agent's generalization ability.
Takashi Furuya, Michael Anthony Puthawala, Matti Lassas, Maarten V. de Hoop
https://openreview.net/forum?id=6cc69ArD3O
Keywords: Deep Learning, Operator Learning, Functional Analysis, Injectivity, Bijectivity, Universal approximation
Compressor summary: This paper investigates when neural networks can learn injective and surjective operators between function spaces and their applications in uncertainty quantification and inverse problems.
Han Shao, Avrim Blum, Omar Montasser
https://openreview.net/forum?id=6cJKcIxPck
Keywords: strategic classification, mistake bound in online learning, PAC learning
Compressor summary: The study analyzes how strategic feature manipulation affects learning in classification problems with different levels of information available to the learner, focusing on ball and non-ball manipulations.
Junkun Yuan, Xinyu Zhang, Hao Zhou, Jian Wang, Zhongwei Qiu, Zhiyin Shao, Shaofeng Zhang, Sifan Long, Kun Kuang, Kun Yao, Junyu Han, Errui Ding, Lanfen Lin, Fei Wu, Jingdong Wang
https://openreview.net/forum?id=6XPPfZkhKi
Keywords: human centric perception, masked image modeling, structural-aware pre-training
Compressor summary: The paper introduces HAP, a pre-training method for human-centric perception tasks that uses masked image modeling with human structure priors to improve performance on various benchmarks.
Lin Wang, Yongxin Guo, Tao Lin, Xiaoying Tang
https://openreview.net/forum?id=6XC5iKqRVm
Keywords: federated learning, client sampling
Compressor summary: DELTA is an unbiased sampling scheme for Federated Learning that considers client diversity and local variance to improve convergence and reduce variance caused by partial client participation.
Mohak Bhardwaj, Tengyang Xie, Byron Boots, Nan Jiang, Ching-An Cheng
https://openreview.net/forum?id=6UCMa0Qgej
Keywords: model based, offline, reinforcement learning, adversarial training
Compressor summary: ARMOR is a novel model-based offline RL framework that learns policies to improve upon any reference policy, optimizing for worst-case performance and being robust to hyperparameters.
Shentao Yang, Shujian Zhang, Congying Xia, Yihao Feng, Caiming Xiong, Mingyuan Zhou
https://openreview.net/forum?id=6SRE9GZ9s6
Keywords: Preference Learning, Training Guidance Learning, Language Model Fine-tuning, Text Sequence Generation
Compressor summary: The paper proposes a method to align language models with preferences by iterating between grounding preferences into token-level training and improving the LM with learned guidance, using a framework that extends pairwise-preference learning and two minimalist learning objectives.
Yue Kang, Cho-Jui Hsieh, Thomas Chun Man Lee
https://openreview.net/forum?id=6RiqluMFNz
Keywords: bandits
Compressor summary: The paper introduces robust Lipschitz bandit algorithms that can handle adaptive adversaries corrupting stochastic rewards up to a budget $C$, achieving sub-linear regret even when $C$ is unknown, and provides lower bounds and experiments.
Chenyang Le, Yao Qian, Long Zhou, Shujie LIU, Yanmin Qian, Michael Zeng, Xuedong Huang
https://openreview.net/forum?id=6Qx7G1xrAk
Keywords: end-to-end speech to text translation, cross-modality learning, joint speech and language training
Compressor summary: ComSL is a data-efficient speech-language model that combines cross-modality learning and multi-task learning to achieve state-of-the-art performance in speech-to-text translation across 21 languages.
Chris Subia-Waud, Srinandan Dasmahapatra
https://openreview.net/forum?id=6Odmtoek02
Keywords: Quantization, compression, bayesian neural networks, accelerators
Compressor summary: The paper proposes a probabilistic method to cluster neural network weights based on their position-specific uncertainty, improving compression and accuracy in large models.
Jialin Chen, Zhitao Ying
https://openreview.net/forum?id=6OOgw4boZI
Keywords: Explainability, Temporal Graph Neural Network
Compressor summary: TempME is a novel approach that uncovers the most pivotal temporal motifs guiding the prediction of temporal graph neural networks, improving their explainability and trustworthiness.
Sharan Vaswani, Amirreza Kazemi, Reza Babanezhad Harikandeh, Nicolas Le Roux
https://openreview.net/forum?id=6MQ5cheYDZ
Keywords: Decision-aware reinforcement learning, Actor-Critic algorithm, Off-policy updates, General function approximation, Theoretical guarantees
Compressor summary: The paper proposes a joint objective for training actor and critic in reinforcement learning that improves policy improvement and uses surrogate functions similar to TRPO and PPO.
Tan Zhu, Fei Dou, Xinyu Wang, Jin Lu, Jinbo Bi
https://openreview.net/forum?id=6JrckqCxtl
Keywords: Feature interaction modeling, model interpretation framework, adptive-order interaction, piece-wise polynomial
Compressor summary: PAM is a new module that allows neural networks to model nonlinear feature interactions adaptively using polyhedrons, improving predictive performance and interpretability.
Yu-Ren Liu, Biwei Huang, Zhengmao Zhu, Honglong Tian, Mingming Gong, Yang Yu, Kun Zhang
https://openreview.net/forum?id=6JJq5TW9Mc
Keywords: Model-based Reinforcement Learning; Causal Representation Learning;
Compressor summary: IFactor is a framework that models four categories of latent state variables to efficiently represent information in high-dimensional, noisy, and non-stationary environments for reinforcement learning.
Yan Scholten, Jan Schuchardt, Aleksandar Bojchevski, Stephan Günnemann
https://openreview.net/forum?id=6IhNHKyuJO
Keywords: Adversarial Robustness, Robustness Certification, Randomized Smoothing, Graph Neural Networks
Compressor summary: Hierarchical randomized smoothing adds targeted noise to subsets of entities in complex objects, improving robustness while maintaining accuracy in classification tasks.
Jiaqi Liu, Jian Lou, Zhan Qin, Kui Ren
https://openreview.net/forum?id=6H8Md75kAw
Keywords: machine unlearning, machin learning privacy, minimax learning, certified removal
Compressor summary: The paper proposes a new certified machine unlearning algorithm for minimax models that uses a total-Hessian-based complete Newton update and Gaussian mechanism, and achieves better generalization and deletion capacity than existing methods.
Yuzheng Hu, Ruicheng Xian, Qilong Wu, Qiuling Fan, Lang Yin, Han Zhao
https://openreview.net/forum?id=6EqUpqMnwl
Keywords: multi-task learning, scalarization, Pareto front
Compressor summary: The paper investigates whether scalarization can fully explore the Pareto front for linear multi-task learning models and finds that it is inherently incapable of doing so, especially for balanced trade-offs between tasks.
Jiaqi Guan, Xingang Peng, PeiQi Jiang, Yunan Luo, Jian Peng, Jianzhu Ma
https://openreview.net/forum?id=6EaLIw3W7c
Keywords: Linker design, generative models
Compressor summary: The paper proposes a new approach to design stable drug-candidate molecules for targeted protein degradation by jointly learning fragment poses and linker structures using a 3D equivariant diffusion model.
Chenyangguang Zhang, Yan Di, Ruida Zhang, Guangyao Zhai, Fabian Manhardt, Federico Tombari, Xiangyang Ji
https://openreview.net/forum?id=6EDHfVHicP
Keywords: hand-held object reconstruction, directed distance field, human-object interaction
Compressor summary: The proposed DDF-HO method uses Directed Distance Fields to represent shapes and capture hand-object interactions more effectively than existing Signed Distance Field methods, achieving better performance on synthetic and real-world datasets.
Zhendong Wang, Yifan Jiang, Yadong Lu, yelong shen, Pengcheng He, Weizhu Chen, Zhangyang Wang, Mingyuan Zhou
https://openreview.net/forum?id=6BZS2EAkns
Keywords: diffusion models, in-context learning
Compressor summary: Prompt Diffusion is a framework that enables in-context learning in diffusion-based generative models using vision-language prompts and a trained diffusion model, achieving high-quality results on various tasks and unseen vision tasks.
Drago Plecko, Elias Bareinboim
https://openreview.net/forum?id=6AAbWSF6Qg
Keywords: Fair Machine Learning, Causal Inference, Decision-Making
Compressor summary: The paper studies outcome control, where automated systems optimize an outcome variable while ensuring fairness and equity by considering sensitive attributes like gender, race, and religion.
Minbo Gao, Zhengfeng Ji, Tongyang Li, Qisheng Wang
https://openreview.net/forum?id=69dAz94zPv
Keywords: Online learning, quantum computing, zero-sum games, linear programming, optimistic multiplicative weight update
Compressor summary: The paper presents a quantum algorithm that finds an approximate Nash equilibrium for zero-sum games quickly and with succinct outputs, using a novel quantum multi-sampling technique.
Chiyuan Zhang, Daphne Ippolito, Katherine Lee, Matthew Jagielski, Florian Tramèr, Nicholas Carlini
https://openreview.net/forum?id=67o9UQgTD0
Keywords: Memorization, Language Models
Compressor summary: The paper proposes a method to study and quantify language model memorization by examining how predictions change when certain documents are removed from training data.
Zhiwei Hao, Jianyuan Guo, Kai Han, Han Hu, Chang Xu, Yunhe Wang
https://openreview.net/forum?id=67MTWzhEOn
Keywords: knowledge distillation, small-data pitfall, vanilla kd
Compressor summary: The paper argues that using larger datasets and stronger data augmentation techniques can improve knowledge distillation approaches for limited-capacity architectures, and demonstrates this with state-of-the-art results on ImageNet.
Yihan Zhou, Eric Price
https://openreview.net/forum?id=66XhNDahk6
Keywords: active learning, binary classification, competitive ratio
Compressor summary: The paper proposes a new agnostic active learning algorithm that is competitive with the optimal algorithm and has a logarithmic overhead in the query complexity, and shows that improving this is NP-hard.
Amit Daniely, Nathan Srebro, Gal Vardi
https://openreview.net/forum?id=65aDEXIhih
Keywords: Learning neural networks, Computational complexity, Hardness of learning, Smoothed analysis, Degenerate weights
Compressor summary: The paper investigates if assumptions about Gaussian inputs and non-degenerate weights are enough to learn deeper neural networks efficiently, and finds that they are not for depth-$3$ ReLU networks.
Eric Neyman, Tim Roughgarden
https://openreview.net/forum?id=639RkUOmW8
Keywords: Logarithmic pooling, online learning, no-regret learning, calibrated experts, online mirror descent, prediction with expert advice
Compressor summary: The paper proposes a novel algorithm for learning expert weights in an online adversarial setting, where experts' forecasts are consistent and calibrated, to minimize log loss using logarithmic pooling and achieve low regret.
Jianzhun Shao, Yun Qu, Chen Chen, Hongchang Zhang, Xiangyang Ji
https://openreview.net/forum?id=62zmO4mv8X
Keywords: multi-agent reinforcement learning, offline reinforcement learning
Compressor summary: The paper proposes CounterFactual Conservative Q-Learning (CFCQL), a new multi-agent offline RL algorithm that estimates conservative values for each agent separately, improving performance and reducing the impact of out-of-distribution actions and value overestimation.
Jacob Portes, Alexander R Trott, Sam Havens, DANIEL KING, Abhinav Venigalla, Moin Nadeem, Nikhil Sardana, Daya Khudia, Jonathan Frankle
https://openreview.net/forum?id=5zipcfLC2Z
Keywords: BERT, Pretraining, Efficiency, FlashAttention, ALiBi
Compressor summary: MosaicBERT is a fast and efficient BERT-style encoder architecture that allows for low-cost pretraining, enabling researchers to customize their own NLP models.
Sagar Vaze, Andrea Vedaldi, Andrew Zisserman
https://openreview.net/forum?id=5ytypAqAsR
Keywords: Category discovery, semi-supervised learning, self-supervised learning, classification
Compressor summary: The paper introduces Clevr-4, a synthetic dataset for Generalized Category Discovery, and proposes a novel method called $\mu$GCD that leverages consistent findings from representation learning to outperform existing approaches on both synthetic and real data.
Kazu Ghalamkari, Mahito Sugiyama, Yoshinobu Kawahara
https://openreview.net/forum?id=5yedZXV7wt
Keywords: Tensor decomposition, Energy based model, Tensor networks
Compressor summary: The paper introduces many-body approximation, a new method for decomposing non-negative tensors that avoids problems with traditional methods by using energy-based modeling and global optimization.
Kevin Course, Prasanth B. Nair
https://openreview.net/forum?id=5yZiP9fZNv
Keywords: variational inference, differential equations, dynamical systems, neural ordinary differential equations, latent stochastic differential equations
Compressor summary: The paragraph describes a novel method for inferring latent stochastic differential equations with low computational cost and memory usage, using amortization and a reparametrization technique.
Hong Sheng Zheng, Yu-Yuan Liu, Chen-Fong Hsu, Tsung Tai Yeh
https://openreview.net/forum?id=5t5u8PQa2T
Keywords: TinyML models, edge AIs, Microcontroller
Compressor summary: StreamNet is a method that improves the performance and memory efficiency of deploying TinyML models on low-power MCUs using 1D and 2D streaming processing and smart parameter selection.
Skander Karkar, Ibrahim Ayed, Emmanuel de Bezenac, patrick gallinari
https://openreview.net/forum?id=5sV53leJCv
Keywords: Deep learning, greedy layerwise training, memory, optimal transport
Compressor summary: The paper introduces TRGL, a module-wise regularization method for neural networks that improves accuracy and reduces memory usage in limited memory settings.
Jinghan Zhang, Shiqi Chen, Junteng Liu, Junxian He
https://openreview.net/forum?id=5r3e27I9Gy
Keywords: Parameter-efficient fine-tuning, module composition
Compressor summary: The paper proposes a method to compose different parameter-efficient modules using linear arithmetic operations in the weight space for various tasks, improving their performance.
Xiao Shou, Debarun Bhattacharjya, Tian Gao, Dharmashankar Subramanian, Oktie Hassanzadeh, Kristin Bennett
https://openreview.net/forum?id=5q8xovQF7r
Keywords: temporal event sequences, causal inference, transformer, causal knowledge graph
Compressor summary: The paper proposes a new method to improve transformer-based models' performance in predicting temporal event sequences using pairwise causal knowledge, and shows its effectiveness in several experiments and applications.
Zongyu Guo, Gergely Flamich, Jiajun He, Zhibo Chen, José Miguel Hernández-Lobato
https://openreview.net/forum?id=5otj6QKUMI
Keywords: Neural Compression, Implicit Neural Representation, Relative Entropy Coding, Bayesian Neural Network
Compressor summary: The authors propose a new method to compress data by overfitting neural networks to their functional representation and encoding the weights using relative entropy coding, which improves rate-distortion performance and allows adjusting trade-offs for different network architectures.
Maofeng Tang, Andrei Liviu Cozma, Konstantinos Georgiou, Hairong Qi
https://openreview.net/forum?id=5oEVdOd6TV
Keywords: Remote Sensting, Self-Supervised Learning
Compressor summary: The paper introduces Cross-Scale MAE, a self-supervised model that learns consistent and meaningful representations for remote sensing image analysis using scale augmentation, cross-scale consistency constraints, contrastive and generative losses, and xFormers library.
Hongwu Peng, Ran Ran, Yukui Luo, Jiahui Zhao, Shaoyi Huang, Kiran Thorat, Tong Geng, Chenghong Wang, Xiaolin Xu, Wujie Wen, Caiwen Ding
https://openreview.net/forum?id=5loV5tVzsY
Keywords: Privacy-Preserving Machine Learning, efficient private inference, machine learning as a service, homomorphic encryption, non-linear pruning, ST-GCN
Compressor summary: LinGCN is a framework that optimizes the performance of homomorphic encryption-based Graph Convolution Networks by reducing multiplication depth and improving latency, accuracy, and scalability for encrypted inference in applications like personal healthcare and financial systems.
Mixue Xie, Shuang Li, Longhui Yuan, Chi Harold Liu, Zehui Dai
https://openreview.net/forum?id=5hVXbiEGXB
Keywords: domain generalization, sequential learning, temporal drift, feature standardization
Compressor summary: EvoS is a method for continual domain generalization over temporal drift that uses multi-scale attention to learn evolving feature distribution patterns and standardize features with generated statistics, enabling models to adapt to gradually shifting data distributions in real-world applications.
Yaroslav Kivva, Saber Salehkaleybar, Negar Kiyavash
https://openreview.net/forum?id=5gz7npbQ6Z
Keywords: Causal inference, Difference-in-Difference, Structural causal models, Potential outcome, Proxy learning
Compressor summary: The paper proposes a new method to estimate the causal effect of a treatment on an outcome using cross moments, when only one proxy variable is available, and shows its advantages over existing methods that require restrictive assumptions or two proxy variables.
Kajetan Schweighofer, Lukas Aichberger, Mykyta Ielanskyi, Günter Klambauer, Sepp Hochreiter
https://openreview.net/forum?id=5eu00pcLWa
Keywords: uncertainty, uncertainty quantification, predictive uncertainty, epistemic uncertainty, out of distribution, mc dropout, deep ensembles, sg-mcmc, adversarial model, adversarial model search, imagenet
Compressor summary: QUAM is a new method to estimate epistemic uncertainty better than current methods by focusing on regions where the product of divergence function and posterior is large, which corresponds to adversarial models.
Amin Karbasi, Grigoris Velegkas, Lin Yang, Felix Zhou
https://openreview.net/forum?id=5cPz5hrjy6
Keywords: Theory, Reinforcement Learning Theory, Statistical Learning Theory, Reproducibility, Replicability
Compressor summary: The authors study replicability in reinforcement learning algorithms and provide efficient algorithms with theoretical guarantees for various settings.
Laura Eline Ruis, Akbir Khan, Stella Biderman, Sara Hooker, Tim Rocktäschel, Edward Grefenstette
https://openreview.net/forum?id=5bWW9Eop7l
Keywords: large language models, pragmatics, natural language processing, communication, conversation, implicature, language model fine-tuning
Compressor summary: The paragraph discusses a study on whether large language models can understand implicatures, a type of inference based on beliefs and knowledge, and finds that instruction-tuned models perform better than others in this task.
Ningyuan Teresa Huang, Ron Levie, Soledad Villar
https://openreview.net/forum?id=5aeyKAZr0L
Keywords: graph neural networks, equivariant machine learning, symmetry, generalization, statistical learning
Compressor summary: The paper explores how active symmetries in GNNs affect their performance on various tasks and proposes a bias-variance formula to choose an appropriate symmetry group for each task.
Chengxu Zuo, Jiawei Fang, Shihui Guo, Yipeng Qin
https://openreview.net/forum?id=5ZMBiS1uMq
Keywords: motion tracking, flexible sensor, on-body displacement, deep learning, domain adaptation
Compressor summary: The authors propose a self-adaptive motion tracking network to deal with the challenges posed by flexible sensors' displacement on human body during different sessions, using a light-weight learnable layer, a Fourier-encoded LSTM network, and a novel sequence discrepancy loss.
Junkang Wu, Jiawei Chen, Jiancan Wu, Wentao Shi, Xiang Wang, Xiangnan He
https://openreview.net/forum?id=5XshcizH9w
Keywords: contrastive learning, distributionally robust optimization, mutual information
Compressor summary: The study shows that contrastive learning is robust to sampling bias, explains its connection to distributionally robust optimization and mutual information, and proposes a new loss function to improve performance and convergence.
Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L. Griffiths, Yuan Cao, Karthik R Narasimhan
https://openreview.net/forum?id=5Xc1ecxO1h
Keywords: large language model, general problem solving, heuristic search, reasoning, planning, decision making
Compressor summary: The Tree of Thoughts framework improves language models' problem-solving abilities by enabling them to explore, plan, and evaluate multiple reasoning paths in tasks requiring non-trivial planning or search.
Zhongjie Yu, Martin Trapp, Kristian Kersting
https://openreview.net/forum?id=5W7cXno10k
Keywords: Characteristic Circuit, Characteristic Function, Probabilistic Circuit, Heterogeneous Data, Density Estimation
Compressor summary: Characteristic circuits are a new type of probabilistic model that can learn high-dimensional distributions over different types of data and perform efficient inference, even when no closed-form density function is available.
Hossein Esfandiari, Amin Karbasi, Vahab Mirrokni, Grigoris Velegkas, Felix Zhou
https://openreview.net/forum?id=5VQFAvUHcd
Keywords: Theory, Clustering Theory, Statistical Learning Theory, Reproducibility, Replicability
Compressor summary: The authors propose replicable algorithms for statistical clustering problems like $k$-medians, $k$-means, and $k$-centers by using approximation routines for their combinatorial versions in a black-box manner.
Jonathan Crabbé, Mihaela van der Schaar
https://openreview.net/forum?id=5UwnKSgY6u
Keywords: interpretability, explainability, robustness, invariance, equivariance, geometric deep learning
Compressor summary: The paper proposes a framework to measure and improve the robustness of interpretability methods for neural networks that are invariant under specific symmetry groups.
Kai Wang, Fei Yang, Shiqi Yang, Muhammad Atif Butt, Joost van de Weijer
https://openreview.net/forum?id=5UXXhVI08r
Keywords: Diffusion Models; Text-guided Image Edit; Textual Inversion; Localization
Compressor summary: Dynamic Prompt Learning (DPL) improves text-to-image generation by focusing cross-attention maps on correct nouns in the prompt, enabling fine-grained image editing without unwanted changes.
Zhenyu Wang, Ya-Li Li, Xi Chen, Hengshuang Zhao, Shengjin Wang
https://openreview.net/forum?id=5UOYGfobhC
Keywords: 3d object detection, unified object detection, point clouds
Compressor summary: Uni3DETR is a unified 3D detector that works well for both indoor and outdoor scenes by using a detection transformer with point-voxel interaction and a mixture of query points.
Chirag Modi, Robert M. Gower, Charles Margossian, Yuling Yao, David Blei, Lawrence K. Saul
https://openreview.net/forum?id=5TTV5IZnLL
Keywords: Variational Inference, score matching, KL projection, polyak stepsize
Compressor summary: The authors propose a new variational inference method called Gaussian score matching VI (GSM-VI), which uses the principle of score matching and is faster and equally or more accurate than black box variational inference (BBVI) for various Bayesian inference problems.
Rie Johnson, Tong Zhang
https://openreview.net/forum?id=5SIz31OGFV
Keywords: Deep neural network training, Generalization gap, Empirical study
Compressor summary: The paper analyzes how inconsistency and instability of model outputs affect the generalization gap in deep neural networks, and shows that reducing inconsistency improves performance.
Alankrita Bhatt, Nika Haghtalab, Abhishek Shetty
https://openreview.net/forum?id=5R9bZlpZKj
Keywords: Online learning, Log loss, Information theory, Smoothed Analysis, Beyond worst case analysis, Oracle Efficient Online Learning
Compressor summary: The study explores smoothed analysis for sequential probability assignment problem with contexts, leading to optimal fast rates and efficient algorithms for minimax learning.
Zhong Yi Wan, Ricardo Baptista, Anudhyan Boral, Yi-Fan Chen, John Anderson, Fei Sha, Leonardo Zepeda-Nunez
https://openreview.net/forum?id=5NxJuc0T1P
Keywords: optimal transport, probabilistic diffusion models, statistical downscaling
Compressor summary: The paper presents a two-stage probabilistic framework that uses unpaired data to downscale low-resolution numerical simulations of fluid flow to high-resolution data, while preserving the physical statistics and addressing the core difficulties in weather and climate modeling.
Chang Deng, Kevin Bello, Pradeep Kumar Ravikumar, Bryon Aragam
https://openreview.net/forum?id=5MG5C5aS6m
Keywords: global optimization, nonconvex optimization, graphical models, directed acyclic graphs, structure learning
Compressor summary: The paper shows how a simple optimization method can find the best solution for learning an acyclic graphical model from data without getting stuck in spurious solutions.
Zhuodong Yu, Ling Dai, Shaohang Xu, Siyang Gao, Chin Pang Ho
https://openreview.net/forum?id=5La4Y8BnQw
Keywords: Markov decision processes, distributionally robust optimization
Compressor summary: The paper introduces a fast algorithm for solving distributionally robust MDPs with Wasserstein ambiguity sets, which reduces computational cost and improves performance over existing methods.
Hyeonsu Kim, Jeheon Woo, SEONGHWAN KIM, Seokhyun Moon, Jun Hyeong Kim, Woo Youn Kim
https://openreview.net/forum?id=5JcKKRX2iH
Keywords: Mutual information, Easy-to-obtain geometry, Denoising, 3D Graph neural network, OC20
Compressor summary: GeoTMI is a new framework that uses denoising to predict quantum chemical properties accurately using easy-to-obtain corrupted geometries.
Tomas Vaskevicius, Lénaïc Chizat
https://openreview.net/forum?id=5HahZRA0fy
Keywords: Wasserstein barycenters, entropic penalization, optimal transport, Sinkhorn's algorithm
Compressor summary: The paper presents an algorithm for computing doubly regularized Wasserstein barycenters with convergence guarantees and extends it to handle discrete point clouds using Monte Carlo sampling.
Anand Paresh Brahmbhatt, Rishi Saket, Aravindan Raghuveer
https://openreview.net/forum?id=5Gw9YkJkFF
Keywords: PAC learning, Learning from label proportions, Linear thresholds
Compressor summary: The paper presents efficient algorithms for learning linear threshold functions from label proportions using Gaussian distributions and shows their effectiveness through experiments.
Guy Gaziv, Michael J. Lee, James J. DiCarlo
https://openreview.net/forum?id=5GmTI4LNqX
Keywords: Vision, Object Recognition, Human, Primate, Ventral Stream, Adversarial Examples, Behavior Modulation, Behavioral Alignment
Compressor summary: The paragraph discusses how robustified artificial neural networks (ANNs) can disrupt and manipulate human visual perception by finding low-norm image perturbations, while human percepts remain stable in the same regime.
Xin Yan, Hui Fang, Qiang He
https://openreview.net/forum?id=5Fr8Nwi5KF
Keywords: Diffusion Model, Graph Inverse Problems, Source Localization, Information Diffusion
Compressor summary: The paper proposes a probabilistic model called DDMSL to locate the sources and reconstruct the paths of information diffusion over complex networks, using Markov chains and a reversible residual network.
Zeyuan Yin, Eric Xing, Zhiqiang Shen
https://openreview.net/forum?id=5Fgdk3hZpb
Keywords: dataset condensation and distillation, ImageNet Scale
Compressor summary: The SRe$^2$L framework creates smaller synthetic data to train models efficiently on different dataset scales and resolutions, achieving state-of-the-art results with fast speed and less memory consumption.
Zhenmei Shi, Junyi Wei, Yingyu Liang
https://openreview.net/forum?id=5F04bU79eK
Keywords: neural networks, gradient descent, feature learning, provable guarantees, theoretical analysis
Compressor summary: This paper presents a unified analysis framework for two-layer neural networks trained by gradient descent, which explains their ability to learn features from gradients and applies to various problems, revealing new insights into network learning.
Zhiding Liu, Mingyue Cheng, Zhi Li, Zhenya Huang, Qi Liu, Yanhu Xie, Enhong Chen
https://openreview.net/forum?id=5BqDSw8r5j
Keywords: Time series forecasting, deep learning, normalization
Compressor summary: The paragraph describes a new method called SAN that improves time series forecasting by adaptively normalizing and denormalizing data at the slice level, accounting for non-stationary changes in statistical properties.
Abdulkadir Canatar, Jenelle Feather, Albert Wakhloo, SueYeon Chung
https://openreview.net/forum?id=5B1ZK60jWn
Keywords: computational neuroscience, neural manifolds, neuro-AI, statistical physics, representational geometry
Compressor summary: The authors analyze how different deep neural networks predict visual cortical activity using a theoretical framework that relates generalization error to spectral properties, and introduce geometrical measures to interpret the neural prediction error.
Xiaoran Hao, Yash Jhaveri, Patrick Shafto
https://openreview.net/forum?id=5AMa9fiyJq
Keywords: Cooperative Communication, Common Ground, Bayesian Theory
Compressor summary: The paragraph discusses a new theory of cooperative communication that accounts for different levels of common ground, connects it to variational autoencoding, and tests it with simulations.
Jinbiao Chen, Jiahai Wang, Zizhen Zhang, Zhiguang Cao, Te Ye, Siyuan Chen
https://openreview.net/forum?id=593fc38lhN
Keywords: neural heuristic, meta learning, deep reinforcement learning, multi-objective combinatorial optimization
Compressor summary: The proposed efficient meta neural heuristic (EMNH) uses a multi-task model for parallel learning and a hierarchical method for fine-tuning to solve MOCOPs faster and better than existing neural heuristics.
Frederik Rahbæk Warburg, Marco Miani, Silas Brack, Søren Hauberg
https://openreview.net/forum?id=58XMiu8kot
Keywords: Laplace approximation, metric learning, uncertainty quantification, weight posterior, bayesian
Compressor summary: The paper introduces a Bayesian encoder for metric learning using the Laplace Approximation, which improves uncertainty, out-of-distribution detection, and predictive performance.
Hui Yuan, Kaixuan Huang, Chengzhuo Ni, Minshuo Chen, Mengdi Wang
https://openreview.net/forum?id=58HwnnEdtF
Keywords: Theory, Diffusion Model, Reward Optimization, Low-dimensional Data, Distribution estimation
Compressor summary: The authors propose a method to generate samples with desired properties using a reward function and a conditional diffusion model, and show its effectiveness in theory and practice.
Hariharan Manikandan, Yiding Jiang, J Zico Kolter
https://openreview.net/forum?id=559NJBfN20
Keywords: language model, prompting, tabular data, summarization, boosting, adaboost
Compressor summary: The paper demonstrates that large language models can act as weak learners in a boosting algorithm for tabular data, achieving better results than traditional tree-based methods and few-shot learning.
Takahiro Mimori, Michiaki Hamada
https://openreview.net/forum?id=54z8M7NTbJ
Keywords: phylogenetic inference, variational inference, control variates, hyperbolic space
Compressor summary: GeoPhy is a new method for phylogenetic inference that uses continuous geometric spaces to represent tree topologies and improve variational Bayesian models without restricting possible tree candidates.
Daniel Freund, Thodoris Lykouris, Wentao Weng
https://openreview.net/forum?id=54hYifmQZU
Keywords: bandits, learning, queueing systems, optimal control
Compressor summary: The paper introduces the Cost of Learning in Queueing (CLQ), a metric to measure how much parameter uncertainty affects queue length in queueing systems, and develops a unified analysis framework for CLQ that combines Lyapunov and bandit analysis.
Biagio La Rosa, Leilani H. Gilpin, Roberto Capobianco
https://openreview.net/forum?id=51PLYhMFWz
Keywords: compositional explanations, network dissection, explainable artificial intelligence, interpretability
Compressor summary: Clustered Compositional Explanations is a method for broadening the spectrum of neuron behavior approximated by logical formulas of concepts, by combining existing techniques with clustering and a novel search heuristic.
Christian Kümmerle, Johannes Maly
https://openreview.net/forum?id=50hs53Zb3w
Keywords: low-rank models, sparsity, iteratively reweighted least squares, non-convex optimization, quadratic convergence, simultaneously structured data
Compressor summary: The paper introduces an algorithm (IRLS) for recovering data with multiple structures (row-sparsity and low-rank) from linear observations, which has better performance and sample complexity than existing methods.
Xun Xian, Ganghua Wang, Jayanth Srinivasa, Ashish Kundu, Xuan Bi, Mingyi Hong, Jie Ding
https://openreview.net/forum?id=4zWEyYGGfI
Keywords: Backdoor attacks, Backdoor Defense, Security for AI
Compressor summary: The study proposes a framework for detecting backdoor attacks in machine learning models that guarantees few false positives and maximizes the accuracy of identifying poisoned samples.
Vinod Raman, UNIQUE SUBEDI, Ambuj Tewari
https://openreview.net/forum?id=4yXnnCK3r9
Keywords: Adversarial Robustness, PAC Learning
Compressor summary: The authors study how relaxing the robust loss in adversarial learning can improve sample complexity and learnability, while showing that some existing relaxations still require more than finite VC dimension for proper learning.
Ryan Singh, Christopher Buckley
https://openreview.net/forum?id=4xckZu4MPG
Keywords: Attention, Structural Inference, Variational Inference, Predictive Coding, Graphical Models
Compressor summary: The paragraph discusses how attention mechanisms in Transformers can be viewed as inference over adjacency structures in graphical models, unifying different attentional architectures and suggesting potential modifications and generalizations.
Zhongwei Wan, Che Liu, Mi Zhang, Jie Fu, Benyou Wang, Sibo Cheng, Lei Ma, César Quilodrán-Casas, Rossella Arcucci
https://openreview.net/forum?id=4vpsQdRBlK
Keywords: Medical Vision Langauge Pretraining, Cross-lingual, Language bias
Compressor summary: This paper introduces Med-UniC, a framework to integrate medical data from English and Spanish, using Cross-Lingual Text Alignment Regularization (CTR) to reduce community bias and improve performance in various medical vision-language tasks.
Hao Hu, Yiqin Yang, Jianing Ye, Ziqing Mai, Chongjie Zhang
https://openreview.net/forum?id=4vGVQVz5KG
Keywords: offline RL, reward-free, behavior extraction
Compressor summary: UBER is an unsupervised method that uses diverse pseudo-rewards from random neural networks to extract useful behaviors from offline reward-free datasets and improve online RL performance.
Muxi Chen, YU LI, Qiang Xu
https://openreview.net/forum?id=4sDHLxKb1L
Keywords: model debugging, error slice discovery
Compressor summary: HiBug is a framework that uses pre-trained models like chatGPT to automatically discover and explain model bugs in computer vision tasks by identifying underperforming data slices and suggesting human-understandable attributes.
TaeHo Yoon, Kibeom Myoung, Keon Lee, Jaewoong Cho, Albert No, Ernest K. Ryu
https://openreview.net/forum?id=4qG2RKuZaA
Keywords: Generative models, Diffusion probabilistic models, Controlled generation, Human Feedback, RLHF
Compressor summary: The paper proposes censored image generation using a pre-traired diffusion model and minimal human feedback, which is very efficient in preventing undesirable images.
Sangwoong Yoon, Young-Uk Jin, Yung-Kyun Noh, Frank C. Park
https://openreview.net/forum?id=4nSDDokpfK
Keywords: Energy-based Models, Anomaly Detection, Generative Models, Out-of-Distribution Detection, Recovery Likelihood
Compressor summary: The proposed algorithm MPDR trains energy-based models for anomaly detection by perturbing data points along low-dimensional manifolds and generating near-manifold negative samples.
David Loiseaux, Luis Scoccola, Mathieu Carrière, Magnus Bakke Botnan, Steve Oudot
https://openreview.net/forum?id=4mwORQjAim
Keywords: topological data analysis, multiparameter persistent homology, kernel methods, optimal transport
Compressor summary: The paper introduces a new method for stable vectorization of multiparameter persistent homology descriptors using signed Radon measures and demonstrates its effectiveness on different data sets.
Yigit Efe Erginbas, Thomas Courtade, Kannan Ramchandran, Soham Rajesh Phade
https://openreview.net/forum?id=4mXYJzoPhf
Keywords: revenue, price, offer, online
Compressor summary: The study proposes online algorithms to price items efficiently and learn user valuations from feedback, and compares their performance under three user valuation models.
Guanren Qiao, Guiliang Liu, Pascal Poupart, zhiqiang xu
https://openreview.net/forum?id=4mPiqh4pLb
Keywords: Inverse Constrained Reinforcement Learning, Learning from Demonstrations, Muti-Modal Learning
Compressor summary: MMICRL is a new algorithm that can learn multiple constraints from diverse expert agents by using flow-based density estimation and contrastive learning, leading to better imitation policies and behavior diversity.
Louis Serrano, Lise Le Boudec, Armand Kassaï Koupaï, Thomas X Wang, Yuan Yin, Jean-Noël Vittaut, patrick gallinari
https://openreview.net/forum?id=4jEjq5nhg1
Keywords: PDEs, Physics, Operator Learning, Deep Learning, Spatiotemporal
Compressor summary: CORAL is a new method that uses coordinate-based networks to solve PDEs on any spatial sampling and geometry, achieving robust performance across multiple resolutions and problem domains.
Tzu-Heng Huang, Harit Vishwakarma, Frederic Sala
https://openreview.net/forum?id=4iV26fZPUD
Keywords: Parameter Market, Pricing, Efficient Model Training
Compressor summary: The authors propose a framework for trading model parameters as commodities, which could improve large-scale model training and create value for agents.
Chung-En Tsai, Ying-Ting Lin, Yen-Huan Li
https://openreview.net/forum?id=4iTAUsyisM
Keywords: Online portfolio selection, small-loss bound, gradual-variation bound, second-order bound, optimistic FTRL with self-concordant regularizers
Compressor summary: The paper presents new regret bounds for online portfolio selection that depend on the data and apply to non-Lipschitz, non-smooth losses, using novel smoothness characterizations and analysis techniques.
Neel Guha, Mayee F Chen, Kush Bhatia, Azalia Mirhoseini, Frederic Sala, Christopher Re
https://openreview.net/forum?id=4iMpwAlza1
Keywords: language models, prompting, embeddings, weak supervision
Compressor summary: Embroid is a method that improves language models' prompt-based learning by using consistency between predictions for similar samples to generate corrected predictions without additional labeled data.
Yann Dubois, Xuechen Li, Rohan Taori, Tianyi Zhang, Ishaan Gulrajani, Jimmy Ba, Carlos Guestrin, Percy Liang, Tatsunori Hashimoto
https://openreview.net/forum?id=4hturzLcKX
Keywords: Instruction-Following, Reinforcement Learning from Human Feedback, Artificial General Intelligence, Large Language Models
Compressor summary: AlpacaFarm is a simulator for training large language models with human feedback, which is cheaper, more trustworthy, and easier to evaluate than real-world data.
Zhongqi Yue, Qianru Sun, Hanwang Zhang
https://openreview.net/forum?id=4hYIxI8ds0
Keywords: unsupervised domain adaptation, transfer learning
Compressor summary: The paper proposes ICON, a method for unsupervised domain adaptation that removes spurious correlations between features by learning an invariant classifier that is consistent with labels in the source domain and clusters in the target domain.
Ke Jiang, Jia-Yu Yao, Xiaoyang Tan
https://openreview.net/forum?id=4gLWjSaw4o
Keywords: Offline reinforcement learning, state distributional shift, state recovery, inverse dynamics model
Compressor summary: The paper proposes a method to improve offline reinforcement learning by using an inverse dynamics model to guide actions that recover the state distribution, leading to better performance on benchmark tasks.
Hanlin Zhu, Paria Rashidinejad, Jiantao Jiao
https://openreview.net/forum?id=4e0NJbkkd8
Keywords: offline RL, actor-critic, l_2 single-policy concentrability, average bellman error
Compressor summary: A-Crab is an offline reinforcement learning algorithm that uses importance sampling and actor-critic framework, achieving optimal convergence rates and outperforming existing methods in complex environments with insufficient data coverage.
Nicolas Emmenegger, Mojmir Mutny, Andreas Krause
https://openreview.net/forum?id=4anryczeED
Keywords: confidence sets, uncertainty quantification, bandits, active learning, testing
Compressor summary: The paper proposes a new likelihood-based method to construct valid uncertainty estimates for sequential decision-making algorithms that works well for problems with well-specified likelihoods and can handle different noise distributions and estimators.
Yotam Alexander, Nimrod De La Vega, Noam Razin, Nadav Cohen
https://openreview.net/forum?id=4aIpgq1nuI
Keywords: Deep Learning, Locally Connected Neural Networks, Data Distributions, Quantum Entanglement, Tensor Networks
Compressor summary: The authors apply quantum physics concepts to study the suitability of locally connected neural networks for different data distributions, and propose a preprocessing method based on low quantum entanglement.
Richeng Jin, Zhonggen Su, Caijun Zhong, Zhaoyang Zhang, Tony Quek, Huaiyu Dai
https://openreview.net/forum?id=4ZaPpVDjGQ
Keywords: Differential privacy, federated data analytics, discrete valued-mechanism, distributed mean estimation
Compressor summary: The paper explores how discrete-valued mechanisms with finite output space can provide local differential privacy guarantees for federated data analytics while also improving communication efficiency and accuracy.
Peng Cui, Dan Zhang, Zhijie Deng, Yinpeng Dong, Jun Zhu
https://openreview.net/forum?id=4WPhXYMK6N
Keywords: uncertainty calibration, sample difficulty, reliable prediction
Compressor summary: The authors propose a method to use large pre-trained models to improve the reliability and calibration of downstream models by incorporating sample difficulty into entropy regularization, leading to better accuracy and uncertainty estimation.
Chris Lu, Yannick Schroecker, Albert Gu, Emilio Parisotto, Jakob Nicolaus Foerster, Satinder Singh, Feryal Behbahani
https://openreview.net/forum?id=4W9FVg1j6I
Keywords: Reinforcement Learning, Meta-Learning, State Space Models
Compressor summary: The authors propose a modified S4 model for reinforcement learning tasks, which initializes and resets the hidden state in parallel, resulting in faster speeds and better performance than Transformers and RNNs on various tasks.
Rui M. Castro, Fredrik Hellström, Tim van Erven
https://openreview.net/forum?id=4VAF3d5jNg
Keywords: Online learning, prediction with experts, selective sampling, active learning
Compressor summary: The text describes a method for online prediction with expert advice that uses selective sampling to reduce label queries while still achieving optimal regret guarantees in some cases.
Zhiyu Jin, Xuli Shen, Bin Li, Xiangyang Xue
https://openreview.net/forum?id=4ULTSBBY4U
Keywords: Text-to-Image Synthesis, Variable-Sized Image Synthesis, Entropy
Compressor summary: The paper proposes a scaling factor for diffusion models to handle various image sizes and aspect ratios, improving visual fidelity, quality, and alignment without extra training.
Tianyu Liu, Yuge Wang, Zhitao Ying, Hongyu Zhao
https://openreview.net/forum?id=4UCktT9XZx
Keywords: Multimodal Learning; Representation Learning; Graph Neural Network; Similarity Learning; Contrastive Learning; Computational Biology and Bioinformatics; Single-cell genomics
Compressor summary: The novel Multimodal Similarity Learning Graph Neural Network model combines multimodal machine learning and deep graph neural networks to learn gene representations from diverse data, enabling better analysis of gene functions and related biomedical phenomena.
Anthony Gruber, Kookjin Lee, Nathaniel Trask
https://openreview.net/forum?id=4SoTUaTK8N
Keywords: graph neural networks, structure preserving machine learning, neural ordinary differential equations, hamiltonian dynamics, metriplectic dynamics
Compressor summary: This paper introduces new graph neural networks (GNNs) based on physics principles that conserve energy or generate dissipation, and explains their performance in relation to reversibility and irreversibility.
Edward Kim, Yohan Karunanayake, Hanna Kurniawati
https://openreview.net/forum?id=4Sn2vUs0zA
Keywords: POMDP, planning under uncertainty, long horizon
Compressor summary: The paper proposes a modified POMDP problem called Reference-Based POMDP, which balances expected reward and similarity to a reference policy, and shows that it improves long-horizon navigation problems compared to the standard POMDP approach.
Muhammad Akhtar Munir, Salman Khan, Muhammad Haris Khan, Mohsen Ali, Fahad Khan
https://openreview.net/forum?id=4SkPTD6XNP
Keywords: Model Calibration, Object Detection, Detection Transformers, Uncertainty
Compressor summary: The paragraph discusses a new mechanism, Cal-DETR, for calibrating uncertainty in DNN-based object detectors using train-time approaches, which improves their performance and reliability in safety-critical applications.
Wei Wang, Lei Feng, Yuchen Jiang, Gang Niu, Min-Ling Zhang, Masashi Sugiyama
https://openreview.net/forum?id=4RoD1o7yq6
Keywords: Weakly supervised learning, binary classification, unbiased risk estimator
Compressor summary: The paper proposes a new weakly supervised binary classification method called ConfDiff classification, which uses confidence differences between unlabeled data pairs instead of pointwise labeling confidence, and shows that it improves model performance and mitigates overfitting issues.
Jiwen Yu, Xuanyu Zhang, Youmin Xu, Jian Zhang
https://openreview.net/forum?id=4R2Y5B12jm
Keywords: Diffusion models, image steganography, Stable Diffusion, coverless steganography
Compressor summary: The authors propose a new image steganography method, CRoSS, that uses diffusion models to improve security and natural robustness without additional training.
Chao Ma, Cheng Zhang
https://openreview.net/forum?id=4PkBhz18in
Keywords: causality, causal inference, causal model evaluation
Compressor summary: The pairs estimator is a new method to improve causal model evaluation in real-world experiments by reducing variance and achieving near-RCT performance.
Moshe Shenfeld, Katrina Ligett
https://openreview.net/forum?id=4L9g1jUDtO
Keywords: Differential Privacy, Adaptive Data Analysis
Compressor summary: The paper proposes a novel approach to prevent overfitting in adaptive data analysis by using noise-addition algorithms and introducing a new data-dependent stability notion.
Ziyu Wang, Mike Zheng Shou, Mengmi Zhang
https://openreview.net/forum?id=4L3RfWnDzL
Keywords: object representation learning, slot attention, object-centric, contrastive random walks
Compressor summary: The authors propose a method to learn object-centric representations from scenes using cyclic walks between perceptual features and object entities, which enables reasoning abilities in both humans and machines.
Youguang Chen, George Biros
https://openreview.net/forum?id=4L2OlXhiTM
Keywords: statistical learning, active learning, logistic regression, regret minimization
Compressor summary: The paper studies how to use pool-based active learning with multinomial logistic regression for multiclass classification, proves theoretical bounds on excess risk using Fisher Information Ratio (FIR), proposes a regret minimization algorithm based on FIR, and shows experimental results that beat other methods on synthetic and real datasets.
Peiyao Xiao, Hao Ban, Kaiyi Ji
https://openreview.net/forum?id=4Ks8RPcXd9
Keywords: Multi-objective optimization, multi-task leaning, stochastic algorithms, convergence and complexity, Pareto stationarity
Compressor summary: The paper proposes SDMGrad, a stochastic optimization method that improves sample complexity and convergence for multi-objective machine learning problems with multiple criteria and tasks.
Wittawat Jitkrittum, Neha Gupta, Aditya Krishna Menon, Harikrishna Narasimhan, Ankit Singh Rawat, Sanjiv Kumar
https://openreview.net/forum?id=4KZhZJSPYU
Keywords: cascades, deferral rules, adaptive computation, model confidence
Compressor summary: The paper investigates when confidence-based deferral in cascade classifiers may fail and proposes alternative deferral strategies for such scenarios.
Gil Kur, Eli Putterman, Alexander Rakhlin
https://openreview.net/forum?id=4KV2xLeqPN
Keywords: empirical risk minimization, bias-variance decomposition, admissibility
Compressor summary: The paper proves that ERM's suboptimality is due to its bias, not variance, and provides proofs and extensions for various settings and models, as well as stability results and a discussion on the irregular loss landscape of ERM.
Akhil Bagaria, Ben M Abbatematteo, Omer Gottesman, Matt Corsaro, Sreehari Rammohan, George Konidaris
https://openreview.net/forum?id=4JCVw8oMlf
Keywords: hierarchical reinforcment learning
Compressor summary: The paragraph discusses the challenges of learning options in hierarchical reinforcement learning, especially the issue of learning initiation sets, and proposes a new method that uses off-policy value estimation and classification to improve performance on various tasks.
Samantha Chen, Yusu Wang
https://openreview.net/forum?id=4JB42GBxGs
Keywords: neural networks, Wasserstein distance, universal approximation, optimal transport
Compressor summary: The paper proposes a neural network architecture that approximates Wasserstein distance between point sets with low complexity, independent of the input size, and shows its superior performance over other models.
Mateo Espinosa Zarlenga, Katherine M. Collins, Krishnamurthy Dj Dvijotham, Adrian Weller, Zohreh Shams, Mateja Jamnik
https://openreview.net/forum?id=4ImZxqmT1K
Keywords: Explainable Artificial Intelligence, Concept Bottleneck Models, Concept-based Explainability, Interpretability, XAI, Concept Interventions
Compressor summary: IntCEMs are a new type of neural architecture that allows users to improve the model's performance by correcting mispredicted concepts during training, based on an learned intervention policy.
Haoxuan Li, Kunhan Wu, Chunyuan Zheng, Yanghao Xiao, Hao Wang, Zhi Geng, Fuli Feng, Xiangnan He, Peng Wu
https://openreview.net/forum?id=4IWJZjbRFj
Keywords: Debiased recommender system, Multi-task learning, Causal inference
Compressor summary: The paper analyzes the limitations of existing debiasing methods for recommender systems and proposes a new multi-task learning approach to address hidden confounding.
Darshan Chakrabarti, Jelena Diakonikolas, Christian Kroer
https://openreview.net/forum?id=4AmJVaJ78I
Keywords: extensive-form games, first-order methods, coordinate descent
Compressor summary: The paper presents a new cyclic coordinate-descent method for solving sequential games that exploits the recursive structure of dilated regularizers, achieving fast convergence and sometimes outperforming state-of-the-art algorithms.
Manli Shu, Jiongxiao Wang, Chen Zhu, Jonas Geiping, Chaowei Xiao, Tom Goldstein
https://openreview.net/forum?id=4AQ4Fnemox
Keywords: Trustworthy machine learning, Large language models, Supervised fine-tuning, instruction tuning
Compressor summary: The paper introduces AutoPoison, a tool that can change the behavior of large language models by injecting specific examples into their training data.
Hyungjin Chung, Jeongsol Kim, Jong Chul Ye
https://openreview.net/forum?id=497CevPdOg
Keywords: Diffusion models, Inverse problems, Diffusion bridge
Compressor summary: The paper proposes a new method called data Consistent DDB that improves the performance of Direct Diffusion Bridges for solving inverse problems by ensuring data consistency and is open-sourced.
Qihang Fan, Huaibo Huang, Xiaoqiang Zhou, Ran He
https://openreview.net/forum?id=492Hfmgejy
Keywords: Vision Transformer, Lightweight Vision Backbone, Convolution Neural Network
Compressor summary: The paper proposes FASA, a mechanism for vision transformers that adaptively extracts local and global information and their bidirectional interaction, leading to improved performance and efficiency in multiple vision tasks.
Yuyang Qiu, Uday Shanbhag, Farzad Yousefian
https://openreview.net/forum?id=46x3zvYCyQ
Keywords: Federated Learning, Nonsmooth Optimization, Nonconvex Optimization, Bilevel Optimization
Compressor summary: The paper proposes a framework for communication-efficient decentralized training that handles nonconvex optimization, bilevel optimization, and minimax problems in federated learning using randomized smoothing and implicit zeroth-order methods.
Wengong Jin, Siranush Sarkizova, Xun Chen, Nir Hacohen, Caroline Uhler
https://openreview.net/forum?id=46gYakmj4e
Keywords: Energy-based Models, Denoising Score Matching, Equivariant Neural Networks
Compressor summary: The paper proposes a new generative modeling approach for predicting binding affinity of antibodies, using a rotation prediction network that models forces between protein and ligand atoms.
Xiaohan Zhao, Hualin Zhang, Zhouyuan Huo, Bin Gu
https://openreview.net/forum?id=45RBLZBJid
Keywords: asynchronous algorithm, one-device learning, forward gradient descent, directional derivative, forward algorithms
Compressor summary: The paper introduces AsyncFGD, an asynchronous framework for on-device learning that reduces memory usage and improves hardware efficiency by decoupling dependencies and utilizing stale parameters.
Christopher Williams, Fabian Falck, George Deligiannidis, Christopher C. Holmes, Arnaud Doucet, Saifuddin Syed
https://openreview.net/forum?id=43ruO2fMjq
Keywords: U-Net, ResNet, Multi-ResNet, Generalised U-Net, Wavelets, Diffusion models, Generative modelling, PDE Modelling, Image Segmentation
Compressor summary: The paper presents a framework for analysing and designing general U-Net architectures, proposes Multi-ResNets with a simplified encoder and novel U-Net designs that incorporate function constraints and data geometry, and shows their improved performance in various tasks.
Dario Paccagnan, Marco Campi, Simone Garatti
https://openreview.net/forum?id=40L3viVWQN
Keywords: Statistical learning theory, Compression theory, Generalization bounds
Compressor summary: The authors propose Pick-to-Learn, a meta-algorithm that improves compression properties and generalization bounds for learning algorithms on MNIST and a synthetic regression problem.
Yang Cai, Haipeng Luo, Chen-Yu Wei, Weiqiang Zheng
https://openreview.net/forum?id=3xSwxlB0fd
Keywords: two-player zero-sum Markov game, last-iterate convergence, path convergence, learning in games
Compressor summary: The paper proposes an efficient and rational algorithm for learning in two-player zero-sum Markov games with various feedback types and convergence rates, outperforming previous methods in terms of assumptions and synchronization.
Amin Nejatbakhsh, Isabel Garon, Alex H Williams
https://openreview.net/forum?id=3ucmcMzCXD
Keywords: Noise Correlations, Wishart Process, Variational Inference
Compressor summary: The authors propose a new method to estimate the covariance structure of neural noise across naturalistic behaviors using Wishart process models, which can provide information on signal fidelity and noise correlations in unseen conditions.
Eric J Michaud, Ziming Liu, Uzay Girit, Max Tegmark
https://openreview.net/forum?id=3tbTw2ga8K
Keywords: scaling laws, emergence, language models, science of deep learning
Compressor summary: The Quantization Model explains neural scaling laws by describing how network abilities are divided into discrete chunks and learned in order of decreasing use frequency.
Sungik Choi, Hankook Lee, Honglak Lee, Moontae Lee
https://openreview.net/forum?id=3qHlPqzjM1
Keywords: Novelty detection, out-of-distribution detection, consistency models, diffusion models, score-based generative models
Compressor summary: The paper proposes Projection Regret (PR), a novelty detection method for diffusion models that mitigates background bias and outperforms existing methods.
Ben Dai, Yixuan Qiu
https://openreview.net/forum?id=3pEBW2UPAD
Keywords: coordinate descent, linear convergence, primal-dual methods, empirical risk minimization, linear constraints, quantile regression
Compressor summary: ReHLine is a novel algorithm for minimizing regularized ERMs with convex piecewise linear-quadratic loss functions and optional linear constraints, which handles diverse problems and has provable linear convergence rate and linear computational complexity.
Tao Yang, Yuwang Wang, Yan Lu, Nanning Zheng
https://openreview.net/forum?id=3ofe0lpwQP
Keywords: Diffusion Probabilistic Model, Disentangled representation
Compressor summary: The authors propose a new task called disentanglement of diffusion probabilistic models (DPMs), which aims to discover and represent inherent factors behind observations, and introduce an unsupervised approach named DisDiff that achieves this task for the first time.
Haonan Yan, Wenjing Zhang, Qian Chen, Xiaoguang Li, Wenhai Sun, HUI LI, Xiaodong Lin
https://openreview.net/forum?id=3n8PNUdvSg
Keywords: Federated Learning, Model Poisoning Attacks, Proactive Detection, Robust Aggregation, Benign Outlier Identification
Compressor summary: RECESS is a novel defense for federated learning against model poisoning attacks that uses trust scoring based aggregation and proactive querying to detect and protect against malicious clients.
Jia-Qi Yang, De-Chuan Zhan, Le Gan
https://openreview.net/forum?id=3kitbpEZZO
Keywords: Uncertainty calibration, Deep neural networks
Compressor summary: The paper introduces Partitioned Calibration Error (PCE) as a generalized definition of calibration error in deep networks, showing that accurate models should be calibrated across any data partition, and proposes a method to learn semantic-aware partitioning functions for improved calibration.
Ali TehraniJamsaz, Quazi Ishtiaque Mahmud, Le Chen, Nesreen K. Ahmed, Ali Jannesari
https://openreview.net/forum?id=3jAsfo8x8k
Keywords: program representation, graph representation, program analysis, graph neural networks, performance optimization
Compressor summary: PERFOGRAPH is a novel graph-based program representation that captures numerical information and aggregate data structure, improving machine learning methods' ability to reason about programs and achieving state-of-the-art results in various applications.
Xuanjie Liu, Daniel Chin, Yichen Huang, Gus Xia
https://openreview.net/forum?id=3iSj4l8ZGT
Keywords: Physics Symmetry, Time series data, Self-supervised Learning, Representation Augmentation
Compressor summary: The study proposes using physical symmetry as a self-consistency constraint for learning interpretable music representations, leading to a linear pitch factor and representation augmentation.
Ali Younis, Erik B. Sudderth
https://openreview.net/forum?id=3gxiOEf2D6
Keywords: particle, filter, mixture, belief propagation, nonparametric, deep learning, generative, discriminative, graphical model, multiple modes, mutli-modal
Compressor summary: The paragraph describes a novel particle filter method that uses deep neural networks to learn nonparametric representations of uncertainty in latent object states from arbitrary observations, improving tracking and localization performance.
Haotong Qin, Yulun Zhang, Yifu Ding, Yifan liu, Xianglong Liu, Martin Danelljan, Fisher Yu
https://openreview.net/forum?id=3gamyee9Yh
Keywords: Super Resolution, Model Quantization, Deep Learning
Compressor summary: The paper proposes a novel quantized image super-resolution network called QuantSR that achieves accurate and efficient processing under low-bit quantization using the Redistribution-driven Learnable Quantizer (RLQ) and Depth-dynamic Quantized Architecture (DQA).
Tam Minh Nguyen, Tan Minh Nguyen, Richard Baraniuk
https://openreview.net/forum?id=3fd776zKmo
Keywords: transformers, self-attention, total variation, nonlocal functionals, over-smoothing
Compressor summary: NeuTRENO is a new type of transformer model that reduces token uniformity by penalizing the difference between input and output tokens in self-attention layers.
Taesik Gong, Yewon Kim, Taeckyung Lee, Sorn Chottananurak, Sung-Ju Lee
https://openreview.net/forum?id=3bdXag2rUd
Keywords: test-time adaptation, domain adaptation, deep learning, machine learning
Compressor summary: The paper introduces SoTTA, a novel test-time adaptation algorithm that is robust to noisy data by using high-confidence uniform-class sampling and entropy-sharpness minimization.
Sara Babakniya, Zalan Fabian, Chaoyang He, Mahdi Soltanolkotabi, Salman Avestimehr
https://openreview.net/forum?id=3b9sqxCW1x
Keywords: federated learning, class incremental learning, generative models, data-free, continual learning
Compressor summary: The paper proposes a framework for federated class incremental learning that uses generative models to synthesize data from past distributions and train the model without requesting data or storing old information from clients, addressing privacy and resource limitations in federated learning.
Kang Han, Wei Xiang, Lu Yu
https://openreview.net/forum?id=3aVZhMfsyz
Keywords: neural rendering, volume rendering, view synthesis, 3D reconstruction
Compressor summary: NeRFs use neural networks to generate realistic images from multiple views, but this can be slow due to many color network evaluations. This paper proposes a new method called VFR that combines queries into one vector and uses a larger color network for better quality with less training time.
Yuka Hashimoto, Masahiro Ikeda, Hachem Kadri
https://openreview.net/forum?id=3ZrGmenVM2
Keywords: kernel method. generalization bound. C*-algebra. Perron-Frobenius operator and Koopman operator.
Compressor summary: The paper introduces deep RKHM, a deep learning framework for kernel methods using reproducing kernel Hilbert module and Perron-Frobenius operator, with a new generalization bound and insights on benign overfitting.
Yixun Liang, Hao He, Ying-Cong Chen
https://openreview.net/forum?id=3ZICE99e6n
Keywords: 3D vision, 3D reconstruction, Generalizable Neural Surface Reconstruction
Compressor summary: The paper introduces ReTR, a novel framework that uses transformer architecture to improve neural surface reconstruction by redesigning the rendering process with a learnable meta-ray token and cross-attention mechanism, resulting in more accurate and confident surface assessment.
Kaidi Cao, Phitchaya Mangpo Phothilimthana, Sami Abu-El-Haija, Dustin Zelle, Yanqi Zhou, Charith Mendis, Jure Leskovec, Bryan Perozzi
https://openreview.net/forum?id=3YDukx2cpr
Keywords: Graph Neural Networks, Graph Property Prediction
Compressor summary: The Graph Segment Training framework uses a divide-and-conquer approach to learn large graph property prediction with constant memory, and introduces techniques to handle input distribution shift and stale embeddings.
Ilyes Batatia, Mario Geiger, Jose M Munoz, Tess Smidt, Lior Silberman, Christoph Ortner
https://openreview.net/forum?id=3XStpETaO8
Keywords: equivariance, point clouds, machine learning, particle physics
Compressor summary: The paper introduces a new neural network architecture that respects symmetries in various scientific fields, such as high energy physics and computer vision, using reductive Lie groups.
Alberto Bietti, Vivien Cabannes, Diane Bouchacourt, Herve Jegou, Leon Bottou
https://openreview.net/forum?id=3X2EbBLNsk
Keywords: transformers, language models, deep learning theory, interpretability
Compressor summary: The paragraph discusses how transformers learn from global and context-specific information and the importance of understanding their internal mechanisms for reliability.
Paul Viallard, Maxime Haddouche, Umut Simsekli, Benjamin Guedj
https://openreview.net/forum?id=3Wrolscjbx
Keywords: Wasserstein, PAC-Bayes, Generalisation Bound, Algorithm
Compressor summary: The paper proposes new generalization bounds for PAC-Bayesian learning using Wasserstein distance, which are stronger than previous ones and can be used in structural risk minimization.
Zixiang Chen, Junkai Zhang, Yiwen Kou, Xiangning Chen, Cho-Jui Hsieh, Quanquan Gu
https://openreview.net/forum?id=3WAnGWLpSQ
Keywords: Sharpness Aware Algorithm, Deep Learning Theory
Compressor summary: The paper explores why Sharpness-Aware Minimization (SAM) works better than Stochastic Gradient Descent (SGD) for certain neural network settings, especially in preventing overfitting and noise learning.
Hua Wang, Sheng Gao, Huanyu Zhang, Weijie J Su, Milan Shen
https://openreview.net/forum?id=3Py8A1j5N3
Keywords: Differential Privacy, Hyperparameter Tuning, Deep Learning
Compressor summary: DP-HyPO is a novel framework for private and adaptive hyperparameter optimization that bridges the gap between private and non-private methods, providing a thorough differential privacy analysis and showing its effectiveness on various real-world datasets.
Peter Shaw, Mandar Joshi, James Cohan, Jonathan Berant, Panupong Pasupat, Hexiang Hu, Urvashi Khandelwal, Kenton Lee, Kristina Toutanova
https://openreview.net/forum?id=3PjCt4kmRx
Keywords: instruction following, web tasks, user interface tasks, vision and language, representation learning, reinforcement learning, imitation learning, tree search, language grounding, web agents, computer control
Compressor summary: The paper proposes pixel-based agents that use keyboard and mouse actions to interact with GUIs and shows they can perform better than humans on GUI instruction following tasks.
Zerui Tao, Toshihisa Tanaka, Qibin Zhao
https://openreview.net/forum?id=3NWWgB2SuF
Keywords: Tensor decomposition, tensor completion, probabilistic methods
Compressor summary: The paper introduces a flexible tensor decomposition framework using energy-based models and neural networks to learn underlying structures and distributions without prior assumptions, and proposes a variational objective for efficient training.
Guohao Li, Hasan Abed Al Kader Hammoud, Hani Itani, Dmitrii Khizbullin, Bernard Ghanem
https://openreview.net/forum?id=3IyL2XWDkG
Keywords: Communicative Agents, Large Language Models, AI Society, Role-Playing, Society of Mind
Compressor summary: This paper introduces role-playing, a framework for enabling autonomous cooperation among chat-based agents using inception prompting, and demonstrates its potential for studying multi-agent systems.
Kangyang Luo, Shuai Wang, Yexuan Fu, Xiang Li, Yunshi Lan, Ming Gao
https://openreview.net/forum?id=3H9QH1v6U9
Keywords: Federated Learning, Data Heterogeneity, Model Heterogeneity, Data-Free Distillation
Compressor summary: DFRD is a new federated learning method that uses data-free knowledge distillation and a conditional generator to learn a robust global model from privately decentralized clients.
Satyapriya Krishna, Jiaqi Ma, Dylan Z Slack, Asma Ghandeharioun, Sameer Singh, Himabindu Lakkaraju
https://openreview.net/forum?id=3H37XciUEv
Keywords: Machine Learning Explainability, Large Language Models
Compressor summary: AMPLIFY is a framework that uses post hoc explanations to generate rationales for large language models, improving their performance on complex tasks without human involvement.
Artyom Gadetsky, Maria Brbic
https://openreview.net/forum?id=3GpIeVYw8X
Keywords: unsupervised learning, deep learning, generalization, self-supervised learning, clustering
Compressor summary: HUME is a simple framework that can infer human labels on a dataset without supervision, using linear classifiers on top of pretrained representations, and achieves state-of-the-art performance on several image classification benchmarks.
Hyojun Go, Jinyoung Kim, Yunsung Lee, Seunghyun Lee, Shinhyeok Oh, Hyeongdon Moon, Seungtaek Choi
https://openreview.net/forum?id=3G2ec833mW
Keywords: Diffusion Models, Multi-Task Learning
Compressor summary: The paper analyzes diffusion training in multi-task learning, identifies challenges like negative transfer and task affinity, and proposes interval clustering to improve it.
Ruofan Wu, Jiawei Qiao, Mingzhe Wu, Wen Yu, Ming Zheng, Tengfei LIU, Tianyi Zhang, Weiqiang Wang
https://openreview.net/forum?id=3Fc9gnR0fa
Keywords: Survival Analysis, Theory, Semiparametric statistics
Compressor summary: The paper introduces neural frailty machine (NFM), a neural model for survival regressions that extends the proportional hazard assumption and uses multiplicative frailty, with theoretical guarantees and empirical results showing its effectiveness.
Zeren Tan, Yang Tian, Jian Li
https://openreview.net/forum?id=3FJaFElIVN
Keywords: Explanation, LIME, Stability, Local fidelity, Interpretability
Compressor summary: The paper proposes \textsc{Glime}, a framework that improves the stability and local fidelity of explanations for black-box machine learning models by modifying LIME's sampling approach and formulation.
Abhishek Singh, Praneeth Vepakomma, Vivek Sharma, Ramesh Raskar
https://openreview.net/forum?id=3DMDNwd7ND
Keywords: privacy, deep learning, neural networks, adversarial learning, reconstruction guarantees, collaborative inference, MLaaS
Compressor summary: The paper proposes a new framework to provide formal privacy guarantees for cloud-based machine learning inference using local sensitivity and extends an existing method for neural network queries.
Peiyan Dong, Zhenglun Kong, Xin Meng, Pinrui Yu, Yifan Gong, Geng Yuan, Hao Tang, Yanzhi Wang
https://openreview.net/forum?id=3Cj67k38st
Keywords: Multi-view 3D detection, Hardware efficiency, Autonomous driving
Compressor summary: This paper proposes a latency-aware model design for bird's-eye-view detection in autonomous driving systems, using efficient building operators and a hardware-oriented backbone to achieve significant speedups while maintaining accuracy.
Yi Feng, Hu Fu, Qun Hu, Ping Li, Ioannis Panageas, bo peng, Xiao Wang
https://openreview.net/forum?id=3CJOaJugMG
Keywords: zero sum game, time-varying game, optimistic gradient, extra gradient, momentum method
Compressor summary: The paper investigates how OGDA and EG perform in time-varying bilinear zero-sum games and finds that they have different last-iterate behaviors.
Sivaramakrishnan Swaminathan, Antoine Dedieu, Rajkumar Vasudeva Raju, Murray Shanahan, Miguel Lazaro-Gredilla, Dileep George
https://openreview.net/forum?id=3AreDQZ8eO
Keywords: mechanistic interpretability, in-context learning, emergence, large language models
Compressor summary: This paper proposes an interpretable alternative method for in-context learning using causal graphs and shows how it works at different levels of complexity.
Paul Mineiro, Steven R Howard
https://openreview.net/forum?id=39cFjnRpYm
Keywords: off-policy evaluation, anytime-valid
Compressor summary: The paper proposes bounds on the distribution of a sequence of real-valued random variables and their importance-weighted extension for counterfactual distributions in randomized experiments.
Bhishma Dedhia, Michael Chang, Jake Snell, Thomas L. Griffiths, Niraj Jha
https://openreview.net/forum?id=38o372YoYt
Keywords: in-context learning, compositionality, generative models
Compressor summary: This paper explores how analogical reasoning can help in-context visual learners generalize to new tasks and domains, and introduces a meta-learning framework called Im-Promptu that trains agents with different levels of compositional granularity.
Loukas Kavouras, Konstantinos Tsopelas, Giorgos Giannopoulos, Dimitris Sacharidis, Eleni Psaroudaki, Nikolaos Theologitis, Dimitrios Rontogiannis, Dimitris Fotakis, Ioannis Emiris
https://openreview.net/forum?id=38dQv3OwN3
Keywords: subgroup fairness, recourse, counterfactual explanations
Compressor summary: FACTS is a framework to audit subgroup fairness using counterfactual explanations that consider individual and group-level difficulties in achieving desired outcomes while being efficient and explainable.
Ev Zisselman, Itai Lavie, Daniel Soudry, Aviv Tamar
https://openreview.net/forum?id=37cADkATD0
Keywords: Reinforcement Learning, Generalization, State Space Maximum Entropy Exploration
Compressor summary: ExpGen is a reinforcement learning algorithm that uses exploration to achieve zero-shot generalization on unseen tasks, outperforming previous invariance-based methods and achieving state-of-the-art results on ProcGen challenge domains.
Clayton Sanford, Daniel Hsu, Matus Telgarsky
https://openreview.net/forum?id=36DxONZ9bA
Keywords: self-attention, approximation theory, communication complexity
Compressor summary: The text discusses the benefits and drawbacks of attention layers in transformers, focusing on their representation power and complexity in different scenarios, using proof techniques that highlight communication complexity and sparse averaging as a key attention task.
Liming Wu, Zhichao Hou, Jirui Yuan, Yu Rong, Wenbing Huang
https://openreview.net/forum?id=35nFSbEBks
Keywords: Equivariance, Spatio-Temporal GNNs, Physical Dynamics
Compressor summary: The paper introduces ESTAG, an equivariant spatio-temporal graph neural network that uses a novel Equivariant Discrete Fourier Transform and forward attention to simulate non-Markovian physical systems.
Yanghao Li, Tongda Xu, Yan Wang, Jingjing Liu, Ya-Qin Zhang
https://openreview.net/forum?id=35dOU92OJM
Keywords: learned image compression, idempotent compression, right-inverse
Compressor summary: The paper proposes a new idempotent image compression codec using blocked convolution and null-space enhancement, which achieves state-of-the-art performance and has less quality decay after re-compression.
Guan Wang, Yuhao Sun, Sijie Cheng, Sen Song
https://openreview.net/forum?id=30o4ARmfC3
Keywords: neuromorphic computing, spiking neural networks, evolutionary algorithms, inference-only approach, hardware-friendly, robotic locomotion tasks
Compressor summary: The evolving connectivity (EC) framework trains recurrent spiking neural networks without gradients by optimizing connection probability distributions with natural evolution strategies, achieving comparable performance to deep neural networks and outperforming gradient-trained RSNNs on robotic locomotion tasks.
Xiao Ma, Bingyi Kang, Zhongwen Xu, Min Lin, Shuicheng YAN
https://openreview.net/forum?id=2z8noau98f
Keywords: Mutual Information, Offline Reinforcement Learning
Compressor summary: MISA is a novel framework for offline RL that maximizes mutual information between states and actions to improve policy evaluation and constrain the policy improvement direction.
Matthew Niedoba, Jonathan Wilder Lavington, Yunpeng Liu, Vasileios Lioutas, Justice Sefas, Xiaoxuan Liang, Dylan Green, Setareh Dabiri, Berend Zwartsenberg, Adam Scibior, Frank Wood
https://openreview.net/forum?id=2yXExAl0FW
Keywords: Diffusion Models, Trajecotry Forecasting, Autonomous Vehicles, Motion Forecasting, Simulation
Compressor summary: The paper introduces DJINN, a diffusion method that generates realistic traffic scenarios for autonomous vehicle simulations using past, present, or future state observations.
Bing Li, Jiaxin Chen, Xiuguo Bao, Di Huang
https://openreview.net/forum?id=2vADOf3K00
Keywords: Compressed video, Action Recognition, Prompt Tuning
Compressor summary: The paper introduces Compressed Video Prompt Tuning (CVPT), a novel approach to adapt pre-trained raw video models to compressed video understanding tasks, using conditional prompts and cross-modal complementary blocks to improve efficiency and accuracy.
Muhammad A Shah, Aqsa Kashaf, Bhiksha Raj
https://openreview.net/forum?id=2tfG9QaFA7
Keywords: adversarial robustness, computer vision, biologically-inspired, retina, blurring
Compressor summary: The authors propose a method called RBlur that simulates peripheral vision blurring and color desaturation to improve the robustness of deep neural networks against various types of image corruptions, including adversarial attacks.
Shi Chen, Ming Jiang, Qi Zhao
https://openreview.net/forum?id=2rq4LwwjfE
Keywords: Saliency prediction, human attention, low-level vision
Compressor summary: The paper introduces a method to analyze how deep saliency models predict human visual attention by breaking down their implicit features into interpretable components related to semantics and applying it to different scenarios.
Soumya Basu, Abishek Sankararaman
https://openreview.net/forum?id=2nTpPxJ5Bs
Keywords: Double Auction, Markets, Bandits, Regret
Compressor summary: The paper studies how buyers and sellers can efficiently learn their valuations and discover prices in Double Auction markets with bandit feedback, and shows that certain regrets are unachievable.
Namkyeong Lee, Heewoong Noh, Sungwon Kim, Dongmin Hyun, Gyoung S. Na, Chanyoung Park
https://openreview.net/forum?id=2lWh1G1W1I
Keywords: ML4Materials, AI4Science, Graph Neural Networks
Compressor summary: The paper proposes a multi-modal transformer that predicts density of states (DOS) in crystalline materials by considering both energy levels and the relationships between atoms, and shows its effectiveness on two types of DOS.
Jeonghoon Kim, Jung Hyun Lee, Sungdong Kim, Joonsuk Park, Kang Min Yoo, Se Jung Kwon, Dongsoo Lee
https://openreview.net/forum?id=2jUKhUrBxP
Keywords: Large Language Models, Parameter-Efficient Fine-Tuning, Neural Network Quantization
Compressor summary: PEQA is a method that combines parameter-efficient fine-tuning and quantization to reduce memory demands and computational costs of large language models while maintaining or improving their performance.
Nathaniel Lahn, Sharath Raghvendra, Kaiyi Zhang
https://openreview.net/forum?id=2izFpGERjU
Keywords: Optimal Transport, Combinatorial Optimization
Compressor summary: The paper introduces a parallel combinatorial algorithm to compute an additive epsilon-approximation of the optimal transport distance with improved complexity and speed compared to existing methods.
Zenan Li, Yunpeng Huang, Zhaoyu Li, Yuan Yao, Jingwei Xu, Taolue Chen, Xiaoxing Ma, Jian Lu
https://openreview.net/forum?id=2ioRi2uwLR
Keywords: Neuro-symbolic learning, logical constraint learning, symbol grounding, difference-of-convex relaxation
Compressor summary: The paper presents a framework that integrates neural network training, symbol grounding, and logical constraint synthesis for end-to-end learning of neuro-symbolic systems, using techniques to relax and maintain logical constraints.
Zhongzhou Liu, Yuan Fang, Min Wu
https://openreview.net/forum?id=2hhIDEHhkk
Keywords: recommendation systems, causal effect, propensity score, propensity estimation
Compressor summary: The paper proposes PropCare, a framework that estimates exposure and propensity from interaction data for causality-based recommendation systems without additional input.
Khashayar Gatmiry, Zhiyuan Li, Tengyu Ma, Sashank J. Reddi, Stefanie Jegelka, Ching-Yao Chuang
https://openreview.net/forum?id=2hQ7MBQApp
Keywords: Sharpness minimization, Deep learning, Matrix factorization, Deep linear networks, Implicit bias, SGD, Trace of Hessian regularizer
Compressor summary: This paper investigates how minimizing the sharpness of the loss function improves generalization in deep linear networks trained with certain measurements, and shows that it is equivalent to minimizing a matrix norm.
Christoph Hertrich, Yixin Tao, László A. Végh
https://openreview.net/forum?id=2gn9WFlqJ4
Keywords: Differentiable Economics, Mechanism Design, Neural Network Theory, Mode Connectivity, RochetNet
Compressor summary: The paper analyzes mode connectivity of neural networks for optimal auction design and provides the first theoretical justification for their empirical success in this area.
Pum Jun Kim, Yoojin Jang, Jisu Kim, Jaejun Yoo
https://openreview.net/forum?id=2gUCMr6fDY
Keywords: GAN, Evaluation, Support Estimation
Compressor summary: TopP&R is a new and better way to evaluate generative models by measuring the significance and confidence of features in samples, unlike existing metrics that rely on unreliable support estimates.
Minseon Kim, Hyeonjeong Ha, Sooel Son, Sung Ju Hwang
https://openreview.net/forum?id=2f0dlMZlNb
Keywords: Adversarial self supervised learning, targeted attack, self supervised learning, contrastive learning, positive mining
Compressor summary: The paragraph discusses a new method for improving adversarial robustness in unsupervised learning by using positive mining for targeted attacks instead of general adversarial attacks.
Fanqing Meng, Wenqi Shao, zhanglin peng, Chonghe Jiang, Kaipeng Zhang, Yu Qiao, Ping Luo
https://openreview.net/forum?id=2ep5PXEZiw
Keywords: transfer learning, model selection, foundation model
Compressor summary: The paper proposes EMMS, a fast and effective method to estimate the transferability of pre-trained neural networks on multiple multi-modal tasks without fine-tuning them, using large-scale foundation models and unified label embeddings.
Alan Q. Wang, Minh Nguyen, Mert R. Sabuncu
https://openreview.net/forum?id=2ePf1sBgLU
Keywords: Invariant representations, causality, domain generalization
Compressor summary: The paper proposes a nonparametric method using Nadaraya-Watson heads to learn invariant representations across different environments, which can improve domain generalization in machine learning models.
Nathanael Bosch, Philipp Hennig, Filip Tronarp
https://openreview.net/forum?id=2dx5MNs2Ip
Keywords: Probabilistic numerics, differential equations, exponential integrators, Kalman filters, Gaussian processes
Compressor summary: The paper presents a new class of probabilistic integrators for dynamical systems that improve stability and efficiency in stiff problems by including fast, linear dynamics in the prior, and generalizing them to non-linear systems using piece-wise semi-linearity.
Jie Hao, Kaiyi Ji, Mingrui Liu
https://openreview.net/forum?id=2dtU9ZbgSN
Keywords: Coreset Selection, Continual Learning, Bilevel Optimization
Compressor summary: The paper proposes a new bilevel formulation for coreset selection in rehearsal-based continual learning, using a novel regularizer and an efficient optimization algorithm that outperforms baselines.
Dimitris Christou, EFSTRATIOS PANTELEIMON SKOULAKIS, Volkan Cevher
https://openreview.net/forum?id=2doqt9r0r0
Keywords: Online Learning, Regret Analysis, Clustering, k-Median
Compressor summary: The paper proposes an online learning algorithm for clustering facilities with moving costs that achieves logarithmic regret and beats the best fixed solution's average cost by a factor of poly-log(n).
Shuyao Li, Yu Cheng, Ilias Diakonikolas, Jelena Diakonikolas, Rong Ge, Stephen Wright
https://openreview.net/forum?id=2ccH4zjKVs
Keywords: low rank matrix sensing, non-convex optimization, high-dimensional robust statistics, second-order optimization, statistical query model
Compressor summary: The paper proposes a general framework to find approximate second-order stationary points in nonconvex optimization with dimension-independent accuracy guarantees, and applies it to low rank matrix sensing with robustness to data corruption.
Ruixiang Tang, Jiayi Yuan, Yiming Li, Zirui Liu, Rui Chen, Xia Hu
https://openreview.net/forum?id=2cYxNWNzk3
Keywords: Backdoor Defense, Honeypot
Compressor summary: The study proposes an honeypot module to absorb backdoor information and prevent backdoor attacks during fine-tuning of pretrained language models, achieving substantial reduction in attack success rate.
Zhijie Deng, Peng Cui, Jun Zhu
https://openreview.net/forum?id=2bRG4Hj8qd
Keywords: data selection, training acceleration, probabilistic modeling, Bayesian methods
Compressor summary: The paper proposes an efficient algorithm that selects data based on its impact on generalization loss, using Bayesian treatment and zero-shot predictors, and shows improved training efficiency on noisy and imbalanced datasets compared to existing methods.
Cristina Menghini, Andrew Delworth, Stephen Bach
https://openreview.net/forum?id=2b9aY2NgXE
Keywords: vision-language models, prompt-tuning, pseudolabels, self-training
Compressor summary: The study explores using zero-shot pseudolabels and prompt tuning strategies to enhance CLIP's performance on image classification tasks without requiring additional labeled data.
Maksim Zhdanov, Nico Hoffmann, Gabriele Cesa
https://openreview.net/forum?id=2YtdxqvdjX
Keywords: equivariance; group convolutions; implicit kernels; physical simulations
Compressor summary: The paper introduces a flexible way to implement Steerable CNNs using MLPs to parameterize G-steerable kernels, which generalizes to different groups G and performs well on diverse tasks.
Jessica Dai, Paula Gradu, Christopher Harshaw
https://openreview.net/forum?id=2Xqvk2KVAq
Keywords: causal inference, randomized experiments, online optimization
Compressor summary: The paper studies how to create adaptive designs for causal inference that are almost as efficient as non-adaptive ones and proposes new performance measures and an example design that performs well.
Jie Ma, Tianyi Zhou, Guodong Long, Jing Jiang, Chengqi Zhang
https://openreview.net/forum?id=2XT3UpOv48
Keywords: Federated Learning
Compressor summary: The paper proposes Clustered Additive Modeling (CAM) for heterogeneous federated learning, which combines global and cluster models to capture shared features and prevent "clustering collapse."
Richard Antonello, Aditya Vaidya, Alexander Huth
https://openreview.net/forum?id=2W4LxJbgec
Keywords: Encoding Models, Language Models, Neuroscience, Scaling Laws
Compressor summary: Larger language models like OPT and LLaMA can better predict brain responses to natural language using fMRI, with performance scaling logarithmically with model size and nearing a theoretical maximum for certain brain areas.
Ming Xu, Timothy L Molloy, Stephen Gould
https://openreview.net/forum?id=2URr3mkagy
Keywords: implicit differentiation, bi-level optimization; constrained learning and control; safe learning for control
Compressor summary: The paper presents a new method for solving non-convex optimal control problems using the implicit function theorem, which improves scalability, numerical stability, and parallelization compared to previous methods.
Vivek Ramanujan, Thao Nguyen, Sewoong Oh, Ali Farhadi, Ludwig Schmidt
https://openreview.net/forum?id=2SScUiWUbn
Keywords: robustness, out-of-distribution shifts, finetuning, pretraining
Compressor summary: The main factor affecting the robustness of fine-tuned models after pre-training is data quantity, while other factors like label space, image diversity, and domain have limited impact.
Ahmed Khaled, Konstantin Mishchenko, Chi Jin
https://openreview.net/forum?id=2RQhgx1WLA
Keywords: normalized gradient descent, gradient descent, adagrad, adaptive optimization, parameter-free, smooth optimization, convex optimization, edge of stability
Compressor summary: DoWG is a new gradient-based optimizer that works efficiently and universally without parameters and achieves this by using a distance-based weighted average of gradients.
Christos Tsirigotis, Joao Monteiro, Pau Rodriguez, David Vazquez, Aaron Courville
https://openreview.net/forum?id=2OcNWFHFpk
Keywords: out-of-distribution generalization, robustness, fairness, spurious correlations, systematic generalization, model selection
Compressor summary: The paper proposes a new bias-unsupervised method to improve group robustness in ERM by using pretrained self-supervised models and logit adjustment training loss, addressing the limitations of existing approaches that depend on group information or bias labels.
Amila Silva, Spencer Whitehead, Chris Lengerich, Hugh James Leather
https://openreview.net/forum?id=2NncD8AaFK
Keywords: Audio Understanding, Contrastive Learning, Audio-Language Grounding
Compressor summary: The proposed method $CoLLAT$ enhances audio understanding and provides fine-grained audio grounding using a novel pretraining objective, achieving state-of-the-art results in various downstream tasks.
Jiazhong Cen, Zanwei Zhou, Jiemin Fang, chen yang, Wei Shen, Lingxi Xie, Dongsheng Jiang, XIAOPENG ZHANG, Qi Tian
https://openreview.net/forum?id=2NkGfA66Ne
Keywords: Segmentation, NeRF, 3D segmentation
Compressor summary: SA3D is an efficient method that uses SAM and NeRF to segment 3D objects from 2D images with minimal manual input.
Tobias Leemann, Martin Pawelczyk, Gjergji Kasneci
https://openreview.net/forum?id=2NUFe4TZMS
Keywords: Privacy, Membership Inference Attacks
Compressor summary: The paper introduces $f$-MIP, a novel privacy notion for machine learning models that considers realistic adversaries and offers better utility, and proposes $\mu$-GMIP, an enhancement of $f$-MIP with added noise to gradient updates.
Luning Sun, Xu Han, Han Gao, Jian-Xun Wang, Liping Liu
https://openreview.net/forum?id=2JtwuJtoa0
Keywords: AI4Science, Fluid Dynamics, Generative Models, Graph Neural Network
Compressor summary: The paper proposes a new model that combines generative and sequential networks to accurately predict both deterministic and stochastic dynamical systems, using an autoencoder and a conditional normalizing flow model.
Xi Yu, Xiang Gu, Haozhi Liu, Jian Sun
https://openreview.net/forum?id=2Ibp83esmb
Keywords: score-based diffusion model, non-isotropic Gaussian diffusion model, image editing
Compressor summary: The Non-isotropic Gaussian Diffusion Model (NGDM) is a novel technique for image editing that uses different noises and diffusion times for different pixels, achieving high quality results while preserving the source image.
Dian Wang, Xupeng Zhu, Jung Yeon Park, Mingxi Jia, Guanang Su, Robert Platt, Robin Walters
https://openreview.net/forum?id=2FMJtNDLeE
Keywords: Equivariance, Deep Learning, Error Bound, Symmetry
Compressor summary: The paper presents a theory to analyze and quantify various types of equivariance in neural networks when the ground truth function is only partially symmetric, and studies its impact on model error.
Ayush Pande, Gaurav Sharma
https://openreview.net/forum?id=2EiqizElGO
Keywords: Video style transfer
Compressor summary: The paper presents a new approach to video stylization using 3D CNN that disentangles motion and appearance, stylizes the appearance, and then adds back the motion, as well as introducing a new dataset for training and testing video stylization networks.
Ahmadreza Moradipari, Mohammad Pedramfar, Modjtaba Shokrian Zini, Vaneet Aggarwal
https://openreview.net/forum?id=2EVTB1idyR
Keywords: Thompson Sampling, Reinforcement Learning, Bayesian Regret
Compressor summary: The paper shows new Bayesian regret bounds for Thompson Sampling in various reinforcement learning scenarios, with improved upper bounds for the information ratio.
Zineng Tang, Ziyi Yang, Chenguang Zhu, Michael Zeng, Mohit Bansal
https://openreview.net/forum?id=2EDqbSCnmF
Keywords: Generative AI, Diffusion Model, Multimodal Generation, Audio-Video Generation
Compressor summary: CoDi is a novel generative model that can create any combination of language, image, video, or audio from any input modalities, using a unique composable generation strategy and multimodal space alignment.
Jesse Mu, Xiang Lisa Li, Noah Goodman
https://openreview.net/forum?id=2DtxPCL3T5
Keywords: language models, instruction finetuning, prompt compression, distillation, context distillation, prompting, soft prompting, efficiency
Compressor summary: The paper proposes gisting, a method that trains language models to compress prompts into smaller sets of tokens for compute efficiency and other benefits.
Taehyeon Kim, Eric Lin, Junu Lee, Christian Lau, Vaikkunth Mugunthan
https://openreview.net/forum?id=2D7ou48q0E
Keywords: Federated Learning, Semi-Supervised Learning, Object Detection
Compressor summary: The paper introduces FedSTO, a semi-supervised federated object detection framework for autonomous driving that uses selective training and orthogonality regularization to handle data shift and limited labeled data.
Yuling Yao, Justin Domke
https://openreview.net/forum?id=2Cmdh5z6ph
Keywords: simulation based calibration, simulation based inference, Bayesian computation, diagnostics, classifier two-sample test, likelihood-free
Compressor summary: The paper proposes a classification approach that learns test statistics from data to improve Bayesian calibration accuracy and provide interpretable divergence measures, which can be used for simulation-based or traditional inference methods.
Sihui Dai, Wenxin Ding, Arjun Nitin Bhagoji, Daniel Cullina, Haitao Zheng, Ben Y. Zhao, Prateek Mittal
https://openreview.net/forum?id=2CRaOpEKWh
Keywords: adversarial robustness, graph theory, fundamental bounds
Compressor summary: The paper develops a framework to find information-theoretic lower bounds on robust loss for multi-class classifiers under test-time attacks and compares them with state-of-the-art methods.
Wenxin Tai, Yue Lei, Fan Zhou, Goce Trajcevski, Ting Zhong
https://openreview.net/forum?id=2C2WZfCfo9
Keywords: speech enhancement, diffusion models, adaptive prior, dropout, generalization
Compressor summary: The paper proposes DOSE, a model-agnostic method that uses two techniques to incorporate condition information into DDPMs for speech enhancement, improving speech quality and stability.
Ren Li, Benoît Guillard, Pascal Fua
https://openreview.net/forum?id=2BrHBj1Puu
Keywords: garment modeling, draping, deformation, human body modeling
Compressor summary: The paper introduces a new model for draping garments on human body models that can handle multi-layered clothing and various poses, using a combination of 2D and 3D parameterization.
Benjamin Holzschuh, Simona Vegetti, Nils Thuerey
https://openreview.net/forum?id=2BpoGPSDCR
Keywords: inverse problems, diffusion models, learned corrections, score matching
Compressor summary: The authors propose a method to solve inverse problems in physics using diffusion models that combines an approximate inverse simulator and a learned correction function, achieving high accuracy and temporal stability.
Mathieu Chalvidal, Thomas Serre, Rufin VanRullen
https://openreview.net/forum?id=2BFZ8cPIf6
Keywords: Meta-learning, Neural Operators, Kernel methods, In-context learning
Compressor summary: The paper proposes a hybrid approach that combines transductive and inductive methods using vector-valued Reproducing Kernel Banach Spaces to create a meta-learned neural approximator called Transducer, which can efficiently capture new functional relationships and model physical systems with little data.
Maksym Andriushchenko, Dara Bahri, Hossein Mobahi, Nicolas Flammarion
https://openreview.net/forum?id=29WbraPk8U
Keywords: sharpness-aware minimization, low-rank features, understanding feature learning
Compressor summary: Sharpness-aware minimization (SAM) reduces feature rank by pruning activations across various neural network architectures and objectives.
Yixiao Zhou, Ruiqi Jia, Hongxiang Lin, Hefeng Quan, Yumeng Zhao, Xiaoqing Lyu
https://openreview.net/forum?id=28RTu9MOT6
Keywords: Graph Matching, Positional Encoding
Compressor summary: The paper introduces a new graph matching method, PREGM, that leverages spatial information of keypoints to establish correspondence between keypoint sets in images using a positional reconstruction encoder-decoder (PR-EnDec).
Johanna Emilia Immonen, Amauri H Souza, Vikas Garg
https://openreview.net/forum?id=27TdrEvqLD
Keywords: graph representation learning, topological deep learning, persistent homology, graph neural networks
Compressor summary: The text introduces a new concept of color-separating sets to improve the representation of graphs using persistent homology and proposes RePHINE, a model that combines vertex- and edge-level information, achieving better results than existing methods on graph classification tasks.
Chi Gao, Zidong Zhou, Luping Shi
https://openreview.net/forum?id=27CRbwewyb
Keywords: Schema Learning, Temporal Regularity, Event Embedding
Compressor summary: The authors propose Noether Embedding (NE), an efficient method to learn temporal regularities from event embeddings, which outperforms existing methods in detecting and querying valid temporal patterns.
Lingwei Zhu, Zheng Chen, Matthew Kyle Schlegel, Martha White
https://openreview.net/forum?id=26qqUHi9XF
Keywords: reinforcement learning, entropy regularization, Tsallis KL divergence
Compressor summary: The paper proposes a new policy optimization approach in reinforcement learning using Tsallis KL divergence, which generalizes the standard KL divergence and can improve performance in Atari games.
Paweł Czyż, Frederic Grabowski, Julia E Vogt, Niko Beerenwinkel, Alexander Marx
https://openreview.net/forum?id=25vRtG56YH
Keywords: Mutual Information, Invariance, Benchmark, Geometric Machine Learning
Compressor summary: The paper introduces a diverse family of distributions with known ground-truth mutual information and a language-independent benchmarking platform for evaluating mutual information estimators in various settings.
Yulei Qin, Xingyu Chen, Yunhang Shen, Chaoyou Fu, Yun Gu, Ke Li, Xing Sun, Rongrong Ji
https://openreview.net/forum?id=25HiFHPcXg
Keywords: webly supervised learning, representation learning, visual-semantic alignment, collective bootstrapping
Compressor summary: CAPro is a method that combats label noise in webly supervised learning by using textual prototypes, visual feature enhancement, and collective bootstrapping to learn semantically aligned visual representations.
Tsun-Hsuan Wang, Juntian Zheng, Pingchuan Ma, Yilun Du, Byungchul Kim, Andrew Everett Spielberg, Joshua B. Tenenbaum, Chuang Gan, Daniela Rus
https://openreview.net/forum?id=1zo4iioUEs
Keywords: soft robot, diffusion model, co-design
Compressor summary: The paper introduces DiffuseBot, a model that generates soft robot morphologies for various tasks by combining physics simulation with diffusion models.
Duligur Ibeling, Thomas Icard
https://openreview.net/forum?id=1zKRwh5Rl2
Keywords: potential outcomes framework, structural causal model, causal inference, logic, probability, graphical causal models, causal abstraction, causal machine learning
Compressor summary: The paper clarifies how Rubin causal models and structural causal models relate and shows that all Rubin models can be represented by structural models, despite some differences in their principles.
Lore Goetschalckx, Lakshmi Narasimhan Govindarajan, Alekh Karkada Ashok, Aarit Ahuja, David Sheinberg, Thomas Serre
https://openreview.net/forum?id=1xPsn2gCOe
Keywords: alignment, RNNs, reaction times, equilibrium dynamics, perceptual grouping, decision making
Compressor summary: The authors present a method to measure reaction times in recurrent vision models, which helps align them with human visual decision-making patterns across different tasks.
Hadi Vafaii, Jacob L. Yates, Daniel A. Butts
https://openreview.net/forum?id=1wOkHN9JK8
Keywords: NeuroAI, VAE, Dorsal stream, Hierarchical Bayesian Inference
Compressor summary: The paper evaluates a new hierarchical VAE model for motion perception tasks and shows how it aligns with brain function in interpreting causal relationships between stimuli and neuronal responses.
Samuel Dooley, Rhea Sanjay Sukthanker, John P Dickerson, Colin White, Frank Hutter, Micah Goldblum
https://openreview.net/forum?id=1vzF4zWQ1E
Keywords: Bias Mitigation, Fairness, Facial Recognition
Compressor summary: The paragraph discusses how biases in face recognition systems are inherent to neural network architectures, and presents a new method for finding fairer architecture designs that outperforms existing methods on accuracy and fairness.
Sebastian Ament, Sam Daulton, David Eriksson, Maximilian Balandat, Eytan Bakshy
https://openreview.net/forum?id=1vyAG6j9PE
Keywords: Bayesian Optimization, Gaussian Process, Multi-Objective Optimization
Compressor summary: LogEI is a new acquisition function family that is easier to optimize numerically than EI and its variants, improving their performance in Bayesian optimization.
Richard Nock, Ehsan Amid, Manfred K Warmuth
https://openreview.net/forum?id=1vvsIJtnnr
Keywords: Boosting, optimization, exponential families
Compressor summary: The paper introduces $t$-AdaBoost, a generalization of AdaBoost that uses tempered exponential measures to improve convergence rates and bound leveraging coefficients for decision tree induction.
Galen Pogoncheff, Jacob Granley, Michael Beyeler
https://openreview.net/forum?id=1uirUsR9E7
Keywords: NeuroAI, Neuroscience, Visual Stream, Convolutional Neural Networks, Biologically inspired deep learning
Compressor summary: The authors improve CNNs by incorporating neuroscience-derived components to better explain V1 neural activity and properties, advancing NeuroAI research.
Babak Esmaeili, Robin Walters, Heiko Zimmermann, Jan-Willem van de Meent
https://openreview.net/forum?id=1tviRBNxI9
Keywords: representation learning, variational autoencoders, homeomorphism, topological, equivariant, lie groups, normalizing flows
Compressor summary: The paper studies the challenges of training models with geometric latent spaces and proposes a new flow-based model that improves interpretability and generalization by mapping data to multimodal distributions over geometric spaces.
Yanhui Guo, Xinxin Zuo, Peng Dai, Juwei Lu, Xiaolin Wu, Li Cheng, Youliang Yan, Songcen Xu, Xiaofei Wu
https://openreview.net/forum?id=1recIOnzOF
Keywords: Texture Generation, Text-Driven, 3D-Consistent Editing, Neural Radiance Field
Compressor summary: Decorate3D is a method that uses neural networks to create and edit 3D objects from images, allowing users to edit or generate textures with high quality.
Jiawei Liu, Chunqiu Steven Xia, Yuyao Wang, LINGMING ZHANG
https://openreview.net/forum?id=1qvx610Cu7
Keywords: LLM4Code, ChatGPT, Automated Test Generation
Compressor summary: EvalPlus is a framework that uses automatic test input generation to rigorously evaluate the functional correctness of code synthesized by large language models, revealing previously undetected errors and improving programming benchmarks.
Tobias Schröder, Zijing Ou, Jen Ning Lim, Yingzhen Li, Sebastian Josef Vollmer, Andrew Duncan
https://openreview.net/forum?id=1qFnxhdbxg
Keywords: Energy-based models, statistical discrepancy, latent-variable model, density estimation
Compressor summary: Energy Discrepancy is a new loss function for training energy-based models that is faster, more accurate, and has theoretical guarantees compared to existing methods.
Liangyu Chen, Bo Li, Sheng Shen, Jingkang Yang, Chunyuan Li, Kurt Keutzer, Trevor Darrell, Ziwei Liu
https://openreview.net/forum?id=1q0feiJ2i4
Keywords: visual reasoning, large language models
Compressor summary: The paper introduces Cola, a method that uses a large language model (LLM) to coordinate multiple vision-language models (VLMs) for better visual reasoning by enabling natural language communication between them.
Xiaoying Zhang, Junpu Chen, Hongning Wang, Hong Xie, Yang Liu, John C.S. Lui, Hang Li
https://openreview.net/forum?id=1pWNhmbllE
Keywords: off-policy learning, uncertainty
Compressor summary: The paragraph discusses off-policy learning, a procedure used in applications like search engines and recommender systems, and proposes a new estimator (UIPS) that models uncertainty to improve the learning process.
Cenk Baykal, Dylan J Cutler, Nishanth Dikkala, Nikhil Ghosh, Rina Panigrahy, Xin Wang
https://openreview.net/forum?id=1p6teT6F73
Keywords: efficiency, efficient transformers
Compressor summary: AltUp is a method that increases a model's capacity by widening token embeddings without increasing latency much, and it can be combined with other techniques for even better results.
Raunak Kumar, Sarah Dean, Robert Kleinberg
https://openreview.net/forum?id=1osmdAfD4P
Keywords: online learning, online convex optimization, online linear control
Compressor summary: This paper introduces Online Convex Optimization with Unbounded Memory (OCO-UM), a generalization of OCO that accounts for long-term dependence on past decisions, and provides upper and lower bounds on policy regret.
Paweł A. Pierzchlewicz, Konstantin Friedrich Willeke, Arne Nix, Pavithra Elumalai, Kelli Restivo, Tori Shinn, Cate Nealley, Gabrielle Rodriguez, Saumil Patel, Katrin Franke, Andreas S. Tolias, Fabian H. Sinz
https://openreview.net/forum?id=1moStpWGUj
Keywords: most exciting inputs, diffusion models, energy guidance, attention, macaque V4
Compressor summary: The authors propose a new method called Energy Guidance (EGG) for generating most exciting inputs (MEIs) that improve predictions of neuronal activity in macaque V4, reduce computational costs, and can be used for other tasks like studying invariances.
Yuxin Pan, Yize Chen, Fangzhen Lin
https://openreview.net/forum?id=1mdTYi1jAW
Keywords: online 3D bin packing problem, combinatorial optimization problem, reinforcement learning
Compressor summary: The paragraph discusses a new reinforcement learning framework, AR2L, which balances average and worst-case performance for solving online 3D bin packing problems, using a permutation-based attacker to evaluate its robustness.
Yu Cao, Jingrun Chen, Yixin Luo, Xiang ZHOU
https://openreview.net/forum?id=1mJQq6zYaE
Keywords: diffusion models; stochastic differential equations; score-based generative models; asymptotic analysis
Compressor summary: The paper analyzes when ODE or SDE models are more suitable for computer vision tasks by studying error accumulation in two limiting scenarios and comparing numerical results with different data distributions.
George Ma, Yifei Wang, Yisen Wang
https://openreview.net/forum?id=1mAYtdoYw6
Keywords: Graph Neural Networks, Positional Encoding, Spectral Embedding, Laplacian Eigenvectors
Compressor summary: The paper proposes Laplacian Canonization, a minimal and efficient pre-processing method for graph embedding that enhances the invariance properties of spectral embedding techniques.
Joowon Lee, Hanbaek Lyu, Weixin Yao
https://openreview.net/forum?id=1kgK0r8PGg
Keywords: Supervised matrix factorization, multi-objective optimization, global convergence, linear convergence, statistical estimation
Compressor summary: The paper proposes a new method and efficient algorithm for supervised matrix factorization that improves feature extraction and classification tasks in high-dimensional data, such as identifying cancer-related gene groups.
Mathias Schreiner, Ole Winther, Simon Olsson
https://openreview.net/forum?id=1kZx7JiuA2
Keywords: AI4Science, Molecular Dynamics, equivariant neural networks, stochastic dynamics
Compressor summary: Implicit Transfer Operator Learning is a framework to learn surrogates for molecular dynamics simulations with multiple time-resolutions, enabling faster and more accurate modeling of molecular systems using coarse representations.
Shreyas Malakarjun Patil, Loizos Michael, Constantine Dovrolis
https://openreview.net/forum?id=1jhmWkZGy6
Keywords: Neural networks, Hierarchical modularity, Pruning, Sparsity
Compressor summary: The paragraph discusses a method to identify the hierarchical structure of sub-functions in a task using deep neural networks, focusing on the domain of Boolean functions and two vision tasks from the MNIST dataset.
Muhammad Salman Ali, Yeongwoong Kim, Maryam Qamar, Sung-Chang Lim, Donghyun Kim, Chaoning Zhang, Sung-Ho Bae, Hui Yong Kim
https://openreview.net/forum?id=1ihGy9vAIg
Keywords: Image Compression, Correlation
Compressor summary: The paper proposes a novel method to improve learned image compression by introducing a correlation loss that reduces discrepancies between latent features and the assumed distribution, achieving similar performance with significantly less computational complexity.
Hongcheng Wang, Andy Guan Hong Chen, Xiaoqi Li, Mingdong Wu, Hao Dong
https://openreview.net/forum?id=1hZwxBgQ3G
Keywords: Visual Navigation, Demand-Driven Navigation
Compressor summary: The paper introduces Demand-driven Navigation (DDN), a method that allows agents to navigate using the user's demand, not just object names, and leverages common sense knowledge and visual features for better navigation performance.
Ilyas Fatkhullin, Alexander Tyurin, Peter Richtárik
https://openreview.net/forum?id=1h92PmnKov
Keywords: Heavy-ball momentum, Polyak momentum, Error feedback, Federated Learning, Distributed Optimization, Stochastic optimization, Nonconvex optimization
Compressor summary: The paper proposes a simple fix that improves error feedback algorithms for distributed machine learning by applying Polyak's momentum to the latest version, EF21, and shows theoretical and practical benefits.
Nina Balcan, Steve Hanneke, Rattana Pukdee, Dravyansh Sharma
https://openreview.net/forum?id=1h7Uh9zUXc
Keywords: Reliable machine learning, adversarial robustness, distribution shift, theory
Compressor summary: The paper proposes a reliable learner with optimal guarantees in challenging test-time environments like adversarial attacks and distribution shifts, and demonstrates strong performance on examples.
Hongchao Zhang, Junlin Wu, Yevgeniy Vorobeychik, Andrew Clark
https://openreview.net/forum?id=1h2TAUEfc4
Keywords: Safety, Neural Barrier Function, Verification
Compressor summary: The paper proposes a method to verify the safety of neural control barrier functions with ReLU activation for nonlinear systems using piecewise linear segments and interval bound propagation.
Yunhua Zhang, Hazel Doughty, Cees G. M. Snoek
https://openreview.net/forum?id=1g0A9kE8Id
Keywords: Multimodal Learning
Compressor summary: The paper proposes a method to learn from different modalities without assuming all combinations are available during training and shows its effectiveness on various tasks like video classification, robot state regression, and multimedia retrieval.
Han Zhong, Tong Zhang
https://openreview.net/forum?id=1bTG4sJ7tN
Keywords: policy optimization, adversarial lienar MDPs, RL theory
Compressor summary: The paper proposes an optimistic PPO variant for linear MDPs and proves a state-of-the-art regret bound, while introducing novel algorithm design and analysis techniques.
Zhepei Wei, Chuanhao Li, Haifeng Xu, Hongning Wang
https://openreview.net/forum?id=1aQivXgZKj
Keywords: contextual bandit, federated learning, incentive mechanism
Compressor summary: The paper introduces a new federated bandit learning problem that considers self-interested clients and proposes a near-optimal incentivized communication protocol called Inc-FedUCB.
Jayadev Acharya, Clement Louis Canonne, Ziteng Sun, Himanshu Tyagi
https://openreview.net/forum?id=1ZzG6td0el
Keywords: statistical estimation; interactivity; local differential privacy; communication constraint
Compressor summary: The paper develops a framework to derive lower bounds on distributed parameter estimation under various constraints and for different types of distributions using interactive protocols and shows its versatility and effectiveness.
Jimmy T.H. Smith, Shalini De Mello, Jan Kautz, Scott Linderman, Wonmin Byeon
https://openreview.net/forum?id=1ZvEtnrHS1
Keywords: spatiotemporal modeling, ConvLSTM, RNN, state spaces, SSM, S4, S5, long-range dependencies, video prediction
Compressor summary: ConvS5 is a fast and efficient model that combines ConvLSTM tensor modeling with state space methods for long-range spatiotemporal sequence generation.
Sifan Liu
https://openreview.net/forum?id=1YEF6TA8Di
Keywords: Completely uniformly distributed; log-concave sampling; low-discrepancy; MCMC;
Compressor summary: LMC can reduce estimation error by using quasi-random samples from CUD sequences for generating Gaussian perturbations.
Dongjin Kim, Woojeong Kim, Suhyun Kim
https://openreview.net/forum?id=1WpmOipyYI
Keywords: Batch Normalization, Activation Functions, Saturation, Sparsity
Compressor summary: Swapping the order of Batch Normalization and bounded activation functions like Tanh improves performance due to increased asymmetry, sparsity, and a modified Tanh function with consistent asymmetry.
Jiawei Huang, Niao He
https://openreview.net/forum?id=1WMdoiVMov
Keywords: Reinforcement Learning Theory, Transfer RL, Tiered RL
Compressor summary: The paper explores how to transfer knowledge between different tasks in parallel reinforcement learning, focusing on a condition called Optimal Value Dominance and proposing novel algorithms for near-optimal performance.
Naoki Nishikawa, Yuichi Ike, Kenji Yamanishi
https://openreview.net/forum?id=1TJaITmK2Q
Keywords: point cloud, persistence homology, isometry-invariant networks, filtration learning
Compressor summary: The paper proposes a framework that learns an adaptive filtration for point clouds using neural networks and shows its effectiveness in enhancing machine learning accuracy for various applications.
Guangyao Zhai, Evin Pinar Örnek, Shun-Cheng Wu, Yan Di, Federico Tombari, Nassir Navab, Benjamin Busam
https://openreview.net/forum?id=1SF2tiopYJ
Keywords: Scene Graph, Scene Synthesis, Diffusion Model, Graph Convolution Network
Compressor summary: CommonScenes is a model that converts scene graphs into realistic 3D scenes with controllable objects, using variational auto-encoder and latent diffusion, and improves generation consistency, quality, and diversity compared to other methods.
Hanwen Jiang, Santhosh Kumar Ramakrishnan, Kristen Grauman
https://openreview.net/forum?id=1SAzP7W43j
Keywords: Visual Query Localization, Egocentric Video, Spatial-Temporal Correspondence, Episodic Memory
Compressor summary: VQLoC is a single-stage framework that uses joint query-to-frame and frame-to-frame correspondences to perform visual query Localization on long-form egocentric videos faster and more accurately than prior methods.
Praveen Venkatesh, Corbett Bennett, Sam Gale, Tamina K. Ramirez, Greggory Heller, Severine Durand, Shawn R Olsen, Stefan Mihalas
https://openreview.net/forum?id=1PnSOKQKvq
Keywords: partial information decomposition, estimation, bias, inter-area interaction, neuroscience
Compressor summary: The paper presents a new method for efficiently computing partial information decompositions on multivariate Gaussian distributions and evaluates its performance on simulated and real mouse brain data.
Yao Liu, Pratik Chaudhari, Rasool Fakoor
https://openreview.net/forum?id=1MUxtSBUox
Keywords: reinforcement learning, offline reinforcement learning, counterfactual reasoning
Compressor summary: The paper proposes a method for offline reinforcement learning that uses dynamic programming to limit counterfactual reasoning and extrapolation errors, achieving better performance than existing approaches on D4RL benchmarks.
Tingting Dan, Jiaqi Ding, Ziquan Wei, Shahar Z Kovalsky, Minjeong Kim, Won Hwa Kim, Guorong Wu
https://openreview.net/forum?id=1M8nDkUU9b
Keywords: graph neural networks (GNNs), total variation (TV), Euler–Lagrange equation, calculus of variations, over-smoothing, min-max optimization
Compressor summary: The paper presents a general framework for enhancing graph neural networks (GNNs) based on continuous diffusion and variational analysis, which improves their ability to model long-range dependencies and global patterns in graphs, as well as overcomes the over-smoothing problem. The paper also introduces a novel GAN to predict spreading flows in graphs using neural transport equation.
Yang Deng, Weibin Wu, Jianping Zhang, Zibin Zheng
https://openreview.net/forum?id=1JlAV2paGu
Keywords: Transferable adversarial example
Compressor summary: The paper proposes a new method called Blurred-Dilated (BD) that modifies the source model to generate adversarial examples for DNNs in black-box settings, improving transferability and effectiveness against other target models.
Congyue Deng, Jiahui Lei, Bokui Shen, Kostas Daniilidis, Leonidas Guibas
https://openreview.net/forum?id=1IOU2329Za
Keywords: 3D deep learning, equivariant network, pointcloud segmentation, multi-body system
Compressor summary: Banana is a Banach fixed-point network that enables equivariant segmentation by co-evolving part assignment and per-part SE(3)-equivariance in complex systems.
Simina Branzei, Mahsa Derakhshan, Negin Golrezaei, Yanjun Han
https://openreview.net/forum?id=1HKJ3lPz6m
Keywords: multi-unit auctions, repeated auctions, online learning, collusion, games and learning, lower bounds, multiplicative weight updates, bandit learning
Compressor summary: The paper studies repeated multi-unit auctions with uniform pricing for CO2 emissions licenses, and analyzes their efficiency, regret, and collusion-proneness.
Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Adam Sealfon
https://openreview.net/forum?id=1GxKVprbwM
Keywords: differential privacy, local differential privacy, pairwise statistics
Compressor summary: The authors develop methods to compute pairwise statistics of user inputs while preserving differential privacy in a local model setting.
Erik Arakelyan, Pasquale Minervini, Daniel Daza, Michael Cochez, Isabelle Augenstein
https://openreview.net/forum?id=1G7CBp8o7L
Keywords: complex query answering, neural link prediction, knowledge graph embeddings, knowledge graphs, relational learning, adapters
Compressor summary: CQD$^{\mathcal{A}}$ is a score adaptation model that improves the accuracy of complex query answering on incomplete knowledge graphs by recalibrating neural link prediction scores with minimal parameter increase and supporting negation reasoning.
Lorenzo Giambagli, Lorenzo Buffoni, Lorenzo Chicchi, Duccio Fanelli
https://openreview.net/forum?id=1FVmMlifl7
Keywords: Network Slimming, Spectral Analysis, Node Pruning, Teacher-Student
Compressor summary: The paper proposes a new optimization method for student networks in machine learning that allows identifying and isolating an invariant subnetwork that matches the complexity of the teacher network, without degrading performance.
David Skrill, Samuel Victor Norman-Haignere
https://openreview.net/forum?id=1EYKYJeZtR
Keywords: language modeling, temporal integration, transformers, timescales, model interpretation
Compressor summary: The paragraph discusses how language models integrate linguistic information over time and how their integration patterns resemble those of human brains, with structure-dependent exponential and power-law windows across different layers.
Xiao Han, Yukang Cao, Kai Han, Xiatian Zhu, Jiankang Deng, Yi-Zhe Song, Tao Xiang, Kwan-Yee K. Wong
https://openreview.net/forum?id=1DmP6ySKYq
Keywords: 3D generative model, head avatar, diffusion models, neural rendering
Compressor summary: The paragraph discusses a new method called HeadSculpt that can create and edit high-quality 3D head avatars from textual descriptions by incorporating 3D awareness and fine-grained editing capabilities.
Hyunjun Choi, Rajan Udwani, Min-hwan Oh
https://openreview.net/forum?id=1DTCoyAFiV
Keywords: cascade bandit, assortment bandit, upper confidence bound, exploration and exploitation, combinatorial optimization
Compressor summary: The text introduces a new combinatorial bandit model called cascading contextual assortment bandit and two UCB-based algorithms for it that improve existing regret bounds and dependence on problem constants.
Hammaad Adam, Fan Yin, Mary Hu, Neil Tenenholtz, Lorin Crawford, Lester Mackey, Allison Koenecke
https://openreview.net/forum?id=1CpVHL10fh
Keywords: Randomized experiments, heterogeneous effects, causal machine learning, fairness, sequential testing, clinical trials, A/B testing
Compressor summary: The paper introduces CLASH, a machine learning method to stop randomized experiments early when the treatment harms different groups of participants unevenly.
Timothy Fei Truong Jr, Tristan Bepler
https://openreview.net/forum?id=1CJ8D7P8RZ
Keywords: protein fitness prediction, transformer, retrieval, language model, MSA, generative model, protein engineering
Compressor summary: PoET is a generative model that learns to create related proteins from any family by using a Transformer layer and attention mechanism, improving variant function prediction and sequence generation.
Xiuyuan Hu, Guoqing Liu, Yang Zhao, Hao Zhang
https://openreview.net/forum?id=1B6YKnHYBb
Keywords: De novo drug design, Molecular generation, Multi-agent reinforcement learning, GPT
Compressor summary: The paper introduces MolRL-MGPT, a reinforcement learning algorithm with multiple GPT agents for diverse molecular generation in drug design, and shows its effectiveness in designing SARS-CoV-2 inhibitors.
Po-han Li, Sravan Kumar Ankireddy, Ruihan Zhao, Hossein Nourkhiz Mahjoub, Ehsan Moradi Pari, ufuk topcu, Sandeep P. Chinchali, Hyeji Kim
https://openreview.net/forum?id=1A4ZqTmnye
Keywords: Data Compression, Distributed Source Coding, Semantic Communication, Multi-sensor Networks, Bandwidth Allocation, Information Theory
Compressor summary: The authors propose a novel distributed compression framework, NDPCA, that learns low-rank task representations and flexibly adapts to varying available bandwidth in multi-sensor networks, improving performance on tasks like robotic arm manipulation and object detection.
Wenjie Qiu, Wensen Mao, He Zhu
https://openreview.net/forum?id=19AgWnmyoV
Keywords: Goal-Conditioned Reinforcement Learning, Linear Temporal Logic
Compressor summary: The paper proposes a new method for teaching reinforcement learning agents to follow complex temporal logic instructions without extra training, using a simple technique that works with any regular expression.
Yutong Xia, Yuxuan Liang, Haomin Wen, Xu Liu, Kun Wang, Zhengyang Zhou, Roger Zimmermann
https://openreview.net/forum?id=17Zkztjlgt
Keywords: Spatio-temporal forecasting
Compressor summary: CaST is a novel framework that uses causal treatments to improve Spatio-Temporal Graph Neural Networks' forecasting performance and address temporal out-of-distribution issues and dynamic spatial causation.
Tianle Liu, Promit Ghosal, Krishna Balasubramanian, Natesh S. Pillai
https://openreview.net/forum?id=14ZM7FfPx8
Keywords: Stein variational gradient descent, Gaussian variational inference, Rates of Convergence
Compressor summary: SVGD is a nonparametric sampling algorithm whose Gaussian variant has theoretical properties and practical advantages, including convergence to the target distribution in certain cases.
Yangqing Fu, Ming Sun, Buqing Nie, Yue Gao
https://openreview.net/forum?id=0zeLTZAqaJ
Keywords: reinforcement learning, mento carlo tree search, state abstraction
Compressor summary: The PTSA algorithm improves the efficiency of MCTS-based algorithms like AlphaGo and MuZero by reducing the search space and providing theoretical guarantees, leading to faster training on various tasks.
Jun Chen, Hong Chen, Bin Gu, Hao Deng
https://openreview.net/forum?id=0ycX03sMAT
Keywords: Federated zeroth-order optimization, stability analysis, theoretical guarantee, non-convex optimization, sub-Weibull distribution
Compressor summary: The paper develops a theory for FedZO, a black-box optimization method that combines zeroth-order optimization and federated learning, by analyzing its stability, generalization error bound, and convergence rate in both synchronous and asynchronous settings.
Ioannis Anagnostides, Ioannis Panageas, Gabriele Farina, Tuomas Sandholm
https://openreview.net/forum?id=0x2Ou3xHbH
Keywords: no-regret learning, optimistic gradient descent, time-varying games, dynamic regret
Compressor summary: The paper studies how optimistic gradient descent learns in time-varying multiagent settings and provides convergence bounds for equilibrium gaps and second-order variations.
Ziheng Sun, Chris Ding, Jicong Fan
https://openreview.net/forum?id=0vdEHDwamk
Keywords: Lovász Number, graph-level representation learning, unsupervised learning, semi-supervised learning
Compressor summary: The paper introduces a new graph representation method, Lovász principle, based on graph theory concepts that can be improved with neural networks and outperforms existing methods in some tasks.
Natalie Frank, Jonathan Niles-Weed
https://openreview.net/forum?id=0uARg5G04K
Keywords: Adversarial learning, surrogate risks, optimal transport
Compressor summary: The paper investigates which surrogate losses can replace the original loss function without changing the minimizing sequences for robust binary classification under adversarial attacks.
Xinyi HU, Jasper C.H. Lee, Jimmy H.M. Lee
https://openreview.net/forum?id=0tnhFpyWjb
Keywords: Constraint optimization, Predict+Optimize
Compressor summary: The paper introduces Two-Stage Predict+Optimize, a new framework for end-to-end training of supervised learning models to handle unknown parameters in both optimization objectives and constraints, with improved prediction performance over existing methods.
Subin Kim, Kyungmin Lee, June Suk Choi, Jongheon Jeong, Kihyuk Sohn, Jinwoo Shin
https://openreview.net/forum?id=0tEjORCGFD
Keywords: Score Distillation Sampling, Diffusion model, Editing
Compressor summary: Collaborative Score Distillation (CSD) is a novel method that uses Stein Variational Gradient Descent to achieve consistent visual synthesis across multiple images for various editing tasks.
Parshin Shojaee, Kazem Meidani, Amir Barati Farimani, Chandan K. Reddy
https://openreview.net/forum?id=0rVXQEeFEL
Keywords: Symbolic Regression, Transformers, Planning, Deep Learning
Compressor summary: TPSR is a new method for symbolic regression that combines transformers with Monte Carlo Tree Search, improving equation accuracy, complexity, extrapolation, and robustness.
Denizalp Goktas, Arjun Prakash, Amy Greenwald
https://openreview.net/forum?id=0rEJx5QAxt
Keywords: Stackelberg games, Equilibrium Computation, Policy Gradient
Compressor summary: The paper proposes policy gradient methods for solving zero-sum stochastic Stackelberg games with noisy gradients, proves convergence to Stackelberg equilibrium for convex-concave games, and shows that this approach leads to safer and more effective solutions for reach-avoid problems.
Deyu Bo, Yuan Fang, Yang Liu, Chuan Shi
https://openreview.net/forum?id=0kz5RmHxmE
Keywords: Graph Contrastive Learning, Spectral Embedding
Compressor summary: EigenMLP is a spectral encoder for graph contrastive learning that is stable, scalable, and invariant to transformations, while Sp$^{2}$GCL fuses spatial and spectral views for better representation learning.
Jinghan Jia, Jiancheng Liu, Parikshit Ram, Yuguang Yao, Gaowen Liu, Yang Liu, Pranay Sharma, Sijia Liu
https://openreview.net/forum?id=0jZH883i34
Keywords: Machine unlearning, model pruning
Compressor summary: The text introduces model sparsification via weight pruning as a novel approach to improve machine unlearning efficiency and performance in various scenarios, such as defending against backdoor attacks and enhancing transfer learning.
Chi Xie, Zhao Zhang, Yixuan Wu, Feng Zhu, Rui Zhao, Shuang Liang
https://openreview.net/forum?id=0hwq2vOHT4
Keywords: open-vocabulary object detection, referring expression comprehension, multi-modal detection
Compressor summary: The paper introduces a new task called Described Object Detection (DOD) that extends Open-Vocabulary object Detection (OVD) and Referring Expression Comprehension (REC) by using more flexible language expressions, and presents a new dataset ($D^3$) to evaluate it.
Liang Yan, Gengchen Wei, Chen Yang, Shengzhong Zhang, Zengfeng Huang
https://openreview.net/forum?id=0gvtoxhvMY
Keywords: Imbalanced Node Classification, Bias-Variance Decomposition, Graph Neural Networks
Compressor summary: The paper proposes a method to handle class imbalance in graph neural networks using data augmentation, variance decomposition, and regularization, and shows its effectiveness on several benchmark datasets.
Jung Yeon Park, Lawson L.S. Wong, Robin Walters
https://openreview.net/forum?id=0eXniewIvr
Keywords: message passing, dynamics, mesh, symmetry, equivariance
Compressor summary: The paragraph discusses a new gauge equivariant architecture using nonlinear message passing for modeling surface PDEs on meshes that leverages surface geometry and outperforms existing methods for complex and nonlinear dynamics.
Jiaxin Shi, Lester Mackey
https://openreview.net/forum?id=0eRDQQK2TW
Keywords: Stein Variational Gradient Descent, SVGD, variational inference, sampling, optimization, Stein's method
Compressor summary: The paper studies SVGD's convergence rate and shows it has an order of ${1/}{\sqrt{\log\log n}}$ when the target distribution is sub-Gaussian with a Lipschitz score.
Abdullah Omar Alomar, Munther A. Dahleh, Sean Mann, Devavrat Shah
https://openreview.net/forum?id=0e4eiXoUn5
Keywords: Time series, System Identification, Singular Spectrum Analysis
Compressor summary: The paper proposes a two-stage algorithm, SAMoSSA, that combines multivariate Singular Spectrum Analysis and Autoregressive models to estimate non-stationary trends and stationary noise in time series data, and provides theoretical guarantees for its forecasting performance.
Eunbi Yoon, Keehun Park, Sungwoong Kim, Sungbin Lim
https://openreview.net/forum?id=0Wp3VHX0Gm
Keywords: Generative Model, Score-based Method, Lévy processes
Compressor summary: The Lévy-Itō Model (LIM) is a novel score-based generative model that uses heavy-tailed Lévy processes to overcome the limitations of Brownian motion, achieving faster and more diverse sampling with high fidelity on various image datasets.
Tiancheng Jin, Junyan Liu, Chloé Rouyer, William Chang, Chen-Yu Wei, Haipeng Luo
https://openreview.net/forum?id=0WLMVDdvDF
Keywords: reinforcement Learning, best of both worlds, MDP, robust RL, adversarial corruption
Compressor summary: The paper presents algorithms that can learn in adversarial Markov Decision Processes with smoothly increasing regret depending on the degree of maliciousness of the adversary, while handling adversarial transitions and losses.
Anders Aamand, Justin Y. Chen, Huy Nguyen, Sandeep Silwal, Ali Vakilian
https://openreview.net/forum?id=0VcvYQ3uPh
Keywords: learning-augmented algorithms, algorithms with predictions, data-driven algorithms, sublinear, streaming, frequency estimation, sketching
Compressor summary: The paper proposes a novel frequency estimation algorithm that outperforms existing methods without predictions and further improves with heavy-hitter predictions.
Rui Wang, Yanyan Ouyang, Panpan Yu, Wangli Xu
https://openreview.net/forum?id=0Tq1RGJBid
Keywords: Big data, Data averaging, Order statistic, Sampling method, Sketching method.
Compressor summary: This paper studies how to estimate linear models with large data sets and varying dimensions, using sketching techniques and a new averaging-based method that improves upon existing methods in terms of speed and accuracy.
Bingyi Kang, Xiao Ma, Chao Du, Tianyu Pang, Shuicheng YAN
https://openreview.net/forum?id=0P6uJtndWu
Keywords: Offline Reinforcement Learning, Diffusion Models
Compressor summary: Efficient diffusion policy (EDP) improves offline reinforcement learning by addressing the limitations of diffusion models and enabling compatibility with maximum likelihood-based RL algorithms, achieving state-of-the-art results on D4RL benchmark tasks.
Andrey Kuzmin, Markus Nagel, Mart Van Baalen, Arash Behboodi, Tijmen Blankevoort
https://openreview.net/forum?id=0OU1ZXXxs5
Keywords: Neural network quantization, neural network pruning, magnitude pruning, post-training quantization, quantization-aware training
Compressor summary: This paper compares neural network quantization and pruning techniques for compressing deep neural networks and shows that quantization usually performs better than pruning, except in rare cases where extreme compression is needed.
Guangchen Lan, Han Wang, James Anderson, Christopher Brinton, Vaneet Aggarwal
https://openreview.net/forum?id=0ORqsMY6OL
Keywords: reinforcement learning, federated learning
Compressor summary: FedNPG-ADMM is a framework that reduces communication complexity in federated reinforcement learning by using the alternating direction method of multipliers (ADMM) to approximate global policy gradient directions efficiently.
Sarp Aykent, Tian Xia
https://openreview.net/forum?id=0OImBCFsdf
Keywords: geometric deep learning, molecule property prediction, geometric representation learning
Compressor summary: The SaVeNet framework is a more efficient and effective way to learn geometric features in molecules, overcoming challenges like computational inefficiency and limited generalizability.
Hanzhong Allan Guo, Cheng Lu, Fan Bao, Tianyu Pang, Shuicheng YAN, Chao Du, Chongxuan Li
https://openreview.net/forum?id=0NuseeBuB4
Keywords: Diffusion models, SDE-based solver, Gaussian mixture, Stroke-based synthesis
Compressor summary: The paragraph introduces a new method called Gaussian Mixture Solvers (GMS) for improving sampling efficiency and quality in diffusion models, especially for image generation and stroke-based synthesis tasks.
Zhongzhan Huang, Pan Zhou, Shuicheng YAN, Liang Lin
https://openreview.net/forum?id=0N73P8pH2l
Keywords: Diffusion Model, Stable Training, Network architectures
Compressor summary: The authors analyze the instability of UNet in diffusion models due to its long skip connects and propose a coefficient scaling framework that improves stability and robustness, leading to faster training.
Taoran Fang, Yunchao Mercer Zhang, Yang Yang, Chunping Wang, Lei CHEN
https://openreview.net/forum?id=0LmWBhIYLi
Keywords: graph neural networks, prompt tuning
Compressor summary: The paper introduces Graph Prompt Feature (GPF), a universal prompt-based tuning method for graph neural networks that works under any pre-training strategy and outperforms fine-tuning and specialized methods in various scenarios.
Kiarash Banihashem, Leyla Biabani, Samira Goudarzi, MohammadTaghi Hajiaghayi, Peyman Jabbarzade, Morteza Monemizadeh
https://openreview.net/forum?id=0K1ZTfHZ0N
Keywords: Non-monotone submodular maximization, dynamic algorithm, oracle query, video summarization
Compressor summary: The paper presents the first dynamic algorithms to solve non-monotone submodular maximization under cardinality constraints and demonstrates their effectiveness on real-world data sets.
Weichao Mao, Haoran Qiu, Chen Wang, Hubertus Franke, Zbigniew Kalbarczyk, Ravi Iyer, Tamer Basar
https://openreview.net/forum?id=0Iw2dLh8uq
Keywords: Reinforcement learning, game theory, multi-agent systems, meta-learning
Compressor summary: This paper explores how meta-learning can improve multi-agent reinforcement learning in various game settings and provides theoretical and empirical evidence for its benefits.
YIMING WANG, Ming Yang, Renzhi Dong, Binbin Sun, Furui Liu, Leong Hou U
https://openreview.net/forum?id=0FhKURbTyF
Keywords: Reinforcement learning, reward shaping, potential-based exploration, inverse dynamic bisimulation metric
Compressor summary: LIBERTY is a potential-based exploration bonus for deep RL that uses bisimulation metric to measure state discrepancy, enhancing agent's discovery of novel states and improving training efficiency.
Tong Wu, Zhihao Fan, Xiao Liu, Hai-Tao Zheng, Yeyun Gong, yelong shen, Jian Jiao, Juntao Li, zhongyu wei, Jian Guo, Nan Duan, Weizhu Chen
https://openreview.net/forum?id=0EG6qUQ4xE
Keywords: text generation, diffusion model, auto-regression, sequential dependency
Compressor summary: AR-Diffusion is a new diffusion model for text generation that uses dynamic denoising steps to account for the sequential dependency of natural language and achieves faster and better results than existing models on various tasks.
Tsai Hor Chan, Kin Wai Lau, Jiajun Shen, Guosheng Yin, Lequan Yu
https://openreview.net/forum?id=0DpKUzl1Se
Keywords: Bayesian deep learning, high-dimensional testing, uncertainty estimation, out-of-distribution detection
Compressor summary: The proposed framework uses data-adaptive high-dimensional hypothesis testing for uncertainty estimation in deep neural networks, without retraining the feature encoder, and improves performance on OOD detection and task-specific prediction.
Yuheng Jia, Fuchao Yang, Yongqiang Dong
https://openreview.net/forum?id=0CbmvZPBGB
Keywords: partial label learning, dissimilarity propagation, candidate label shrinkage
Compressor summary: The paper proposes a method to disambiguate candidate labels in partial label learning using adversarial similarity and dissimilarity matrices, which improves performance and has theoretical guarantees.
Zi Wang, Alexander Ku, Jason Michael Baldridge, Thomas L. Griffiths, Been Kim
https://openreview.net/forum?id=0BwB03qA5T
Keywords: Interpretability, probing, Bayesian, Gaussian process, transparency
Compressor summary: Gaussian process probes (GPP) is a framework that allows measuring uncertainty about concepts represented by models using a distribution over classifiers without needing access to training data or model details, which can be applied to any pre-trained model and helps understand and evaluate their capabilities.
David Simchi-Levi, Zeyu Zheng, Feng Zhu
https://openreview.net/forum?id=0BfQT652sC
Keywords: multi-armed bandit, worst-case optimality, instance-dependent consistency, light-tailed risk
Compressor summary: The paper studies how to design policies for the stochastic multi-armed bandit problem with three desired properties: worst-case optimality, instance-dependent consistency, and light-tailed risk, and proposes a novel policy that achieves optimal trade-offs among them.
Guillermo Ortiz-Jimenez, Alessandro Favero, Pascal Frossard
https://openreview.net/forum?id=0A9f2jZDGW
Keywords: model editing, transfer learning, neural tangent kernel, vision-language pre-training, deep learning science
Compressor summary: Task arithmetic improves vision-language model performance by leveraging weight disentanglement and linearizing the model's tangent space.
Jason Cheuk Nam Liang, Haihao Lu, Baoyu Zhou
https://openreview.net/forum?id=09bZyE9tfp
Keywords: autobidding, online advertising, bandit online convex optimization, constrained optimization
Compressor summary: The paragraph describes an online learning framework for optimizing ad platform lever decisions in a realistic bandit feedback environment with non-stationary procurement outcomes, and presents a primal-dual algorithm that achieves low regret across various scenarios.
Sangha Park, Jisoo Mok, Dahuin Jung, Saehyung Lee, Sungroh Yoon
https://openreview.net/forum?id=090ORrOAPL
Keywords: Out-of-distribution detection
Compressor summary: The paper explores using textual outliers instead of visual ones for Out-of-Distribution (OoD) detection in neural networks, introducing new methods to generate them and showing their competitive performance on benchmarks.
George Stein, Jesse C. Cresswell, Rasa Hosseinzadeh, Yi Sui, Brendan Leigh Ross, Valentin Villecroze, Zhaoyan Liu, Anthony L. Caterini, Eric Taylor, Gabriel Loaiza-Ganem
https://openreview.net/forum?id=08zf7kTOoh
Keywords: generative models, generative model evaluation, self-supervised learning, representation learning, metrics
Compressor summary: The authors systematically evaluate generative models using human perception of image realism and 17 modern metrics, finding that existing metrics are flawed and propose improvements and a new dataset and library for future work.
Tao Huang, Yuan Zhang, Mingkai Zheng, Shan You, Fei Wang, Chen Qian, Chang Xu
https://openreview.net/forum?id=08hStXdT1s
Keywords: knowledge distillation, diffusion models
Compressor summary: The paper proposes DiffKD, a novel knowledge distillation method that uses diffusion models to denoise and match features between teacher and student models, achieving state-of-the-art performance on various tasks.
Iosif Sakos, Emmanouil-Vasileios Vlatakis-Gkaragkounis, Panayotis Mertikopoulos, Georgios Piliouras
https://openreview.net/forum?id=05P1U0jk8r
Keywords: Nash Equilibrium, Games, Gradient, Non-monotone VI, Natural Gradient, Precondition
Compressor summary: The paper introduces PHGD, a first-order method that exploits latent convex structures in non-cooperative games to achieve convergence under minimal assumptions.
Poorya Mianjy, Raman Arora
https://openreview.net/forum?id=02Uc0G2Cym
Keywords: Adversarial training, neural networks, robustness, guarantees
Compressor summary: The paper proposes a method to train robust neural networks against adversarial attacks by maximizing a lower bound on the loss function using projected gradient descent and provides convergence guarantees and empirical evidence.
Tianyang Hu, Fei Chen, Haonan Wang, Jiawei Li, Wenjia Wang, Jiacheng Sun, Zhenguo Li
https://openreview.net/forum?id=00EKYYu3fD
Keywords: generative model, latent space, distance between distributions, generative adversarial network, vqgan
Compressor summary: The study proposes a novel method to find the optimal latent space for generative models by considering model complexity and improves the performance of various models with less complex ones.
Mehrdad Ghadiri, David Arbour, Tung Mai, Cameron N Musco, Anup Rao
https://openreview.net/forum?id=009LK0vLcY
Keywords: regression adjustment; treatment effect estimation; average treatment effect
Compressor summary: The paper explores regression adjustment in finite populations, using randomized numerical linear algebra to select subsets of subjects for experiments and providing non-asymptotic accuracy bounds for estimating sample mean, individual treatment effects, and average treatment effect.
Fereshte Khani, Marco Tulio Ribeiro
https://openreview.net/forum?id=f39Q3JyoIi
Keywords: alignment, collaborative alignment, debugging, nlp, interference, multi-user interaction
Compressor summary: CoAligment is a framework that helps multiple users align NLP models with their values by learning local and global models and generating instances to resolve disagreements.
Tianxiang Gao, Xiaokai Huo, Hailiang Liu, Hongyang Gao
https://openreview.net/forum?id=Z2he2Y0MoH
Keywords: Gradient descent, deep equilibrium model, Gaussian processes, kernel methods, NNGP, NTK
Compressor summary: The paper analyzes deep equilibrium models (DEQs), infinite-depth neural networks that converge to Gaussian processes, and shows they have non-degenerate kernels, enabling training and generalization.
Rohan Alur, Loren Laine, Darrick K Li, Manish Raghavan, Devavrat Shah, Dennis Shung
https://openreview.net/forum?id=VEpU9rFaQr
Keywords: hypothesis testing, human-AI complementarity, machine learning for healthcare
Compressor summary: The authors propose a statistical test to determine if human experts add value in high-stakes prediction tasks beyond what an algorithm can capture, using the example of admissions decisions for patients with acute gastrointestinal bleeding.
Jiaze Qiu
https://openreview.net/forum?id=OXhymu6MeN
Keywords: Variational Bayes; Naive Mean Field; Gaussian comparison inequalities; High-dimensional statistics; Proportional asymptotic.
Compressor summary: The paper derives asymptotic characterizations for the NMF approximation in high-dimensional linear regression and shows its limitations in practice.
Haotian Xue, Alexandre Araujo, Bin Hu, Yongxin Chen
https://openreview.net/forum?id=JTwxylP6U9
Keywords: Robustness, Adversarial Samples, Diffusion Model
Compressor summary: The paper proposes a new method called Diff-PGD for creating realistic adversarial samples that stay close to natural images while being effective against neural networks.