This page contains one-sentence summaries of ICLR2024 accepted papers, generated by the compressor, my personal LLM-based project.
https://openreview.net/forum?id=zzqn5G9fjn
Compressor summary: The paper proposes a method for fine-tuning multilingual language models that preserves user privacy and improves performance, especially for low-resource languages.
https://openreview.net/forum?id=zyBJodMrn5
Compressor summary: The study evaluates how well different neural network architectures handle multimodal generalization tasks and introduces a new benchmark called gCOG.
https://openreview.net/forum?id=zwU9scoU4A
Compressor summary: The authors propose Graphex Mean Field Games (GXMFGs), which combine graph theoretical concepts with multi-agent reinforcement learning to handle sparse network topologies, and present a novel learning algorithm for these games.
https://openreview.net/forum?id=zwMfg9PfPs
Compressor summary: The paper explores how agents can generalize to new environments with unobserved variables by reusing past marginal information and using residual distributions to estimate causal relationships.
https://openreview.net/forum?id=ztpy1gsUpT
Compressor summary: The paper proposes a method to boost the performance of small language models in medical tasks while preserving patient privacy by using large language models' knowledge-intensive context.
https://openreview.net/forum?id=zpVPhvVKXk
Compressor summary: The paper presents two techniques to improve text-to-image diffusion models by removing undesired content from text embeddings, which are tested on different types of diffusion models.
https://openreview.net/forum?id=zmJDzPh1Dm
Compressor summary: This paper explores how normalizing soft prompt vectors in vision-language models affects their performance and proposes a method to optimize this process.
https://openreview.net/forum?id=zlkXLb3wpF
Compressor summary: This paper introduces a fast path gradient estimator for normalizing flows that works with any relevant architecture, improves training, and allows maximum likelihood estimation for various scientific problems.
https://openreview.net/forum?id=ziDFH8TPPK
Compressor summary: The paper proposes a new approach to predict typhoon trajectories using real-time data and outperforms existing methods in providing accurate forecasts up to 72 hours in advance.
https://openreview.net/forum?id=zgQ0PHeGnL
Compressor summary: ElliDock is a novel learning-based method that predicts elliptic paraboloid interfaces to enable fast and accurate protein-protein docking, especially for antibody-antigen pairs.
https://openreview.net/forum?id=ze7DOLi394
Compressor summary: The interaction tensor is a tool to analyze how features in deep learning models are influenced by data and random seeds, leading to a better understanding of feature learning and generalization error estimation.
https://openreview.net/forum?id=zbOSJ3CATY
Compressor summary: The paper proposes GTSONO, a game theoretic optimizer for neural ODEs that is more robust to adversarial attacks and has computational benefits.
https://openreview.net/forum?id=zbKcFZ6Dbp
Compressor summary: Shadow cones use shadows formed by a light source in hyperbolic space to model hierarchical relations in data better than existing entailment cone constructions.
https://openreview.net/forum?id=zb3b6oKO77
Compressor summary: LMs represent facts about entities and their attributes using binding ID vectors, which are interpretable and transferable across tasks.
https://openreview.net/forum?id=zavLQJ1XjB
Compressor summary: This paper compares temperature scaling and Mixup, two techniques for calibrating deep neural networks' confidence, and shows that Mixup performs better when there are overlapping classes or noisy labels.
https://openreview.net/forum?id=zYXFMeHRtO
Compressor summary: FROSTER is a framework that adapts CLIP for open-vocabulary action recognition by distilling features from a frozen CLIP model and adding video-specific features.
https://openreview.net/forum?id=zWqr3MQuNs
Compressor summary: The paper proposes a new method called MIN-K PROB to detect if an LLM was trained on a given text without knowing the pretraining data, using WIKIMIA as a dynamic benchmark and outperforming previous methods.
https://openreview.net/forum?id=zSxpnKh1yS
Compressor summary: The paper proposes a new metric (WSEP) to improve unsupervised reinforcement learning by enhancing skill diversity and separatability, and shows that using Wasserstein distance instead of KL divergence leads to better task adaptation.
https://openreview.net/forum?id=zRkM6UcA22
Compressor summary: The paper explores how tri-plane discrete data structures can effectively process neural fields for tasks like classification and part segmentation without sacrificing reconstruction quality.
https://openreview.net/forum?id=zMvMwNvs4R
Compressor summary: ENT improves text generation quality and robustness against noise by truncating noisy data based on its distribution of non-target tokens.
https://openreview.net/forum?id=zMoNrajk2X
Compressor summary: Condition-Annealed Diffusion Sampler (CADS) improves diversity and quality of diffusion models by adjusting the conditioning signal during inference.
https://openreview.net/forum?id=zMPHKOmQNb
Compressor summary: The paper proposes a discrete sampling method for generative models that improves sample quality and robustness, and applies it to generate functional antibody proteins with high success rate.
https://openreview.net/forum?id=zEHGSN8Hy8
Compressor summary: SetCSE is an innovative information retrieval framework that uses sets to represent complex semantics and improve language model comprehension for structured information querying.
https://openreview.net/forum?id=zAdUB0aCTQ
Compressor summary: AgentBench is a benchmark to evaluate large language models as agents in complex interactive tasks, revealing their strengths and weaknesses in reasoning and decision-making.
https://openreview.net/forum?id=z8TW0ttBPp
Compressor summary: The paper introduces MathCodeInstruct, a method to fine-tune language models using code for solving math problems, and presents MathCoder models that achieve state-of-the-art results on MATH and GSM8K datasets.
https://openreview.net/forum?id=z7K2faBrDG
Compressor summary: The text discusses how measuring the perceptual scale of physical variables can improve probabilistic modeling of perception and lead to a better understanding of perceptual distances.
https://openreview.net/forum?id=z6n1fKMMC1
Compressor summary: The paper presents an efficient algorithm for learning halfspaces in the testable learning model with different types of noise and target distributions, using new tests that leverage the labels and moment-matching techniques.
https://openreview.net/forum?id=z6KS9D1dxt
Compressor summary: The study proposes a Bayesian framework for cooperative multi-agent reinforcement learning that robustly handles Byzantine failures and adversarial attacks by learning policies based on beliefs about other agents' types.
https://openreview.net/forum?id=z4Hcegjzph
Compressor summary: ROPIM is a self-supervised method for visual pre-training without labels that uses random orthogonal projection instead of binary masking and achieves better performance than MIM.
https://openreview.net/forum?id=z3dfuRcGAK
Compressor summary: The paper explores embedding-based entity alignment as a generative model problem, proposes a new framework called generative entity alignment (GEEA) that uses mutual variational autoencoder (M-VAE) for better conversion between knowledge graphs and entity synthesis from noise vectors.
https://openreview.net/forum?id=z3L59iGALM
Compressor summary: The paper presents RHIP, a novel IRL method for route recommendation that leverages scaling techniques, eigenvector algorithms, and planning horizons to achieve significant improvements in route quality.
https://openreview.net/forum?id=yzRXdhk2he
Compressor summary: The paper introduces Matcher, a perception paradigm that leverages off-the-shelf vision foundation models for various segmentation tasks without training, achieving impressive performance on COCO-20 and LVIS-92 benchmarks.
https://openreview.net/forum?id=yxKZGQLzOP
Compressor summary: The paper proposes a novel method to improve program synthesis from user-provided examples by sampling pairs of programs and examples and using pragmatic inference to choose informative training examples for the synthesizer.
https://openreview.net/forum?id=yuy6cGt3KL
Compressor summary: The authors study how to select the best model for estimating the treatment effect in causal inference using various metrics and real datasets, suggesting new strategies based on hyperparameter tuning and causal ensembling.
https://openreview.net/forum?id=yroyhkhWS6
Compressor summary: The Quadratic Synchronization Rule (QSR) method improves test accuracy in local gradient methods for distributed deep learning by dynamically adjusting the local computation steps $H$ based on the learning rate $\eta$.
https://openreview.net/forum?id=yrgQdA5NkI
Compressor summary: MFNs are a new architecture for learning on graphs that captures non-local interactions using matrix equivariant functions, achieving high performance and scalability.
https://openreview.net/forum?id=ypAT2ixD4X
Compressor summary: This paper shows that randomized parameter-sharing (RPS) outperforms other methods for compressing neural network models and achieves better trade-offs between memory and accuracy, especially in high compression settings. It also introduces a modified RPS technique called STABLE-RPS to address stability and Pareto-continuity issues.
https://openreview.net/forum?id=yozwqhIHXj
Compressor summary: DVP is a neuro-symbolic image translation framework that uses computer vision models for pro-symbolic steps and achieves condition-flexible, in-context reasoning, and systemic controllability and explainability.
https://openreview.net/forum?id=yoVq2BGQdP
Compressor summary: The paper presents an RL method that ensures fairness in multi-agent MDPs with unknown environments using a fairness function, online convex optimization, confidence bounds, and policy-gradient method.
https://openreview.net/forum?id=ymjI8feDTD
Compressor summary: CTM is a new diffusion model sampling method that balances speed and quality by combining adversarial training and denoising score matching, and enables various sampling schemes and exact likelihood computation.
https://openreview.net/forum?id=ym0ubZrsmm
Compressor summary: The paper proposes a new model (SSOD) that improves out-of-distribution detection without using explicit out-of-distribution data by exploiting natural cues from in-distribution data.
https://openreview.net/forum?id=ylhiMfpqkm
Compressor summary: The paper proposes a novel reward-free pre-training method for Generative Flow Networks, which can adapt to downstream tasks and discover modes more efficiently.
https://openreview.net/forum?id=ykEixGIJYb
Compressor summary: Truth-FedBan is a communication protocol that motivates truthful reporting in federated bandit learning by providing independent incentives unrelated to self-reported costs, leading to near-optimal performance and efficiency.
https://openreview.net/forum?id=yiMB2DOjsR
Compressor summary: The paper proposes a new non-Markovian sampling method based on Gaussian kernels, log-concave conditional densities, and empirical Bayes that can tunnel between modes and is compared with existing Langevin MCMC algorithms.
https://openreview.net/forum?id=ygxTuVz9eU
Compressor summary: VDC uses multimodal language models to detect and clean dirty data in AI systems by aligning and reasoning across images and labels.
https://openreview.net/forum?id=yeeVBMDAwy
Compressor summary: VPL is a framework that reduces label degeneracy by using variance-enlarged regularization and improves unsupervised learning on graphs.
https://openreview.net/forum?id=ycv2z8TYur
Compressor summary: EmerNeRF is a method that learns spatial-temporal representations of dynamic scenes using neural fields, achieving state-of-the-art performance in sensor simulation and addressing positional bias in Transformers.
https://openreview.net/forum?id=ycF7mKfVGO
Compressor summary: SharpeRatio@k is a new metric that measures the risk-return tradeoff and efficiency of policy portfolios formed by an OPE estimator under varying online evaluation budgets, aiming to improve the selection of policies for online deployment.
https://openreview.net/forum?id=yarUvgEXq3
Compressor summary: The study develops a scalable and effective recommender system that ensures high quality for less-satisfied users by minimizing the risk over their loss.
https://openreview.net/forum?id=yVJd8lKyVX
Compressor summary: The Hybrid Sharing Query (HSQ) is a transformer-based model that uses a mixture-of-experts architecture to improve image multi-label classification by leveraging label correlations and mitigating heterogeneity.
https://openreview.net/forum?id=yV6wwEbtkR
Compressor summary: The paper proposes MCMI, a new method to estimate BCPD for KD that enhances the teacher's ability to capture contextual information and boosts the student's performance in zero-shot and few-shot scenarios.
https://openreview.net/forum?id=yV6fD7LYkF
Compressor summary: The paper proposes an evaluation framework to systematically assess uncertainty methods for semantic segmentation applications, addressing key research questions and revealing insights on data and model uncertainties.
https://openreview.net/forum?id=yTBXeXdbMf
Compressor summary: This study develops a new theoretical framework for preference-based reinforcement learning that reduces human feedback and can handle linear and low-rank MDPs, as well as action-based comparison feedback.
https://openreview.net/forum?id=ySS7hH1smL
Compressor summary: The authors propose Lingual-SMoE, a multilingual machine translation framework that adapts routing policies based on linguistic features and difficulty levels to improve performance over existing methods.
https://openreview.net/forum?id=yRrPfKyJQ2
Compressor summary: ChatDrug is a framework using LLMs to systematically investigate and improve drug editing tasks, achieving the best performance on all 39 tasks and providing diverse, valid suggestions and explanations.
https://openreview.net/forum?id=yQDFsuG9HP
Compressor summary: The paper proposes a new method to generate realistic human animation from limited input using bidirectional temporal modeling and denoising diffusion models.
https://openreview.net/forum?id=yN4Wv17ss3
Compressor summary: This paper analyzes how supervised pretraining helps transformer models perform in-context reinforcement learning, and shows they can approximate near-optimal algorithms for certain tasks.
https://openreview.net/forum?id=yLClGs770I
Compressor summary: MAmmoTH is a new open-source language model that excels at solving various math problems using a unique combination of chain-of-thought and program-of-thought rationales, achieving significant accuracy gains over existing models.
https://openreview.net/forum?id=yKksu38BpM
Compressor summary: This paper explores how to use simpler ML algorithms to analyze and explain neural networks, focusing on new methods for approximating the empirical neural tangent kernel (eNTK) that balance time, memory, and accuracy.
https://openreview.net/forum?id=yC2waD70Vj
Compressor summary: The paper proves that nonlinear RNNs can approximate certain sequences well if they have an exponential decaying memory structure, and proposes a reparameterization method to address the limitations of RNNs for long-term memory.
https://openreview.net/forum?id=yBIJRIYTqa
Compressor summary: The paper proposes a generalization of the bandits with knapsacks framework that allows resource replenishment and shows its effectiveness in online learning and economic applications.
https://openreview.net/forum?id=y886UXPEZ0
Compressor summary: The authors propose a method to transform raw corpora into reading comprehension texts, which improves large language models' performance on various tasks and domains without requiring domain-specific training.
https://openreview.net/forum?id=y33lDRBgWI
Compressor summary: The paper proposes AdjointDPM, a method to optimize diffusion models with less memory by using probability-flow ODEs and the adjoint sensitivity method for efficient gradient backpropagation.
https://openreview.net/forum?id=y21ZO6M86t
Compressor summary: PolyGCL is a graph contrastive learning technique that uses polynomial filters to capture high-pass information for heterophily, improving performance on both homophilic and heterophilic graphs.
https://openreview.net/forum?id=y0GJXRungR
Compressor summary: This paper analyzes how well large language models like Code Llama, GPT-3.5, and GPT-4 can fix their own mistakes in code generation tasks and finds that self-repair is often limited by the model's feedback ability.
https://openreview.net/forum?id=y01KGvd9Bw
Compressor summary: DreamLLM is a framework that uses direct sampling and interleaved document generation to create versatile multimodal models with better comprehension and creation capabilities.
https://openreview.net/forum?id=xyxU99Nutg
Compressor summary: UnMix-TNS is a novel strategy to adapt batch normalization parameters online using mixed statistics components, improving the robustness of test-time adaptation methods under non-i.i.d. conditions.
https://openreview.net/forum?id=xxaEhwC1I4
Compressor summary: This paper investigates the optimal convergence rate of SGD's last iterate for various optimization scenarios without restrictive assumptions, providing a unified proof method.
https://openreview.net/forum?id=xx0ITyHp3u
Compressor summary: Sparse Model Soups improves neural network compression by combining sparse models with shared connectivity, enhancing generalization and out-of-distribution performance.
https://openreview.net/forum?id=xwKt6bUkXj
Compressor summary: The text discusses how different mechanisms in recurrent neural networks help them solve memory-dependent tasks with varying memory requirements, and how a multi-head curriculum can improve their performance and stability.
https://openreview.net/forum?id=xw5nxFWMlo
Compressor summary: The paper compares two methods to enhance large language models (LLMs) for tasks like question answering and summarization: extending their context window or adding retrieval, and shows that retrieval can improve performance regardless of the context window size.
https://openreview.net/forum?id=xw29VvOMmU
Compressor summary: Our method adapts pretrained language models efficiently by decomposing matrices into high-precision and memory-efficient parts, updating the latter during finetuning.
https://openreview.net/forum?id=xv8iGxENyI
Compressor summary: This paper proposes a hybrid adversarial attack (HART) on spiking neural networks (SNNs) using rate and temporal information, which outperforms previous methods in various scenarios.
https://openreview.net/forum?id=xuY33XhEGR
Compressor summary: ClimODE is a novel deep learning model that incorporates physics-based advection and uncertainty estimation to improve climate prediction.
https://openreview.net/forum?id=xtOydkE1Ku
Compressor summary: The paper presents a new model for multivariate time series prediction using transformer-based attentional copulas, which improves performance and training dynamics while handling various tasks and data characteristics.
https://openreview.net/forum?id=xt9Bu66rqv
Compressor summary: The paper proposes dual RL framework to unify offline imitation learning and reinforcement learning, and improves their performance with new methods ReCOIL and $f$-DVL.
https://openreview.net/forum?id=xsd2llWYSA
Compressor summary: The text introduces a method that learns motion dynamics from sparse data using self-supervised representation and generation, enabling online tracking of various motions with safe action execution.
https://openreview.net/forum?id=xriGRsoAza
Compressor summary: MILLET is a framework that uses Multiple Instance Learning to make Time Series Classification models more interpretable without sacrificing performance.
https://openreview.net/forum?id=xpw7V0P136
Compressor summary: SynTra reduces hallucination in large language models by optimizing their system messages on a synthetic task before applying them to real-world summarization tasks.
https://openreview.net/forum?id=xnhvVtZtLD
Compressor summary: ROAD is a new algorithm that improves local fairness in machine learning by using distributionally robust optimization and adversarial learning.
https://openreview.net/forum?id=xmQMz9OPF5
Compressor summary: The paper proposes a new way to train masked autoencoders without carefully choosing target representations, using a randomly initialized model as the teacher and achieving better results on downstream tasks.
https://openreview.net/forum?id=xkXdE81mOK
Compressor summary: FedRAP is a new federated learning model for recommendations that learns global and personalized item embeddings, improving performance and reducing communication costs.
https://openreview.net/forum?id=xhEN0kJh4q
Compressor summary: The authors propose a new method using a novel VAE to overcome the problem of separation in protein design optimization, which leads to improved sampling and generalization across discrete and continuous spaces.
https://openreview.net/forum?id=xhCZD9hiiA
Compressor summary: This study proposes a batch-normalized MLP with linear activations and batch normalization that has bounded gradients and optimal signal propagation.
https://openreview.net/forum?id=xcMmebCT7s
Compressor summary: The work introduces PPIRef, a large 3D protein-protein interaction dataset, and PPIformer, a new SE(3)-equivariant model that predicts mutation effects on protein-protein interactions and generalizes well across diverse scenarios.
https://openreview.net/forum?id=xbjSwwrQOe
Compressor summary: RLHF alternatives like SLiC and DPO have limitations in using preference data; RSO overcomes these by sourcing data from the optimal policy, and a unified framework improves loss functions for both methods.
https://openreview.net/forum?id=xZDWO0oejD
Compressor summary: PASTA is a method that allows large language models to read text with user-specified emphasis marks, enhancing their ability to follow instructions or integrate new knowledge.
https://openreview.net/forum?id=xUzWmFdglP
Compressor summary: The paper proposes MMCC, a new algorithm that analyzes privacy amplification via sampling for any matrix mechanism, and shows its utility in improving privacy/utility trade-offs for DP-FTRL style algorithms.
https://openreview.net/forum?id=xUO1HXz4an
Compressor summary: NegLabel is a novel post hoc out-of-distribution (OOD) detection method for vision-language models that uses negative labels from corpus databases and achieves state-of-the-art performance with good generalization and robustness.
https://openreview.net/forum?id=xJbsmB8UMx
Compressor summary: The paper proposes SALMON, a novel alignment paradigm for large language models that uses human-defined principles to control the behavior of an AI assistant without relying on extensive human feedback.
https://openreview.net/forum?id=xJEd8PkdNz
Compressor summary: The choice of the quadrature rule in integral reinforcement learning affects control performance due to computational errors, which can be reduced by using Bayesian quadrature with an appropriate kernel function.
https://openreview.net/forum?id=xJ5N8qrEPl
Compressor summary: The paper proposes a new Hessian-free algorithm for solving constrained Bi-Level Optimization problems in machine learning, which uses a smooth proximal Lagrangian value function to handle constraints and performs well in practice.
https://openreview.net/forum?id=xIHi5nxu9P
Compressor summary: Squaring subtractive mixtures in probabilistic circuits allows for better modeling of complex distributions with fewer components.
https://openreview.net/forum?id=xI4yNlkaqh
Compressor summary: 3D-MoLM is a language model that can comprehend 3D molecular structures by integrating a 3D molecular encoder, improving its performance in biomolecular tasks.
https://openreview.net/forum?id=xHmCdSArUC
Compressor summary: The paper proposes a method to improve privacy in machine learning by introducing correlations in the noise, providing analytical bounds for its utility, and demonstrating its effectiveness in experiments on deep learning.
https://openreview.net/forum?id=xGvPKAiOhq
Compressor summary: This paper shows how over-parameterization affects gradient descent for matrix sensing problems, leading to faster convergence rates with an asymmetric parameterization and a novel method to remove dependency on initialization scale.
https://openreview.net/forum?id=xEJMoj1SpX
Compressor summary: The paper investigates the exposure bias problem in diffusion models and proposes Epsilon Scaling, a training-free method that improves generative performance across various frameworks and settings.
https://openreview.net/forum?id=xCRr9DrolJ
Compressor summary: The paper proposes a faster way to learn policies from diffusion models using deterministic inference, improving both speed and performance on offline reinforcement learning tasks.
https://openreview.net/forum?id=xC8xh2RSs2
Compressor summary: Key findings from analyzing 7,433 dataset documentation cards on Hugging Face reveal the importance of completing all sections, prioritizing description and structure, and improving accessibility and reproducibility.
https://openreview.net/forum?id=xBfQZWeDRH
Compressor summary: GeoDiffusion is a framework that uses text prompts to generate high-quality detection data from pre-trained text-to-image diffusion models, encoding geometric conditions like bounding boxes and camera views.
https://openreview.net/forum?id=xAqcJ9XoTf
Compressor summary: SPE is a new method for graph neural networks that uses eigenvalues to softly partition eigenvectors, improving stability and expressivity over existing Laplacian-based positional encodings.
https://openreview.net/forum?id=x8VNtpCu1I
Compressor summary: This paper explores how encoder-only BERT models can perform instruction following and zero-shot learning using XML-R and shows that they achieve competitive results while being faster than autoregressive models.
https://openreview.net/forum?id=x7d1qXEn1e
Compressor summary: The authors propose a method to improve image restoration by using pre-trained neural networks as implicit priors, achieving state-of-the-art performance.
https://openreview.net/forum?id=x6u2BQ7xcq
Compressor summary: Tag2Text is a framework that uses image tags from paired text to improve vision-language models' performance on various tasks without extensive annotations.
https://openreview.net/forum?id=x5txICnnjC
Compressor summary: The assumption of Euclidean geometry in studying synaptic plasticity may be incorrect, and experiments could reveal the true geometry by comparing weight distributions before and after learning.
https://openreview.net/forum?id=x5LvBK43wg
Compressor summary: PROGRAM is a novel test-time adaptation method that uses a prototype graph model for reliable pseudo-label generation and robust self-training for target domain adaptation with noisy pseudo-labels.
https://openreview.net/forum?id=x4OPJ7lHVU
Compressor summary: DP-ICL is a method to protect sensitive information in large language models by generating noisy responses from multiple models.
https://openreview.net/forum?id=x1ptaXpOYa
Compressor summary: ADoPD is a new dataset for analyzing documents that combines large-scale data, human expertise, and diversity to improve document understanding.
https://openreview.net/forum?id=wxJ0eXwwda
Compressor summary: Alignment tuning mostly changes the language style of AI assistants, so an alternative method called Urial can achieve effective alignment with only a few stylistic examples.
https://openreview.net/forum?id=wsWGcw6qKD
Compressor summary: The paper explores how to train teachers for knowledge distillation using empirical risk minimization (ERM) and proposes a method called SoTeacher that improves student accuracy.
https://openreview.net/forum?id=wsRXwlwx4w
Compressor summary: CoPrompt is a new fine-tuning method for vision-language models that uses consistency constraints, perturbed inputs, and adapters/prompts to improve generalization on few-shot tasks.
https://openreview.net/forum?id=wriKDQqiOQ
Compressor summary: This paper studies the effect of batch size on Byzantine-robust distributed learning and proposes a new method (ByzSGDnm) that increases model accuracy under attacks by using larger batch sizes and normalized momentum.
https://openreview.net/forum?id=wprSv7ichW
Compressor summary: The paper proposes a new Federated Domain Generalization benchmark and evaluates 13 methods on it, finding significant performance gaps in this challenging setting.
https://openreview.net/forum?id=wpnlc2ONu0
Compressor summary: The paper proposes EIHL, an algorithm that adjusts network connectivity using excitation-inhibition and switches between local and global learning for better spiking neural network training performance and sparsity.
https://openreview.net/forum?id=wpXGPCBOTX
Compressor summary: This paper studies inverse optimal transport, which infers the ground cost from samples, and derives a sufficient condition for sparse recovery of the cost using $\ell_1$ regularization.
https://openreview.net/forum?id=wmX0CqFSd7
Compressor summary: Inverse design can be improved by optimizing over the learned energy function of a diffusion model, enabling compositional system design and avoiding adversarial examples.
https://openreview.net/forum?id=wm4WlHoXpC
Compressor summary: The authors propose a novel unified representation for generating high-fidelity crystal structures from complex chemical systems using a diffusion probabilistic model and introduce new metrics to evaluate the quality of generated materials for downstream applications like discovering novel stable materials.
https://openreview.net/forum?id=wkbeqr5XhC
Compressor summary: The authors propose a novel approach using a Vision Transformer variant with a learnable under-sampling mask to reduce hyperspectral data acquisition volume and enhance real-time detection applications.
https://openreview.net/forum?id=wg8NPfeMF9
Compressor summary: The proposed NAISR method represents 3D shapes by deforming an atlas using disentangled covariates for interpretable shape discovery and transfer.
https://openreview.net/forum?id=wfzXa8e783
Compressor summary: LyCORIS is an open-source library that helps fine-tune Stable Diffusion, a text-to-image generator, by offering various methodologies and assessing their effects with different metrics and prompts.
https://openreview.net/forum?id=wfgZc3IMqo
Compressor summary: The authors propose a unified method that combines loss reweighting and label correction to improve robustness against label noise in classification tasks using a shifted Gaussian model.
https://openreview.net/forum?id=wcka3bd7P4
Compressor summary: FROND is a graph neural network model that uses fractional calculus to capture long-term memories and mitigate over-smoothing, improving graph representation learning.
https://openreview.net/forum?id=wcaE4Dfgt8
Compressor summary: Uni3D is a 3D foundation model that leverages 2D pretrained models and image-text aligned models to create scalable 3D representations, achieving state-of-the-art results on various 3D tasks.
https://openreview.net/forum?id=wZXlEFO3tZ
Compressor summary: The paper proposes a method to estimate counterfactual distributions for causal effects using kernel Stein discrepancies, which is consistent and asymptotically normal under certain conditions.
https://openreview.net/forum?id=wZWTHU7AsQ
Compressor summary: The text introduces GRAD, a game-theoretic method for robust reinforcement learning that handles temporally-coupled perturbations, outperforming prior methods in experiments on continuous control tasks.
https://openreview.net/forum?id=wYvuY60SdD
Compressor summary: Mowst is a method that combines a lightweight MLP and a GNN to classify nodes on graphs, with a confidence mechanism that controls when to use the GNN for better performance.
https://openreview.net/forum?id=wYmvN3sQpG
Compressor summary: The study shows that neural networks trained with large learning rates and oscillations can learn weak features better than those trained with small learning rates, leading to improved generalization.
https://openreview.net/forum?id=wXpSidPpc5
Compressor summary: CLEX is a novel method that extends the context window of LLMs by using ordinary differential equations to model position embedding scaling, enabling better performance in long-context applications without sacrificing accuracy or increasing latency.
https://openreview.net/forum?id=wXWfvSpYHh
Compressor summary: MVSFormer++ is a transformer-based model with attention mechanisms that enhances different components of the MVS pipeline, achieving state-of-the-art results on challenging benchmarks.
https://openreview.net/forum?id=wR9qVlPh0P
Compressor summary: AutoVP is an end-to-end framework for automating visual prompting design choices and a benchmark for evaluating its performance on 12 image-classification tasks, significantly improving accuracy over existing methods.
https://openreview.net/forum?id=wPhbtwlCDa
Compressor summary: The paper proposes a new class of pseudometrics called STARC metrics for comparing reward functions in reinforcement learning, which can help improve theoretical and empirical evaluation of reward learning algorithms.
https://openreview.net/forum?id=wOSYMHfENq
Compressor summary: The text explains why normalization layers like batch normalization are important for deep learning and shows how using enough random features can achieve universal function approximation in various network architectures.
https://openreview.net/forum?id=wISvONp3Kq
Compressor summary: The paper proposes an algorithm for training sparse generalized linear models online with varying observations, using a bilevel optimization framework and adaptive regularization parameters.
https://openreview.net/forum?id=wHLDHRkmEu
Compressor summary: BarLeRIa is a novel Parameter-Efficient Tuning framework for dense prediction tasks in computer vision that uses bi-directional intertwined vision language adapters and outperforms prior methods by 5.6%.
https://openreview.net/forum?id=wHBfxhZu1u
Compressor summary: YaRN extends the context window of language models using RoPE embeddings, improving generalization and extrapolation beyond the original pre-training.
https://openreview.net/forum?id=wG12xUSqrI
Compressor summary: The paper analyzes how score-based generative models learn sub-Gaussian distributions and proves their approximation accuracy with respect to the target distribution.
https://openreview.net/forum?id=wFf9m4v7oC
Compressor summary: The paper presents a novel method for estimating causal effects from observational data with unobserved confounding variables, using conditional front-door adjustment and deep generative models.
https://openreview.net/forum?id=wFWuX1Fhtj
Compressor summary: The paper analyzes the limitations of primal-dual algorithms for constrained cooperative multi-agent reinforcement learning due to nonconvex constraints and proposes a decentralized primal approach with its own challenges.
https://openreview.net/forum?id=w9tc699w3Z
Compressor summary: The authors propose a method to train VLMs for remote-sensing images without textual annotations by using co-located internet imagery as an intermediary and show that their approach performs better than supervised methods on various tasks.
https://openreview.net/forum?id=w7LU2s14kE
Compressor summary: Transformer language models use linear relations to approximate some of their knowledge, which can be extracted by approximating the model with a single prompt.
https://openreview.net/forum?id=w7BwaDHppp
Compressor summary: The paper proposes a new mapping function for neural radiance fields (NeRFs) to generate novel views of unbounded scenes by adaptively sampling rays based on scene geometry.
https://openreview.net/forum?id=w50MQ9Vfty
Compressor summary: The article proposes a new experimental design for causal inference on networks with interference, which improves performance by controlling interference exposure.
https://openreview.net/forum?id=w4rODxXsmM
Compressor summary: The paper proposes RFCL, a method to improve reinforcement learning with offline data by using a reverse curriculum (generated via state resets) followed by a forward curriculum to efficiently leverage multiple demonstrations and solve complex tasks from sparse rewards.
https://openreview.net/forum?id=w4abltTZ2f
Compressor summary: FLORA is a framework that improves LoRA's ability to handle multiple task-specific adaptations efficiently, enabling fine-tuning of foundation models for diverse requests.
https://openreview.net/forum?id=w4DW6qkRmt
Compressor summary: SuRe is a framework that enhances open-domain question answering with large language models by generating summaries of retrieved passages and using them to select the most plausible answer.
https://openreview.net/forum?id=w3YZ9MSlBu
Compressor summary: The paper proposes a self-supervised learning framework for music audio, using teacher models to provide pseudo labels in acoustic pre-training, achieving state-of-the-art results on 14 music understanding tasks.
https://openreview.net/forum?id=w1JanwReU6
Compressor summary: The study examines how large language models show gender biases in neutral sentences without obvious word-gender associations, and finds high levels of bias across different models.
https://openreview.net/forum?id=vy42bYs1Wo
Compressor summary: Our method improves safety and constraint satisfaction in primal-dual safe RL by reducing cost estimation uncertainty with conservative policy optimization and local policy convexification.
https://openreview.net/forum?id=vuK8MhVtuu
Compressor summary: The paper proposes a new data-driven path planning model called GDP, which learns path patterns from road networks and outperforms traditional search-based methods.
https://openreview.net/forum?id=vtyasLn4RM
Compressor summary: The paper presents a scalable Graph Neural Network (GNN) based Graph Drawing framework that learns to optimize stress using coarsening and positional rewiring techniques, achieving state-of-the-art performance.
https://openreview.net/forum?id=vt5mnLVIVo
Compressor summary: Grokking occurs when a neural network initially tries to fit a simple solution with its initial features before transitioning to feature learning and eventually generalizing.
https://openreview.net/forum?id=vrE2fqAInO
Compressor summary: The paper studies how to find the best arm in a bandit problem with a limited budget and privacy constraints, using a policy based on maximum absolute determinants and deriving upper and lower bounds on the error probability.
https://openreview.net/forum?id=vrBVFXwAmi
Compressor summary: Q-TAPE is a pre-trained deep learning model that learns from diverse quantum systems in an unsupervised way, improving performance and reducing costs for quantum many-body problems.
https://openreview.net/forum?id=vqIH0ObdqL
Compressor summary: The paper introduces Corr2Cause, a new task to test causal inference skills of large language models, and shows their poor performance and lack of generalization on it.
https://openreview.net/forum?id=vpV7fOFQy4
Compressor summary: The paper compares three offline RL algorithms (CQL, BC, and DT) across various data settings and tasks to determine their strengths and weaknesses.
https://openreview.net/forum?id=vpJMJerXHU
Compressor summary: The paper proposes ModernTCN, a modified version of traditional TCN that improves performance and efficiency in time series analysis by increasing effective receptive fields.
https://openreview.net/forum?id=vngVydDWft
Compressor summary: The paper introduces a method to incorporate invariances into neural representations, improving latent similarity and downstream performance across different tasks and modalities.
https://openreview.net/forum?id=vkkHqoerLV
Compressor summary: The Alice benchmarks provide large-scale datasets and evaluation protocols for object re-ID tasks using synthetic data, aiming to train effective models for real-world applications.
https://openreview.net/forum?id=viftsX50Rt
Compressor summary: The paper introduces a new algorithm for fast and scalable estimation of graph kernel functions using random walks and neural networks.
https://openreview.net/forum?id=vfzRRjumpX
Compressor summary: This paper proposes a two-stage pretraining scheme for code representation learning using a large amount of code data, which improves performance on various downstream tasks.
https://openreview.net/forum?id=vePdNU3u6n
Compressor summary: CEMA is a paradigm that enables online adaptation of cloud models using limited computation and data transmission by excluding unreliable and low-informative samples and distilling normalization layers.
https://openreview.net/forum?id=vZfi5to2Xl
Compressor summary: SAS improves accuracy in neural networks by using sparse activation with a specific structure, reducing computation and leveraging NVIDIA's SparseTensorCore.
https://openreview.net/forum?id=vZZ4hhniJU
Compressor summary: The paper proposes a contrastive learning approach to improve communication in multi-agent RL by using messages as encodings that capture global state information.
https://openreview.net/forum?id=vZ6r9GMT1n
Compressor summary: The paper proposes a simple and lightweight defense against black-box attacks on deep neural networks by adding random noise to intermediate layers, enhancing resilience without adversarial training or accuracy loss.
https://openreview.net/forum?id=vYhglxSj8j
Compressor summary: CodeChain is a framework that improves the modularity and correctness of code generated by large language models by guiding them to reuse previously developed sub-modules through a chain of self-revisions.
https://openreview.net/forum?id=vY9nzQmQBw
Compressor summary: Vocos is a new neural vocoding model that generates speech from spectral coefficients using fast Fourier-based time-frequency representations, achieving high audio quality and significant computational efficiency.
https://openreview.net/forum?id=vXxardq6db
Compressor summary: SliceGPT is a post-training sparsification technique that reduces the embedding dimension of large language models, enabling faster inference with fewer GPUs and minimal accuracy loss.
https://openreview.net/forum?id=vW1SkPl4kp
Compressor summary: The paper introduces a risk-sensitive reinforcement learning framework that uses Iterated CVaR objective, human feedback, and provides sample-efficient algorithms with theoretical analysis and lower bound proof.
https://openreview.net/forum?id=vSwu81S33z
Compressor summary: Nonparametric transfer learning (NPTL) is a flexible posterior sampling method for deep neural networks that handles distribution shifts between upstream and downstream data better than existing methods.
https://openreview.net/forum?id=vSh5ePa0ph
Compressor summary: The paper studies how pretraining a simple linear attention model for linear regression with a Gaussian prior enables it to solve unseen tasks using in-context learning, showing that effective pretraining requires only a few tasks and that the model closely matches the Bayes optimal algorithm.
https://openreview.net/forum?id=vRyp2dhEQp
Compressor summary: The paper proposes realistic data-constrained backdoor attacks on DNNs, where attackers can't access all training data, and introduces CLIP-based techniques to improve efficiency in injecting malicious features.
https://openreview.net/forum?id=vNiI3aGcE6
Compressor summary: The paper proposes a memory-efficient model-free self-play algorithm for two-player zero-sum Markov games in multi-agent reinforcement learning, achieving low space and computational complexity, and improving sample and burn-in cost compared to existing methods.
https://openreview.net/forum?id=vN9fpfqoP1
Compressor summary: The authors propose using fine-tuned large language models to generate novel materials with favorable properties for engineering by sampling from text-encoded atomistic data, which is simple, reliable, and captures crystal symmetries better than diffusion models.
https://openreview.net/forum?id=vLJcd43U7a
Compressor summary: SYMBOL is a novel meta-learning framework that uses symbolic equation learning to automate the discovery of black-box optimizers, surpassing state-of-the-art baselines and showing exceptional zero-shot generalization.
https://openreview.net/forum?id=vKViCoKGcB
Compressor summary: The paper explores data attribution methods for diffusion models and finds that theoretically unjustified design choices perform better than previous baselines in attributing diffusion models' outputs to training data.
https://openreview.net/forum?id=vI95kcLAoU
Compressor summary: The paper proposes SkipAT, a method to reuse self-attention computations across layers in vision transformers, improving efficiency and performance.
https://openreview.net/forum?id=vEfmVS5ywF
Compressor summary: The text discusses how employers can strategically manipulate hiring policies to improve outcomes, but also shows that these policies may have unintended negative consequences in certain situations.
https://openreview.net/forum?id=vESNKdEMGp
Compressor summary: The study shows that large language models can produce harmful content when prompted with non-English or jailbreak instructions, especially for low-resource languages, and proposes a self-defense framework to mitigate these risks.
https://openreview.net/forum?id=vE5MyzpP92
Compressor summary: The paper proposes a new metric and loss function to improve image retrieval performance and consistency across different classes and data distributions.
https://openreview.net/forum?id=vE1e1mLJ0U
Compressor summary: The Expressive Leaky Memory (ELM) neuron model, which uses memory-like hidden states and nonlinear integration of synaptic input, accurately mimics cortical neuron computations with under 10,000 parameters and performs well on temporal tasks.
https://openreview.net/forum?id=vBw8JGBJWj
Compressor summary: UnitigBin is a novel metagenomic binning tool that uses representation learning and single-copy marker genes to improve the accuracy of clustering genomic sequences from microbial communities.
https://openreview.net/forum?id=vBo7544jZx
Compressor summary: The PMI framework mimics human brain's memory system to enhance question-answering, relation reasoning, and image classification tasks using Transformers and CNN models.
https://openreview.net/forum?id=v8jdwkUNXb
Compressor summary: The paper proposes a fast and effective policy representation called consistency policy for different reinforcement learning settings, using an actor-critic algorithm with a generative model as a policy.
https://openreview.net/forum?id=v8L0pN6EOi
Compressor summary: The text compares outcome and process supervision methods for training reliable large language models and shows that process supervision with active learning performs better on MATH problems.
https://openreview.net/forum?id=v7ZPwoHU1j
Compressor summary: The paper introduces a novel clustering method that solves a nonnegative low-rank $K$-means problem using NMF, and shows its superior performance and guarantees.
https://openreview.net/forum?id=v63GWletn8
Compressor summary: The paper proposes an improved algorithm for sampling from logconcave distributions on a polytope with constraints, using efficient linear solvers, determinantal estimators, and randomized Taylor series.
https://openreview.net/forum?id=v3XXtxWKi6
Compressor summary: RLCD is a method that uses contrasting prompts to create preference pairs without human feedback, then trains a language model with reinforcement learning to follow alignment principles like harmlessness.
https://openreview.net/forum?id=v3K5TVP8kZ
Compressor summary: The authors propose using TikZ as an intermediate representation for scientific figures, introduce a large dataset (DaTikZ) of TikZ drawings with captions, and fine-tune LLaMA and CLiMA models on it, which outperform GPT-4 and Claude 2 in generating realistic and aligned figures.
https://openreview.net/forum?id=v1VvCWJAL8
Compressor summary: The paper proposes a method to estimate domain counterfactuals, which are what a sample would have looked like in a different environment, by making assumptions on invertibility and sparse interventions, and shows that this approach improves counterfactual estimation over existing methods.
https://openreview.net/forum?id=v0zNCwwkaV
Compressor summary: The text introduces a generalized attention scheme that captures triple-wise correlations using tensors and shows its advantages and limitations in terms of efficiency and expressivity.
https://openreview.net/forum?id=uznKlCpWjV
Compressor summary: This paper analyzes the theoretical convergence properties of PPO-Clip, a popular reinforcement learning algorithm that uses a clipped surrogate objective function.
https://openreview.net/forum?id=uz7d2N2zul
Compressor summary: The paper proposes a personalized coreset weighted federated learning method that reduces training time and improves accuracy by using representative data points instead of the entire client data.
https://openreview.net/forum?id=uwO71a8wET
Compressor summary: The paper proposes a new Bayesian method that can estimate both the treatment effect and its uncertainty in continuous time using neural equations, which can help doctors make better decisions in personalized medicine.
https://openreview.net/forum?id=uvXK8Xk9Jk
Compressor summary: The text discusses how using sparse activations in deep networks can improve efficiency, but shows that some natural sparsifiers have training instability issues and suggests clipping their magnitude to fix this.
https://openreview.net/forum?id=uvFhCUPjtI
Compressor summary: The Evolving Graph Fourier Transform (EFT) is a new method for capturing the evolving spectra of graphs over time, which is efficient, invertible, and effective for downstream tasks.
https://openreview.net/forum?id=usrChqw6yK
Compressor summary: DVDet is a new open-vocabulary detector that uses conditional context prompts and hierarchical textual descriptors to align visual embeddings with fine-grained text descriptions of object parts, achieving state-of-the-art performance.
https://openreview.net/forum?id=uqxBTcWRnj
Compressor summary: The paper presents a new framework for learning symbolic knowledge from images using a game-theoretic diffusion model and proposes two evaluation metrics that align with human judgments.
https://openreview.net/forum?id=up6hr4hIQH
Compressor summary: This paper proposes new fidelity measures for GNN explainability that resist distribution shifts and align better with the information-theoretic definition of explainability.
https://openreview.net/forum?id=uleDLeiaT3
Compressor summary: The paper presents GROOT, an AI agent that learns to follow reference videos as instructions in open-world environments, achieving competitive performance against human players and generalist agents in Minecraft SkillForge.
https://openreview.net/forum?id=ulaUJFd96G
Compressor summary: HOMER is a training-free scheme that splits long inputs into smaller units, fuses them at Transformer layers, and reduces token count to improve performance and memory efficiency of large language models.
https://openreview.net/forum?id=ukidfml68f
Compressor summary: The text proposes a new technique called Pixel-wise Gradient Clipping (PGC) that improves the quality of 3D object generation by controlling the gradients used in latent representation-based models.
https://openreview.net/forum?id=ueTdErd5Ib
Compressor summary: The paper proposes a novel method for solving uncertain optimization problems that combines learning with downstream optimization, discretizes the feasible region into subsets, and integrates individual approximations to prescribe decisions that minimize expected cost and protect against worst case scenarios.
https://openreview.net/forum?id=udO3k28bEw
Compressor summary: The paper explains how hard negative mining works in deep metric learning and shows its equivalence to an optimization problem, providing theoretical justification for its effectiveness in preventing network collapse.
https://openreview.net/forum?id=uZfjFyPAvn
Compressor summary: The paper explores how implicit neural representations using wavelets can better capture high-frequency features of signals in image processing.
https://openreview.net/forum?id=uXjfOmTiDt
Compressor summary: Embodied Active Defense (EAD) is a proactive strategy that uses recurrent feedback mechanisms and active vision to counter adversarial patches in 3D real-world settings, enhancing robustness without compromising accuracy.
https://openreview.net/forum?id=uWvKBCYh4S
Compressor summary: The paper proposes a new method called Mixture of LoRA Experts (MoLE) to effectively combine multiple LoRAs for fine-tuning pre-trained models, overcoming challenges in existing approaches and achieving better results in NLP and V\&L tasks.
https://openreview.net/forum?id=uWVC5FVidc
Compressor summary: This study shows that it is possible to integrate watermarks in large language models without affecting their output quality or performance, enabling responsible AI development.
https://openreview.net/forum?id=uREj4ZuGJE
Compressor summary: The In-context Autoencoder (ICAE) uses a large language model to compress long texts into compact memory slots for efficient response generation.
https://openreview.net/forum?id=uQBW7ELXfO
Compressor summary: UNSB is a new method for unpaired image-to-image translation using Schrödinger Bridge with improved discriminators and regularization.
https://openreview.net/forum?id=uNrFpDPMyo
Compressor summary: The study presents a method to reduce memory usage in generative inference for large language models by adaptively compressing the key-value cache based on attention module profiles.
https://openreview.net/forum?id=uNkKaD3MCs
Compressor summary: PALM is a distance-based OOD detection method that models each class with multiple prototypes to capture sample diversities and enhance OOD detection performance.
https://openreview.net/forum?id=uMAujpVi9m
Compressor summary: The paper proposes a novel pocket pretraining method that leverages knowledge from atomic protein structures and pretrained small molecule representations to improve biomedical applications involving pockets and ligands.
https://openreview.net/forum?id=uKB4cFNQFg
Compressor summary: The study presents BEND, a benchmark for evaluating DNA language models on realistic and biologically meaningful tasks related to genome annotation.
https://openreview.net/forum?id=uJVHygNeSZ
Compressor summary: The paper proposes a geometry-aware attention mechanism for 3D vision tasks that encodes the geometric structure of tokens, leading to improved learning efficiency and performance of transformer-based neural networks.
https://openreview.net/forum?id=uIKZSStON3
Compressor summary: The ICEE algorithm improves the efficiency and speed of in-context policy learning by optimizing exploration-exploitation trade-offs within a Transformer model without Bayesian inference.
https://openreview.net/forum?id=uH0FGECSEI
Compressor summary: EFlowNets improve GFlowNets for stochastic tasks, while AFlowNets enable strong performance in adversarial environments like two-player games.
https://openreview.net/forum?id=uGtfk2OphU
Compressor summary: The paper proposes a Selective Rationalization method that uses shortcuts to improve neural network explanations and provides data augmentation techniques to increase annotated rationales.
https://openreview.net/forum?id=uFbWHyTlPn
Compressor summary: The paper proposes a new differentially private algorithm for training deep learning models with sensitive data, which avoids constant bias from gradient clipping and achieves higher accuracy.
https://openreview.net/forum?id=uELjxVbrqG
Compressor summary: The paper proposes a feature recombination technique and a IIC constraint to improve FR performance by exploiting the feature representation space of different models.
https://openreview.net/forum?id=uDxeSZ1wdI
Compressor summary: The paper proposes a structured method for visual reinforcement learning that can handle multiple object manipulation and goal dependencies, enabling agents to generalize from 3 objects to over 10 objects tasks.
https://openreview.net/forum?id=u859gX7ADC
Compressor summary: ReCAT is a Transformer model that can learn hierarchical syntax without gold trees and use a novel CIO layer for better contextualization and inter-span communication.
https://openreview.net/forum?id=u7559ZMvwY
Compressor summary: AToP is a novel framework that combines random transforms and purifier model fine-tuning to enhance the robustness and generalization of deep neural networks against adversarial attacks.
https://openreview.net/forum?id=u6imHU4Ebu
Compressor summary: The paper presents LLaRP, a method to adapt large language models as generalizable policies for embodied visual tasks, achieving higher success rates than common baselines on unseen tasks and introducing a new benchmark, Language Rearrangement.
https://openreview.net/forum?id=u48tHG5f66
Compressor summary: The authors propose a method to generate high-resolution images from pre-trained diffusion models without training, addressing object repetition and unreasonable structures by adjusting the convolutional perception field with re-dilation and using dispersed convolution and noise-damped guidance.
https://openreview.net/forum?id=u3dX2CEIZb
Compressor summary: The paper proposes a scalable method to enforce hard physical constraints in neural networks using Mixture-of-Experts (MoE), which improves accuracy, convergence, and efficiency for complex dynamical systems.
https://openreview.net/forum?id=u3dHl287oB
Compressor summary: Our paper studies how overparameterization and task similarity in continual learning affect forgetting and presents an analytical expression for it.
https://openreview.net/forum?id=twSnZwiOIm
Compressor summary: The paper proposes an optimization problem over neural networks to learn a state representation for stochastic dynamical systems, which can be used for forecasting and interpreting system dynamics.
https://openreview.net/forum?id=tvhaxkMKAn
Compressor summary: RLHF-trained AI assistants often prioritize matching user beliefs over truthfulness, and this behavior is partly influenced by human preferences for sycophancy.
https://openreview.net/forum?id=tveiUXU2aa
Compressor summary: SWAP-Score is a novel training-free metric that measures network expressivity based on sample-wise activation patterns, outperforming existing metrics in Neural Architecture Search.
https://openreview.net/forum?id=tuzTN0eIO5
Compressor summary: The paper proposes a scheduling strategy to eliminate pipeline bubbles in distributed training, improving throughput and memory efficiency.
https://openreview.net/forum?id=ttXg3SKAg5
Compressor summary: The paper proposes a method to improve cross-modal learning from uni-modal data by bridging the modality gap in multi-modal contrastive space using Connect, Collapse, Corrupt steps, achieving state-of-the-art results on various tasks.
https://openreview.net/forum?id=tsE5HLYtYg
Compressor summary: SafeDreamer is a new algorithm that uses world models to improve safe reinforcement learning, achieving nearly zero-cost performance in various tasks while maintaining safety criteria.
https://openreview.net/forum?id=tr0KidwPLc
Compressor summary: This paper introduces LLMBAR, a benchmark to test how well large language models can evaluate instructions followed by generated text, revealing differences among evaluators and suggesting prompting strategies for improvement.
https://openreview.net/forum?id=tqh1zdXIra
Compressor summary: The paper presents a fast and accurate method to select and finetune pretrained models using a meta-dataset of performance data and a meta-learned hyperparameter predictor.
https://openreview.net/forum?id=tplXNcHZs1
Compressor summary: The paper introduces a new sampling algorithm for diffusion models that solves linear inverse problems without approximations and shows improved performance on tasks like image inpainting, super-resolution, and deblurring.
https://openreview.net/forum?id=tnBaiidobu
Compressor summary: The study examines how CLIP's out-of-distribution generalization is affected by training data similarity and pruning, finding that other factors besides data similarity drive CLIP's performance.
https://openreview.net/forum?id=tmsqb6WpLz
Compressor summary: Finetuning language models on domain-specific data affects their topic and style preferences more than their factual knowledge, and this study uses instruction-following LLMs to generate controlled text examples to analyze these effects.
https://openreview.net/forum?id=tmqOhBC4a5
Compressor summary: The paper proposes a unified framework for learning stochastic policies in cooperative MARL using probabilistic graphical models and maximum entropy objectives, leading to HASAC algorithm that improves sample efficiency, robustness, and exploration.
https://openreview.net/forum?id=tmBKIecDE9
Compressor summary: Motif is a method to use a large language model's knowledge to improve an agent's decision-making and performance in complex environments like NetHack game.
https://openreview.net/forum?id=tm8s3696Ox
Compressor summary: Co-Boosting is a novel framework that improves the quality of synthesized data and ensemble models in One-shot Federated Learning through an adversarial process.
https://openreview.net/forum?id=tjn2YZSHUv
Compressor summary: The paper introduces Social Reward, a reward modeling framework that uses implicit feedback from social network users to evaluate and improve text-to-image models, making AI-generated art more aligned with community preferences.
https://openreview.net/forum?id=tiiAzqi6Ol
Compressor summary: CPMs are a new framework for training Preference Models that decompose preference assessments into interpretable features, use a LM to score these features, and aggregate them with logistic regression, improving generalization and robustness over standard PMs.
https://openreview.net/forum?id=thbtoAkCe9
Compressor summary: The paper proposes a new coreset selection method, $\mathbb{D}^2$ Pruning, that uses graph-based message passing to balance data diversity and difficulty scores, improving model performance on various vision and NLP tasks.
https://openreview.net/forum?id=tbVWug9f2h
Compressor summary: The paper introduces MTOB, a benchmark for testing large language models' ability to learn low-resource languages from a single book of grammar explanations, finding that current models perform similarly to humans but not as well as desired.
https://openreview.net/forum?id=tbFBh3LMKi
Compressor summary: Uni-O4 is a reinforcement learning method that combines offline and online learning seamlessly for efficient and safe learning in real-world environments.
https://openreview.net/forum?id=ta26LtNq2r
Compressor summary: The paper proposes a novel approach for learning to reject classifiers that perform well under imbalanced label distributions and show its effectiveness through experiments on image classification tasks.
https://openreview.net/forum?id=tVTN7Zs0ml
Compressor summary: GraphCare uses external knowledge graphs and language models to generate patient-specific graphs and improve predictions from electronic health records.
https://openreview.net/forum?id=tVMPfEGT2w
Compressor summary: The paper proposes an algorithm for offline preference-based reinforcement learning using human feedback and shows a novel guarantee that allows learning any target policy with general function approximation, as long as it's covered by the offline data.
https://openreview.net/forum?id=tUtGjQEDd4
Compressor summary: The paper introduces a novel phase space-based generative model that leverages velocity information for efficient and realistic data generation, outperforming or rivaling existing diffusion models on image generation benchmarks.
https://openreview.net/forum?id=tUVG9nGzgE
Compressor summary: Non-commutative invariance learning improves domain adaptation by using source domain samples more effectively.
https://openreview.net/forum?id=tUM39YTRxH
Compressor summary: Text2Reward is a framework that uses large language models to generate dense reward functions for reinforcement learning tasks, achieving good performance on robotic manipulation and locomotion benchmarks.
https://openreview.net/forum?id=tPEwSYPtAC
Compressor summary: This paper connects sharpness and robustness to improve out-of-distribution generalization bounds for models.
https://openreview.net/forum?id=tOzCcDdH9O
Compressor summary: Matryoshka Diffusion Model (MDM) is a new framework for generating high-quality images and videos that denoises inputs at multiple resolutions jointly and trains end-to-end using a progressive training schedule, achieving strong zero-shot generalization.
https://openreview.net/forum?id=tMzPZTvz2H
Compressor summary: The paper explores how ResNet generalizes in the limit of infinitely wide and deep neural networks, using a time-variant Gram matrix and analyzing KL divergence.
https://openreview.net/forum?id=tGQirjzddO
Compressor summary: The paper proposes a novel approach that uses latent diffusion to model trajectory sequences as compressed latent skills for offline reinforcement learning, improving credit assignment and reward propagation.
https://openreview.net/forum?id=tEgrUrUuwA
Compressor summary: PMP is a new message passing method for Graph Fraud Detection that adapts to different label neighbors by using node-specific aggregation functions, improving GFD performance.
https://openreview.net/forum?id=tBROYsEz9G
Compressor summary: The paper proposes Constrained Deep Generative Models (C-DGMs) that enforce compliance with background knowledge using a Constraint Layer, improving utility and detection in tabular data generation tasks.
https://openreview.net/forum?id=t9dWHpGkPj
Compressor summary: The paper investigates language model inversion to recover hidden text using next-token probabilities and proposes a method that achieves high BLEU and F1 scores.
https://openreview.net/forum?id=t8eO0CiZJV
Compressor summary: MaRio is a method to improve small language models' ability to generate rationales for question answering by optimizing multiple properties such as plausibility, diversity, and consistency.
https://openreview.net/forum?id=t8cBsT9mcg
Compressor summary: The paper proposes a concept bottleneck model that uses soft concepts and concept confirmation to enhance selective classification in deep learning, improving both coverage and performance under uncertainty.
https://openreview.net/forum?id=t3vnnLeajU
Compressor summary: DA-CLIP is a degradation-aware vision-language model that adapts to different types of image corruptions and achieves state-of-the-art performance on image restoration tasks using cross-attention and a controller that predicts both feature embeddings and degradation features.
https://openreview.net/forum?id=t0FI3Q66K5
Compressor summary: The paper demonstrates that LLMs can effectively encode visual tasks without language, using a frozen transformer block from pre-trained models as an encoder, and proposes a hypothesis to explain this phenomenon.
https://openreview.net/forum?id=sxGugrYhP9
Compressor summary: BatteryML is an open-source platform that simplifies battery life prediction by providing data processing, feature extraction, and various models, while also establishing unified standards for the field.
https://openreview.net/forum?id=svIdLLZpsA
Compressor summary: The text discusses how synthetic training data can be improved for deep learning models by analyzing its principles from a distribution-matching perspective, leading to better performance in various image classification tasks and other applications.
https://openreview.net/forum?id=sqRgz88TM3
Compressor summary: The paper introduces VFLAIR, a framework for vertical federated learning that supports various models, datasets, and protocols, and evaluates 11 attacks and 8 defenses to provide insights for deploying VFL securely.
https://openreview.net/forum?id=spvaV5LELF
Compressor summary: The text introduces a new challenge dataset for testing neural models' STEM skills, requiring understanding of multimodal vision-language information, and reveals that current models perform poorly compared to elementary students.
https://openreview.net/forum?id=sojpn00o8z
Compressor summary: Cascaded models are multi-scale generative models that can also be good likelihood models when using hierarchical volume-preserving maps, leading to improved performance on density estimation, lossless compression, and out-of-distribution detection tasks.
https://openreview.net/forum?id=sn7CYWyavh
Compressor summary: The paper proposes a novel hierarchical language for generating whole-song pop music with high quality and controllability.
https://openreview.net/forum?id=smy4DsUbBo
Compressor summary: The paper proposes a higher-order graph neural network model that can quickly and accurately predict the stiffness of lattices, by using symmetries and energy conservation rules.
https://openreview.net/forum?id=sllU8vvsFF
Compressor summary: The Large Reconstruction Model (LRM) uses a transformer-based architecture with many learnable parameters to predict a neural radiance field from a single image and reconstruct 3D objects in high quality and generalize across various inputs.
https://openreview.net/forum?id=slSmYGc8ee
Compressor summary: This paper explores how the structure of the initial weights in neural networks affects their learning dynamics, finding that high-rank initializations lead to lazier learning while low-rank initializations promote richer learning, with some exceptions related to task and data statistics.
https://openreview.net/forum?id=skcTCdJz0f
Compressor summary: ProSMin is a probabilistic self-supervised learning method that uses two neural networks to enhance representation quality, mitigate dimensional collapse, and improve performance on various downstream tasks.
https://openreview.net/forum?id=shr9PXz7T0
Compressor summary: This paper shows that large language models are biased towards specific option IDs in multiple choice questions due to token bias, and proposes a debiasing method called PriDe that separates the prior bias from the prediction distribution.
https://openreview.net/forum?id=shgx0eqdw6
Compressor summary: ARGS is a novel framework that improves language model alignment with human preferences by adjusting probabilistic predictions using a reward signal during decoding, resulting in better average rewards and semantic diversity.
https://openreview.net/forum?id=sehRvaIPQQ
Compressor summary: CIPIR is a communication protocol that uses raw transformer output embeddings instead of natural language to enhance reasoning ability in large language models, outperforming existing methods by 1-3.5%.
https://openreview.net/forum?id=samyfu6G93
Compressor summary: NeuroBack is a novel approach that uses Graph Neural Networks to predict phases of variables in the SAT problem, enabling faster and more efficient solving on CPU without relying on GPU resources.
https://openreview.net/forum?id=sY5N0zY5Od
Compressor summary: DSPy is a programming model for developing and optimizing language models pipelines using text transformation graphs, enabling sophisticated tasks and outperforming hard-coded prompt templates and few-shot prompting.
https://openreview.net/forum?id=sVEu295o70
Compressor summary: The paper investigates how state-action coverage, reward density, and the number of augmented transitions per update affect data efficiency in reinforcement learning tasks with sparse rewards and dynamics-invisible data augmentation.
https://openreview.net/forum?id=sTYuRVrdK3
Compressor summary: The paper introduces a benchmark suite for evaluating protein structure representation learning methods, comparing various pre-training and downstream tasks to improve understanding of protein structures and functions.
https://openreview.net/forum?id=sSyytcewxe
Compressor summary: SEED is a novel approach for class-incremental learning that selects and fine-tunes the most optimal expert for each task using Gaussian distributions, improving diversity and stability.
https://openreview.net/forum?id=sSaN4gxuEf
Compressor summary: The paper proposes a method to adapt a model using few unlabeled data, by leveraging pre-trained backbones, source domains, and a domain prompt generator that guides the features towards the target domain.
https://openreview.net/forum?id=sPUrdFGepF
Compressor summary: Consistent4D is a method to generate 4D dynamic objects from monocular videos without multi-view data or camera calibration, using a cascade DyNeRF and interpolation-driven consistency loss.
https://openreview.net/forum?id=sP1tCl2QBk
Compressor summary: The paper proposes a differentiable Monte Carlo estimator that samples line segments instead of points for integrals with discontinuous integrands, enabling gradient-based optimization in applications like robotics and graphics.
https://openreview.net/forum?id=sOJriBlOFd
Compressor summary: NeRM is a generative model that uses implicit neural representations to synthesize realistic human motions with any framerate, while being memory-efficient and fast.
https://openreview.net/forum?id=sNFLN3itAd
Compressor summary: The text introduces a novel link prediction model, NCNC, that leverages structural features and graph completion to overcome graph incompleteness, outperforming existing models.
https://openreview.net/forum?id=sMoifbuxjB
Compressor summary: Intra-Fusion combines pruning and fusion methods in neural networks to achieve better compression, accuracy, and training time than standard pruning alone.
https://openreview.net/forum?id=sLregLuXpn
Compressor summary: The paper provides a robust theoretical framework to understand and optimize the role of Gaussian noise injection in image-to-image translation models for better performance in noisy environments.
https://openreview.net/forum?id=sLkj91HIZU
Compressor summary: This paper explores if transformers can learn to make optimal predictions for mixtures of regressions and shows they can achieve low error, sample efficiency, and robustness.
https://openreview.net/forum?id=sLdVl0q68X
Compressor summary: NuwaDynamics is a causal concept for spatio-temporal prediction that identifies important regions in data and enables models to reason causally, improving results on real-world tasks.
https://openreview.net/forum?id=sLQb8q0sUi
Compressor summary: Vertical federated Shapley value (VerFedSV) is a fair and efficient contribution valuation metric for data sources in vertical federated learning scenarios.
https://openreview.net/forum?id=sKPzAXoylB
Compressor summary: UPGD is a novel approach for continual learning of representations that combats catastrophic forgetting and loss of plasticity by applying different size modifications to useful and unuseful units.
https://openreview.net/forum?id=sJ88Wg5Bp5
Compressor summary: The paper proposes ViDA, a visual domain adapter for continual test-time adaptation, which uses high-rank and low-rank features to handle domain-specific and domain-shared knowledge and achieves state-of-the-art performance in classification and segmentation tasks.
https://openreview.net/forum?id=sHEJJmzBIN
Compressor summary: The authors propose a computationally efficient GAN approach for open domain text generation using Transformer models, which improves text quality and avoids sparsity problems, outperforming larger baseline models with fewer parameters.
https://openreview.net/forum?id=sGVmr7KHfn
Compressor summary: MemSPM is a novel method that mines sub-classes within categories to reduce feature gap and improve adaptation performance when there is significant concept shift between samples.
https://openreview.net/forum?id=sFyTZEqmUY
Compressor summary: The authors propose UniSim, a universal simulator for real-world interaction, which can be trained on diverse datasets and enable various intelligent applications.
https://openreview.net/forum?id=sBQwvucduK
Compressor summary: MagicDrive is a framework for generating street views with precise 3D control and high-fidelity synthesis, enabling better 3D perception tasks like object detection.
https://openreview.net/forum?id=s9z0HzWJJp
Compressor summary: SocioDojo is a lifelong learning environment for creating autonomous agents that can analyze and make decisions on societal topics using information sources, knowledge bases, time series, and analysis tools.
https://openreview.net/forum?id=s90VIdza2K
Compressor summary: The paper proposes a stochastic optimization framework for fair machine learning that works with small data batches, has theoretical convergence guarantees, and performs well under distribution shifts.
https://openreview.net/forum?id=s8cMuxI5gu
Compressor summary: This paper proposes an algorithm to recover augmented soft labels and input features from gradients in federated learning, considering realistic label smoothing and mixup techniques, and shows its effectiveness in attacking the system.
https://openreview.net/forum?id=s5hSp7EdL3
Compressor summary: The paper studies how learning from a strategic labeler who selectively abstains from labeling can affect the learning process and proposes a near-optimal algorithm for this scenario.
https://openreview.net/forum?id=s56xikpD92
Compressor summary: The paper proposes a defense method called BaDExpert that finetunes a backdoored DNN to recognize only its malicious inputs, thus detecting and filtering out the backdoors.
https://openreview.net/forum?id=s2NjWfaYdZ
Compressor summary: KPrune is a retraining-free algorithm that accurately compresses pretrained encoder-based language models by preserving useful knowledge and improving F1 score on the SQuAD benchmark.
https://openreview.net/forum?id=rzF0R6GOd4
Compressor summary: The paper proposes a method for 3D reconstruction of dynamic scenes from multi-view videos using Signed Distance Function flow, which handles topology changes and outperforms existing methods.
https://openreview.net/forum?id=rzBskAEmoc
Compressor summary: CAMIL is a new model for cancer diagnosis from tissue biopsy images that considers neighboring information and achieves better performance than existing methods.
https://openreview.net/forum?id=rxlF2Zv8x0
Compressor summary: The paper proposes a method, Gibbs sampling with Graph-based Smoothing (GGS), to smooth the fitness landscape and improve protein optimization using graph signals and energy-based models.
https://openreview.net/forum?id=rxVBKhyfSo
Compressor summary: SelMix is a technique to optimize pre-trained models for practical performance objectives by mixing features of samples from specific classes.
https://openreview.net/forum?id=rvUq3cxpDF
Compressor summary: LAPO is a method to train deep RL policies from action-free videos by inferring latent actions, which enables web data pre-training and fine-tuning.
https://openreview.net/forum?id=rvDQtdMnOl
Compressor summary: The paper proposes a new graph neural network method (LSR-MP) that improves the simulation of chemical and biological systems using ab initio molecular dynamics, achieving state-of-the-art results with reduced error.
https://openreview.net/forum?id=rulxyXjf46
Compressor summary: The paper proposes a method to use conformal prediction for regression by converting it to a classification problem with a new loss function and modified techniques.
https://openreview.net/forum?id=ruk0nyQPec
Compressor summary: SILO is a language model that uses a combination of a parametric model trained on public domain text and a nonparametric datastore with customizable data attribution to enable compliant use of copyrighted or restricted data during inference.
https://openreview.net/forum?id=ruGY8v10mK
Compressor summary: The paper proposes a new data-driven measure of relative uncertainty for detecting misclassified instances in machine learning models based on soft-predictions and shows its effectiveness in image classification tasks.
https://openreview.net/forum?id=rtx8B94JMS
Compressor summary: The paper presents a new method for inferring parameters in neural stochastic differential equations using fractional Brownian motion and variational methods, which can capture long-term dependencies and be applied to video prediction.
https://openreview.net/forum?id=rtl4XnJYBh
Compressor summary: This paper analyzes two mechanisms behind multimodal contrastive learning's success in robust representation learning, and shows how rich captions help improve zero-shot classification under distribution shift.
https://openreview.net/forum?id=rsg1mvUahT
Compressor summary: FedWad is a principled algorithm that computes Wasserstein distance between two distributions in a federated manner by manipulating and exchanging geodesics, with convergence guarantees and empirical benefits.
https://openreview.net/forum?id=rsY6J3ZaTF
Compressor summary: DistillSpec is a method that uses knowledge distillation to improve the alignment between a fast draft and a slow target language model for speculative decoding, achieving significant speedups and latency control.
https://openreview.net/forum?id=rpH9FcCEV6
Compressor summary: The paper proposes a query-based membership inference attack on diffusion models for image and audio generation tasks and analyzes their privacy robustness in the text-to-speech task.
https://openreview.net/forum?id=rp0EdI8X4e
Compressor summary: The authors propose a new concept called Faithful Vision-Language Concept (FVLC) models that improve interpretable machine learning by addressing the instability issue of label-free bottleneck models and maintaining good performance.
https://openreview.net/forum?id=rnHNDihrIT
Compressor summary: SORL is a novel method for offline RL that leverages policy diversity by alternating between policy learning and trajectory clustering.
https://openreview.net/forum?id=rmg0qMKYRQ
Compressor summary: The study compares text-to-image classifiers based on generative models with discriminative models and human perception, finding that generative models have impressive properties similar to human object recognition.
https://openreview.net/forum?id=rmXXKxQpOR
Compressor summary: This paper provides a theoretical understanding of why unsupervised pretraining helps in machine learning systems by studying a generic framework with latent variable models and empirical risk minimization.
https://openreview.net/forum?id=rkplYfqUr0
Compressor summary: Gen-Z is a generative framework for zero-shot text classification that uses natural language labels and improves robustness to prompt variations and allows personalization.
https://openreview.net/forum?id=riNuqYiD66
Compressor summary: The paper proposes a branching decoder for generating sets of text without order dependence, which improves effectiveness and efficiency over the existing sequential decoder.
https://openreview.net/forum?id=rhgIgTSSxW
Compressor summary: TabR is a new tabular deep learning model that uses a novel attention-like mechanism to retrieve and utilize nearest neighbors for better prediction, outperforming existing methods on several benchmarks.
https://openreview.net/forum?id=rhaQbS3K3R
Compressor summary: The authors study how well object recognition models perform on crowdsourced global datasets and find that progress on these datasets lags behind standard benchmarks and may increase geographic disparities in model performance.
https://openreview.net/forum?id=rUf9G9k2im
Compressor summary: The pixel spread model combines GANs' optimization efficiency with probabilistic models' prediction tractability to improve image inpainting quality and efficiency, achieving state-of-the-art results.
https://openreview.net/forum?id=rTDyN8yajn
Compressor summary: The paper introduces Octavius, a framework for studying multimodal learning with large language models, which combines MoE and LoRA to reduce interference and improve performance on various tasks.
https://openreview.net/forum?id=rTBL8OhdhH
Compressor summary: The authors propose a new algorithm for Dataset Distillation that matches the difficulty of the generated patterns with the size of the synthetic dataset, enabling lossless distillation for larger datasets than previous methods.
https://openreview.net/forum?id=rR03qFesqk
Compressor summary: FIRE is a new position encoding method for Transformers that improves their ability to handle longer inputs and generalizes better to different tasks.
https://openreview.net/forum?id=rM9VJPB20F
Compressor summary: The paper reveals that existing group robustness methods in machine learning unintentionally boost adversarial success by amplifying poison samples, and current poisoning defenses damage group robustness by eliminating minority samples along with poisons.
https://openreview.net/forum?id=rL7xsg1aRn
Compressor summary: The paper introduces MSG, a method to accelerate pre-training of large language models by growing the model in an optimal way with function-preserving operators that work regardless of weight initialization.
https://openreview.net/forum?id=rIx1YXVWZb
Compressor summary: The paper analyzes a simple Transformer model for integer addition, explaining its algorithms, execution, and rare failures, to help understand and ensure safe and ethical AI use.
https://openreview.net/forum?id=rINBD8jPoP
Compressor summary: This text describes a new quantum architecture search algorithm (CRLQAS) that improves simulation speed, explores the search space efficiently, and optimizes for shorter circuits in noisy environments.
https://openreview.net/forum?id=rHzapPnCgT
Compressor summary: This paper introduces Progressive Conditional Diffusion Models (PCDMs), which incrementally synthesize realistic person images with different poses by using three stages of conditional diffusion models.
https://openreview.net/forum?id=rGFrRMBbOq
Compressor summary: PFNR is a novel method to encode multiple complex videos using adaptive sub-modules in Fourier space, enabling efficient and scalable continual learning.
https://openreview.net/forum?id=rDH7dIFn20
Compressor summary: The paper proposes a new algorithm for contextual dueling bandits that accounts for uncertainty in pairwise comparisons and achieves a better regret bound than existing methods.
https://openreview.net/forum?id=rBH7x87VfJ
Compressor summary: The authors propose a framework to train overparameterized neural networks with gradient flow, showing that this method can achieve arbitrarily low loss and generalize common deep learning models by sparsifying their weight tensors.
https://openreview.net/forum?id=rAHcTCMaLc
Compressor summary: S$^2$AC is a MaxEnt RL algorithm using SVGD that computes entropy efficiently for EBM policies, achieving better performance than SQL and SAC in various tasks.
https://openreview.net/forum?id=r9FsiXZxZt
Compressor summary: The paper proposes an object-centric architecture that uses weak supervision from sparse perturbations to disentangle latent variables of multiple objects, addressing the limitations of existing causal representation learning methods.
https://openreview.net/forum?id=r65xfUb76p
Compressor summary: The paper presents a method to train smaller and more efficient named entity recognition (NER) models from large language models (LLMs) by using mission-focused instruction tuning, achieving state-of-the-art results without direct supervision.
https://openreview.net/forum?id=r5njV3BsuD
Compressor summary: We present new linear convergence bounds for denoising diffusion models with finite second moment assumptions and show how to extend Girsanov-based methods and use stochastic localization theory in the proof.
https://openreview.net/forum?id=r42tSSCHPh
Compressor summary: The text describes a new attack method that exploits language models' generation methods to cause malicious behavior, highlighting the need for better alignment and safety evaluation of open-source large language models.
https://openreview.net/forum?id=qz3mcn99cu
Compressor summary: The text describes a study that improves robustness of neural networks against adversarial attacks using Lipschitz-based methods and novel techniques, achieving significant performance gains.
https://openreview.net/forum?id=qxLVaYbsSI
Compressor summary: Twinsight is a novel FSSL method that trains two models simultaneously with different objectives to reduce gradient conflicts and improve performance.
https://openreview.net/forum?id=qxGXjWxabq
Compressor summary: The paper presents Canonpipe, a method for computing Shapley-based data importance over ML pipelines, which can efficiently discover data errors and sometimes perform better than existing methods.
https://openreview.net/forum?id=qwYKE3VB2h
Compressor summary: This paper studies how using graphs instead of hypergraphs to model interconnected systems can lose higher-order information, and proposes a learning method to recover this lost information from real-world datasets.
https://openreview.net/forum?id=qup9xD8mW4
Compressor summary: The paper introduces behaviour distillation and HaDES, a method to create small synthetic datasets for reinforcement learning tasks without expert data.
https://openreview.net/forum?id=qtE9K23ISq
Compressor summary: Batch Styles Standardization (BSS) is a Fourier-based method that standardizes image styles within a batch, enhancing domain-invariance in self-supervised learning models without requiring domain labels or specific architectures, and improving their performance on unseen domains.
https://openreview.net/forum?id=qr4ECbGcSj
Compressor summary: The text compares 17 reinforcement learning objective-specification formalisms, revealing their expressivity limitations and implications for policy optimisation and reward learning.
https://openreview.net/forum?id=qoYogklIPz
Compressor summary: The paper proposes a method to make embeddings more interpretable by using large language models to generate narratives from them, enabling querying and exploration of complex data in various tasks.
https://openreview.net/forum?id=qoHeuRAcSl
Compressor summary: This paper proposes a method to use large language models to guide learning structures and constraints in robot manipulation tasks, using mode families as an intermediate layer between high-level language and low-level physical trajectories.
https://openreview.net/forum?id=qo21ZlfNu6
Compressor summary: The authors propose a new data extraction attack called "neural phishing" that exploits large language models trained on private data to extract sensitive information with high success rates.
https://openreview.net/forum?id=qnWtw3l0jb
Compressor summary: Imitation learning using RL foundation models can quickly mimic expert behavior with few demonstrations, outperforming standard approaches in speed and efficiency.
https://openreview.net/forum?id=qmXedvwrT1
Compressor summary: LEGO bricks are a new adaptable network backbone for iterative refinement in diffusion models, enabling efficient generation of higher-resolution images with variable resolutions and faster convergence.
https://openreview.net/forum?id=qiduMcw3CU
Compressor summary: The text proposes a framework for agents to learn skill primitives and compose them flexibly to achieve temporal logic specifications in various environments.
https://openreview.net/forum?id=qi5Xa2cOZg
Compressor summary: LGA is a method that uses natural language to automatically create state abstractions for imitation learning, improving policy generalization and robustness.
https://openreview.net/forum?id=qhkEOCcVX9
Compressor summary: The study introduces a Newborn Embodied Turing Test (NETT) for object segmentation, comparing the learning abilities of newborn chicks and machine-learning algorithms in recognizing objects across different backgrounds and viewpoints.
https://openreview.net/forum?id=qgWJkDiI5p
Compressor summary: This paper investigates the Fast Equilibrium conjecture for normalization layers in deep learning and proves it under more general assumptions than previous works.
https://openreview.net/forum?id=qg5JENs0N4
Compressor summary: The paper studies how reinforcement learning algorithms can use past experience to solve new tasks without explicit stitching mechanisms, showing that this property depends on a form of generalization different from i.i.d. generalization, and proposes data augmentation as a way to improve it.
https://openreview.net/forum?id=qg2boc2AwU
Compressor summary: The paper presents EBMDock, a geometric deep learning framework that uses statistical potential and Langevin dynamics to sample protein-protein docking poses from a probability distribution, achieving superior results on two benchmarks.
https://openreview.net/forum?id=qe49ybvvPs
Compressor summary: Distributional RL algorithms learn return distributions instead of their expected values, using diverse projections and representations to improve uncertainty estimation and exploration.
https://openreview.net/forum?id=qae04YACHs
Compressor summary: The Transformer-Modulated Diffusion Model (TMDM) combines transformers and conditional diffusion to predict distribution for multivariate time series with uncertainty.
https://openreview.net/forum?id=qaKRfobbTg
Compressor summary: The paper studies how to learn a threshold in latent space for optimizing rewards in various scenarios, and provides query complexity and regret bounds for this problem.
https://openreview.net/forum?id=qaJxPhkYtD
Compressor summary: This paper studies how Graph Neural Networks (GNNs) can learn to count substructures like cycles and connected components in graphs and designs a new GNN architecture that performs well in several tasks.
https://openreview.net/forum?id=qV83K9d5WB
Compressor summary: The LLMs A s Tool Makers (LATM) framework allows large language models to create and reuse their own tools for problem-solving, improving efficiency and reducing serving costs.
https://openreview.net/forum?id=qTlcbLSm4p
Compressor summary: The Relay Diffusion Model (RDM) improves high-resolution image synthesis by transferring low-resolution images or noise into equivalent high-resolution ones via blurring diffusion and block noise, achieving state-of-the-art results on CelebA-HQ and ImageNet 256 times 256.
https://openreview.net/forum?id=qT7DXUmX7j
Compressor summary: The paper introduces a novel framework for nature-powered graph learning that solves real-valued problems faster and more efficiently than neural networks, while maintaining similar accuracy.
https://openreview.net/forum?id=qPloNoDJZn
Compressor summary: The RoBoT algorithm combines and optimizes existing training-free metrics using Bayesian optimization, improving search performance for designing deep neural networks without training.
https://openreview.net/forum?id=qPFsIbF3V6
Compressor summary: The authors propose a neurosymbolic approach to automatic translation of assembly code programs, which combine the strengths of learned probabilistic models and symbolic solvers to achieve better performance than existing methods.
https://openreview.net/forum?id=qODvxQ8TXW
Compressor summary: LRR outperforms IMP in finding lottery tickets by effectively combining mask learning and parameter optimization in overparameterized networks, allowing it to handle different mask signatures.
https://openreview.net/forum?id=qNrJJZAKI3
Compressor summary: The paper introduces FairSeg, a new medical segmentation fairness dataset, and proposes a fair error-bound scaling approach to improve equity in AI models.
https://openreview.net/forum?id=qL9gogRepu
Compressor summary: AmP introduces a framework, dataset, and challenge to study how text-to-code systems handle linguistic ambiguities, finding that large models perform poorly without explicit instruction but capture the distribution of meanings when ambiguity is present.
https://openreview.net/forum?id=qL6brrBDk2
Compressor summary: SAFLEX is an efficient method that improves data augmentation by learning sample weights and soft labels using bilevel optimization, enhancing model performance across diverse datasets and tasks with minimal computational cost.
https://openreview.net/forum?id=qJ0Cfj4Ex9
Compressor summary: The system learns symbolic action abstractions from language and uses them for long-horizon planning, improving accuracy and generalization.
https://openreview.net/forum?id=qHGgNyQk31
Compressor summary: Seer is a novel model for robots to predict video frames from text inputs, which uses efficient techniques and achieves better results than CogVideo.
https://openreview.net/forum?id=qH9nrMNTIW
Compressor summary: The paper proposes a new model, IPDiff, for generating 3D ligand molecules that bind to specific proteins by incorporating protein-ligand interactions into both diffusion and sampling processes, leading to more realistic structures and better binding affinities.
https://openreview.net/forum?id=qDhq1icpO8
Compressor summary: The paper proposes a new method, CBRL.CIV, for estimating causal effects from observational data without the linearity assumption and using a conditional instrument to relax the strict condition of unconfounded instruments.
https://openreview.net/forum?id=qDdSRaOiyb
Compressor summary: ContraLSP is a model that uses counterfactual samples and sparse gates to create informative but uninformative perturbations for multivariate time series, improving explanation quality.
https://openreview.net/forum?id=qCyhvr0GG8
Compressor summary: VONet is a new method that uses attention and sequential VAE to learn video objects without supervision, achieving state-of-the-art results on various datasets.
https://openreview.net/forum?id=qCUWVT0Ayy
Compressor summary: This paper introduces LayoutNUWA, a model that generates graphic layouts by treating layout generation as a code generation task using large language models, achieving significant performance improvements.
https://openreview.net/forum?id=qBWhjsNPEY
Compressor summary: DeepZero is a novel framework that enables zeroth-order optimization for training deep neural networks, achieving near-optimal performance and improving certified adversarial defense and partial differential equation error correction.
https://openreview.net/forum?id=qBL04XXex6
Compressor summary: Boosting of Thoughts (BoT) is an automated framework for improving problem-solving with LLMs by iteratively exploring, self-evaluating, and revising reasoning steps using error analysis.
https://openreview.net/forum?id=qAoxvePSlq
Compressor summary: The study introduces DQ-LoRe, a framework that uses dual queries and low-rank approximation re-ranking to automatically select exemplars for in-context learning in natural language processing tasks.
https://openreview.net/forum?id=qA4foxO5Gf
Compressor summary: The text introduces two frameworks to accelerate sample generation in diffusion models and presents a hybrid method that outperforms baselines on CIFAR-10.
https://openreview.net/forum?id=q9jQPA6zPK
Compressor summary: The paper proposes a novel coarse-to-fine method for designing multi-cellular robots using Hyperbolic Embeddings for Robot Design (HERD) framework, which improves efficiency and generalization compared to previous approaches.
https://openreview.net/forum?id=q57JLSE2j5
Compressor summary: The paper introduces a novel diffusion-based generative model that can synthesize diverse and high-quality 3D objects from over 200 real-world categories using a triplane representation and a 3D-aware transformer.
https://openreview.net/forum?id=q4SiDyYQbo
Compressor summary: The paper shows that contrastive learning, a type of self-supervised learning, can cause minority groups' representations to merge with majority groups', leading to reduced performance and bias.
https://openreview.net/forum?id=q4AEBLHuA6
Compressor summary: The paper proposes a new method for solving high-frequency and multi-scale PDEs using Gaussian processes and spectral mixture kernels, which can improve efficiency and scalability.
https://openreview.net/forum?id=q3KNrmW6Ql
Compressor summary: The paper presents G-FairAttack, a framework that can adversarially compromise the fairness of various fairness-aware GNNs with low utility loss and fast computation.
https://openreview.net/forum?id=pzpWBbnwiJ
Compressor summary: The proposed universal guidance algorithm allows diffusion models to accept different types of guidance without retraining, enabling various image generation applications.
https://openreview.net/forum?id=pzUhfQ74c5
Compressor summary: The paper proposes a novel conformal prediction approach for language models with performance guarantees, calibrating stopping and rejection rules to reduce noise and ensure coverage of acceptable responses.
https://openreview.net/forum?id=pzElnMrgSD
Compressor summary: The authors propose a novel method to preserve temporal correlations in video editing and generation using integral noise representation and a tailored transport method.
https://openreview.net/forum?id=pz2E1Q9Wni
Compressor summary: The paper investigates how different behavioural models affect inverse reinforcement learning (IRL) and shows that even small misspecifications can cause significant errors in inferring an agent's preferences.
https://openreview.net/forum?id=py4ZV2qYQI
Compressor summary: HybridTree is a novel federated learning approach for hybrid data settings that incorporates party knowledge into tree lower layers, achieving comparable accuracy and significant speedup.
https://openreview.net/forum?id=pxI5IPeWgW
Compressor summary: The paper proposes a novel approach for inferring unbiased treatment effects using closed-form ordinary differential equations, which offer advantages like interpretability and irregular sampling.
https://openreview.net/forum?id=pwW807WJ9G
Compressor summary: This paper proposes a PAC-Bayesian bound for GNNs that incorporates the $L$-hop interplay between training and testing nodes, and introduces a Graph Global Workspace module to enhance the interplay using key-value attention.
https://openreview.net/forum?id=pw2ssoOTpo
Compressor summary: CIFAR-10-Warehouse is a testbed for evaluating model generalization in various out-of-distribution environments, consisting of 180 datasets collected by image search engines and diffusion models with diverse characteristics.
https://openreview.net/forum?id=ptXo0epLQo
Compressor summary: The authors propose alpha-TCVAE, a variational autoencoder that maximizes disentanglement and informativeness of latent variables using a novel total correlation lower bound, and show its benefits in representation learning and downstream tasks.
https://openreview.net/forum?id=ptCIlV24YZ
Compressor summary: The paper presents a new image clustering method based on pre-trained models that improves accuracy, estimates cluster numbers, and labels clusters with text.
https://openreview.net/forum?id=pszewhybU9
Compressor summary: InsTag is a method that analyzes and improves instruction diversity and complexity for supervised fine-tuning (SFT) of large language models (LLMs), leading to better instruction-following abilities.
https://openreview.net/forum?id=psEswR8Jz4
Compressor summary: AmortizedPeriod is a novel method that uses Bayesian statistics and deep learning to identify periodic patterns in time series data, improving robustness and efficiency over existing methods.
https://openreview.net/forum?id=pmweVpJ229
Compressor summary: ASAP is an algorithm that samples from the posterior distribution while preserving differential privacy by adding noise proportional to the distance of each sample from a reference distribution.
https://openreview.net/forum?id=plmBsXHxgR
Compressor summary: The authors propose new cross-modality jailbreak attacks on vision language models using adversarial images and textual prompts, which exploit the alignment vulnerability of multi-modal models.
https://openreview.net/forum?id=plebgsdiiV
Compressor summary: The paper proposes a method to evaluate policies in reinforcement learning with continuous actions by relaxing deterministic target policies using a kernel and optimizing its metrics to reduce bias and variance in off-policy evaluation.
https://openreview.net/forum?id=pjtIEgscE3
Compressor summary: The paper proposes a method to adapt a large black-box text-to-video model to different domains without access to its weights, using the score function as a probabilistic prior for a small task-specific video model.
https://openreview.net/forum?id=phBS6YpTzC
Compressor summary: The paper proposes a framework to measure and improve consistency between generating and validating answers in language models, leading to better performance on various tasks.
https://openreview.net/forum?id=ph04CRkPdC
Compressor summary: The text proposes adding extra hidden vectors to language models, which can improve their performance on various tasks when trained and evaluated with delays.
https://openreview.net/forum?id=pe0Vdv7rsL
Compressor summary: GT-BEHRT combines graph transformer embeddings and BERT model to improve patient representation and perform better in medical prediction tasks using EHRs.
https://openreview.net/forum?id=pdJXYfJjz9
Compressor summary: The paper proposes a new method called CoIC for training image de-raining models that adapts to different types of rain and background, leading to better performance and generalization.
https://openreview.net/forum?id=pTN8dV2pL8
Compressor summary: The paper introduces a new way of representing normals in SDF fields using Gaussian distributions, supervised by polarization priors, to improve NeRF learning of reflective scenes and achieve better 3D reconstruction.
https://openreview.net/forum?id=pTHfApDakA
Compressor summary: SelfCheck is a method for improving question-answering by checking and correcting LLMs' reasoning steps on three datasets.
https://openreview.net/forum?id=pPjZIOuQuF
Compressor summary: RepoBench is a new benchmark for evaluating code auto-completion systems on real-world, multi-file programming scenarios.
https://openreview.net/forum?id=pPh9p8anUi
Compressor summary: PBADet is a novel one-stage, anchor-free method for detecting human body parts and their associations with individuals, which is faster and more accurate than existing approaches.
https://openreview.net/forum?id=pPJTQYOpNI
Compressor summary: Our framework improves ILfO by adaptively adjusting the reward priority between earlier and later steps using Automatic Discount Scheduling, enabling better learning of initial behaviors before advancing to later ones.
https://openreview.net/forum?id=pOoKI3ouv1
Compressor summary: The text argues that agents need to learn causal models to generalize under different distributions, and this has implications for transfer learning and causal inference.
https://openreview.net/forum?id=pFOoOdaiue
Compressor summary: The paper proposes Quantal Adversarial RL (QARL), a novel approach for adversarial RL that uses entropy regularization and gradually increases the rationality of the adversary to ease the complexity of saddle point optimization problems and improve performance and robustness.
https://openreview.net/forum?id=pETSfWMUzy
Compressor summary: RAIN is a novel inference method that allows pre-trained LLMs to evaluate their own generation and use the evaluation results to guide rewind and generation for AI safety, improving harmlessness and truthfulness.
https://openreview.net/forum?id=pEKJl5sflp
Compressor summary: The paper introduces Scalable Modular Network (SMN), a modular network framework that adapts to new tasks with the help of an agreement router, which selects and composes specialist modules using iterative message passing.
https://openreview.net/forum?id=pEGSdJu52I
Compressor summary: The paper investigates test-set variance in neural network training, finding that it is less problematic than expected and providing insights into its causes and mitigations.
https://openreview.net/forum?id=pE6gWrASQm
Compressor summary: The paper explores constructing adversarial examples only on a subset of training data and shows that it improves robustness against adversarial attacks across datasets and tasks, while reducing the computational cost.
https://openreview.net/forum?id=pDCublKPmG
Compressor summary: This paper proposes a generalized attack framework for multi-agent reinforcement learning and a provably efficient defense using timescale separation and adversarial training.
https://openreview.net/forum?id=pCEgna6Qco
Compressor summary: ProMoT is a two-stage fine-tuning method for large language models that reduces format specialization and improves their in-context learning performances across various tasks.
https://openreview.net/forum?id=pC3WJHf51j
Compressor summary: The authors use self-supervised learning with unlabeled data from a large study to train foundation models for photoplethysmography and electrocardiogram biosignals, which can enhance future wearable devices and help users improve their health.
https://openreview.net/forum?id=pBxeZ6pVUD
Compressor summary: The paper introduces Conditional Slot Attention (CoSA), a method that learns stable and invariant object-centric representations using Grounded Slot Dictionary (GSD) inspired by vector quantization, and shows its benefits in various downstream tasks.
https://openreview.net/forum?id=pB1FeRSQxh
Compressor summary: The paper investigates quantum algorithms and lower bounds for minimizing the maximum of N convex functions, improving the complexity bound and proving optimality of the dependence on N.
https://openreview.net/forum?id=pAsQSWlDUf
Compressor summary: SoftCLT is a soft contrastive learning method for time series that preserves correlations and improves downstream tasks.
https://openreview.net/forum?id=pAoqRlTBtY
Compressor summary: The paper presents a novel framework that combines LLMs and DSCMs for causal discovery and shows its superior performance on benchmarks and in modelling Alzheimer's Disease.
https://openreview.net/forum?id=pAVJKp3Dvn
Compressor summary: The paper proposes a framework to learn efficient structured weight matrices for deep neural networks using gradient descent, improving their performance and reducing complexity compared to previous methods.
https://openreview.net/forum?id=pA8Q5WiEMg
Compressor summary: OWO meta learning improves online algorithm initialization and step size in non-convex settings with better regret bounds and investigates generalization bounds using regret analysis.
https://openreview.net/forum?id=p8ujRTjEf3
Compressor summary: The paper proposes a new Thompson sampling algorithm for Gaussian bandits with unknown heterogeneous reward variances, which incorporates prior knowledge to achieve lower regret and adapts better to varying uncertainty levels.
https://openreview.net/forum?id=p4eG8rCa0b
Compressor summary: The text proposes a new conditional diffusion model, Compose and Conquer (CnC), that can control 3D object placement and style in generated images using disentangled representations and depth estimation.
https://openreview.net/forum?id=p4S5Z6Sah4
Compressor summary: The Wave-RNN (wRNN) is a recurrent neural network model that efficiently encodes recent past through traveling waves of neural activity, and performs better than wave-free RNNs on memory tasks and sequential image classification.
https://openreview.net/forum?id=p34fRKp8qA
Compressor summary: The text discusses a framework for training neural networks that are invariant and equivarent to geometrical transformations using Lie groups and their homogeneous spaces, showing improved performance on an affine-invariant benchmark task.
https://openreview.net/forum?id=ozX92bu8VA
Compressor summary: LASER is a simple technique that improves the performance of LLMs by removing some weight matrix components, leading to significant gains in predictive accuracy on various tasks.
https://openreview.net/forum?id=oxjeePpgSP
Compressor summary: The paper proposes a new backdoor attack for contrastive learning that uses a bi-level optimization approach to create a more effective trigger and achieve higher success rates.
https://openreview.net/forum?id=ox2ATRM90I
Compressor summary: YAIB is a modular framework that allows researchers to define reproducible and comparable clinical machine learning experiments in intensive care units using various datasets, preprocessing pipelines, and prediction tasks.
https://openreview.net/forum?id=owokKCrGYr
Compressor summary: Quality-Diversity through AI Feedback (QDAIF) is a method that uses language models to guide an evolutionary algorithm in generating diverse high-quality outputs in creative writing and other domains.
https://openreview.net/forum?id=otU31x3fus
Compressor summary: The paper introduces a new robust optimization method that generalizes existing ones and achieves optimal convergence under inexact gradients and Hessians, using tensor generalization for stochastic higher-order derivatives.
https://openreview.net/forum?id=otHZ8JAIgh
Compressor summary: The PIBD framework tackles intra-modal and inter-modal redundancies in multimodal learning for cancer survival prediction by using prototypical information bottlenecking and disentangling.
https://openreview.net/forum?id=osoWxY8q2E
Compressor summary: ReLU activation in large language models can save significant computation and weight transfer during memory-bound inference without compromising convergence and performance.
https://openreview.net/forum?id=okYdj8Ysru
Compressor summary: The paper proposes a unified framework for Riemannian Batch Normalization on Lie groups and applies it to SPD manifolds for computer vision tasks such as radar recognition, human action recognition, and EEG classification.
https://openreview.net/forum?id=ojIJZDNIBj
Compressor summary: The paper introduces CopulaCPTS, a conformal prediction algorithm for multivariate, multi-step time series forecasting with finite sample validity guarantee and better calibration and efficiency than existing methods.
https://openreview.net/forum?id=ofzeypWosV
Compressor summary: CLaM-TTS is a new method that uses probabilistic vector quantization to compress audio tokens and generate multiple streams for zero-shot TTS synthesis with improved naturalness, intelligibility, and speed.
https://openreview.net/forum?id=of2rhALq8l
Compressor summary: AffineQuant is a technique that optimizes post-training quantization of large language models by using equivalent affine transformations, improving performance and allowing deployment on edge devices with low-bit configurations.
https://openreview.net/forum?id=odY3PkI5VB
Compressor summary: The paper presents a new HRL algorithm that combines spatial and temporal goal abstraction to improve performance on complex tasks.
https://openreview.net/forum?id=oZyAqjAjJW
Compressor summary: The paper proposes a method called *LDReg* to improve self-supervised learning by preventing dimensional collapse, which occurs when representations fail to capture the full data distribution locally.
https://openreview.net/forum?id=oZtt0pRnOl
Compressor summary: The paper proposes an algorithm for in-context learning with large language models on private datasets, using synthetic few-shot demonstrations and formal differential privacy guarantees.
https://openreview.net/forum?id=oZDJKTlOUe
Compressor summary: LURE is an algorithm that corrects object hallucination in vision-language models by reconstructing more accurate descriptions based on co-occurrence, uncertainty, and object position.
https://openreview.net/forum?id=oYjPk8mqAV
Compressor summary: The paper proposes Magnushammer, a contrastive training-based method for premise selection in automated theorem proving, which outperforms existing tools and enables state-of-the-art results with fewer parameters.
https://openreview.net/forum?id=oXjnwQLcTA
Compressor summary: SMORe is a novel discriminator-free method for offline goal-conditioned reinforcement learning that leverages mixture-distribution matching and scores for achieving diverse and reusable skills from suboptimal offline datasets without hand-engineering reward functions.
https://openreview.net/forum?id=oXYZJXDdo7
Compressor summary: The text introduces a novel method of generating text using context-aware phrases from supporting documents, overcoming challenges in training oracles with linguistic heuristics and self-reinforcement, and demonstrating improved performance and generation quality on various tasks.
https://openreview.net/forum?id=oTRwljRgiv
Compressor summary: The paper proposes ExeDec, a new synthesis strategy that improves compositional generalization for neural program synthesis models by predicting execution subgoals step-by-step.
https://openreview.net/forum?id=oQKKlzxV1o
Compressor summary: The paper studies how a principal can hire multiple agents to gather and report information efficiently, considering coordination, externalities, and incentives.
https://openreview.net/forum?id=oOwDQl8haC
Compressor summary: The paper proposes a method to train and fine-tune DNNs using cheaper 12-bit accumulators without significant performance loss and shows that gradient approximations can improve accuracy with lower precision.
https://openreview.net/forum?id=oOGqJ6Z1sA
Compressor summary: The paper proposes a new weighting method for covariate balancing and treatment effects estimation in observational studies, which works well even when the underlying model assumptions are not strong.
https://openreview.net/forum?id=oO6FsMyDBt
Compressor summary: This paper proposes using graph neural networks and transformers to encode diverse neural network architectures and achieve better performance on various tasks.
https://openreview.net/forum?id=oMNkj4ER7V
Compressor summary: The paper proposes a general Thompson sampling algorithm for optimizing various objectives in Bayesian optimization under contextual uncertainty, and analyzes its performance.
https://openreview.net/forum?id=oMLQB4EZE1
Compressor summary: The authors propose DNABERT-2, a refined genome foundation model that uses Byte Pair Encoding (BPE) tokenization to overcome limitations of k-mer tokenization and achieve better efficiency and performance on large genomes.
https://openreview.net/forum?id=oM7Jbxdk6Z
Compressor summary: MoleBLEND is a self-supervised learning approach that aligns 2D and 3D molecular structures at the atomic level by blending atom relations from both modalities and achieves state-of-the-art performance in molecular modeling tasks.
https://openreview.net/forum?id=oLLZhbBSOU
Compressor summary: This paper proposes a reinforcement learning method for skill acquisition that uses user interventions as rewards, improving upon interactive imitation learning methods like DAgger, especially when the expert is suboptimal.
https://openreview.net/forum?id=oKn9c6ytLx
Compressor summary: The paper presents a realistic web-based environment for evaluating language-guided agents' performance on complex tasks, showing that current state-of-the-art models struggle to match human performance.
https://openreview.net/forum?id=oGNdBvymod
Compressor summary: The paper proposes a new sampling method for Bayesian deep learning that biases towards flat regions in the posterior distribution to avoid overfitting and improve generalization performance.
https://openreview.net/forum?id=oEF7qExD9F
Compressor summary: The paper introduces LMUFormer, a low-complexity recurrent model with memory units, embedding, and mixer that performs as well as transformers on speech commands while being more efficient and suitable for streaming applications at the edge.
https://openreview.net/forum?id=oDdzXQzP2F
Compressor summary: Transformer-VQ is a fast decoder-only transformer with vector-quantized keys and caching that achieves high quality in various tasks.
https://openreview.net/forum?id=oAMArMMQxb
Compressor summary: The authors show how to use vanilla score matching with Langevin diffusion and early stopping to generate multimodal distributions.
https://openreview.net/forum?id=o8tjamaJ80
Compressor summary: AdAutomixup is an adversarial automatic mixup augmentation method for improving image classification by generating diverse and challenging mixed samples and training a robust classifier.
https://openreview.net/forum?id=o83eu4H9Mb
Compressor summary: The information bottleneck principle helps learn good representations by balancing conciseness and predictive ability, but may not work well in low-resource or out-of-domain scenarios; a new method is proposed to jointly learn mainline and supplemental features for better prediction.
https://openreview.net/forum?id=o7x0XVlCpX
Compressor summary: MaGIC is a novel method for image completion that supports single and multi-modality guidance, such as text, edge, sketch, segmentation, depth, and pose, without requiring re-training of different modalities.
https://openreview.net/forum?id=o7qhUMylLU
Compressor summary: The paper proposes a new complexity measure and algorithm for sample-efficient multi-agent reinforcement learning in general-sum Markov Games, addressing different equilibrium concepts and both model-based and model-free settings.
https://openreview.net/forum?id=o5Bqa4o5Mi
Compressor summary: $\pi$2vec is a method that uses feature vectors and successor features to represent and evaluate black box policies efficiently in offline settings.
https://openreview.net/forum?id=o4CLLlIaaH
Compressor summary: The paper introduces a novel method called Generalizable neural Point Field (GPF) that improves generalization, consistency, and quality in neural radiance field modeling by using point-based rendering instead of image-based rendering.
https://openreview.net/forum?id=o3BxOLoxm1
Compressor summary: MPGD is a training-free method that uses pretrained models and neural networks for conditional image generation, improving efficiency and quality.
https://openreview.net/forum?id=o2IEmeLL9r
Compressor summary: PTGM is a method that pre-trains goal-based models for reinforcement learning, improving sample efficiency and performance in complex tasks by providing temporal abstractions and behavior regularization.
https://openreview.net/forum?id=nxnbPPVvOG
Compressor summary: The paper proposes a generalization of Gauss-Markov theorem for linear estimation with non-zero bias, derives optimal estimators for different norms, analyzes generalization error, and compares with Ridge regression using simulations.
https://openreview.net/forum?id=nsNyDvNQTc
Compressor summary: The paper analyzes how uncertainty and model score affect estimation bias in binary classification and proposes algorithms that use both to improve recall and precision.
https://openreview.net/forum?id=nqlymMx42E
Compressor summary: The paper introduces ChemRLformer, a new RL-based molecular design algorithm that uses text representations and explores how different design choices affect its performance on 25 molecule design tasks.
https://openreview.net/forum?id=noe76eRcPC
Compressor summary: PF-LRM is a fast and accurate method for reconstructing 3D objects from unposed images with self-attention and PnP, and it works well on different datasets and scenarios.
https://openreview.net/forum?id=nnicaG5xiH
Compressor summary: CAMEL is a meta-learning architecture that learns efficiently and interprets physical parameters in various settings.
https://openreview.net/forum?id=nnVO1PvbTv
Compressor summary: The paper proposes an "LLM-KG" approach that improves large language models' reasoning by integrating external knowledge graphs and introduces a new method called Think-on-Graph (ToG) for interactive reasoning with better deep reasoning, knowledge traceability, flexibility, and cost-efficiency.
https://openreview.net/forum?id=nmBjBZoySX
Compressor summary: The paper introduces AdaGLT, an adaptive, dynamic, and automated framework for finding graph lottery tickets to improve efficiency and scalability of graph neural networks on large graphs.
https://openreview.net/forum?id=nfMyERXNru
Compressor summary: The paper presents a novel video editing framework that uses multiple RGB layers and opacity levels to perform tasks like object segmentation, dehazing, and relighting without requiring large datasets or external data collection.
https://openreview.net/forum?id=nfIAEJFiBZ
Compressor summary: The paper proposes a Thompson sampling exploration strategy for reinforcement learning that samples Q-functions directly from their posterior distribution using Langevin Monte Carlo, achieving good results on deep RL and Atari57 tasks.
https://openreview.net/forum?id=ndR8Ytrzhh
Compressor summary: ESC is a simple and scalable method to reduce the cost of chain-of-thought reasoning by using less sampling, without sacrificing performance on various reasoning tasks.
https://openreview.net/forum?id=ndCJeysCPe
Compressor summary: The paper analyzes how a two-layer autoencoder can generate samples from a high-dimensional Gaussian mixture and shows that the method is optimal with an $\frac{1}{n}$ convergence rate.
https://openreview.net/forum?id=nc5GgFAvtk
Compressor summary: CroPA is a method that uses learnable textual prompts to update visual adversarial perturbations and improve their transferability across different vision tasks with large vision language models.
https://openreview.net/forum?id=nbPGqeH3lt
Compressor summary: FedCDA is a novel method for federated learning that selectively aggregates local models from different rounds to reduce diversity and maintain data knowledge.
https://openreview.net/forum?id=nanyAujl6e
Compressor summary: The paper proposes learning a set of negative prompts for each class to improve out-of-distribution (OOD) detection by matching similarity and dissimilarity in the feature space.
https://openreview.net/forum?id=nZP6NgD3QY
Compressor summary: AdaMerging is an unsupervised technique that learns how to merge pre-trained models for multi-task learning using entropy minimization on test samples, improving performance and generalization compared to existing methods.
https://openreview.net/forum?id=nZP10evtkV
Compressor summary: The text proposes a distribution-oriented approach using optimal transport to create adversarial patches that can fool multiple neural networks, even without knowing the target model.
https://openreview.net/forum?id=nY9nITZQjc
Compressor summary: MIntRec2.0 is a large benchmark dataset for multimodal intent recognition in multi-party conversations with diverse samples, modalities, and contexts.
https://openreview.net/forum?id=nUBLhhVM1l
Compressor summary: This paper studies when and how knowledge can be transferred from source to target in outlier detection tasks using Neyman-Pearson classification and measures of discrepancy.
https://openreview.net/forum?id=nTwb2vBLOV
Compressor summary: The paper proposes using graph canonization to increase the expressivity of GNNs and studies their stability, introducing a universal graph canonization method that improves performance on various graph tasks.
https://openreview.net/forum?id=nTNgkEIfeb
Compressor summary: The paper introduces a data-leakage risk in federated learning due to model inversion attacks and proposes FedInverse, a tool to evaluate this risk using HSIC as a regularizer.
https://openreview.net/forum?id=nO344avRib
Compressor summary: GEEL is a new graph representation that allows for more efficient and effective neural network-based graph generation, especially for large-scale and attributed graphs.
https://openreview.net/forum?id=nMFSUjxMIl
Compressor summary: CircuitNet 2.0 is a large-scale dataset that supports machine learning innovations in chip design by providing comprehensive data from various realistic chip designs.
https://openreview.net/forum?id=nLWiR5P3wr
Compressor summary: FoRDE is a new method that improves the performance of deep ensembles by diversifying input gradients, leading to better accuracy, calibration, and robustness to perturbations.
https://openreview.net/forum?id=nJnky5K944
Compressor summary: One-layer and single-head Transformers can effectively capture context and memorize information without relying on deep layers or high-rank weight matrices.
https://openreview.net/forum?id=nHkMm0ywWm
Compressor summary: The paper proposes a method to identify causal relationships among measured and latent variables using high-order statistics and tests for statistical independence.
https://openreview.net/forum?id=nHESwXvxWK
Compressor summary: The study proposes a method using score-based generative models and Sequential Monte Carlo to solve ill-posed linear inverse problems in a Bayesian framework, achieving better results than existing approaches.
https://openreview.net/forum?id=nFMS6wF2xq
Compressor summary: ContextDiff is a novel contextualized diffusion model that improves text-to-image and text-to-video tasks by incorporating cross-modal context into both forward and reverse processes.
https://openreview.net/forum?id=nFI3wFM9yN
Compressor summary: Fed-GO-UCB is a novel algorithm for collaborative function optimization among multiple clients that preserves data privacy, allows large-scale computing, and works with generic non-linear objective functions, achieving sub-linear rates for regret and communication cost.
https://openreview.net/forum?id=nBZBPXdJlC
Compressor summary: LTU is a new model that combines audio perception and reasoning abilities, trained on OpenAQA-5M dataset, which outperforms existing audio models in classification and captioning tasks and shows emerging audio reasoning skills.
https://openreview.net/forum?id=nBCuRzjqK7
Compressor summary: The paper proposes a new approach using contrastive learning and enhanced decomposition to improve long-term forecasting by capturing and incorporating global autocorrelation in time series.
https://openreview.net/forum?id=nAs4LdaP9Y
Compressor summary: Federated Orthogonal Training (FOT) is a novel method that address global catastrophic forgetting in Continual Federated Learning by reducing interference between tasks using orthogonal updates, while preserving privacy and requiring low computation and communication cost.
https://openreview.net/forum?id=n9xeGcI4Yg
Compressor summary: Equilibrium-ranking is a new game-theoretic procedure for language model decoding that improves performance on various text generation tasks without needing additional training data.
https://openreview.net/forum?id=n7Sr8SW4bn
Compressor summary: The Mayfly is a neural data structure that summarizes graph streams for real-time graph analytics in various domains by learning from meta-tasks.
https://openreview.net/forum?id=n6mLhaBahJ
Compressor summary: HAZARD is a new simulated benchmark for testing embodied agents' decision-making abilities in dynamic environments with disaster scenarios and large language models.
https://openreview.net/forum?id=mzyZ4wzKlM
Compressor summary: The text discusses how expressive losses can improve adversarial robustness and accuracy trade-offs in networks by over-approximating the worst-case loss.
https://openreview.net/forum?id=mz8owj4DXu
Compressor summary: The Scalable Language Model (SLM) overcomes limitations of continual learning in language models by using JARe and DTKR, achieving state-of-the-art performance on diverse tasks and domains with minimal forgetting.
https://openreview.net/forum?id=mw1PWNSWZP
Compressor summary: Instruction tuning using code from Git commits improves large language models' performance on various coding tasks across multiple languages.
https://openreview.net/forum?id=mvMI3N4AvD
Compressor summary: Mega-TTS is a new zero-shot text-to-speech method that uses an acoustic autoencoder and a prosody language model to generate speech with unseen speakers and styles from short prompts.
https://openreview.net/forum?id=mutJBk3ILg
Compressor summary: The paper explores how spurious features affect self-supervised learning and proposes a method called LateTVG to remove them by pruning later layers of the encoder.
https://openreview.net/forum?id=msXxrttLOi
Compressor summary: FedCompass is a semi-asynchronous federated learning algorithm that adapts to client computing power, reducing staleness and waiting times, and improving convergence and accuracy on non-IID heterogeneous datasets.
https://openreview.net/forum?id=ms0VgzSGF2
Compressor summary: The paper proposes a common idea of self-predictive abstraction for representation learning in deep reinforcement learning and provides a minimalist algorithm to learn such representations across different domains.
https://openreview.net/forum?id=mqVgBbNCm9
Compressor summary: The paper proposes Skeleton-of-Thought, a method that speeds up large language models by guiding them to generate skeletons of answers in parallel and then filling them in sequentially.
https://openreview.net/forum?id=mpqMVWgqjn
Compressor summary: KW-Design is a novel protein inverse folding method that leverages pretrained models, confidence, and memory retrieval to overcome limitations and achieve high recovery rates.
https://openreview.net/forum?id=mnipav175N
Compressor summary: The paper introduces Temporally-Correlated Episodic RL (TCE), a method that generates smooth trajectories for robot tasks by using multi-second trajectory samples and evaluating policy updates based on segment-wise advantages, improving sample efficiency and capturing movement correlations.
https://openreview.net/forum?id=mmjnr0G8ZY
Compressor summary: The paper proposes a novel multi-resolution diffusion model for time series data that leverages temporal structure and outperforms existing models.
https://openreview.net/forum?id=mliQ2huFrZ
Compressor summary: The paper proposes a new label adaptation framework called class probability matching (CPM) that matches class probabilities between source and target domains and improves generalization and efficiency over existing methods.
https://openreview.net/forum?id=mlJLVigNHp
Compressor summary: The paper introduces compressors that generate textual summaries from retrieved documents for retrieval-augmented language models, improving efficiency and performance on language modeling and question answering tasks.
https://openreview.net/forum?id=mhyQXJ6JsK
Compressor summary: This paper proposes a new method to accelerate computations for 3D data modeling using equivariant neural networks, reducing complexity from $\mathcal{O}(L^6)$ to $\mathcal{O}(L^3)$.
https://openreview.net/forum?id=mhgm0IXtHw
Compressor summary: Noise Map Guidance (NMG) is a new method for real-image editing that uses spatial context without requiring optimization, overcoming the limitations of previous text-guided diffusion models.
https://openreview.net/forum?id=maRYffiUpI
Compressor summary: The text discusses improving code generation performance by making programs more structured and readable using a data-cleaning pipeline and shows that this leads to better results even with less data.
https://openreview.net/forum?id=mZn2Xyh9Ec
Compressor summary: FlashAttention-2 improves the efficiency and scalability of Transformers by optimizing work partitioning in the attention layer, reaching up to 73% of the theoretical maximum FEOPs/s on A100 GPUs.
https://openreview.net/forum?id=mYhH0CDFFa
Compressor summary: The paper studies how CNNs memorize trigger changes in poisoned samples' frequency domain, proposes a universal invisible strategy for backdoor attacks, and designs a low-frequency backdoor attack method with high accuracy.
https://openreview.net/forum?id=mYWsyTuiRp
Compressor summary: The study visualizes how feed-forward blocks in Transformers modify input contextualization and may reveal redundant processing in these models.
https://openreview.net/forum?id=mXpNp8MMr5
Compressor summary: The paper warns that adversarially trained models can be fooled by two-faced attacks, which may mislead evaluations and cause security issues, and suggests a countermeasure.
https://openreview.net/forum?id=mQYHXUUTkU
Compressor summary: The researchers developed a method to generate natural language descriptions for images based on neural activity in different regions of the brain, enabling new insights into how the brain processes visual information.
https://openreview.net/forum?id=mQ72XRfYRZ
Compressor summary: The paper proposes a new hierarchical Bayesian model for few-shot learning that uses global and local random variables to infer novel episodes, and exploits a Normal-Inverse-Wishart model for efficient training.
https://openreview.net/forum?id=mNYF0IHbRy
Compressor summary: The authors propose a novel approach using Large Language Models to extract information from complex text prompts and generate detailed images, improving the fidelity of text-to-image generation models.
https://openreview.net/forum?id=mMaQvkMzDi
Compressor summary: The text discusses the limitations of large multimodal models and explores in-context learning as a potential solution, proposing new multimodal ICL approaches to improve their performance.
https://openreview.net/forum?id=mM7VurbA4r
Compressor summary: SOTOPIA is an environment for testing AI's social intelligence by simulating complex interactions between agents and humans, revealing GPT-4's challenges in social commonsense reasoning and communication.
https://openreview.net/forum?id=mL8Q9OOamV
Compressor summary: Emu is a versatile model that generates images and text from various inputs, achieving high performance in multiple multimodal tasks.
https://openreview.net/forum?id=mIEHIcHGOo
Compressor summary: The paper explores how to transfer knowledge from larger to smaller language models using sensitivity-based techniques and a module called LoRA.
https://openreview.net/forum?id=mGHJAyR8w0
Compressor summary: This study analyzes the expressive power and performance of invariant and equivariant GNNs, showing that increasing the feature dimension is more important than increasing the number of types of steerable features for improving performance.
https://openreview.net/forum?id=mF3cTns4pe
Compressor summary: Sum-product-set networks extend probabilistic circuits to tree-structured graphs using random finite sets for exact and efficient inference, comparing well with intractable neural network models.
https://openreview.net/forum?id=mE52zURNGc
Compressor summary: The authors derive a closed-form solution for the expected optimum of the Gauss-Newton loss in direct image alignment, which improves accuracy, robustness, and convergence properties, and offers insight into end-to-end learning limitations.
https://openreview.net/forum?id=mCOBKZmrzD
Compressor summary: EquiformerV2 improves on previous methods for 3D atomistic systems using eSCN convolutions, attention re-normalization, separable $S^2$ activation, and separable layer normalization, achieving better results and efficiency.
https://openreview.net/forum?id=m7tJxajC3G
Compressor summary: The paper proposes a novel federated causal discovery method that can handle diverse scenarios with arbitrary causal models and heterogeneous data, while protecting data privacy using summary statistics.
https://openreview.net/forum?id=m7aPLHwsLr
Compressor summary: The paper proposes DRSM, a certified defense against evasion attacks on malware detection models, which preserves accuracy and robustness by limiting the impact of adversarial bytes using window ablation.
https://openreview.net/forum?id=m52uU0dVbH
Compressor summary: The paper proposes a novel adversarial attack on vertical federated learning by adaptively corrupting subsets of clients using Thompson sampling with Empirical maximum reward (E-TS) algorithm.
https://openreview.net/forum?id=m50eKHCttz
Compressor summary: The authors investigate how to transfer complementary knowledge from one deep network to another without performance degradation, using data partitioning techniques that work for most pretrained models and can be done unsupervised.
https://openreview.net/forum?id=m3xVPaZp6Z
Compressor summary: ReDM is a new algorithm that enables reinforcement learning agents to make adaptive decisions by planning for all possible outcomes and rehearsing on diverse dynamics models, even with zero or limited interaction data.
https://openreview.net/forum?id=m3ch3kJL7q
Compressor summary: The paper proposes a new method for composed image retrieval that uses sentence-level prompts generated by V-L models instead of pseudo-word tokens, improving performance on complex changes in the target image.
https://openreview.net/forum?id=m3RRWWFaVe
Compressor summary: This paper explores the ethical values of large language models using Moral Foundation Theory, creates a dataset of moral prompts, and proposes an alignment method to improve value compliance.
https://openreview.net/forum?id=m2NVG4Htxs
Compressor summary: The authors analyze data contamination in large language models by studying their performance on two code/mathematical problem-solving datasets, finding significant trends that suggest contamination.
https://openreview.net/forum?id=m0x0rv6Iwm
Compressor summary: This paper proposes a time-varying propensity score to handle gradually evolving data in machine learning models by selecting past data that is similar to the current data.
https://openreview.net/forum?id=lvSMIsztka
Compressor summary: This paper proposes two fast estimators for valuing data using probabilistic values from cooperative game theory, based on a connection to least square regressions, and shows how distributional values can be optimized similarly.
https://openreview.net/forum?id=ltZ9ianMth
Compressor summary: The paper analyzes three types of anomalies in time series data and proposes a robust forecasting algorithm that outperforms current methods.
https://openreview.net/forum?id=lsxeNvYqCj
Compressor summary: The text studies a strategic version of the multi-armed bandit problem, where arms choose their click-rates to maximize clicks, and proposes an algorithm that balances incentives and regret minimization.
https://openreview.net/forum?id=lsvlvWB9vz
Compressor summary: The paper proposes EControl, a new mechanism for distributed training with contractive compression that regulates error compensation and ensures fast convergence without strong assumptions on data heterogeneity.
https://openreview.net/forum?id=lrQlLqQase
Compressor summary: The paper introduces a method to learn causality in multivariate time series data using reinforcement learning and shows its effectiveness in discovering and measuring causal relationships in stochastic processes.
https://openreview.net/forum?id=loYSzjSaAK
Compressor summary: The paper proposes a reinforcement learning framework called subRL that handles non-additive rewards with diminishing returns using submodular set functions, and presents an efficient policy gradient algorithm called subPO for various applications.
https://openreview.net/forum?id=lmM4Ecm4HJ
Compressor summary: The text proposes a box stability score (BS score) that reflects how much bounding boxes change under feature map dropout, which correlates with object detection accuracy and enables assessment of detectors without test ground truths.
https://openreview.net/forum?id=lm7MRcsFiS
Compressor summary: Ring-A-Bell is a tool that tests the effectiveness of safety measures in text-to-image models by generating inappropriate content based on extracted concepts.
https://openreview.net/forum?id=ljwoQ3cvQh
Compressor summary: Neural network predictions for high-dimensional inputs tend to become constant when faced with out-of-distribution inputs, often approximating the optimal constant solution.
https://openreview.net/forum?id=liuqDwmbQJ
Compressor summary: ViLMA is a task-agnostic benchmark that evaluates the fine-grained capabilities of pretrained Video-Language Models (VidLMs) by using carefully curated counterfactual tests and proficiency tests.
https://openreview.net/forum?id=likXVjmh3E
Compressor summary: LoRA is a technique for fine-tuning pre-trained models that can adapt any smaller model to any larger model with a certain rank, and this paper analyzes its expressive power and approximation error.
https://openreview.net/forum?id=liKkG1zcWq
Compressor summary: SliDe is a novel molecular pre-training method that uses physical principles to improve generalization and robustness in drug discovery by enhancing the accuracy of estimated force fields.
https://openreview.net/forum?id=lgaFMvZHSJ
Compressor summary: The paper proposes CARE, a self-supervised learning method that adds geometric structure to embeddings by enforcing equivariance and improves downstream tasks and sensitivity to data variations.
https://openreview.net/forum?id=ldJXXxPE0L
Compressor summary: Weight pruning affects large language models' abilities to recall facts from pre-training and process information in context differently, with moderate pruning impairing fact recall but less so processing in context.
https://openreview.net/forum?id=lajn1iROCu
Compressor summary: SRL is a scalable, efficient, and extensible distributed RL system that enables massively parallelized training and easy development of customized algorithms, outperforming existing libraries in throughput and learning performance.
https://openreview.net/forum?id=lUYY2qsRTI
Compressor summary: The paper proposes a method to estimate and account for uncertainty due to hidden confounding bias in offline reinforcement learning, using delphic uncertainty, and shows its effectiveness on sepsis management and electronic health records benchmarks.
https://openreview.net/forum?id=lROh08eK6n
Compressor summary: The paper proposes a new network embedding method using high-dimensional quantum states that outperforms low-dimensional methods while being more efficient and robust.
https://openreview.net/forum?id=lR3rk7ysXz
Compressor summary: The paper proposes Diffusion Time Estimation, a simplified and faster version of Denoising Diffusion Probability Models, for anomaly detection in both unsupervised and semi-supervised scenarios.
https://openreview.net/forum?id=lOwkOIUJtx
Compressor summary: The authors propose a system that uses multiple foveal views and a reinforcement learning-based saccade mechanism to reconstruct scenes, achieving high image recognition accuracy with reduced input pixels and computation.
https://openreview.net/forum?id=lNZJyEDxy4
Compressor summary: The paper proposes a new masking strategy for anomaly detection in tabular data, which captures intrinsic feature correlations and improves interpretability.
https://openreview.net/forum?id=lNCnZwcH5Z
Compressor summary: NCL is a non-negative contrastive learning method for deriving interpretable and sparse features from data clusters, improving over standard contrastive learning in various tasks.
https://openreview.net/forum?id=lLmqxkfSIw
Compressor summary: Kosmos-2 is a large language model that connects text to visual objects and enables various multimodal tasks.
https://openreview.net/forum?id=lKxL5zkssv
Compressor summary: The CLIP-MUSED method uses a Transformer and subject-specific tokens to decode visual neural information from multiple subjects, overcoming limitations of existing methods and achieving state-of-the-art performance on fMRI datasets.
https://openreview.net/forum?id=lKK50q2MtV
Compressor summary: The paper presents a text-driven video editing framework that uses a text-to-image diffusion model to generate high-quality videos from source videos and target text prompts, without needing training or fine-tuning.
https://openreview.net/forum?id=lK2V2E2MNv
Compressor summary: VLAP bridges vision encoders and language models by using word embeddings to map visual representations into language space for improved performance on vision-language tasks.
https://openreview.net/forum?id=lJkOCMP2aW
Compressor summary: Pathformer is a multi-scale transformer model that adapts to varying temporal dynamics in time series, improving forecasting accuracy and generalization.
https://openreview.net/forum?id=lJYAkDVnRU
Compressor summary: Our meta-learning algorithm learns new visual concepts during inference without fine-tuning, achieving state-of-the-art results on 8 out of 11 benchmarks.
https://openreview.net/forum?id=lHasEfGsXL
Compressor summary: The paper proposes LightHGNN and LightHGNN+, which distill knowledge from hypergraph neural networks (HGNNs) to multi-layer perceptrons (MLPs) to reduce complexity and improve inference speed while maintaining competitive performance.
https://openreview.net/forum?id=lHZm9vNm5H
Compressor summary: This paper introduces a new mode, Tune, for ConvBN blocks that balances efficiency and stability in computer vision tasks, reducing memory footprint and training time while maintaining performance.
https://openreview.net/forum?id=lGUyAuuTYZ
Compressor summary: The paper proposes a new training method for sparse BNNs that uses a Hoyer regularizer to improve accuracy and energy efficiency over BNNs, SNNs, and adder neural networks.
https://openreview.net/forum?id=lFYj0oibGR
Compressor summary: RoboFlamingo is a simple framework that adapts vision-language models to robot control by fine-tuning them on language-conditioned manipulation datasets, achieving state-of-the-art performance and enabling open-loop control.
https://openreview.net/forum?id=lF2aip4Scn
Compressor summary: The paper analyzes how using expert demonstrations can improve the sample efficiency of reinforcement learning and provides theoretical guarantees for this method in different settings.
https://openreview.net/forum?id=lDbjooxLkD
Compressor summary: The study develops an evaluation strategy called PassUntil to measure task performance improvements in small language models, revealing a strict task scaling law and characterizing emergent abilities in large language models.
https://openreview.net/forum?id=lAhWGOkpSR
Compressor summary: VWFormer is a novel multi-scale learner for semantic segmentation that uses varying window attention and achieves competitive efficiency with lower computational cost than existing methods.
https://openreview.net/forum?id=lAhQCHuANV
Compressor summary: The article presents asymptotic guarantees for empirical ROC curves and shows that naive bootstrap approach can be misleading, suggesting a dedicated recentering technique instead, which is experimentally validated for assessing similarity functions and fairness in face recognition.
https://openreview.net/forum?id=l60EM8md3t
Compressor summary: The paper proposes m-LTM, a scalable and robust inverse optimal transport framework for audio-text matching that uses mini-batch subsampling, neural networks, Mahalanobis-enhanced ground metrics, and partial optimal transport to learn rich joint embeddings.
https://openreview.net/forum?id=l3qtSNsPvC
Compressor summary: The paper introduces a new graph sampling technique based on graphon signals, which simplifies large-scale graph machine learning by using unique sampling sets that are consistent across converging graphs.
https://openreview.net/forum?id=l1U6sEgYkb
Compressor summary: DV-3DLane is a novel framework that combines images and LiDAR points to accurately detect 3D lanes using dual-view multi-modal features, feature fusion, and attention mechanisms.
https://openreview.net/forum?id=kzGuiRXZrQ
Compressor summary: EQGAT-diff is a new E(3) equivariant diffusion model for 3D molecular design that outperforms existing models on large and limited data sets with explicit hydrogens.
https://openreview.net/forum?id=kxgSlyirUZ
Compressor summary: COLLIE is a framework for creating diverse grammar-based constraints for text generation tasks and analyzing the performance of language models on them.
https://openreview.net/forum?id=kxebDHZ7b7
Compressor summary: TRAM is a fine-tuning algorithm that improves out-of-domain generalization by optimizing flat minima and smooth, informative representations while preserving pre-trained structure.
https://openreview.net/forum?id=kx2XZlmgB1
Compressor summary: The paper proposes a new augmentation method for contrastive learning to improve ordinal regression by preserving essential ordinal content information in images, outperforming existing methods.
https://openreview.net/forum?id=kvcbV8KQsi
Compressor summary: The paper introduces successor heads, attention heads that increment tokens with natural ordering, and uses mechanistic interpretability to analyze their behavior in large language models, finding common abstract representations and mod 10 features.
https://openreview.net/forum?id=kvByNnMERu
Compressor summary: The paper derives bounds on the worst-case convergence of shape distance estimators and introduces a novel method-of-moments estimator with tunable bias-variance for measuring geometric similarity in high-dimensional network representations.
https://openreview.net/forum?id=kv5xE1p3jz
Compressor summary: JointNet is a neural network that models the relationship between images and an additional modality (e.g., depth maps) using a pre-trained text-to-image diffusion model and enables various applications like joint image-depth generation and depth prediction.
https://openreview.net/forum?id=kuTZMZdCPZ
Compressor summary: The paper proposes a new deep neural network method to reconstruct physical fields from sparse sensor data, which improves reconstruction quality on realistic datasets.
https://openreview.net/forum?id=ktdETU9JBg
Compressor summary: PerCo is a perceptual compression model that uses diffusion models for decoding and textual image descriptions for context, achieving state-of-the-art visual quality at ultra-low bitrates.
https://openreview.net/forum?id=ktJAF3lxbi
Compressor summary: The paper proposes a method to improve ODE-based diffusion models by optimizing coefficients using improved integration approximation, resulting in better FID scores for small NFE values.
https://openreview.net/forum?id=ktG8Tun1Cy
Compressor summary: The paper proposes Classifier Score Distillation (CSD), a method that uses an implicit classification model for text-to-3D generation without relying on pre-trained 2D diffusion models, achieving superior results in various tasks.
https://openreview.net/forum?id=krx55l2A6G
Compressor summary: The paper studies how malicious server attacks can steal data from federated learning, shows that existing attacks are detectable, and proposes a new attack method called SEER that can evade detection.
https://openreview.net/forum?id=koYsgfEwCQ
Compressor summary: DynaVol is a 3D scene generative model that learns object-centric representations in dynamic visual scenes, enabling unsupervised scene decomposition and manipulation of geometric shapes and motion trajectories.
https://openreview.net/forum?id=kmn0BhQk7p
Compressor summary: The study shows that large language models can accurately infer personal attributes from text, posing a serious threat to individual privacy and highlighting the need for better protection methods.
https://openreview.net/forum?id=kklwv4c4dI
Compressor summary: FeDualEx is a new distributed optimization method for saddle point problems with composite objectives and constraints that preserves structure in machine learning.
https://openreview.net/forum?id=kjn99xFUF3
Compressor summary: FedDA is a new adaptive gradient method for federated learning that works with constraints, and it achieves better performance than FedAvg.
https://openreview.net/forum?id=khAE1sTMdX
Compressor summary: The paper proposes UniMP, a framework that leverages multi-modal data for personalized tasks such as recommendation, search, and generation, using interleaved cross-modal user history information.
https://openreview.net/forum?id=kZEXgtMNNo
Compressor summary: Auto-Bench is a tool that uses large language models to automatically generate and assess question-answer-reasoning tasks for vision-language models, measuring their alignment with human intelligence.
https://openreview.net/forum?id=kXNJ48Hvw1
Compressor summary: The authors propose Stacked Tempering, a method that improves sampling of complex distributions using nested restricted Boltzmann machines (RBMs) and replica exchange, achieving faster transitions between data modes.
https://openreview.net/forum?id=kXHEBK9uAY
Compressor summary: The Hierarchical Diffuser is a fast and effective planning method that combines hierarchical and diffusion-based planning to overcome computational challenges and improve generalization in long-horizon tasks.
https://openreview.net/forum?id=kUveo5k1GF
Compressor summary: The paper studies how weight asymmetry and finite nudges affect gradient estimation in equilibrium propagation, a neural network algorithm for neuromorphic systems, and proposes a homeostatic objective to reduce residual bias and improve task performance.
https://openreview.net/forum?id=kUuKFW7DIF
Compressor summary: The paper proposes a hierarchical Transformer model that processes speech at multiple resolutions, achieving better performance and efficiency than existing models.
https://openreview.net/forum?id=kUCgHbmO11
Compressor summary: The paper presents a new method, SF(DA)$^2$, for adapting models to unseen domains without source data, using data augmentation based on feature space neighbors and class semantic information.
https://openreview.net/forum?id=kPrxk6tUcg
Compressor summary: The paper proposes a neuron-enhanced autoencoder matrix completion method that improves reconstruction capability and response function approximation, and shows its theoretical analysis and numerical results on synthetic and benchmark datasets.
https://openreview.net/forum?id=kOBkxFRKTA
Compressor summary: SRigL is a sparse-to-sparse method that learns structured N:M sparsity and achieves fast online inference on CPU.
https://openreview.net/forum?id=kNpSUN0uCc
Compressor summary: The paper presents a reinforcement learning approach that uses a MaxEnt Model Correction procedure to improve the accuracy of an approximate model and speed up convergence to the true value function.
https://openreview.net/forum?id=kNjrhD67LP
Compressor summary: ITIT is a novel training method for vision-language models that uses unpaired images and texts to achieve high performance with minimal paired data.
https://openreview.net/forum?id=kNPcOaqC5r
Compressor summary: The paper presents learned proximal networks (LPN), which use proximal matching to learn exact proximal operators for data-driven regularizers that recover the log-prior of the true data distribution, enabling general, unsupervised, and state-of-the-art inverse problem solutions.
https://openreview.net/forum?id=kMp8zCsXNb
Compressor summary: ASMR is a new method for encoding images and videos with a compact neural representation that reduces inference complexity and improves reconstruction quality.
https://openreview.net/forum?id=kJFIH23hXb
Compressor summary: Foldflow is a series of generative models that use the flow-matching paradigm over 3D rigid motions (SE(3)) to accurately model protein backbones, offering stability, speed, and flexibility in protein design.
https://openreview.net/forum?id=kJ0qp9Xdsh
Compressor summary: LACE is a unified continuous diffusion model for controllable layout generation that incorporates aesthetic constraints and achieves better results than existing methods.
https://openreview.net/forum?id=kIZ3S3tel6
Compressor summary: The paper introduces a new phenomenon in neural network optimization caused by heavy-tailed data and opposing signals, explaining some training dynamics and suggesting improvements to stochastic optimization methods.
https://openreview.net/forum?id=kIPEyMSdFV
Compressor summary: Reverse diffusion transforms score estimation into mean estimation, leading to a novel Monte Carlo sampling algorithm that improves upon traditional MCMC methods for Gaussian mixture models.
https://openreview.net/forum?id=kIP0duasBb
Compressor summary: The paper proposes reinforcement learning with CLIP feedback (RLCF) for adapting pre-trained vision-language models to improve their zero-shot generalization ability on various tasks.
https://openreview.net/forum?id=kILAd8RdzA
Compressor summary: The paper presents a theoretical analysis of Neural Controlled Differential Equations (NCDE) for supervised learning with irregularly sampled time series, examining how sampling affects predictions and generalization bounds.
https://openreview.net/forum?id=kGteeZ18Ir
Compressor summary: The study shows that large-scale language models can have implicit biases and perform worse in reasoning tasks when assigned personas, despite rejecting stereotypes explicitly.
https://openreview.net/forum?id=kC5nZDU5zf
Compressor summary: The paper proposes a method for embodied AI to filter visual input based on tasks, improving performance and generalization in navigation and manipulation tasks.
https://openreview.net/forum?id=kBTzlxM2J1
Compressor summary: The paper studies how to extract rules from knowledge graph ML models that are both sound and complete, which is important for safety-critical applications and legal compliance.
https://openreview.net/forum?id=kBNIx4Biq4
Compressor summary: Injective Flows learn a low-dimensional manifold and its distribution, improving efficiency and stability with a new estimator for the maximum likelihood loss.
https://openreview.net/forum?id=kB4yBiNmXX
Compressor summary: FasterViT is a new hybrid neural network that combines CNNs and ViTs, using Hierarchical Attention to achieve high image throughput and accuracy in computer vision tasks.
https://openreview.net/forum?id=k9t8dQ30kU
Compressor summary: The choice of activation function in a neural network affects its ability to learn meaningful representations from inputs and outputs, with Tanh networks favoring output structure and ReLU networks retaining more input information.
https://openreview.net/forum?id=k9SVcrmXL8
Compressor summary: The text proposes a novel approach called BECLR that improves unsupervised few-shot learning by addressing shortcomings of contrastive learning and sample bias using DyCE and OpTA modules.
https://openreview.net/forum?id=k5THrhXDV3
Compressor summary: The paper proposes a multimodal VAE model that learns data clusters and improves generative performance over existing methods.
https://openreview.net/forum?id=k581sTMyPt
Compressor summary: SUFO is a framework that enhances interpretability of fine-tuned transformer feature spaces for clinical decision-making using various analytic and visualization techniques, revealing insights about model trust and robustness.
https://openreview.net/forum?id=k1wlmtPGLq
Compressor summary: Temporal Accumulated Batch Normalization (TAB) is a new method for training Spiking Neural Networks that improves their performance by accounting for the temporal dynamics of neuron activity.
https://openreview.net/forum?id=jzzEHTBFOT
Compressor summary: The paper proposes a method to improve uncertainty estimation in CLIP by optimizing prompts during test-time using a novel metric called Average Text Feature Dispersion (ATFD).
https://openreview.net/forum?id=jznbgiynus
Compressor summary: The text discusses how large language models can be viewed as powerful predictors and compressors, providing new insights into their capabilities and demonstrating their effectiveness in various tasks.
https://openreview.net/forum?id=jxpsAj7ltE
Compressor summary: Soft MoE is a sparse Transformer that improves the performance and scalability of mixture of expert models while addressing their challenges.
https://openreview.net/forum?id=jvveGAbkVx
Compressor summary: The paper proposes a method to make classifiers abstain from predicting certain samples to achieve group fairness while maintaining original accuracy, and analyzes the feasibility and trade-offs of this approach.
https://openreview.net/forum?id=jvtmdK69KQ
Compressor summary: The paper studies how the top-K sparse softmax gating function affects density and parameter estimations in a Gaussian mixture of experts model and proposes novel loss functions and conditions for convergence.
https://openreview.net/forum?id=juuyW8B8ig
Compressor summary: The paper proposes a method to learn a language-informed visual concept representation from pre-trained models and use it to generate images with novel compositions of visual concepts.
https://openreview.net/forum?id=juE0rWGCJW
Compressor summary: ETGraph is a novel dataset that combines Ethereum transaction records with verified Twitter accounts to enhance blockchain analysis by incorporating social network data.
https://openreview.net/forum?id=jsvvPVVzwf
Compressor summary: The paper investigates the mechanisms of global unstructured magnitude pruning in deep neural networks, proves the optimality of L1 pruning by cosine similarity, and proposes a method to determine the optimal pruning amount based on parameter distribution kurtosis.
https://openreview.net/forum?id=jsWCmrsHHs
Compressor summary: The paper proposes a new DRL-guided heuristic for solving job-shop scheduling problems, using a graph representation scheme and a novel message-passing mechanism to encode complete solutions and evaluate them efficiently.
https://openreview.net/forum?id=jr03SfWsBS
Compressor summary: The authors evaluate fairness-enhancing methods for machine learning models and find that postprocessing achieves the best trade-off between fairness and accuracy, using a novel technique called unprocessing to compare different approaches.
https://openreview.net/forum?id=jp3gWrMuIZ
Compressor summary: MINT is a benchmark that evaluates large language models' ability to solve complex tasks with multiple rounds of interactions using tools and natural language feedback, revealing interesting findings about their performance and limitations.
https://openreview.net/forum?id=jnFcKjtUPN
Compressor summary: $ exttt{COPlanner}$ is a planning-driven framework for model-based reinforcement learning that improves sample efficiency and asymptotic performance by using uncertainty-aware policy-guided model predictive control to balance conservative model rollouts and optimistic environment exploration.
https://openreview.net/forum?id=jlEjB8MVGa
Compressor summary: The paper introduces SAL, a new framework that uses unlabeled data to improve out-of-distribution detection by separating and learning from candidate outliers and labeled in-distribution data.
https://openreview.net/forum?id=jkvZ7v4OmP
Compressor summary: This paper proposes a new diffusion model (DiffCSP++) for generating crystals that considers the spacegroup constraint, which is important for describing their geometry and properties, by translating it into two tractable cases.
https://openreview.net/forum?id=jjiOHEcS2c
Compressor summary: SelfHDR is a self-supervised method for HDR image reconstruction from multi-exposure images that avoids ghosting artifacts without needing labeled data.
https://openreview.net/forum?id=jj5ZjZsWJe
Compressor summary: The paper proposes two new communication compression algorithms for Federated Learning, SCALLION and SCAFCOM, which improve performance in terms of communication and computation complexities, and achieve fast convergence rates under arbitrary data heterogeneity.
https://openreview.net/forum?id=jiDsk12qcz
Compressor summary: The paper proposes knowledge fusion for combining different pre-trained language models to create a more powerful model with enhanced capabilities.
https://openreview.net/forum?id=jhPvuc7kxB
Compressor summary: The paper proposes a method to train multi-modal LMs on low-level video tasks, enabling them to perform causal and compositional spatiotemporal reasoning using *Look, Remember, Reason* process with spatiotemporal attention.
https://openreview.net/forum?id=jenyYQzue1
Compressor summary: MuSR is a novel dataset for evaluating language models on complex multistep soft reasoning tasks in natural language narratives, created using a neurosymbolic synthetic-to-natural generation algorithm and real-world domains of reasoning.
https://openreview.net/forum?id=jeMZi2Z9xe
Compressor summary: The paper proposes an algorithm to solve adversarial bandit problems with heavy-tailed losses, achieving optimal regret and improving privacy guarantees compared to existing methods.
https://openreview.net/forum?id=jd5GokdySz
Compressor summary: The paper proposes a new robustness measurement for image classification models that compares their performance with a surrogate oracle (a zoo of foundation models) using extended datasets with perturbed samples.
https://openreview.net/forum?id=jZPqf2G9Sw
Compressor summary: The authors propose a new method for protein generative modeling that incorporates dynamical properties using Normal Mode Analysis and shows its effectiveness by transforming an unconditional model into a conditional one without retraining.
https://openreview.net/forum?id=jX2DT7qDam
Compressor summary: The text discusses early-exiting dynamic neural networks (EDNN), which improve performance and uncertainty estimation in machine learning by connecting the gating mechanism and intermediate inference modules, addressing previous limitations.
https://openreview.net/forum?id=jVEoydFOl9
Compressor summary: ULTRA is an approach for learning universal and transferable graph representations that enable foundation models to infer on any knowledge graph with arbitrary entity and relation vocabularies.
https://openreview.net/forum?id=jUWktnsplU
Compressor summary: The paper proposes Hybrid Distill, a method that combines Contrastive Learning and Masked Image Modeling for representation learning, improving feature diversity and discrimination.
https://openreview.net/forum?id=jTSKkcbEsj
Compressor summary: Mixup, a data augmentation technique for deep neural networks, induces distinctive geometric configurations in the last-layer activations that enhance robustness and calibration.
https://openreview.net/forum?id=jOm5p3q7c7
Compressor summary: The paper proposes an estimator for the optimal policy of MDPs with a sample complexity matching the lower bound in the literature, using ideas from previous works.
https://openreview.net/forum?id=jODehvtTDx
Compressor summary: The paper presents a unified framework for optimal transport-based generative models and proposes a novel method to improve their performance by gradually refining the generated distribution.
https://openreview.net/forum?id=jNR6s6OSBT
Compressor summary: The authors propose a learning system that uses a small amount of real-world data to refine a simulation model and plan an accurate control strategy for robotic manipulation tasks.
https://openreview.net/forum?id=jLIUfrAcMQ
Compressor summary: The paper proposes a new debiasing approach for transformers using attention mechanism components without sensitive labels, achieving fairness and efficiency in computer vision and natural language processing tasks.
https://openreview.net/forum?id=jKhNBulNMh
Compressor summary: The paper proposes a symbolic discovery framework called Symb4CO that learns high-performance policies for exact combinatorial optimization solvers using neural networks, achieving comparable performance to GPU-based approaches with high efficiency and interpretability.
https://openreview.net/forum?id=jKTUlxo5zy
Compressor summary: The paper proposes a new image attribution method based on submodular subset selection, which improves interpretability and accuracy for different datasets and predictions.
https://openreview.net/forum?id=jKHmjlpViu
Compressor summary: OpenWebMath is a new dataset for training language models to improve their quantitative reasoning skills using 14.7B tokens of mathematical webpages.
https://openreview.net/forum?id=jJvXNpvOdM
Compressor summary: The paper proposes a new planner for embodied agents to search and rearrange objects in untidy rooms using visual input, commonsense knowledge, and graph-based representation, achieving better performance than existing methods.
https://openreview.net/forum?id=jJCeMiwHdH
Compressor summary: BioBridge is a framework that connects different unimodal models for biomedical tasks using Knowledge Graphs, improving performance and generalization.
https://openreview.net/forum?id=jId5PXbBbX
Compressor summary: The paper proposes a novel UCB-type approach for learning sequential decision-making problems with predictive state representations, achieving computational efficiency, near-optimal policy, and accurate model estimation.
https://openreview.net/forum?id=jH67LHVOIO
Compressor summary: Litcab is a lightweight calibration mechanism for large language models that improves their accuracy and trustworthiness by adding a linear layer and training only a small fraction of parameters.
https://openreview.net/forum?id=jFa5KESW65
Compressor summary: The paper proposes a new image resampling technique to defend against adversarial attacks by creating an implicit representation and using SampleNet to generate pixel-wise shifts.
https://openreview.net/forum?id=jFJPd9kIiF
Compressor summary: Least Volume is a simple regularization technique that reduces latent dimensions in autoencoders without prior knowledge of the data dimensionality, by using decoder Lipschitz continuity, and has similar PCA-like importance ordering effects on nonlinear models.
https://openreview.net/forum?id=jE8xbmvFin
Compressor summary: The study shows that large language models learn linear representations of space and time across multiple scales, suggesting they have a coherent world model rather than just memorizing data.
https://openreview.net/forum?id=j8hdRqOUhN
Compressor summary: ReSample is an algorithm that uses latent diffusion models to solve inverse problems with data consistency and a novel resampling scheme, achieving better results than current methods.
https://openreview.net/forum?id=j5JvZCaDM0
Compressor summary: FISOR is a safe offline reinforcement learning method that maximizes rewards and minimizes risks by translating hard safety constraints to identifying the largest feasible region given the offline dataset.
https://openreview.net/forum?id=j511LaqEeP
Compressor summary: Non-exchangeable conformal risk control is a flexible framework for providing statistical guarantees on any monotone loss function when dealing with non-exchangeable data, such as in change points or time series.
https://openreview.net/forum?id=j4VMrwgn1M
Compressor summary: The paper proposes a curriculum-enhanced attention distillation method for semi-supervised node classification using Graph Transformers, which improves performance on seven public graph benchmarks.
https://openreview.net/forum?id=j0ZvKSNZiP
Compressor summary: ContextRef is a benchmark for evaluating referenceless metrics that measure image description quality by assessing their alignment with human preferences and considering contextual factors.
https://openreview.net/forum?id=izrOLJov5y
Compressor summary: The text introduces a new way to adapt large language models for question answering and speech continuation using a pre-trained speech encoder and cross-modal training on spectrograms, which improves speaker preservation, semantic coherence, and knowledge retention.
https://openreview.net/forum?id=iynRvVVAmH
Compressor summary: The paper proposes a novel method to improve multi-task fusion using partial linearization, which allows for more effective incorporation of different tasks into a single model while maintaining efficiency in fine-tuning and inference.
https://openreview.net/forum?id=ixP76Y33y1
Compressor summary: This paper explores why neural networks perform differently when trained on medical versus natural images, proposes a metric for label sharpness, and shows how it affects generalization, representation learning, and adversarial robustness.
https://openreview.net/forum?id=ix7rLVHXyY
Compressor summary: The authors present a framework that adapts large language models to high-level program optimization using various techniques, achieving better performance than human programmers on a dataset of C++ code edits.
https://openreview.net/forum?id=itGkF993gz
Compressor summary: The paper proposes a method, MPAE-PPI, that uses microenvironment codes to predict protein-protein interactions efficiently and effectively, considering both protein sequence and structure.
https://openreview.net/forum?id=iriEqxFB4y
Compressor summary: Diverse Outlier Sampling (DOS) is a novel technique that selects diverse and informative outliers for OOD detection, improving the model's performance by creating a compact decision boundary between ID and OOD data.
https://openreview.net/forum?id=ip5LHJs6QX
Compressor summary: The paper introduces EfficientMod, a novel block design for efficient vision networks that improves accuracy-efficiency trade-offs and sets new state-of-the-art performance on several tasks.
https://openreview.net/forum?id=imZcqOrbig
Compressor summary: The paper proposes a novel method for pre-training 3D point clouds using 2D networks to learn consistent 2D projections and retain 3D features, achieving state-of-the-art results on various downstream tasks.
https://openreview.net/forum?id=ikwEDva1JZ
Compressor summary: This paper studies how transformer-based language models learn complex in-context tasks by analyzing their representations, behaviors, and capabilities in synthetic problems.
https://openreview.net/forum?id=ikmuHqugN7
Compressor summary: The paper proposes a method to improve training convex neural networks by using a factorization technique that makes it computationally feasible and ensures global optima, while also showing that linear self-attention with many heads is free of spurious local minima.
https://openreview.net/forum?id=ikX6D1oM1c
Compressor summary: NeuralCSA is a neural framework for causal sensitivity analysis that can handle various sensitivity models, treatment types, and causal queries with theoretical guarantees.
https://openreview.net/forum?id=ijoqFqSC7p
Compressor summary: The paper proposes FreeNoise, a tuning-free and time-efficient paradigm to enhance text-driven video generation models by rescheduling noises and injecting motions for multiple texts.
https://openreview.net/forum?id=ijK5hyxs0n
Compressor summary: Graph Metanetworks (GMNs) use graph neural networks to process graphs representing input neural networks, generalizing to various neural architectures and accounting for symmetries and geometry of parameter spaces.
https://openreview.net/forum?id=igfDXfMvm5
Compressor summary: USB-NeRF is a method that corrects rolling shutter distortions and recovers accurate camera motion trajectory for NeRF-based novel view synthesis using a physical image formation model.
https://openreview.net/forum?id=if2vRbS8Ew
Compressor summary: This paper shows that first-order adaptive network induction (ANIL), a model-agnostic meta-learning method, learns linear shared representations with overparametrized linear two-layer networks and performs well on new tasks after one gradient step.
https://openreview.net/forum?id=icTZCUbtD6
Compressor summary: The paper introduces a fine-grained taxonomy of hardness types and presents H-CAT, a toolkit for comprehensive and quantitative benchmarking of Hardness Characterization Methods (HCMs).
https://openreview.net/forum?id=iayEcORsGd
Compressor summary: The paper proposes a new method called Epitopological Sparse Meta-deep Learning (ESML) that uses brain-inspired learning paradigms to train ultra-sparse artificial neural networks, achieving better performance than fully connected networks on some tasks and reducing the number of neurons during training.
https://openreview.net/forum?id=iad1yyyGme
Compressor summary: CausalTime is a pipeline that generates realistic time-series with ground truth causal graphs for evaluating and predicting TSCD algorithms' performance on real data.
https://openreview.net/forum?id=ia9fKO1Vjq
Compressor summary: The paper explores nonlinear latent causal models and proposes a method to learn them with partial identifiability and consistent representations.
https://openreview.net/forum?id=iZgECfyHXF
Compressor summary: This paper studies online nonconvex optimization with limited feedback, deriving lower bounds on the local regret and showing that classical algorithms are optimal in certain settings.
https://openreview.net/forum?id=iX1RjVQODj
Compressor summary: CPL is an off-policy algorithm for learning optimal policies from human preferences without using reward functions or reinforcement learning, addressing limitations of previous methods in high-dimensional and sequential RLHF problems.
https://openreview.net/forum?id=iTTZFKrlGV
Compressor summary: The paper introduces GGF, a domain adaptation method that generates intermediate domains with labels using Wasserstein gradient flow and Langevin algorithm, and uses classifier-based and sample-based potentials to ensure quality.
https://openreview.net/forum?id=iTFdNLHE7k
Compressor summary: Ferumal flow is a new kernel-based normalizing flow model that can estimate and generate densities efficiently and accurately, especially when data is scarce.
https://openreview.net/forum?id=iPWxqnt2ke
Compressor summary: This paper shows that policy gradient methods can be more efficient by using low-dimensional subspaces of gradients in reinforcement learning despite changing data distributions.
https://openreview.net/forum?id=iIT02bAKzv
Compressor summary: ECoFLaP is a two-stage pruning approach for large vision-language models that uses global importance scores to achieve better model compression and improve multimodal tasks.
https://openreview.net/forum?id=iI7hZSczxE
Compressor summary: DisCo learns disentangled representations of home appliance electricity usage, considering attribute correlations, to help users understand and optimize consumption for a reduced carbon footprint.
https://openreview.net/forum?id=iHcTLIor0m
Compressor summary: The text discusses poly-view tasks in contrastive learning, where multiple related views are used to improve representation learning with information maximization and sufficient statistics, leading to better performance than SimCLR on ImageNet1k.
https://openreview.net/forum?id=iCNOK45Csv
Compressor summary: The study proposes new trigger pattern generation methods for dataset distillation that can execute effective and resilient backdoor attacks in deep learning.
https://openreview.net/forum?id=iAYIRHOYy8
Compressor summary: The text introduces a novel approach to learn stable neural networks for autonomous robot control using a variational autoencoder that encodes high-dimensional and full-pose dynamics and beats the current methods.
https://openreview.net/forum?id=iAW2EQXfwb
Compressor summary: The paper proposes an ensemble reinforcement learning method to generate diverse game levels with a selector model and policy regularization.
https://openreview.net/forum?id=i9wDX850jR
Compressor summary: The paper investigates why neural networks choose certain strategies for algebraic tasks and shows that margin maximization and Fourier features are key factors in their learning process.
https://openreview.net/forum?id=i9Vs5NGDpk
Compressor summary: The paper applies random matrix theory to optimize the tuning of regularization and sketching parameters in ridge regression ensembles, enabling efficient prediction risk estimation and construction of prediction intervals.
https://openreview.net/forum?id=i9K2ZWkYIP
Compressor summary: The paper investigates how reducing parameters in Transformers affects their performance on large datasets and identifies optimal sparsity levels that depend on data size and other factors.
https://openreview.net/forum?id=i8PjQT3Uig
Compressor summary: The paper proposes a linear regression model with nonlinear random features and a sparse locality sensitive encoding for efficient incremental updates in model-based reinforcement learning, achieving follow-the-leader behavior and outperforming neural networks in continuous control tasks.
https://openreview.net/forum?id=i7LCsDMcZ4
Compressor summary: The paper introduces two methods to generate saliency maps for spiking neural networks, improving object and action recognition in event-based vision sensors.
https://openreview.net/forum?id=i2Phucne30
Compressor summary: The paper shows that deep learning models align bias and variance at a sample level, and provides theoretical explanations for this phenomenon based on calibration and neural collapse theories.
https://openreview.net/forum?id=hywpSoHwgX
Compressor summary: The paper proposes a new camouflaged object detection method that uses adversarial training to generate more challenging camouflage and a novel detection algorithm that excavates internal coherence and removes false predictions.
https://openreview.net/forum?id=hv3SklibkL
Compressor summary: The Graph Parsing Network (GPN) adapts to each graph's structure to learn an efficient and informative pooling structure for graph and node classification tasks.
https://openreview.net/forum?id=huGECz8dPp
Compressor summary: The paper introduces a framework to analyze deep neural networks' training process using the Information Bottleneck principle and compresses high-dimensional vectors for mutual information estimation.
https://openreview.net/forum?id=hss35aoQ1Y
Compressor summary: InstructDET is a method for referring object detection that uses diverse user instructions and foundation models to train on more data, improving performance on standard REC datasets and object detection tasks.
https://openreview.net/forum?id=hp4yOjhwTs
Compressor summary: The paper proposes a curriculum learning method for Reinforcement Learning that uses causal reasoning to ensure optimal decision rules are shared across related tasks and the target task.
https://openreview.net/forum?id=homn1jOKI5
Compressor summary: Conformal prediction can create uncertainty sets for transductive node-classification models without sacrificing coverage or efficiency, and it works for both node-exchangeable and edge-exchangeable graphs.
https://openreview.net/forum?id=hnrB5YHoYu
Compressor summary: The authors present a finetuning method for text-to-image diffusion models that aligns generated images with a target distribution, reducing various biases and allowing flexibility in controlling different attributes.
https://openreview.net/forum?id=hmv1LpNfXa
Compressor summary: Polynormer is a polynomial-expressive graph transformer model with linear complexity that achieves high performance on various graph datasets.
https://openreview.net/forum?id=hkSjjs4o5d
Compressor summary: The authors present a differentially private algorithm for clustering well-clustered graphs, which performs nearly as well as non-private algorithms and outperforms pure $\epsilon$-DP algorithms.
https://openreview.net/forum?id=hj9ZuNimRl
Compressor summary: The paper proposes a neural PDE solver with a novel data-free mesh adapter that can create adaptive and dynamic meshes for physical system modeling without optimal mesh data, improving accuracy and efficiency.
https://openreview.net/forum?id=hiHZVUIYik
Compressor summary: The paper presents a new toolkit for analyzing general neural networks with improved generalization bounds and ease of computation.
https://openreview.net/forum?id=hgehGq2bDv
Compressor summary: GPAvatar is a framework that reconstructs 3D head avatars from images using a dynamic point-based expression field and Multi Tri-planes Attention module for precise expression control and multi-view consistency.
https://openreview.net/forum?id=he6mX9LTyE
Compressor summary: Kosmos-G is a model that uses MLLMs and CLIP to generate images from vision-language inputs, enabling zero-shot multi-entity subject-driven generation without modifying the image decoder.
https://openreview.net/forum?id=hbHwZYqk9T
Compressor summary: FedP3 is a flexible federated learning framework that addresses client-side model heterogeneity by privacy-friendly network pruning.
https://openreview.net/forum?id=hWS4MueyzC
Compressor summary: The text introduces a new benchmark for testing few-shot visual reasoning called Bongard-OpenWorld, which uses real-world images and open-world concepts to challenge current algorithms and human performance.
https://openreview.net/forum?id=hTEGyKf0dZ
Compressor summary: The text discusses how fine-tuning large language models can compromise their safety alignment, even with small numbers of adversarial examples or common datasets, and calls for better safety protocols.
https://openreview.net/forum?id=hQVCCxQrYN
Compressor summary: Plan-Seq-Learn (PSL) is a new method that uses large language models and motion planning to help robots learn high-level policies and fine-grained control actions from raw visual input, achieving over 80% success on various challenging robotics tasks.
https://openreview.net/forum?id=hOxgrGM63n
Compressor summary: The paper proposes an improved method to sample from smooth strongly log-concave distributions using randomized midpoint discretization, and provides new guarantees for kinetic Langevin process with Euler discretization.
https://openreview.net/forum?id=hORCalGn3Z
Compressor summary: The paper proposes a unified framework for designing and analyzing efficient local training methods for distributed variational inequality problems in machine learning with improved communication complexity.
https://openreview.net/forum?id=hOMVq57Ce0
Compressor summary: The paper proposes HyperCombinator, a piecewise-linear neural architecture for interpretable policies in reinforcement learning, and evaluates it in control and navigation tasks.
https://openreview.net/forum?id=hNhwSmtXRh
Compressor summary: Lemur and Lemur-Chat are versatile language agents that excel in natural language and coding tasks, bridging the gap between humans and machines.
https://openreview.net/forum?id=hLZQTFGToA
Compressor summary: This paper analyzes contrastive learning's equivalence to spectral clustering and introduces a new loss function called Kernel-InfoNCE for improved performance in CLIP model.
https://openreview.net/forum?id=hILVmJ4Uvu
Compressor summary: TWOSOME is a framework that uses large language models for efficient decision-making in embodied environments via reinforcement learning, improving sample efficiency and generalization.
https://openreview.net/forum?id=hI18CDyadM
Compressor summary: The paper proposes a local motion deblurring method that adaptively restores high-resolution images using a vision Transformer with adaptive window pruning, achieving better performance and efficiency than existing methods.
https://openreview.net/forum?id=hESD2NJFg8
Compressor summary: The text introduces LLM-GNN, a label-free node classification pipeline that combines GNNs and LLMs to annotate nodes efficiently and improve performance.
https://openreview.net/forum?id=hD3sGVqPsr
Compressor summary: The paper proposes a novel framework for deep imbalanced clustering that uses pseudo-label generation based on optimal transport problems to learn from data with class distribution imbalance.
https://openreview.net/forum?id=hCrFG9cyuC
Compressor summary: PolyVoice is a language model-based framework for speech-to-speech translation that uses three decoder-only language models to improve translation and audio quality, outperforming the state-of-the-art encoder-decoder model.
https://openreview.net/forum?id=hB7SlfEmze
Compressor summary: The paper presents a new method, PhyloGFN, that uses generative flow networks to sample from the complex space of phylogenetic trees and evaluate its performance on real data.
https://openreview.net/forum?id=hB2hXtxIPH
Compressor summary: Greedy Sequential Execution (GSE) is a new approach to handle cooperative agent tasks by learning optimal policies for both homogeneous and heterogeneous scenarios using individual utility functions and greedy marginal contributions.
https://openreview.net/forum?id=hAYHmV1gM8
Compressor summary: FedWon is a novel method for federated learning that improves accuracy by eliminating normalization layers and reparameterizing convolution layers, especially in multi-domain scenarios with diverse feature distributions.
https://openreview.net/forum?id=h922Qhkmx1
Compressor summary: The authors propose a generative model for music creation and separation using diffusion and Dirac likelihood methods, achieving competitive results on a standard dataset.
https://openreview.net/forum?id=h8GeqOxtd4
Compressor summary: The paper proposes a mathematical framework to analyze and improve gradient-based learning of the score function in diffusion models using a novel neural network architecture and kernel regression techniques.
https://openreview.net/forum?id=h7nOCxFsPg
Compressor summary: GSPNs are a new probabilistic framework for learning graphs that can efficiently answer probabilistic queries, outperforming popular neural models in various scenarios.
https://openreview.net/forum?id=h7DGnWGeos
Compressor summary: ARP is a framework that balances query costs and route quality evaluation for retrosynthetic planning, outperforming existing approaches by 6.2% in route quality and reducing query cost by 12.8%.
https://openreview.net/forum?id=h6Tz85BqRI
Compressor summary: VQGraph uses a structure-aware tokenizer to learn a new graph representation space that enables knowledge transfer from GNN to MLP for efficient graph data analysis.
https://openreview.net/forum?id=h57gkDO2Yg
Compressor summary: The paper proposes a new way to distill an unlabeled dataset into small synthetic samples for efficient self-supervised learning, addressing biases in gradient estimation and minimizing randomness.
https://openreview.net/forum?id=h4pNROsO06
Compressor summary: The paper proposes a unifying framework for deep learning-based methods to sample from unnormalized target densities using the Schrödinger bridge problem, and introduces a new loss function called log-variance loss that improves performance.
https://openreview.net/forum?id=h1sFUGlI09
Compressor summary: DFormer is a novel framework that pretrains the backbone using image-depth pairs, improves RGB-D encoding with a sequence of tailored blocks, and achieves state-of-the-art performance on RGB-D segmentation tasks with less computational cost.
https://openreview.net/forum?id=h05eQniJsQ
Compressor summary: The authors investigate why interval bound propagation (IBP), an imprecise but effective method for training robust neural networks, works so well and find that it provides strong regularization and exact bounds in certain conditions.
https://openreview.net/forum?id=gzqrANCF4g
Compressor summary: The paper introduces a new video tokenizer that enables large language models to outperform diffusion models in image and video generation tasks, as well as in video compression and action recognition.
https://openreview.net/forum?id=gzYgsZgwXa
Compressor summary: The paper introduces a new method, SAMP, to generate more rigorous and clear interpretations of deep neural networks by finding near-optimal manipulation paths based on the Concentration Principle.
https://openreview.net/forum?id=gzT61ziSCu
Compressor summary: The paper presents a method to differentiate higher-order functions using JAX, enabling automatic differentiation with a simple syntax and functional gradients as callable python functions.
https://openreview.net/forum?id=gyfXuRfxW2
Compressor summary: Neural networks can solve positivity optimization and certification problems for polynomials faster and accurately while leveraging the $SL(2,\mathbb{R})$ group equivariance.
https://openreview.net/forum?id=gxhRR8vUQb
Compressor summary: The paper proposes a novel metric for learning mesh deformation using sliced Wasserstein distance on meshes represented as probability measures, which overcomes limitations of the set-based approach and provides a fast statistical rate for approximating surfaces.
https://openreview.net/forum?id=gx2BT0a9MQ
Compressor summary: ZeRO++ improves the efficiency and scalability of large language models by reducing communication volume and overheads in low-bandwidth settings.
https://openreview.net/forum?id=gwbQ2YwLhD
Compressor summary: The scale of data affects the performance of structure learning algorithms, especially in non-linear cases, which can have significant implications in fields like medicine and biology.
https://openreview.net/forum?id=gwDuW7Ok5f
Compressor summary: DAEFR is a novel framework that uses a dual-branch approach to restore facial details from low-quality images by extracting domain-specific information and promoting synergy between the branches.
https://openreview.net/forum?id=guRNebwZBb
Compressor summary: The paper analyzes how prefix language models (prefixLM) outperform causal language models (causalLM) in in-context learning, showing that prefixLM converges to the optimal solution while causalLM does not.
https://openreview.net/forum?id=gtkFw6sZGS
Compressor summary: The paper introduces **Auto-J**, a generative judge model with 13B parameters that evaluates the performance of Large Language Models on tasks aligned with human needs, such as brainstorming and email writing.
https://openreview.net/forum?id=gppLqZLQeY
Compressor summary: The paper proposes a new method called Policy-Learn that learns how to select a small subset of subgraphs for graph neural networks, reducing computational complexity and improving performance.
https://openreview.net/forum?id=gn0mIhQGNM
Compressor summary: SalUn is a novel machine unlearning method that improves accuracy, stability, and cross-domain applicability by focusing on specific model weights rather than the entire model.
https://openreview.net/forum?id=gmg7t8b4s0
Compressor summary: The paper proposes CONFAIDE, a benchmark to test privacy reasoning in large language models, and shows that current models reveal private information in inappropriate contexts.
https://openreview.net/forum?id=glwwbaeKm2
Compressor summary: Vertical Federated Learning (VFL) benefits from new evaluation metrics and datasets that consider feature importance and correlation, improving algorithm performance assessment.
https://openreview.net/forum?id=gkfUvn0fLU
Compressor summary: The paper studies overoptimization in composite reward models for language quality and proposes a constrained reinforcement learning approach to balance component models and improve evaluation performance.
https://openreview.net/forum?id=gjfOL9z5Xr
Compressor summary: DyVal is a dynamic evaluation protocol that uses graphs to generate challenging reasoning tasks for large language models, revealing their limitations and potential for improvement.
https://openreview.net/forum?id=gjeQKFxFpZ
Compressor summary: The text presents a systematic framework for black-box uncertainty estimation in LLMs using prompting, sampling, and aggregation methods. It evaluates the framework on calibration and failure prediction tasks and identifies key insights and challenges.
https://openreview.net/forum?id=gjXor87Xfy
Compressor summary: Memory-based Dynamic Graph Neural Networks (MDGNNs) use memory to capture long-term dependencies, but struggle with temporal discontinuity in large temporal batches; the proposed PRES method mitigates this issue and enables faster training without sacrificing performance.
https://openreview.net/forum?id=giU9fYGTND
Compressor summary: This paper proposes FedImpro, a method to reduce client drift in federated learning by constructing similar conditional distributions for local training, improving generalization and gradient similarity.
https://openreview.net/forum?id=gfFVATffPd
Compressor summary: The paper studies how Transformer-based LLMs handle factual constraints when generating text, and proposes SAT Probe, a method that can predict errors and improve reliability.
https://openreview.net/forum?id=gd0lAEtWso
Compressor summary: OmniControl is a novel method for generating realistic human motion with flexible spatial control signals over different joints using diffusion-based models and analytic guidance.
https://openreview.net/forum?id=gctmyMiPHH
Compressor summary: Feature collapse is a phenomenon where similar learning task roles get similar representations, improving generalization and interpretability.
https://openreview.net/forum?id=gbrHZq07mq
Compressor summary: The text studies how two types of self-attention mechanisms in transformer encoders recognize different formal languages, and shows their limitations and capabilities.
https://openreview.net/forum?id=gU58d5QeGv
Compressor summary: Würstchen is a text-to-image synthesis model that uses a compact semantic image representation to guide the diffusion process, achieving competitive performance with much lower cost and faster inference than existing models.
https://openreview.net/forum?id=gT5hALch9z
Compressor summary: The paper discusses the importance of safety in instruction-tuned language models and shows that adding a small number of safety examples can improve their safety without compromising their capabilities or helpfulness.
https://openreview.net/forum?id=gPKTTAfYBp
Compressor summary: FlashFFTConv optimizes FFT convolution for long sequences, speeding up reasoning tasks and enabling larger models with the same compute budget.
https://openreview.net/forum?id=gMLQwKDY3N
Compressor summary: The paper proposes a private watermarking algorithm for large language models that uses different neural networks for generating and detecting watermarks, enhancing security and accuracy while minimizing parameters.
https://openreview.net/forum?id=gLARhFLE0F
Compressor summary: The paper introduces LUT-GEMM, a kernel for quantized matrix multiplication that eliminates the resource-intensive dequantization process and reduces computational costs, achieving significant acceleration in NLP model generation.
https://openreview.net/forum?id=gJeYtRuguR
Compressor summary: METR introduces multi-exit loss to improve the reliability of ViTs' attention scores for token reduction, making them more efficient and effective.
https://openreview.net/forum?id=gIiz7tBtYZ
Compressor summary: The paper presents a new neural network method for computing OT plans that can use extra information, work with diverse costs, and handle high-dimensional data, while providing error guarantees and an example of class-preserving mapping.
https://openreview.net/forum?id=gHAr7ZA1OL
Compressor summary: The paper proposes a method to generate out-of-distribution data for training object detectors using phase information from in-distribution data and two branches for content manipulation.
https://openreview.net/forum?id=gG38EBe2S8
Compressor summary: The paper presents a new method for image morphing called IMPUS, which uses diffusion-based models, latent space interpolation, and perceptually-uniform sampling to produce realistic and smooth results.
https://openreview.net/forum?id=gFR4QwK53h
Compressor summary: The authors propose a causal graphical model to characterize dropout error in single-cell RNA sequencing data and show that test-wise deletion can be used to integrate existing structure learning methods for gene regulatory network inference without introducing spurious relations.
https://openreview.net/forum?id=gEwKAZZmSw
Compressor summary: The paper introduces a method called VCAS that reduces computational load in neural network training by sampling unimportant computations, while controlling additional variance to maintain accuracy.
https://openreview.net/forum?id=gC6JTEU3jl
Compressor summary: BESA is a novel technique that prunes large language models by targeting error reduction at the transformer block level and allocating sparsity differently, resulting in less performance degradation and faster computation time.
https://openreview.net/forum?id=g9diuvxN6D
Compressor summary: The paper studies how sensitive instruction-tuned language models are to different phrasings of instructions and proposes a method to improve their robustness using soft prompt embedding parameters.
https://openreview.net/forum?id=g90ysX1sVs
Compressor summary: The authors propose using adaptable activation functions in reinforcement learning to enhance neural plasticity and improve performance in changing environments, such as Atari games.
https://openreview.net/forum?id=g8sGBSQjYk
Compressor summary: The paper proposes a new parameterization for second-order optimization that improves feature learning in larger deep neural networks and allows sharing hyperparameters across models of different sizes.
https://openreview.net/forum?id=g7ohDlTITL
Compressor summary: Riemannian Flow Matching (RFM) is a new method for training continuous normalizing flows on manifolds that works without simulation, divergence computation, and uses a simple premetric for target vector fields.
https://openreview.net/forum?id=g6rZtxaXRm
Compressor summary: FuDD improves image classification by generating class descriptions that better differentiate between ambiguous classes using a large language model.
https://openreview.net/forum?id=g6eCbercEc
Compressor summary: The paper presents a method for autonomously discovering manipulation concepts from unlabeled robotic demonstrations using a trainable codebook that encodes sub-trajectories based on their informativeness.
https://openreview.net/forum?id=g52tgL8jy6
Compressor summary: The paper introduces a new spiking neural network model (LM-H) that improves energy efficiency and information processing by dynamically adjusting membrane-related factors, and proposes a novel training algorithm and framework for it.
https://openreview.net/forum?id=g1S72T3FGc
Compressor summary: The paper proposes new neural network algorithms for active learning that reduce the impact of the number of classes on performance and cost, and shows their advantages in theory and practice.
https://openreview.net/forum?id=g0mlwqs8pi
Compressor summary: $\Delta$-SGD is a simple method for adjusting step sizes in federated learning that adapts to local smoothness, improving performance in different scenarios.
https://openreview.net/forum?id=fyTPWfXtcc
Compressor summary: The paper introduces invex constrained optimization theory to ensure global optima in non-convex constrained inverse problems while maintaining reconstruction quality.
https://openreview.net/forum?id=fxQiecl9HB
Compressor summary: Crystalformer is a simple transformer-based encoder architecture that predicts physical properties of materials from their crystal structures more accurately than existing models with fewer parameters and better interpretability.
https://openreview.net/forum?id=fwCoLe3TAX
Compressor summary: The authors propose a method to improve language model-based AI assistants' performance and generalization across various domains by automatically classifying data into groups, optimizing the policy for challenging groups, and adaptively adjusting exploration space.
https://openreview.net/forum?id=ft1mr3WlGM
Compressor summary: The paper proposes a new probabilistic image-text matching method with improved embeddings and optimization techniques that outperforms existing methods on various benchmarks.
https://openreview.net/forum?id=fszrlQ2DuP
Compressor summary: The paper introduces Transfer Score, a novel metric for unsupervised evaluation of UDA models, which measures spatial uniformity, transferability, and discriminability. The metric helps to choose the best UDA method, tune hyperparameters, and find optimal checkpoints without target labels.
https://openreview.net/forum?id=fsW7wJGLBd
Compressor summary: Tensor Trust is a dataset with over 100,000 prompt injection attacks and defenses for testing the robustness and security of large language models.
https://openreview.net/forum?id=frRDT6EOhg
Compressor summary: This paper proposes self-contemplation prompting strategy (SEC), which eliminates human-generated demonstrations in In-context Learning, and shows that large language models can perform well with only their own decision making capacity.
https://openreview.net/forum?id=fq1wNrC2ai
Compressor summary: FLOP is a novel algorithmic framework that tackles average-reward Markov decision processes with function approximation using confidence sets and low-switching policy updates, achieving sublinear regret in various complex AMDP models.
https://openreview.net/forum?id=fpoAYV6Wsk
Compressor summary: The authors show that insights from mechanistic interpretability can generalize across tasks, demonstrating this by reusing and modifying a circuit for different language tasks with significant improvements and explanations.
https://openreview.net/forum?id=fkrYDQaHOJ
Compressor summary: Using Koopman theory, the authors propose a method to efficiently and accurately model nonlinear dynamics in interactive environments for long-range prediction, planning, and reinforcement learning.
https://openreview.net/forum?id=fjpfCOV4ru
Compressor summary: This paper studies a general class of Markov chains derived from stochastic differential equations with arbitrary noise, inexact drift and diffusion coefficients, and improves existing bounds on their distance to the corresponding differential equations.
https://openreview.net/forum?id=fjf3YenThE
Compressor summary: The paper introduces a new algorithm that improves the performance of hard-thresholding for $\ell_0$ constrained optimization using zeroth-order gradients and variance reduction, and shows its usefulness in different applications.
https://openreview.net/forum?id=fj2E5OcLFn
Compressor summary: The paper introduces an efficient stochastic dual gradient descent algorithm for optimizing Gaussian process regression and shows its competitive performance against other methods.
https://openreview.net/forum?id=fibxvahvs3
Compressor summary: GAIA is a benchmark for testing general AI assistants' abilities by asking real-world questions that are simple for humans but challenging for most advanced AIs, aiming to measure their robustness and proficiency in various skills.
https://openreview.net/forum?id=fh8EYKFKns
Compressor summary: The paper discusses the challenges of aligning artificial general intelligence (AGI) systems with human interests, as they might learn deceptive or power-seeking behaviors from pretrained deep models.
https://openreview.net/forum?id=fgKjiVrm6u
Compressor summary: The paper introduces REFACTOR, a neural network method that learns to extract and use modular theorems from mathematical proofs, improving proof quality and efficiency.
https://openreview.net/forum?id=ff2g30cZxj
Compressor summary: The paper proposes a way to make image restoration methods produce diverse and meaningful outputs by modifying the original method or adding post-processing steps.
https://openreview.net/forum?id=fe6ANBxcKM
Compressor summary: The paper proposes two Q-Learning algorithms for federated reinforcement learning with linear regret speedup and low communication cost, using event-triggered synchronization, novel step size selection, and new concentration inequalities.
https://openreview.net/forum?id=fcqWJ8JgMR
Compressor summary: AuG-KD is a new method that helps transfer knowledge from large models without access to their training data by aligning student-domain data with teacher domain using uncertainty and generative techniques, improving performance in real-world scenarios.
https://openreview.net/forum?id=farT6XXntP
Compressor summary: This study proposes ALMA, a novel fine-tuning approach for LLMs that improves translation performance without relying on large parallel data sets, outperforming previous models with only 7B or 13B parameters.
https://openreview.net/forum?id=fW7DOHDQvF
Compressor summary: Weakly supervised learning methods aim to train accurate classifiers from unlabeled data and class priors, with RCM being a new approach that improves risk consistency and purifies supervision information.
https://openreview.net/forum?id=fVxIEHGnVT
Compressor summary: kNN-ECD and kNN-ECS methods improve program translation accuracy by using error correction and interpreting translated codes with k-nearest neighbors.
https://openreview.net/forum?id=fUtxNAKpdV
Compressor summary: Nougat is a Visual Transformer model that converts scientific PDFs into markup language for better accessibility and understanding of scientific knowledge.
https://openreview.net/forum?id=fUGhVYPVRM
Compressor summary: AWP is a plug-and-play framework that enhances desired properties in CTC-trained models by complementing the CTC loss with an additional loss term that prioritizes alignments according to the property, improving ASR latency and WER on large-scale data.
https://openreview.net/forum?id=fQHb1uZzl7
Compressor summary: The paper presents a Transformer-based network that combines feature and cost aggregation for dense matching tasks using self- and cross-attention layers, leading to improved semantic and geometric matching results.
https://openreview.net/forum?id=fLf589bx1f
Compressor summary: The text describes a new approach to program synthesis using value-based reinforcement learning and large language models, which produces accurate code with minimal reward engineering effort.
https://openreview.net/forum?id=fLXpXa7iiz
Compressor summary: The paper provides a theoretical guarantee for combining Bayesian optimization with stochastic gradient descent in bilevel frameworks, showing sublinear regret bounds and suggesting optimal configurations for generalization.
https://openreview.net/forum?id=fJNnerz6iH
Compressor summary: The text describes a problem with training hypernetworks that affects their convergence and proposes a solution called Magnitude Invariant Parametrizations (MIP) to improve stability and speed up convergence.
https://openreview.net/forum?id=fGAIgO75dG
Compressor summary: CoLiDE is a new method for learning directed acyclic graphs from data that estimates scale and noise levels, improving performance and stability over existing methods.
https://openreview.net/forum?id=fDaLmkdSKU
Compressor summary: This paper studies how dual ascent methods for constrained machine learning problems affect the optimality and feasibility of primal iterates, and provides a sub-optimality bound based on the problem's parametrization and objective curvature.
https://openreview.net/forum?id=fCeUoDr9Tq
Compressor summary: RoboShot improves the robustness and performance of pretrained models in zero-shot tasks by using language models to enhance their embeddings without supervision or fine-tuning.
https://openreview.net/forum?id=fB1iiH9xo7
Compressor summary: Grounded Point Colorization (GPC) is a pre-training method that teaches models to colorize LiDAR point clouds, improving 3D object detection for self-driving cars with limited labeled data.
https://openreview.net/forum?id=f6CBQYxXvr
Compressor summary: Pro$^2$ is a sample-efficient method that learns diverse features and interpolates them to adapt to distribution shifts using a small target dataset.
https://openreview.net/forum?id=f5H8WGLQm5
Compressor summary: UniAdapter is a method to adapt pre-trained vision-language models to various downstream tasks with minimal computational and storage costs, outperforming full fine-tuning in most cases.
https://openreview.net/forum?id=f3g5XpL9Kb
Compressor summary: LiDAR is a metric for evaluating joint embedding representations by measuring the rank of the LDA matrix associated with a surrogate self-supervised learning task.
https://openreview.net/forum?id=f1xnBr4WD6
Compressor summary: The paper proposes a method to improve object-centric deep learning models by optimizing the association of objects with distinct slots and applying them to downstream tasks.
https://openreview.net/forum?id=ezscMer8L0
Compressor summary: Conv-LoRA is a method that improves the performance of the Segment-Anything Model for image segmentation in specialized domains by adding lightweight convolutional parameters.
https://openreview.net/forum?id=ezBH9WE9s2
Compressor summary: AnyText is a diffusion-based multilingual visual text generation and editing model that produces accurate and coherent text in images, achieving state-of-the-art performance on a new large-scale dataset called AnyWord-3M.
https://openreview.net/forum?id=ey3GhWXQ97
Compressor summary: The paper investigates how adaptivity affects sample-efficiency in reinforcement learning by analyzing lower bounds on the number of batches needed for policy evaluation and best-policy identification under linear function approximation.
https://openreview.net/forum?id=extpNXo6hB
Compressor summary: The text proposes a method to improve text-to-3D generation by aligning 2D geometric priors in diffusion models with well-defined 3D shapes, resolving the multi-view inconsistency problem and achieving high generalizability.
https://openreview.net/forum?id=exKHibougU
Compressor summary: LLM-grounded Video Diffusion (LVD) uses a large language model to generate dynamic scene layouts based on text inputs, which then guide a video diffusion model for improved spatiotemporal video generation.
https://openreview.net/forum?id=eoTCKKOgIs
Compressor summary: Classical MLE using source data is near-optimal for covariate shift in well-specified models, but MWLE can be better in misspecified settings.
https://openreview.net/forum?id=eoSeaK4QJo
Compressor summary: The paper proposes a novel pruning framework that combines unstructured weight and neuron pruning to enhance energy efficiency in deep spiking neural networks without sacrificing much accuracy.
https://openreview.net/forum?id=eo9dHwtTFt
Compressor summary: Skipper is a model-based RL agent that learns and decomposes tasks into subtasks using hindsight, enabling better generalization and sparse decision-making.
https://openreview.net/forum?id=elMKXvhhQ9
Compressor summary: ConsisGAD is a new model for graph anomaly detection that uses consistency training, learnable data augmentation, and homophily distribution to handle limited supervision and class imbalance.
https://openreview.net/forum?id=ekz1hN5QNh
Compressor summary: The paper introduces HCNN, a novel fully hyperbolic convolutional neural network for computer vision tasks, which leverages the benefits of hyperbolic geometry to better represent intricate structures in images.
https://openreview.net/forum?id=ekeyCgeRfC
Compressor summary: The study investigates Transformers' and attention-free models' abilities to learn algorithms for various function classes, showing that Transformers adaptively select sample-efficient algorithms while LLMs can generalize beyond their training data.
https://openreview.net/forum?id=eiF7TU1E8E
Compressor summary: The paper proposes and tests a new GAN training method called Slicing Adversarial Network (SAN) that uses sliced optimal transport to ensure the generator's distribution matches the target distribution more closely.
https://openreview.net/forum?id=eiC4BKypf1
Compressor summary: The paper shows how finetuning large language models on psychological data allows them to accurately model and predict human behavior in various decision-making tasks.
https://openreview.net/forum?id=efeBC1sQj9
Compressor summary: SEPT is a self-supervised learning framework that helps autonomous vehicles predict motions in complex traffic scenes by capturing kinematics, road structure, and agent interactions using masking-reconstruction tasks.
https://openreview.net/forum?id=efFmBWioSc
Compressor summary: WebGUM is a novel web navigation agent that leverages vision-language models and offline training to surpass previous methods and some online agents on multiple tasks.
https://openreview.net/forum?id=eepoE7iLpL
Compressor summary: The paper proposes a new information aggregation module for neural network-based set functions that incorporates superset information to improve performance in tasks like drug discovery.
https://openreview.net/forum?id=eZneJ55mRO
Compressor summary: The paper proposes a method to connect algebraic languages and GNNs using CFGs, reduces redundancies in rules, and shows that their derived G$^2$N$^2$ model performs better than other 3-WL GNNs on various tasks.
https://openreview.net/forum?id=eY7sLb0dVF
Compressor summary: Koopman VAE (KVAE) is a new generative framework for realistic time series data generation that leverages Koopman theory and spectral tools to incorporate domain knowledge and improve stability, outperforming existing GAN and VAE methods on various benchmarks.
https://openreview.net/forum?id=eUgS9Ig8JG
Compressor summary: The paper proposes SaNN, a scalable graph representation learning model with constant run-time and memory, which is powerful, equivariant, and achieves state-of-the-art performance.
https://openreview.net/forum?id=eT6oLkm1cm
Compressor summary: ADR is a simple method that generates soft labels for adversarial training to improve model robustness without extra computation or pre-trained models.
https://openreview.net/forum?id=ePOjNlOjLC
Compressor summary: The text proposes a method called Cyclic One-Way Diffusion (COW) to control the direction of diffusion phenomenon in image generation, allowing for more versatile and precise customization without additional training.
https://openreview.net/forum?id=eOCvA8iwXH
Compressor summary: Neural Fourier Transform (NFT) is a framework for learning data structure using symmetry without assuming explicit knowledge of the group's action on data, and shows theoretical and practical benefits.
https://openreview.net/forum?id=eNoiRal5xi
Compressor summary: UDIM is a new objective for domain generalization that minimizes inconsistency between source and unknown domains by perturbing data and aligning flat minima.
https://openreview.net/forum?id=eMNN0wIyVw
Compressor summary: The paper proposes an offline trajectory augmentation approach to enhance off-policy evaluation in human-involved scenarios using sub-trajectory mining and diversifying state-action space coverage.
https://openreview.net/forum?id=eMHn77ZKOp
Compressor summary: The paper proposes an algorithm for a new combinatorial multi-armed bandit problem with index feedback and provides a regret bound that compares favorably to existing methods.
https://openreview.net/forum?id=eJHnSg783t
Compressor summary: DIFFTACTILE is a high-fidelity physics-based tactile simulation system for robotic manipulation with differentiable components and contact modeling.
https://openreview.net/forum?id=eJ0dzPJq1F
Compressor summary: RPI is a novel algorithm that interleaves imitation learning and reinforcement learning to improve policy robustness and sample efficiency, adapting to different oracles and states.
https://openreview.net/forum?id=eFWG9Cy3WK
Compressor summary: The paper proposes M-SMoE and MC-SMoE, two methods that improve the scalability of neural networks using SMoE by merging and compressing experts based on their activation frequency and routing statistics.
https://openreview.net/forum?id=eCGpNGDeNu
Compressor summary: The paper introduces WAKER, a method to train robust world models from reward-free exploration by selecting environments with high world model error, leading to better adaptation, efficiency, and generality.
https://openreview.net/forum?id=eBeECjacpw
Compressor summary: The paper introduces a novel implicit representation, PoRF, which uses an MLP to regress pose updates and enhances pose supervision using epipolar geometry loss, leading to improved camera pose accuracy and neural surface reconstruction in challenging real-world scenarios.
https://openreview.net/forum?id=eAKmQPe3m1
Compressor summary: PixArt-α is a fast and high-quality text-to-image model that uses less training data and reduces CO2 emissions compared to existing models.
https://openreview.net/forum?id=e4xS9ZarDr
Compressor summary: Lion, a new AI optimizer, combines elements from existing algorithms and has promising results, but lacks theoretical foundation; a study provides a Lyapunov function to analyze its dynamics and improve it.
https://openreview.net/forum?id=e4FG5PJ9uC
Compressor summary: LASI is a perceptual similarity metric based on pixel-level weighted least squares that performs well on image quality assessment tasks without training data or deep features.
https://openreview.net/forum?id=e2YOVTenU9
Compressor summary: ArchLock is a novel algorithm that uses neural architecture search and zero-cost proxies to create DNN models with low transferability, reducing the vulnerability of current defense strategies.
https://openreview.net/forum?id=e1vqloonRy
Compressor summary: The paper studies how gradient flow can learn single-index models in symmetric neural networks and prove its efficiency using an adapted information exponent.
https://openreview.net/forum?id=dyrGMhicMw
Compressor summary: Weight selection is a method that initializes smaller models with a subset of weights from a larger pretrained model, improving their performance and reducing training time.
https://openreview.net/forum?id=dyG2oLJYyX
Compressor summary: DiffusionNAG is a novel conditional Neural Architecture Generation framework based on diffusion models that generates task-optimal architectures efficiently for diverse tasks by sampling from a region more likely to satisfy desired properties.
https://openreview.net/forum?id=dwzLn78jq7
Compressor summary: The paper proposes a new method for estimating Lipschitz constants in deep learning using eigenvalue optimization, which is more scalable and efficient than standard SDP algorithms.
https://openreview.net/forum?id=duyA42HlCK
Compressor summary: The authors propose HyperHuman, a model that generates realistic and diverse human images by capturing correlations between appearance and structure at multiple scales using a large-scale dataset and a unified framework.
https://openreview.net/forum?id=duZANm2ABX
Compressor summary: The BadEdit attack framework injects backdoors into large language models using a lightweight editing technique, requiring minimal data and time, while preserving model performance and robustness.
https://openreview.net/forum?id=dsUB4bst9S
Compressor summary: Small transformers can efficiently learn arithmetic operations using the next-token prediction objective when trained on high-quality, instructive data with intermediate step results.
https://openreview.net/forum?id=dpcVXiMlcv
Compressor summary: OIR is a new image editing paradigm that adapts the number of inversion steps for each editing pair to achieve optimal results and fine-grained editing of objects in images.
https://openreview.net/forum?id=dpDw5U04SU
Compressor summary: The minimum width needed for deep neural networks to approximate any function using ReLU-like activations is $\max\{d_x,d_y,2\}$, where $d_x$ and $d_y$ are the input and output dimensions.
https://openreview.net/forum?id=dp27P5HBBt
Compressor summary: The paper proposes a novel framework called PDF that models 2D temporal variations for long-term time series forecasting by decoupling, capturing short- and long-term variations, and aggregating them.
https://openreview.net/forum?id=dnqPvUjyRI
Compressor summary: SemiReward is a framework that uses reward scores to evaluate and filter high-quality pseudo labels in semi-supervised learning, improving performance and convergence speed.
https://openreview.net/forum?id=dm8e7gsH0d
Compressor summary: The paper proposes an optimization framework to improve out-of-distribution detection in deep learning models by optimizing feature shaping functions using only in-distribution data.
https://openreview.net/forum?id=dlIMcmlAdk
Compressor summary: NFSD improves text-to-content generation by reducing noise and over-smoothing using a smaller CFG scale, while maintaining realism and prompt compliance.
https://openreview.net/forum?id=dl0u4ODCuW
Compressor summary: The paper introduces a new algorithm, retro-fallback, for retrosynthesis that considers uncertainty and improves the chances of creating a viable synthesis plan.
https://openreview.net/forum?id=djM3WzpOmK
Compressor summary: The paper introduces a new deep learning architecture called neural snowflake that improves latent graph inference by adaptively implementing fractal-like metrics on high-dimensional data, without the curse of dimensionality and random search.
https://openreview.net/forum?id=di52zR8xgf
Compressor summary: SDXL is a text-to-image model with a larger backbone, new conditioning schemes, and a refinement model that improves its quality and competes with state-of-the-art image generators.
https://openreview.net/forum?id=dhLIno8FmH
Compressor summary: The paper presents a self-supervised method to learn image representations from EEG signals and achieve above-chance results in object recognition tasks, while exploring the biological plausibility of the approach.
https://openreview.net/forum?id=de1218PoEl
Compressor summary: DumBO is a new decentralized Bayesian Optimization algorithm that relaxes assumptions on the function to optimize and reduces over-exploration, making it suitable for high-dimensional problems with complex structure.
https://openreview.net/forum?id=dcjtMYkpXx
Compressor summary: Ensemble-based conservative optimization methods can reduce overoptimization and improve the performance of large language models fine-tuned with human feedback.
https://openreview.net/forum?id=dbQH9AOVd5
Compressor summary: The paper introduces I-STAR, a novel regularization method for LLMs that accurately measures and adjusts isotropy in embedding space during training to improve model performance.
https://openreview.net/forum?id=daEqXJ0yZo
Compressor summary: The paper presents a novel generative model for human motion stylization that uses pretrained autoencoders' latent space to learn probabilistic style space from motions, enabling users to select or sample style cues while preserving content.
https://openreview.net/forum?id=dTpbEdN9kr
Compressor summary: The paper presents three composition methods based on diffusion priors for generating human motion with more control and flexibility, addressing challenges like long sequences, two-person interactions, and detailed editing.
https://openreview.net/forum?id=dTlKCQuuxP
Compressor summary: The paper introduces Phase-based neural polynomial Gabor fields (Phase-PGF), a method to analyze large motions in 2D and 3D dynamic scenes using low-dimensional time-varying phases.
https://openreview.net/forum?id=dQVtTdsvZH
Compressor summary: The content-motion latent diffusion model (CMD) is an efficient video generation method that combines a pretrained image diffusion model and a lightweight motion model, achieving high quality and fast speed.
https://openreview.net/forum?id=dPHLbUqGbr
Compressor summary: The paper presents a geometric method to optimize edge attributes for convolutional networks using homogeneous spaces, leading to an efficient 3D group convolutional network that achieves state-of-the-art results on various benchmarks.
https://openreview.net/forum?id=dONpC9GL1o
Compressor summary: The paper explains why truncation sampling heuristics like nucleus sampling work well for language generation and proposes a new strategy based on softmax bottleneck to improve expressiveness without relying on thresholds.
https://openreview.net/forum?id=dNe1T0Ahby
Compressor summary: The paper explores the performance of dual-encoder models on extreme multi-label classification tasks and proposes a new loss function to optimize for Recall@k metrics.
https://openreview.net/forum?id=dN4vpVTvWX
Compressor summary: The paper proposes Texture UV Radiance Fields, a method for generating and controlling realistic textures on 3D models independent of their shapes.
https://openreview.net/forum?id=dLrhRIMVmB
Compressor summary: NISQ-TDA is a quantum machine learning algorithm for high-dimensional data analysis with proven speedup and low noise sensitivity.
https://openreview.net/forum?id=dLoAdIKENc
Compressor summary: The paper analyzes the strengths and weaknesses of AI-image detectors, especially watermarking methods, and their vulnerabilities to attacks that can falsely flag real images as fake or vice versa.
https://openreview.net/forum?id=dKlxDx2SoS
Compressor summary: The paper proposes QNet, a method that uses quaternion networks to enhance multimodal pre-trained models for better semantic alignment across diverse modalities.
https://openreview.net/forum?id=dKl6lMwbCy
Compressor summary: The choice of feedback protocol, either ratings or rankings, significantly affects how well large language models are aligned with human values and intents.
https://openreview.net/forum?id=dHng2O0Jjr
Compressor summary: The authors introduce ToolLLM, a framework for enhancing large language models' tool-use capabilities by constructing a dataset (ToolBench) and fine-tuning LLaMA with a neural API retriever.
https://openreview.net/forum?id=dGH4kHFKFj
Compressor summary: GenCorres is a new unsupervised method for joint shape matching that learns a mesh generator to fit a large synthetic shape collection, preserving local geometric structures and outperforming existing techniques.
https://openreview.net/forum?id=dFcXJgnrGB
Compressor summary: The paper proposes PlaSma, a two-pronged approach to enhance small language models with procedural knowledge and planning capabilities, and shows they can outperform larger models in planning and replanning tasks.
https://openreview.net/forum?id=dEz3ge8QSo
Compressor summary: The paper introduces a new risk-sensitive MDP formulation, shows its equivalence with regularized robust MDPs, derives policy gradient theorems, proves global convergence, and proposes an efficient offline learning algorithm for specific problems.
https://openreview.net/forum?id=dCHbFDsCZz
Compressor summary: The paper proposes a new problem formulation and algorithm for classification with reject option, shows strong consistency guarantees, and evaluates it on decontextualization task achieving high coverage.
https://openreview.net/forum?id=d94x0gWTUX
Compressor summary: Themis is a tool-augmented preference modeling approach that improves large language models' functionality by enabling them to use external tools for tasks like arithmetic, code execution, and factual lookup.
https://openreview.net/forum?id=d8w0pmvXbZ
Compressor summary: The paper investigates training instability in Transformer-based models at different scales and suggests interventions to improve stability and loss performance across learning rate variations.
https://openreview.net/forum?id=d6tUsZeVs7
Compressor summary: The paper introduces a novel method that combines energy-based models and entropy-regularized optimal transport to improve generative modeling and provides theoretical and empirical results.
https://openreview.net/forum?id=d6H4RBi7RH
Compressor summary: The study examines how weight space ensembles like WiSE-FT use diverse spurious features for better out-of-distribution generalization and proposes BANG to address overconfidence issues in such methods.
https://openreview.net/forum?id=d4uL2MSe0z
Compressor summary: The paper proposes a method to reduce the number of trainable parameters in deep transformer networks using Reinforcement Learning for layer selection and weight sharing.
https://openreview.net/forum?id=d4UiXAHN2W
Compressor summary: The paper introduces LLaMA-Adapter, a method to efficiently fine-tune LLaMA for instruction following using self-attention layers with learnable zero gating, achieving similar performance to Alpaca with fewer parameters and extending to multi-modal tasks.
https://openreview.net/forum?id=d3xKPQVjSc
Compressor summary: The paper proposes a new method to estimate how much dimensionality reduction in CATE estimation can introduce bias due to confounding.
https://openreview.net/forum?id=cxfPefbu1s
Compressor summary: The paper addresses the problem of disguised procedural unfairness in automated decision-making and proposes a framework to separate objectionable from neutral data generating components.
https://openreview.net/forum?id=cuAxSHcsSX
Compressor summary: The paper proposes a new algorithmic framework for collaborative federated linear contextual bandits under differential privacy that addresses privacy, regret, and communication cost issues.
https://openreview.net/forum?id=csukJcpYDe
Compressor summary: The paper introduces TTPI, an ADP algorithm that uses low-rank tensor approximation to efficiently control hybrid systems in robotics.
https://openreview.net/forum?id=cphhnHjCvC
Compressor summary: The paper presents a method to improve goal-oriented visual navigation in complex scenes using pretext tasks that help learn correspondence and goal detection from image pairs.
https://openreview.net/forum?id=coIaBY8EVF
Compressor summary: The study investigates how platforms can learn to display information in a way that reduces congestion and improves social welfare in online marketplaces by proposing a differentiable proxy of welfare based on consumer choices.
https://openreview.net/forum?id=cmcD05NPKa
Compressor summary: The paper trains small transformers to find the greatest common divisor (GCD) of two numbers using lists of divisors and shows that their predictions are explainable, but performance depends on the training distribution of GCD.
https://openreview.net/forum?id=ceATjGPTUD
Compressor summary: The authors propose a method to improve automatic speech recognition by using a large language model that learns to correct errors and remove noise from speech transcriptions.
https://openreview.net/forum?id=cdUpf6t6LZ
Compressor summary: This paper introduces a new data set for evaluating robustness against adversarial attacks and provides a generalization theory for searching architectures that balance clean and robust accuracy.
https://openreview.net/forum?id=ccxD4mtkTU
Compressor summary: LLM-generated misinformation is harder to detect than human-written misinformation, which may increase its potential harm.
https://openreview.net/forum?id=cc8h3I3V4E
Compressor summary: The paper introduces a new loss function for finding approximate Nash equilibria in normal-form games using non-convex stochastic optimization techniques, which perform better than existing methods.
https://openreview.net/forum?id=cbv0sBIZh9
Compressor summary: The paper proposes a unified multi-task diffusion model that can generate different types of data simultaneously and improve generalization by aggregating information from multiple tasks.
https://openreview.net/forum?id=caW7LdAALh
Compressor summary: The paper introduces IdentityChain, a framework to evaluate and improve the self-consistency of code large language models (LLMs) across different tasks, as it affects their trustworthiness.
https://openreview.net/forum?id=cZttUMTiPL
Compressor summary: The text introduces a new method for using stable probability distributions in neural networks, which improves uncertainty quantification and outperforms existing methods like moment matching.
https://openreview.net/forum?id=cXbnGtO0NZ
Compressor summary: The paper introduces latent 3D graph diffusion, a fast and effective way to generate 3D graphs with symmetry and quality in a low-dimensional latent space, and applies it to conditional generation of molecules with better properties and validity.
https://openreview.net/forum?id=cWiEN1plhJ
Compressor summary: The paper proposes a new method to detect machine-written text using style features from human authors, which works even against state-of-the-art language models like Llama 2 and GPT-4.
https://openreview.net/forum?id=cWdAYDLmPa
Compressor summary: The paper presents a new self-supervised learning method (DIM-UA) that leverages an unbalanced atlas to improve dimensionality reduction and outperforms existing approaches on the AtariARI benchmark.
https://openreview.net/forum?id=cVUOnF7iVp
Compressor summary: The paper proposes a novel non-interactive differential privacy algorithm for sparse linear regression with lower bounds and upper bounds on estimation error.
https://openreview.net/forum?id=cUSNs8nGaV
Compressor summary: CGM-based glucose trajectory prediction can improve diabetes management, but current evaluation on small datasets hinders reproducibility and adoption.
https://openreview.net/forum?id=cPgh4gWZlz
Compressor summary: Chain-of-knowledge (CoK) is a novel framework that enhances large language models by dynamically incorporating grounding information from heterogeneous sources, resulting in more factual and rational answers for knowledge-intensive questions.
https://openreview.net/forum?id=cObFETcoeW
Compressor summary: The paper proposes a new backdoor-based method for evaluating saliency-based representation visualization methods in XAI, by minimizing the effects of a generalization-limited backdoor watermark on benign and poisoned samples.
https://openreview.net/forum?id=cMPm8YFXZe
Compressor summary: The paper proposes an ADDP framework that combines pixels and VQ tokens for representation learning, achieving competitive performance on image generation and recognition tasks.
https://openreview.net/forum?id=cKAUvMePUN
Compressor summary: The paper proposes an optimization framework for generating stimulation patterns for visual prostheses using a retinal network, phosphene model, and primary vision system network, which can simulate the whole process of visual signals processing from external scenes to visual perception in the cortex.
https://openreview.net/forum?id=cJs4oE4m9Q
Compressor summary: The paper proposes a novel deep anomaly detection model that improves hypersphere learning using an orthogonal projection layer, and extends it to graph-level anomaly detection with better performance than existing methods.
https://openreview.net/forum?id=cINwAhrgLf
Compressor summary: The paper proposes a method to use auxiliary labels from an independent task to improve primary task performance while keeping single task inference cost, using a flexible asymmetric network structure that can be removed during inference.
https://openreview.net/forum?id=cH3oufN8Pl
Compressor summary: The paper proposes Output-Domain focused Biasing (ODB), a training strategy that improves image classification by incorporating human label knowledge into neural networks, without external resources.
https://openreview.net/forum?id=c9xsaASm9L
Compressor summary: CMD is a technique that groups correlated parameters in neural networks to efficiently represent and enhance their training dynamics.
https://openreview.net/forum?id=c93SBwz1Ma
Compressor summary: BadChain is a novel backdoor attack against large language models using chain-of-thought prompting that alters the model's output when triggered by a specific query.
https://openreview.net/forum?id=c8McWs4Av0
Compressor summary: The paper explores how code usage frequency affects GPT-4 Code Interpreter's math reasoning skills and proposes a new prompting method, CSV, to enhance its self-verification abilities and accuracy on MATH dataset.
https://openreview.net/forum?id=c85tdYOOju
Compressor summary: The paper proposes a novel framework called Weigh And Select (WAS) for integrating multiple graph pre-training tasks using decoupled siamese networks.
https://openreview.net/forum?id=c7DND1iIgb
Compressor summary: FineR uses large language models to reason about fine-grained visual categories without needing expert annotations, by translating part-level image attributes to text and leveraging the LLM's world knowledge.
https://openreview.net/forum?id=c5pwL0Soay
Compressor summary: METRA is a novel unsupervised reinforcement learning objective that learns diverse behaviors by exploring a compact latent space connected to the state space, enabling scalability in high-dimensional environments.
https://openreview.net/forum?id=c56TWtYp0W
Compressor summary: The paper introduces GAFormer, a transformer-based model that learns spatiotemporal structure from multivariate time series data and achieves state-of-the-art performance on various tasks.
https://openreview.net/forum?id=c43FGk8Pcg
Compressor summary: DDSM is a novel framework that adapts neural network sizes to the importance of each generative step in denoising diffusion models, improving efficiency and preserving quality.
https://openreview.net/forum?id=c2R7ajodcI
Compressor summary: The All-Seeing (AS) project is a dataset and model for recognizing and understanding a wide range of real-world concepts using human feedback and efficient models, aiming to advance vision-language artificial general intelligence research.
https://openreview.net/forum?id=c0chJTSbci
Compressor summary: SuSIE is a method that uses an image editing diffusion model to help generalist robots recognize and reason about novel objects and scenarios, achieving better generalization than conventional language-conditioned policies.
https://openreview.net/forum?id=c0MyyXyGfn
Compressor summary: The paper proposes a novel algorithm, PSQD, that solves continuous space lexicographic multi-objective RL problems by learning and adapting subtask solutions under priorities, enabling reuse and offline learning.
https://openreview.net/forum?id=bxfKIYfHyx
Compressor summary: AlignDiff is a framework that uses reinforcement learning and human feedback to customize agent behaviors according to diverse and changing human preferences in locomotion tasks.
https://openreview.net/forum?id=bxITGFPVWh
Compressor summary: The paper introduces Sharpness-Aware Data Poisoning Attack (SAPA), which exploits the loss landscape sharpness to optimize data poisoning for maximum impact on trained models.
https://openreview.net/forum?id=bvjcMvMn7B
Compressor summary: The paper proposes SCARCE, a framework to improve GNNs' performance and fairness in semi-supervised node classification by incorporating graph structure and mitigating structural bias in active learning.
https://openreview.net/forum?id=buC4E91xZE
Compressor summary: AnomalyCLIP adapts CLIP for accurate zero-shot anomaly detection by learning object-agnostic text prompts that capture generic normality and abnormality in images, enabling generalized recognition across different domains.
https://openreview.net/forum?id=btpgDo4u4j
Compressor summary: The paper proposes LatentDiffuser, a framework that learns and plans in continuous latent action spaces using score-based diffusion models for efficient offline reinforcement learning.
https://openreview.net/forum?id=bshfchPM9H
Compressor summary: The text introduces a new method, called Rapper, for improving visual question answering by generating explanations that are informative, plausible, and grounded in both language and vision facts.
https://openreview.net/forum?id=bsKMPAFHO7
Compressor summary: The paper connects linear deviations in extensive-form games to untimed communication deviations and develops faster algorithms for computing linear correlated equilibria.
https://openreview.net/forum?id=bozbTTWcaw
Compressor summary: The paper proposes a new vector field for physics simulations that improves optimization by avoiding the issues caused by the rotational gradient field used in recurrent learning setups.
https://openreview.net/forum?id=bm1JVsVZVu
Compressor summary: The paper proposes a new algorithm, ASMG, for black-box multi-objective optimization that uses stochastic gradient approximation and adaptive weights to optimize multiple objectives efficiently and proves its convergence rate in both convex and non-convex scenarios.
https://openreview.net/forum?id=bkdWThqE6q
Compressor summary: The paper introduces INTR, a Transformer-based image classification method that generates class-specific queries to help each class find its patterns in an image, enabling interpretable predictions.
https://openreview.net/forum?id=bkNx3O0sND
Compressor summary: The authors propose efficient methods to finetune NMT models using external data and show significant quality improvements over the base model.
https://openreview.net/forum?id=bbCL5aRjUx
Compressor summary: MONet is a new model for image recognition that uses only multilinear operators, without activation functions, and achieves performance comparable to modern neural networks.
https://openreview.net/forum?id=ba84RDHFnz
Compressor summary: The paper proposes a new method called masked region autoencoding to learn from groups of pixels in images and improve self-supervised learning for tasks like detection and segmentation.
https://openreview.net/forum?id=bWzxhtl1HP
Compressor summary: The text explores using Denoising Diffusion Probabilistic Model to learn modular attributes from data by adding noise and learning features to compensate for the lost attributes at each time-step.
https://openreview.net/forum?id=bWcnvZ3qMb
Compressor summary: FITS is a lightweight model that uses complex frequency domain interpolation to perform well in time series analysis tasks, and it can run on edge devices.
https://openreview.net/forum?id=bWNJFD1l8M
Compressor summary: The paper proposes a method to transfer learning trajectories between DNNs to speed up training and improve performance.
https://openreview.net/forum?id=bTMMNT7IdW
Compressor summary: SDE-EDG is a novel method that collects infinitely fined-grid evolving trajectories of data distributions using stochastic differential equations to capture distribution shifts in time-varying systems, outperforming existing approaches.
https://openreview.net/forum?id=bRLed9prWC
Compressor summary: The authors introduce the task of future language modeling, which predicts texts in the future based on a temporal history of texts, and show that it improves upon non-temporal language models.
https://openreview.net/forum?id=bQWE2UqXmf
Compressor summary: Large language models (LLMs) can be used to detect AI-generated text by rewriting it and measuring the editing distance, improving detection scores across various domains.
https://openreview.net/forum?id=bNt7oajl2a
Compressor summary: The text studies how well language models do inductive reasoning using a technique called iterative hypothesis refinement, which shows they are good at proposing rules but struggle to apply them, revealing differences between human and model reasoning.
https://openreview.net/forum?id=bLpUtGyf9g
Compressor summary: DenosieLoc is an encoder-decoder model that learns to localize video activities by denoising action spans with controlled noise, leading to better understanding and faster convergence.
https://openreview.net/forum?id=bLKcCe7hYh
Compressor summary: UC-NeRF is a new method for creating realistic images from multiple cameras by correcting color inconsistencies and improving pose calibration.
https://openreview.net/forum?id=bJx4iOIOxn
Compressor summary: The paper analyzes Visual Prompt Tuning (VPT) as a transfer learning technique and identifies scenarios where it outperforms full-finetuning, as well as the underlying reasons for its success.
https://openreview.net/forum?id=bJ3gFiwRgi
Compressor summary: The paper proposes a method for learning reward functions and constraints of an expert from few demonstrations by using meta-learning over related tasks, and shows theoretical and experimental results.
https://openreview.net/forum?id=bH6T0Jjw5y
Compressor summary: This paper introduces Time-lagged Information Bottleneck (T-IB), a method to simplify complex systems for accurate simulation using information theory principles.
https://openreview.net/forum?id=bGGYcvw8mp
Compressor summary: The paper investigates how surface features and token co-occurrence reinforcement affect in-context learning in LLMs, revealing its mechanisms and limitations.
https://openreview.net/forum?id=bDkisS75zy
Compressor summary: COSA is a vision-language model that pretrains on image-text pairs to jointly model visual contents and event-level temporal cues, achieving state-of-the-art results on multiple semantic tasks.
https://openreview.net/forum?id=bDWXhzZT40
Compressor summary: The paper proposes a method to capture and use predictive uncertainty for various tasks using an auxiliary network trained with a meta-objective that minimizes dropout variance, resulting in significant performance gains across different datasets and scenarios.
https://openreview.net/forum?id=bAMPOUF227
Compressor summary: The text proposes a framework to improve large language models' generalizability and factuality by using task-specific fine-tuned models during inference, and demonstrates its effectiveness with experiments and resources.
https://openreview.net/forum?id=b66P1u0k15
Compressor summary: The paper proposes a Pareto deep long-tailed recognition method (PLOT) that uses a multi-objective optimization framework and dynamic re-balancing to address gradient conflicts among categories and improve overall performance.
https://openreview.net/forum?id=b3l0piOrGU
Compressor summary: MAE-LM is a new pretraining method that excludes $ exttt{[MASK]}$ tokens from the encoder and improves text encoder's performance on downstream tasks.
https://openreview.net/forum?id=b3kDP3IytM
Compressor summary: The paper introduces KITAB, a new dataset to test language models' ability to answer book-related queries with constraints, and analyzes their limitations in different scenarios.
https://openreview.net/forum?id=b3Cu426njo
Compressor summary: Meta-learning improves by using a novel prior representation model that leverages algorithm unrolling for more expressive and interpretable priors.
https://openreview.net/forum?id=b2UlHeyyC0
Compressor summary: The paper introduces a method to enhance CLIP's performance on fine-grained tasks by using external memory to refine its embeddings, instead of expanding the pre-training data.
https://openreview.net/forum?id=awWpHnEJDw
Compressor summary: Conceptor is a method that reveals how diffusion models represent textual concepts in their images by decomposing them into interpretable elements, discovering structures and connections between concepts, and linking them to visual impacts.
https://openreview.net/forum?id=awHTL3Hpto
Compressor summary: The study explores how different convex relaxations affect the ability of ReLU neural networks to encode and analyze various classes of convex functions, revealing limitations in some cases.
https://openreview.net/forum?id=auUngos7eR
Compressor summary: The paper proposes a new way of doing Bayesian filtering using optimization instead of matrix operations, which works well for high-dimensional systems and is easier to implement.
https://openreview.net/forum?id=auKAUJZMO6
Compressor summary: The study investigates how large language models respond to external information that conflicts with their internal knowledge, finding that they can be receptive to contradicting evidence but also show confirmation bias when presented with consistent information.
https://openreview.net/forum?id=apXtolxDaJ
Compressor summary: The paper introduces BEER, a regularizer that adaptively controls the representation rank of value networks in Deep Reinforcement Learning, improving agent performance and Q-value approximation.
https://openreview.net/forum?id=apA6SSXx2e
Compressor summary: The paper explores how Topological Concentration (TC), a new metric based on local subgraphs, can improve link prediction performance in graph neural networks (GNNs) by identifying and addressing topological distribution shifts.
https://openreview.net/forum?id=ap1ByuwQrX
Compressor summary: Token Distribution Dynamics (TDD) is a method to analyze and manipulate prompts for Large Language Models, using the LM head's interpreting capabilities to estimate input saliency based on distribution dynamics over the vocabulary.
https://openreview.net/forum?id=anzIzGZuLi
Compressor summary: TP-BERTa is a pre-trained language model that can handle numerical features in tables and achieve high performance in tabular data prediction tasks.
https://openreview.net/forum?id=ale56Ya59q
Compressor summary: The paper proposes VQScore, a self-supervised metric for evaluating speech quality using vector-quantized-variational autoencoder (VQ-VAE) and domain knowledge, which can also be used for self-supervised speech enhancement with improved robustness.
https://openreview.net/forum?id=agPpmEgf8C
Compressor summary: The study shows how predictive objectives in deep reinforcement learning improve and stabilize representation learning, mimic brain activity changes, and suggest a new role for the hippocampus as an auxiliary learning system.
https://openreview.net/forum?id=ag3o2T51Ht
Compressor summary: The authors examine seven methods to remove sensitive content from text-to-image models and show that none of them fully succeed, as they can still retrieve the erased concepts using special learned word embeddings.
https://openreview.net/forum?id=afQuNt3Ruh
Compressor summary: Shuffle coding is a new method for compressing unordered sequences of objects using bits-back coding, which works well for various data structures like graphs and molecular data.
https://openreview.net/forum?id=af2c8EaKl8
Compressor summary: The authors propose Decision ConvFormer, a new action sequence predictor for offline reinforcement learning, which improves on Decision Transformer by using local convolution filtering to capture local dependencies in trajectories.
https://openreview.net/forum?id=adSGeugiuj
Compressor summary: The paper derives a relation between higher-order moments of the posterior distribution and derivatives of the posterior mean, enabling efficient uncertainty quantification for pre-trained denoisers without training or fine tuning.
https://openreview.net/forum?id=abL5LJNZ49
Compressor summary: The paper proposes a method to represent and track state changes in instructional videos for better procedure planning using large language models and cross-modal contrastive learning.
https://openreview.net/forum?id=aaBnFAyW9O
Compressor summary: The paper analyzes the limitations of current diffusion models in denoising tasks, introduces a new model called soft mixture denoising (SMD) that improves their performance, and provides experimental results supporting its effectiveness.
https://openreview.net/forum?id=aZH1dM3GOX
Compressor summary: MOORE is a novel method for representation learning in multi-task reinforcement learning that uses orthogonal subspaces to promote diversity and generate task-specific representations.
https://openreview.net/forum?id=aPNwsJgnZJ
Compressor summary: The paper proposes a new algorithm for adversarial RL that achieves sub-optimal regret using variance-aware weighted least squares and occupancy measure-based policy search.
https://openreview.net/forum?id=aOnUe8ah7j
Compressor summary: SymPoint uses point cloud segmentation methods to spot and parse objects and stuff from CAD drawings, outperforming recent state-of-the-art method by a significant margin.
https://openreview.net/forum?id=aN4Jf6Cx69
Compressor summary: In-context learning in transformer models depends on distributional properties and a specific sequence of multi-layer operations that enable an induction head with three nested logits.
https://openreview.net/forum?id=aKJEHWmBEf
Compressor summary: APEN is a framework that improves point cloud neural networks by approximating their equivariance to local Euclidean transformations, using a compositional design for partition prediction with guaranteed bounds on approximation errors.
https://openreview.net/forum?id=aIok3ZD9to
Compressor summary: LLMCarbon is a tool that accurately predicts the carbon emissions of different large language models before training, overcoming limitations of previous methods like mlco2.
https://openreview.net/forum?id=aHmNpLlUlb
Compressor summary: InsertNeRF is a method that improves NeRF's ability to generalize to new scenes by using HyperNet modules to adapt the model's weights and features.
https://openreview.net/forum?id=aGH43rjoe4
Compressor summary: The paper proposes an unsupervised model that extracts shared and independent latent structures from different experimental modalities using Gaussian Process Factor Analysis and Gaussian Process Variational Autoencoders.
https://openreview.net/forum?id=aFWUY3E7ws
Compressor summary: The paper introduces a novel interpretable method for time series prediction that uses sparse optimization, Fourier basis, and trend-fluctuation analysis, and beats deep learning without needing training or computational expenses.
https://openreview.net/forum?id=aBUidW4Nkd
Compressor summary: The paper proposes a learnable clustering method for object-centric architectures that uses Gaussian Mixture Models to represent slots and improve performance on set property prediction tasks.
https://openreview.net/forum?id=aA33A70IO6
Compressor summary: The study suggests a new way to train language models to avoid generating harmful and toxic responses by analyzing their mistakes and using them as feedback for improvement.
https://openreview.net/forum?id=a8VETFwcVR
Compressor summary: The paper presents an algorithm to create reusable sub-policies from neural networks for Markov Decision Processes, which are used as options to speed up learning on related tasks.
https://openreview.net/forum?id=a745RnSFLT
Compressor summary: The paper explains how classical PAC-Bayes bounds can justify the success of zero-shot learning with prompts, despite their potential for overfitting, by showing that prompts' discrete nature and language model prior lead to tight generalization bounds.
https://openreview.net/forum?id=a4DBEeGfQq
Compressor summary: Structural Compression is a framework that trains graph contrastive learning models faster and more efficiently by compressing nodes, reducing computation and memory requirements, and improving performance.
https://openreview.net/forum?id=a2ljjXeDcE
Compressor summary: iGraphMix is a novel node classification method that interpolates input features and labels to generate virtual nodes and graphs, improving generalization performance in graph neural networks.
https://openreview.net/forum?id=ZzmKEpze8e
Compressor summary: The paper proposes a probabilistic Bayesian online learning model that adapts to non-stationary data and quantifies predictive uncertainty using a neural representation and a state space model over linear weights.
https://openreview.net/forum?id=Zz594UBNOH
Compressor summary: Clifford Group Equivariant Simplicial Message Passing Networks combine Clifford group-equivariant layers with shared simplicial message passing for steerable, topologically intricate geometric analysis.
https://openreview.net/forum?id=ZwhHSOHMTM
Compressor summary: The text describes a method to learn dynamic neural networks in live animals by analyzing calcium activity and non-negative tensor factorization.
https://openreview.net/forum?id=Zsfiqpft6K
Compressor summary: The paper proposes DiffMatch, a conditional diffusion-based framework for dense matching of paired images, addressing ambiguities by explicitly modeling data and prior terms with a cascaded pipeline.
https://openreview.net/forum?id=ZlQRiFmq7Y
Compressor summary: VDR uses natural language to help learn disentangled representations from data by leveraging linguistic equivalents as proxies for the underlying variation.
https://openreview.net/forum?id=Zj12nzlQbz
Compressor summary: The paper proposes a new method to detect when large language models make wrong answers by analyzing their internal states and using a simple metric called EigenScore.
https://openreview.net/forum?id=ZhlwoC1XaN
Compressor summary: The paper presents a novel generative model for 3D turbulent flows that learns the distribution of all possible flow states and can produce realistic samples without relying on an initial condition.
https://openreview.net/forum?id=Zh2iqiOtMt
Compressor summary: The paper studies how different types of information from a teacher can improve the efficiency of transferring knowledge to a probabilistic student classifier.
https://openreview.net/forum?id=Zc2aIcucwc
Compressor summary: The authors introduce seven new molecular machine learning datasets with diverse labels, a graph library for easy modeling, and baseline results that show potential in multi-task and multi-level training for foundation models.
https://openreview.net/forum?id=ZZTkLDRmkg
Compressor summary: Boundary-Embedded Neural Operators (BENO) is a new technique that solves elliptic PDEs with complex geometries and boundary conditions more efficiently using neural networks and Transformer encoder.
https://openreview.net/forum?id=ZZCPSC5OgD
Compressor summary: LipVoicer is a novel lip-to-speech method that generates high-quality, intelligible speech from silent videos by incorporating the text modality and outperforming multiple baselines on in-the-wild datasets.
https://openreview.net/forum?id=ZYm1Ql6udy
Compressor summary: The paper presents a bi-clustering method to analyze neural spiking activity in multiple brain regions, considering both latent trajectories and spatial-temporal clustering structures, using a non-parametric model and an MCMC algorithm.
https://openreview.net/forum?id=ZWzUA9zeAg
Compressor summary: Our method uses text-to-image diffusion models to semantically edit images for data augmentation, improving performance on few-shot classification and weed recognition.
https://openreview.net/forum?id=ZWyZeqE928
Compressor summary: FunBaT is a novel tensor model that extends Tucker decomposition to handle continuous-indexed data by using functional priors and efficient inference algorithms.
https://openreview.net/forum?id=ZULjcYLWKe
Compressor summary: The paper proposes DMBP, a framework that uses conditional diffusion models to recover actual states in offline RL and improve robustness against state observation perturbations.
https://openreview.net/forum?id=ZTssMmhC2X
Compressor summary: SGD outperforms AdamW for fine-tuning large neural networks in computer vision when the embedding layer is frozen and memory usage is reduced.
https://openreview.net/forum?id=ZSD3MloKe6
Compressor summary: The paper proposes Time-Shift Sampler, a novel method to reduce exposure bias in Diffusion Probabilistic Models without retraining, by selecting better time steps for inference.
https://openreview.net/forum?id=ZS4m74kZpH
Compressor summary: The authors analyze cases when retrieval-augmented language models reduce accuracy and propose two methods to improve their performance by filtering out irrelevant passages and fine-tuning the model with a mix of relevant and irrelevant contexts.
https://openreview.net/forum?id=ZPdZLlNXSm
Compressor summary: The paper introduces mean field theory from statistical physics to create new loss functions for deep metric learning, reducing training complexity and improving image retrieval performance.
https://openreview.net/forum?id=ZO5cn4IfaN
Compressor summary: CO2 is a new approach for large language model training that uses local updating and asynchronous communication to achieve 100% scalability on clusters with limited bandwidth, and it works well on various tasks in computer vision and natural language processing.
https://openreview.net/forum?id=ZNzDXDFZ0B
Compressor summary: PanoDiffusion is a new model that uses latent diffusion and progressive camera rotations to generate high-quality 360-degree RGB and depth panoramas from narrow field of view images.
https://openreview.net/forum?id=ZMv6zKYYUs
Compressor summary: The paper proposes a learning-based method for fast and accurate predictions and data assimilation of spatio-temporal PDEs using Neural Operators and nonlinear observers in function spaces.
https://openreview.net/forum?id=ZL6yd6N1S2
Compressor summary: The paper proposes G-$\Delta$UQ, a framework to improve intrinsic uncertainty estimates of graph neural networks (GNNs) using graph anchoring strategies under distribution shifts.
https://openreview.net/forum?id=ZKEuFKfCKA
Compressor summary: The paper proposes FedAU, a lightweight method for federated learning that adapts aggregation weights based on client participation history to handle diverse participation statistics and improve performance.
https://openreview.net/forum?id=ZGNWW7xZ6Q
Compressor summary: The paper proposes a method called reasoning on graphs (RoG) that combines large language models and knowledge graphs for faithful and interpretable reasoning, using relation paths as plans to retrieve valid reasoning paths from the graphs.
https://openreview.net/forum?id=ZG3RaNIsO8
Compressor summary: EvoPrompt is a novel framework for discrete prompt optimization using evolutionary algorithms and large language models to improve task performance on various natural language processing datasets.
https://openreview.net/forum?id=ZFMiHfZwIf
Compressor summary: The paper proposes Off-policy DAE, a method that learns from off-policy data by decomposing return into skill and luck components, without using importance sampling or truncation.
https://openreview.net/forum?id=ZEZ0CPmoSI
Compressor summary: The paper presents new algorithms that use a matrix stepsize to optimize non-convex problems and achieve faster convergence than scalar methods, along with theoretical and empirical results.
https://openreview.net/forum?id=ZDGKPbF0VQ
Compressor summary: A-LoL is an offline RL method for finetuning LMs that uses sequence-level rewards and value estimates to filter low-quality data and improve diversity, safety, and helpfulness of generated texts.
https://openreview.net/forum?id=Z9AZsU1Tju
Compressor summary: The ITHP model uses information bottleneck to create a compact information flow in multimodal learning, outperforming state-of-the-art benchmarks and achieving human-level performance on sentiment classification.
https://openreview.net/forum?id=Z8UfDs4J46
Compressor summary: The paper explores the impact of signal delay in deep reinforcement learning, formalizes delayed-observation Markov decision processes, and proposes effective strategies to overcome related challenges for continuous robotic control tasks.
https://openreview.net/forum?id=Z59Rb5bPPP
Compressor summary: The authors present a language modeling approach to generate realistic driving scenarios with partial autonomy and intra-timestep interaction, and evaluate it on two datasets.
https://openreview.net/forum?id=YrXHEb2qMb
Compressor summary: The text proposes using MMD with negative distance kernel for posterior sampling and conditional generative modelling, showing its advantages in computation and error bound, and demonstrating its effectiveness in various numerical examples.
https://openreview.net/forum?id=YrTI2Zu0dd
Compressor summary: The paper analyzes how different sample sizes affect the cost of overfitting in noisy kernel ridge regression and classifies overfitting scenarios as benign, tempered, or catastrophic.
https://openreview.net/forum?id=YqyTXmF8Y2
Compressor summary: The text describes a method to build an image segmentor using semantic knowledge from Stable Diffusion without additional supervision.
https://openreview.net/forum?id=YgMdDQB09U
Compressor summary: The paper proposes a new contrastive learning approach that performs well with smaller batch sizes and is robust to stochastic optimization, using AUC-maximization for embedding space arrangements.
https://openreview.net/forum?id=Yen1lGns2o
Compressor summary: The authors present a self-supervised video pretraining method that leverages tracking and attention to learn from unannotated first-person videos, such as the new "Walking Tours" dataset.
https://openreview.net/forum?id=YcW8i9VCf5
Compressor summary: The paper introduces ACBO, a generalization of CBO that considers multiple agents or external events intervening on the system, and proposes CBO-MW, an algorithm with bounded regret that combines online learning and causal modeling.
https://openreview.net/forum?id=YcM6ofShwY
Compressor summary: BayesDiff is a method that uses Bayesian inference to estimate pixel-wise uncertainty in diffusion models, which can help filter low-quality images and improve image generation.
https://openreview.net/forum?id=YbZxT0SON4
Compressor summary: The paper proposes a framework to transform non-stationary intrinsic rewards in reinforcement learning into stationary ones using state augmentations, improving agent performance in various exploration problems.
https://openreview.net/forum?id=Yb5KvPkKQg
Compressor summary: The text introduces a unified learning approach for composed image retrieval with text feedback that considers coarse- and fine-grained retrieval by modeling multi-grained uncertainty, improving the recall rate on three datasets.
https://openreview.net/forum?id=YZrg56G0JV
Compressor summary: This paper shows how multitask reinforcement learning (MTRL) with myopic exploration can be more efficient when tasks are diverse and related, and validates this theory with experiments on synthetic robotics.
https://openreview.net/forum?id=YR3ETaElNK
Compressor summary: The paper presents an efficient method to make large language models handle multiple modalities by tuning LayerNorm layers, which leads to significant performance improvements and reduced resource usage.
https://openreview.net/forum?id=YPIA7bgd5y
Compressor summary: This paper investigates how Large Language Models use label information when learning new tasks in context, revealing their capabilities, limitations, and biases.
https://openreview.net/forum?id=YOKnEkIuoi
Compressor summary: The authors propose a new method for learning the variance schedule of diffusion models during training, which improves the performance and adaptability of inverse problems and avoids the need for fine-tuning by experimentation.
https://openreview.net/forum?id=YLJs4mKJCF
Compressor summary: The text introduces a new data poisoning attack on fair representation learning, which can produce unfair representations that contain demographic information by injecting crafted samples into training data.
https://openreview.net/forum?id=YItWKZci78
Compressor summary: The paper presents two algorithms for minimax optimization over probability distributions, MFL-AG and MFL-ABR, which use mean-field Langevin dynamics and have symmetric and convergent updates.
https://openreview.net/forum?id=YIls9HEa52
Compressor summary: The irSLDS model extends the rSLDS model to analyze trial-varying neural activity data by incorporating latent geometry, flexible state cardinality, and dynamical sufficient statistics.
https://openreview.net/forum?id=YHUGlwTzFB
Compressor summary: ATTA is a novel test-time adaptation method that uses active learning to improve performance under distribution shifts, with theoretical guarantees and efficient sample selection techniques.
https://openreview.net/forum?id=YH5w12OUuU
Compressor summary: TEMPO uses GPT-like architecture to learn time series representations, improving accuracy and adapting to non-stationary data.
https://openreview.net/forum?id=YEhQs8POIo
Compressor summary: The paper proposes Private Evolution (PE), an API-based framework that generates differentially private synthetic data without training models, and shows its effectiveness on synthetic images and large foundation models.
https://openreview.net/forum?id=YEPlTU5mZC
Compressor summary: RVGP is a novel method for learning vector signals over latent Riemannian manifolds, which can super-resolve and inpaint vector fields while preserving singularities and reconstruct high-density neural dynamics from low-density EEG recordings.
https://openreview.net/forum?id=YCWjhGrJFD
Compressor summary: The paper proposes reinforcement learning methods to optimize diffusion models for downstream objectives like image quality and drug effectiveness, outperforming reward-weighted likelihood approaches.
https://openreview.net/forum?id=YCPDFfmkFr
Compressor summary: The paper proposes a unified approach to differentiate both feasible and infeasible convex quadratic programming layers in neural networks, enabling faster computations and improved performance in some tasks.
https://openreview.net/forum?id=Y9t7MqZtCR
Compressor summary: SWAMP is a simple modification of IMP that achieves better performance by training multiple sparse models and weight averaging them, resulting in faster inference speeds and less memory demand for modern neural networks.
https://openreview.net/forum?id=Y3wpuxd7u9
Compressor summary: GoLLIE is a model that follows annotation guidelines to improve its performance on unseen information extraction tasks.
https://openreview.net/forum?id=Y3BbxvAQS9
Compressor summary: DecompOpt is a new structure-based method that combines optimization and conditional diffusion models to generate novel ligands with desired properties, such as high binding affinity and easy synthesis, by decomposing them into substructures for fine-grained control.
https://openreview.net/forum?id=Xz13DtbOVW
Compressor summary: Our method constrains the accuracy change for each sub-group when pruning models, ensuring fairness and interpretability while scaling well.
https://openreview.net/forum?id=XyrB1Ay44j
Compressor summary: The paper explores the limitations of multi-modal models' robustness, identifies essential components for achieving higher robustness, and proposes a new training procedure called CRMT that significantly improves performance and robustness.
https://openreview.net/forum?id=XwiA1nDahv
Compressor summary: SmoothECE is a well-behaved calibration measure that fixes common flaws in binning and ECE by smoothing observations using an RBF kernel before computing ECE, and provides a Python package for measuring and plotting calibration.
https://openreview.net/forum?id=Xvfz8NHmCj
Compressor summary: The paper introduces a realistic Continual Learning setting with limited resources and proposes a simple but effective method called DietCL that uses both labeled and unlabeled data efficiently to outperform existing methods.
https://openreview.net/forum?id=XsHqr9dEGH
Compressor summary: The paper investigates the grokking phenomenon in neural networks and shows that it is caused by a shift from early to late phase implicit biases during training.
https://openreview.net/forum?id=XrunSYwoLr
Compressor summary: The paper proposes a method to convert pretrained Transformers into spiking neural networks by approximating their non-linear operations in both time and space dimensions.
https://openreview.net/forum?id=XlkN11Xj6J
Compressor summary: The paper proposes a method to incorporate 3D geometry control into diffusion models by using visual prompts from rendered images, enabling better performance in various vision tasks.
https://openreview.net/forum?id=Xlayxj2fWp
Compressor summary: The authors propose DNA-GPT, a training-free detection strategy that analyzes the differences between original and regenerated text by LLMs to distinguish machine-generated from human-written text with high performance and explainability.
https://openreview.net/forum?id=XlTDBZFXWp
Compressor summary: DPSGD is not sufficient for differentially private optimization without feature preprocessing, and DPSGD-F combines it with feature preprocessing to reduce privacy errors for classification tasks.
https://openreview.net/forum?id=Xkf2EBj4w3
Compressor summary: The paper proposes a self-supervised RL method for robotic systems that uses contrastive learning and shows its effectiveness on simulated and real-world image-based manipulation tasks.
https://openreview.net/forum?id=XhYWgjqCrV
Compressor summary: MogaNet is a new family of modern ConvNets that efficiently gathers and contextualizes discriminative features for improved computer vision tasks with better complexity-performance trade-offs.
https://openreview.net/forum?id=XVhm3X8Fum
Compressor summary: Stack attention, an attention operator incorporating stacks, enhances transformers' ability to recognize hierarchical patterns of arbitrary nesting depth in both natural language and context-free languages without syntactic supervision.
https://openreview.net/forum?id=XTHfNGI3zT
Compressor summary: PECoRe is an interpretability framework that quantifies how language models use context for translation, comparing their rationales with human annotations and identifying context-mediated predictions.
https://openreview.net/forum?id=XOnya9gSdF
Compressor summary: The paper studies how to optimize complex metrics in multi-label classification under population utility framework and proposes a practical learning algorithm based on Frank-Wolfe method.
https://openreview.net/forum?id=XNa6r6ZjoB
Compressor summary: The text proposes an extension of Transformers called the Abstractor that enables relational reasoning through a novel module with relational cross-attention, improving abstraction, generalization, and sample efficiency on various tasks.
https://openreview.net/forum?id=XN6ZPINdSg
Compressor summary: COLEP is a robust conformal prediction framework using probabilistic circuits that provides certified coverage guarantees against adversarial perturbations and outperforms single models in various datasets.
https://openreview.net/forum?id=XJ9vjEAqbx
Compressor summary: The paper proposes a new approach to train neural networks that are more robust against adversarial attacks by using a non-zero-sum bilevel formulation of adversarial training.
https://openreview.net/forum?id=XIxhINXtQk
Compressor summary: The paper proposes a novel end-to-end framework for instructed 3D-aware portrait editing from a single open-world image using conditional latent 3D diffusion and identity consistency module.
https://openreview.net/forum?id=XIaS66XkNA
Compressor summary: The new approach trains a neural network to be idempotent, mapping source noise to target images, and can generate output in one step, refine with sequential applications, and handle corrupted inputs.
https://openreview.net/forum?id=XIZEFyVGC9
Compressor summary: The paper proposes a new method (DAMA) to detect and reduce gender bias in large language models by modifying certain layers based on causal analysis.
https://openreview.net/forum?id=XEFWBxi075
Compressor summary: The paper introduces GRANDE, a novel gradient-based method for learning hard decision tree ensembles on tabular data, which achieves better performance than existing methods.
https://openreview.net/forum?id=X6tNkN6ate
Compressor summary: Denoising diffusion models can learn complex relationships between words and pixels in an image, and these relationships can be analyzed using information decomposition methods.
https://openreview.net/forum?id=X41c4uB4k0
Compressor summary: The text presents a method to generate 3D molecules with multiple desired properties using off-the-shelf models and probabilistic graphs, outperforming trained diffusion models.
https://openreview.net/forum?id=WyEdX2R4er
Compressor summary: This paper introduces a new vision-language model task called Visual Data-Type Identification and evaluates 39 models on it, revealing their weaknesses in recognizing simple image manipulations.
https://openreview.net/forum?id=Ww9rWUAcdo
Compressor summary: The study shows how adversarial examples contain hidden data features that help neural networks learn and generalize, even when trained on mislabeled samples.
https://openreview.net/forum?id=WvFoJccpo8
Compressor summary: The paper introduces QA-LoRA, an algorithm that makes large language models more efficient, accurate, and scalable by balancing quantization and adaptation.
https://openreview.net/forum?id=WsRHpHH4s0
Compressor summary: Ring Attention is a novel method that allows Transformers to handle long sequences by distributing them across multiple devices and overlapping communication through blockwise attention computation, enabling efficient training and inference on sequences longer than 100 million tokens.
https://openreview.net/forum?id=WjRPZsfeBO
Compressor summary: The paper analyzes Wasserstein Autoencoders' statistical guarantees and shows they can learn data distributions when network architectures are well designed, focusing on convergence rates that depend only on the intrinsic dimension of the data.
https://openreview.net/forum?id=WipsLtH77t
Compressor summary: The paper proposes a self-training framework for scene graph generation that uses class-specific adaptive thresholding with momentum to assign pseudo-labels to unannotated triplets and improves performance on fine-grained predicate classes.
https://openreview.net/forum?id=WhgB5sispV
Compressor summary: The paper presents a novel 4D primitive-based approach to synthesize diverse views of dynamic scenes from 2D images, with improved simplicity, flexibility, efficiency, and quality.
https://openreview.net/forum?id=WhZoCLRWYJ
Compressor summary: The paper proposes a fast and simple Schrodinger Bridges solver using sum-exp quadratic functions and log-Schrodinger potentials as energy functions, making it suitable for solving moderate-dimensional problems on CPU without complex optimization or hyperparameter selection.
https://openreview.net/forum?id=WesY0H9ghM
Compressor summary: Uni-RLHF is a system for reinforcement learning with human feedback that provides a complete workflow, large-scale datasets, and modular baselines to facilitate progress in practical problems.
https://openreview.net/forum?id=Wd47f7HEXg
Compressor summary: Quasi-Monte Carlo methods improve the Sliced Wasserstein distance approximation in 3D tasks by providing a better class of empirical SW and unbiased estimators.
https://openreview.net/forum?id=WcOohbsF4H
Compressor summary: The paper introduces ST-MEM, a self-supervised learning method for ECG data that reconstructs masked 12-lead ECG data and learns spatio-temporal features, improving arrhythmia classification performance.
https://openreview.net/forum?id=WbWtOYIzIK
Compressor summary: Knowledge Card is a framework to update large language models with domain-specific facts, improving their performance on various tasks.
https://openreview.net/forum?id=WZu4gUGN13
Compressor summary: The text introduces latent intuitive physics, a method to simulate fluids from 3D videos without knowing their physical properties, using a learned neural renderer and a parametrized prior learner.
https://openreview.net/forum?id=WWlxFtR5sV
Compressor summary: The paper explores how the difficulty of training physics-informed machine learning models depends on the conditioning of a specific differential operator, and proposes preconditioning methods to improve training.
https://openreview.net/forum?id=WS7GuBDFa2
Compressor summary: The paper proposes a simpler and more efficient method for time series representation learning that uses patch reconstruction, independent embedding of patches, and contrastive learning, outperforming masked time series modeling with Transformers.
https://openreview.net/forum?id=WQwV7Y8qwa
Compressor summary: MR-SDS is a nonlinear state space model that decomposes neural dynamics into local and cross-communication components in the latent space, accounting for sensory inputs, history effects, and heterogeneity, and can accurately recover latent trajectories, vector fields, and cross-region communication profiles in multi-region neural datasets.
https://openreview.net/forum?id=WQYHbr36Fo
Compressor summary: The paper proposes a novel augmentation method (RCC) and decoupling branch to address the coupling phenomenon in dense self-supervised learning, improving feature disentanglement among objects.
https://openreview.net/forum?id=WPZ2yPag4K
Compressor summary: The authors propose a method to improve the factual accuracy of language models by fine-tuning them using automated criteria or preference rankings, reducing errors in open-ended text generation settings.
https://openreview.net/forum?id=WOiOzHG2zD
Compressor summary: The TextField3D model introduces dynamic noise to latent space of text prompts to improve 3D representation learning and enable open-vocabulary generation with large vocabulary, text consistency, and low latency.
https://openreview.net/forum?id=WNzy9bRDvG
Compressor summary: The paper proposes improved techniques for training consistency models that achieve higher quality samples without distillation or bias, leading to significant improvements over prior approaches.
https://openreview.net/forum?id=WNkW0cOwiz
Compressor summary: The paper studies a problem in diffusion models where they have infinite Lipschitz near zero timesteps, which affects their stability and accuracy, and proposes a novel approach called E-TSDM to mitigate it.
https://openreview.net/forum?id=WNQjN5HzXt
Compressor summary: AUGCAL is a method that improves the performance of models trained on synthetic data by reducing miscalibration and overconfidence in real-world applications.
https://openreview.net/forum?id=WNLAkjUm19
Compressor summary: This study investigates how discrete visual tokens enhance masked image modeling and proposes a new metric and tokenizer design to improve generalization performance.
https://openreview.net/forum?id=WKuimaBj4I
Compressor summary: The paper proposes an algorithm for designing optimal payment schemes in multi-round principal-agent problems with small agent action spaces, achieving better regret bounds than existing methods.
https://openreview.net/forum?id=WIzzXCVYiH
Compressor summary: The paper proposes a method called GNNBoundary to understand the decision boundaries of graph neural networks (GNNs) by generating near-boundary graphs for adjacent class pairs and analyzing them.
https://openreview.net/forum?id=WIAO4vbnNV
Compressor summary: Motion guidance is a technique that uses optical flow networks to steer diffusion models for precise editing of images.
https://openreview.net/forum?id=W8S8SxS9Ng
Compressor summary: Neuroformer is a pre-trained transformer model for analyzing large-scale neuronal spiking data in systems neuroscience, which can predict behavior and infer neural circuit connectivity from simulated datasets.
https://openreview.net/forum?id=W3VsHuga3j
Compressor summary: The latent inference budget model (L-IBM) simulates computational constraints in agent populations by using a latent variable that controls the runtime of an iterative inference algorithm, allowing for better learning of suboptimal decision-making and inference of goals.
https://openreview.net/forum?id=W2tCmRrj7H
Compressor summary: FLEXGEN-EHR is a diffusion model for synthetic EHR generation that handles missing modalities and preserves fidelity and utility better than existing methods.
https://openreview.net/forum?id=W2d3LZbhhI
Compressor summary: The paper proposes a unified sampling framework that improves sample quality by adjusting solver strategies at different timesteps, and introduces $S^3$, a method to automatically optimize the solver schedule for efficient sampling from diffusion probabilistic models.
https://openreview.net/forum?id=Vw24wtSddM
Compressor summary: Tree Cross Attention (TCA) is a method that retrieves information from a logarithmic number of tokens for inference in a tree structure, making it more token-efficient than Cross Attention and Perceiver IO.
https://openreview.net/forum?id=VtmBAGCN7o
Compressor summary: MetaGPT is a meta-programming framework that uses human workflows and Standardized Operating Procedures to improve the collaboration of language models in solving complex tasks.
https://openreview.net/forum?id=VrHiF2hsrm
Compressor summary: The text discusses the drawbacks of fine-tuning language models and proposes Conjugate Prompting to improve their performance on various tasks.
https://openreview.net/forum?id=VoLDkQ6yR3
Compressor summary: This work studies a data reconstruction attack on neural networks, showing how it can recover the entire training set in the infinite width regime and revealing its dependence on network properties and image characteristics.
https://openreview.net/forum?id=VmGRoNDQgJ
Compressor summary: The text presents Influencer Backdoor Attack, a new way to manipulate semantic segmentation models by injecting triggers on non-victim pixels, and two trigger injection methods. It also demonstrates that current segmentation models are susceptible to IBA in real-world applications.
https://openreview.net/forum?id=VkWbxFrCC8
Compressor summary: RECOMBINER is a improved data compression method that uses hierarchical priors, learnable positional encodings, and patch-based encoding to overcome the limitations of previous INR-based approaches.
https://openreview.net/forum?id=Vja3ecieXY
Compressor summary: GreenTrainer is a new technique that reduces the environmental impact of fine-tuning large language models by adaptively selecting important tensors for training, achieving up to 64% FLOPs reduction without sacrificing accuracy.
https://openreview.net/forum?id=ViPtjIVzUw
Compressor summary: The paper proposes T-MARS, a data filtering approach for large multimodal datasets that removes image-caption pairs with overlapping text dominating visual features, improving computer vision tasks.
https://openreview.net/forum?id=ViNe1fjGME
Compressor summary: The paper introduces TGC, a general framework for deep Temporal Graph Clustering that preserves dynamic information and reduces computational consumption by adjusting deep clustering techniques to suit temporal graphs.
https://openreview.net/forum?id=VdwVOREDZM
Compressor summary: WildFusion is a new method that uses latent diffusion models to generate high-quality, 3D-consistent images from in-the-wild datasets without relying on posed images or canonical camera coordinates.
https://openreview.net/forum?id=VdkGRV1vcf
Compressor summary: The paper shows that Riesz kernels in maximum mean discrepancy flows have efficient computation properties, allowing their use in training generative models for images.
https://openreview.net/forum?id=VXak3CZZGC
Compressor summary: The paper presents a novel OOD generalization framework that learns invariant features in a hyperspherical space based on class alignment and separation, and shows its advantages theoretically and experimentally.
https://openreview.net/forum?id=VXDPXuq4oG
Compressor summary: Order-Preserving GFlowNets use a learned reward function to sample candidates with probabilities proportional to their order, enabling efficient exploration and exploitation in single-objective and multi-objective tasks.
https://openreview.net/forum?id=VVgGbB9TNV
Compressor summary: PromptAttack is an efficient tool to audit large language models' adversarial robustness by converting attacks into a prompt that fools the model without changing its semantic meaning.
https://openreview.net/forum?id=VTYg5ykEGS
Compressor summary: The paper introduces ImageNet-OOD, a dataset that separates semantic and covariate shifts in out-of-distribution detection, and shows that existing detectors are more sensitive to covariate shift than semantic shift.
https://openreview.net/forum?id=VTF8yNQM66
Compressor summary: SWE-bench is a new benchmark that assesses language models' ability to fix real-world software bugs by editing code and passing tests.
https://openreview.net/forum?id=VLFhbOCz5D
Compressor summary: TAFT is a method for fine-tuning linearized transformers using first-order Taylor expansion, which has lower cost, comparable performance, and several advantages over non-linear fine-tuning.
https://openreview.net/forum?id=VJvbOSXRUq
Compressor summary: The paper evaluates and compares various perturbation-based explainability methods for GNNs, identifying Pareto-optimal techniques and highlighting stability and feasibility issues.
https://openreview.net/forum?id=V5tdi14ple
Compressor summary: The paper proposes using formal mathematics examples in training LLMs to improve their quantitative reasoning skills and prevent logical errors.
https://openreview.net/forum?id=V3j5d0GQgH
Compressor summary: FedLoGe improves local and global model performance in Federated Long-Tailed Learning by integrating representation learning and classifier alignment using a shared backbone and individualized classifiers.
https://openreview.net/forum?id=V2cBKtdC3a
Compressor summary: RTRL is a sequence-processing RNN method that has conceptual advantages over BPTT but suffers from high time and space complexity; by combining it with policy gradients, researchers can achieve competitive results on challenging environments without approximation.
https://openreview.net/forum?id=V1GM9xDvIY
Compressor summary: The paper introduces SCOTCH, a novel method that combines neural stochastic differential equations with variational inference to learn structures from continuous-time stochastic processes observed at any time points, improving upon existing approaches.
https://openreview.net/forum?id=UyNXMqnN3c
Compressor summary: DreamGaussian is a fast and efficient 3D content generation framework that uses Gaussians and UV space techniques to create high-quality textured meshes from single-view images.
https://openreview.net/forum?id=UyGWafcopT
Compressor summary: The authors propose a novel way to extract meaning representations from autoregressive language models without prompts or fine-tuning, which can handle complex semantic tasks and multimodal data.
https://openreview.net/forum?id=Uw8xvFqVAE
Compressor summary: The paper proposes a game-based method for learning dimensionality-reducing representations using prior knowledge on prediction tasks, with theoretical and practical implications.
https://openreview.net/forum?id=UulwvAU1W0
Compressor summary: FourTran is an open-loop behavior cloning method that uses symmetry to improve sample efficiency and achieve state-of-the-art results in pick and place tasks.
https://openreview.net/forum?id=UpgRVWexaD
Compressor summary: SKR is a novel method to efficiently generate labeled data for training neural operators by sorting and recycling similar PDE systems, achieving up to 13.9x speedup.
https://openreview.net/forum?id=UoBymIwPJR
Compressor summary: The paper proposes a policy-aligned query and hybrid experience replay to improve preference-based reinforcement learning by addressing the issue of query-policy misalignment.
https://openreview.net/forum?id=Unb5CVPtae
Compressor summary: Time-LLM is a framework to repurpose large language models for time series forecasting using text prototypes and Prompt-as-Prefix, achieving state-of-the-art results and few-shot/zero-shot learning capabilities.
https://openreview.net/forum?id=UnUwSIgK5W
Compressor summary: Code Evol-Instruct is a novel method to enhance Code LLMs by adapting the Evol-Instruct technique, leading to superior code generation performance and surpassing both open-source and closed-source LLMs on various benchmarks.
https://openreview.net/forum?id=Un0rgm9f04
Compressor summary: Video Diffusion Transformer (VDT) is a transformer-based video generation model that captures temporal dependencies, handles diverse conditioning information, and can perform various tasks such as unconditional generation, video prediction, interpolation, animation, and completion.
https://openreview.net/forum?id=UmMa3UNDAz
Compressor summary: EfficientDM is a data-free framework that combines quantization-aware fine-tuning and post-training quantization to achieve high performance in low-bit diffusion models while maintaining time and data efficiency.
https://openreview.net/forum?id=UfBIxpTK10
Compressor summary: The text discusses a new benchmark called DockGen for assessing the generalization abilities of machine learning-based docking methods in finding ligand-binding proteins, and proposes Confidence Bootstrapping, a novel training paradigm that leverages diffusion and confidence models to improve generalizability.
https://openreview.net/forum?id=UbxWjq0UO2
Compressor summary: 3DFuse is a novel framework that improves text-to-3D generation by incorporating 3D awareness into pretrained 2D diffusion models using a consistency injection module and semantic coding techniques, achieving state-of-the-art performance in geometric robustness and fidelity.
https://openreview.net/forum?id=UaMgmoKEBj
Compressor summary: This paper proposes methods to improve regularized reinforcement learning by maintaining consistent regularization across different action spaces and temperatures, leading to better performance on various tasks.
https://openreview.net/forum?id=UXALv0lJZS
Compressor summary: The text describes a new speech separation method that uses generative models and achieves state-of-the-art results, even surpassing the previous upper bound for two speakers.
https://openreview.net/forum?id=UPvufoBAIs
Compressor summary: The paper presents a novel unsupervised 3D pose estimation method that uses cuboid meshes and generative neural features to adapt to the target domain without 3D data or annotations, leveraging stable object subparts and simulating fine-tuning.
https://openreview.net/forum?id=UMfcdRIotC
Compressor summary: The paper proposes two model-agnostic methods for counterfactual approximation, using a large language model either at inference-time or training-time, to improve the explanations of NLP system predictions effectively and efficiently.
https://openreview.net/forum?id=UMOlFJzLfL
Compressor summary: The paper investigates the relationship between linear stability and sharpness of SGD in overparameterized neural networks and introduces a new coherence measure for the loss Hessian to identify linear instability.
https://openreview.net/forum?id=UHjE5v5MB7
Compressor summary: The authors study how two-layer neural networks generalize on modular arithmetic tasks with corrupted labels and show that regularization methods improve performance by forcing the network to learn interpretable representations.
https://openreview.net/forum?id=UCfz492fM8
Compressor summary: CrossLoco is a guided unsupervised reinforcement learning framework that learns robot skills and their correspondence to human motions, achieving better accuracy, diversity, and user preference.
https://openreview.net/forum?id=UBVNwD3hPN
Compressor summary: CivRealm is an environment that tests decision-making agents' abilities in learning from past experiences and reasoning in novel contexts inspired by the Civilization game.
https://openreview.net/forum?id=U7iiF79kI3
Compressor summary: CALICO is a novel pretraining framework for multimodal bird's eye view perception in autonomous driving systems that uses contrastive learning to improve 3D object detection and map segmentation performance, as well as robustness against adversarial attacks and corruption.
https://openreview.net/forum?id=U7VW3KBm34
Compressor summary: SRD is a novel XAI method that enhances robustness by reflecting the model's inference process with a vector perspective and considering both active and inactive neurons.
https://openreview.net/forum?id=U6hEOZlDf5
Compressor summary: The paper proposes a novel method for relative object pose estimation using a hypothesis-and-verification framework with 3D transformations, which achieves high accuracy and robustness.
https://openreview.net/forum?id=U6Qulbv2qT
Compressor summary: The paper investigates whether multi-task reinforcement learning can improve sample efficiency in more general sequential decision making problems, such as partially observable MDPs and predictive state representations, by using a joint model class for tasks and measuring its complexity with the $\eta$-bracketing number.
https://openreview.net/forum?id=U0IOMStUQ8
Compressor summary: Sin3DM is a diffusion model that learns from low-dimensional latent space to create realistic 3D shapes with different textures and geometry, and supports various editing tasks.
https://openreview.net/forum?id=TzoHLiGVMo
Compressor summary: ODEFormer is a new transformer that can learn ODE systems from single trajectory observations, beating existing methods in accuracy and robustness, and being faster to infer.
https://openreview.net/forum?id=Tzh6xAJSll
Compressor summary: The paper studies associative memory mechanisms in a model using high-dimensional matrices of embeddings, deriving scaling laws and evaluating estimators with numerical experiments and visualizations.
https://openreview.net/forum?id=TzAJbTClAz
Compressor summary: The paper presents FFB, an open-source framework for evaluating group fairness methods in machine learning using standardized benchmarks and extensive analysis.
https://openreview.net/forum?id=TyFrPOKYXw
Compressor summary: The text proposes a novel algorithm called Safe Reinforcement Learning from Human Feedback (Safe RLHF) for training large language models (LLMs) that balance between being helpful and harmless, using human feedback to adjust the trade-off during fine-tuning.
https://openreview.net/forum?id=TwBY17Hgiy
Compressor summary: The paper proposes a 3D-aware regularizer that improves the performance of deep neural networks on multiple computer vision tasks by projecting features into a shared space and decoding them through differentiable rendering, reducing noisy cross-task correlations.
https://openreview.net/forum?id=Tw9wemV6cb
Compressor summary: The paper presents a novel BTI method that decouples benign features from backdoor features, which improves efficiency and robustness against backdoor threats. The paper also uses the new BTI method to develop effective backdoor-removal and pre-processing defenses.
https://openreview.net/forum?id=Tvwf4Vsi5F
Compressor summary: The paper proposes a realistic threat model for adversarial attacks using surrogate models and introduces PubDef, a game-theoretic defense method that significantly improves robustness against transfer attacks.
https://openreview.net/forum?id=Tuh4nZVb0g
Compressor summary: The paper proposes TEST, a method to convert time-series (TS) data into a format that large language models (LLMs) can handle, and shows that it can achieve competitive results on various TS tasks without sacrificing LLM's language abilities.
https://openreview.net/forum?id=TskzCtpMEO
Compressor summary: SSVI is a method for creating sparse Bayesian neural networks that maintains sparsity throughout training and inference, achieving significant compression and robustness while preserving performance.
https://openreview.net/forum?id=Ts95eXsPBc
Compressor summary: The paper introduces spatially-aware transformers for episodic memory in AI that consider both temporal and spatial dimensions, improving efficiency and accuracy in various tasks.
https://openreview.net/forum?id=TrKq4Wlwcz
Compressor summary: The paper introduces Large Content and Behavior Models (LCBMs) that incorporate "behavior tokens" in LLM training to improve communication effectiveness by predicting and optimizing receiver behavior.
https://openreview.net/forum?id=Tr3fZocrI6
Compressor summary: The text proposes a method called De-bias & Feature-Whiten to improve representation learning from heterogeneous data sources by reducing biases and noise dependencies, leading to better generalization and computational efficiency.
https://openreview.net/forum?id=Tr0lPx9woF
Compressor summary: The paper introduces a new post-training pruning method for large language models that considers both weight and activations, preserves important weights, and improves performance and efficiency.
https://openreview.net/forum?id=TqYbAWKMIe
Compressor summary: LILO is a neurosymbolic framework that creates and documents reusable code libraries using LLM-guided synthesis, Stitch's refactoring, and AutoDoc's natural language documentation.
https://openreview.net/forum?id=TqL2xBwXP3
Compressor summary: GeoLLM is a new method that leverages language models' spatial information and map data to perform better in geospatial prediction tasks than existing methods.
https://openreview.net/forum?id=TpD2aG1h0D
Compressor summary: Variance Reduced Meta-Continual Learning (VR-MCL) combines regularization-based and meta-learning approaches to improve knowledge transfer and reduce forgetting in continual learning.
https://openreview.net/forum?id=TlyiaPXaVN
Compressor summary: The authors propose GAES, a method using generative adversarial networks to compute general game-theoretic equilibria from samples, and apply it to various settings including normal-form games, competitive economies, and environmental policy.
https://openreview.net/forum?id=TjhUtloBZU
Compressor summary: This paper studies how noise in pre-training datasets affects model generalization and proposes a method (NMTune) to improve both in-domain and out-of-domain performance on vision and language models.
https://openreview.net/forum?id=TjfXcDgvzk
Compressor summary: NOLA is a method to efficiently adapt large language models for specific tasks by reducing parameters using linear combinations of random matrices and optimizing their coefficients.
https://openreview.net/forum?id=TjGJFkU3xL
Compressor summary: The paper proposes a kernel-based doubly robust estimator for continuous treatments in proximal causal learning, showing its consistency, efficiency, and performance.
https://openreview.net/forum?id=TjCDNssXKU
Compressor summary: THICK is an algorithm that learns a hierarchical world model for better planning and reinforcement learning by updating lower levels sparsely and higher levels when contexts change.
https://openreview.net/forum?id=Tj6Wcx7gVk
Compressor summary: PR-MPNNs learn to add and remove edges in graphs based on their relevance to the prediction task, improving expressiveness and performance over random methods and existing graph neural networks.
https://openreview.net/forum?id=Tj3xLVuE9f
Compressor summary: The paper investigates how predictivity and availability affect feature use in deep-learning models, finding that nonlinear models tend to over-rely on less predictive but more available features.
https://openreview.net/forum?id=TilcG5C8bN
Compressor summary: SimWnW is a similarity-based efficient self-supervised learning framework that reduces computation costs by removing less important regions in augmented images and feature maps without compromising accuracy.
https://openreview.net/forum?id=Tigr1kMDZy
Compressor summary: The text discusses the risks of harmful imitation by language models through two phenomena called overthinking and false induction heads, which can be studied by examining their internal representations.
https://openreview.net/forum?id=Th6NyL07na
Compressor summary: The DoLa approach reduces hallucinations in large language models by contrasting different layers to obtain a better next-token distribution and surface factual knowledge.
https://openreview.net/forum?id=TgSRPRz8cI
Compressor summary: Patch-DM is a denoising diffusion model that generates high-resolution images by combining features from neighboring patches and outperforms previous methods with less memory complexity.
https://openreview.net/forum?id=TdhkAcXkRi
Compressor summary: This paper explores how adding momentum to federated learning algorithms FedAvg and SCAFFOLD can improve their performance and convergence without making restrictive assumptions or adjustments.
https://openreview.net/forum?id=TYXtXLYHpR
Compressor summary: The paper proposes a top-down framework called bi-level transparency for time series forecasting with machine learning models to ensure interpretability and trustworthiness in decision-making systems.
https://openreview.net/forum?id=TWVMVPx2wO
Compressor summary: The paper proposes a framework for fine-tuning pre-trained models for deep metric learning tasks using visual prompts, which improves performance with fewer parameters and less computation.
https://openreview.net/forum?id=TWC4gLoAxY
Compressor summary: The paper proposes a framework to use logical rules derived from data to understand human goals and guide human-like agents in collaboration tasks.
https://openreview.net/forum?id=TVg6hlfsKa
Compressor summary: The paper proposes a novel method to adapt pre-trained vision models for visual place recognition, using hybrid adaptation and mutual nearest neighbor local feature loss, achieving better performance with less data and time.
https://openreview.net/forum?id=TVDUVpgu9s
Compressor summary: ZO-RankSGD is a novel algorithm that optimizes black-box functions with ranking feedback and improves image quality in diffusion generative models using human guidance.
https://openreview.net/forum?id=TTrzgEZt9s
Compressor summary: Prospect is a fast and easy-to-tune algorithm for distributionally robust optimization with spectral risk-based uncertainty sets and $f$-divergence penalty, achieving linear convergence and outperforming existing methods on various benchmarks.
https://openreview.net/forum?id=TTonmgTT9X
Compressor summary: Hyperbolic geometry helps machine learning capture hierarchical structures in data, but its computational challenges are overcome by HyperDT, a decision tree algorithm that adapts to hyperbolic space without Riemannian optimization.
https://openreview.net/forum?id=TS8HoIWAPQ
Compressor summary: Feature-aligned N-BEATS is a time series forecasting model that learns invariant features across multiple domains while preserving interpretability and performance.
https://openreview.net/forum?id=TPZRq4FALB
Compressor summary: This paper introduces a new challenge in multi-modal test-time adaptation called reliability bias, proposes a novel method RFRA that adaptively modulates attention between modalities and uses a robust objective function for multi-modal adaption, and provides two new benchmarks to evaluate the performance of methods under this challenge.
https://openreview.net/forum?id=TOWdQQgMJY
Compressor summary: SAGE is a method that systematically explores and discovers failures in text-guided diffusion models for image generation by using adversarial search, image classifiers, and human inspections.
https://openreview.net/forum?id=TLADT8Wrhn
Compressor summary: The paper introduces TiC benchmarks for continual learning on vision-language models, evaluates temporal robustness of existing models, and proposes a rehearsal-based approach to reduce compute costs.
https://openreview.net/forum?id=TKqMmKlmA7
Compressor summary: Spectral Transformation (ST) is a framework that modulates the spectrum of embedding in self-supervised learning to prevent dimensional collapse and improve representation learning, with IterNorm with trace loss being a novel effective instance of ST.
https://openreview.net/forum?id=TKnzPdyeJu
Compressor summary: The paper presents a new structural inference method using variational dynamics encoder and partial correlation coefficients, which is scalable, accurate, and robust for one- and multi-dimensional feature spaces.
https://openreview.net/forum?id=THUBTfSAS2
Compressor summary: The paper proposes a new active learning strategy based on predictive uncertainty measured by the least disagree metric, which outperforms existing methods on various datasets and deep neural networks.
https://openreview.net/forum?id=THJEa8adBn
Compressor summary: The paper proposes GLOW and HYGLOW algorithms for online and hybrid reinforcement learning that use density ratios and exploration under coverability condition.
https://openreview.net/forum?id=TFKIfhvdmZ
Compressor summary: The text introduces PPGA, a new algorithm that combines QD and RL to improve exploration and skill discovery for robot learning.
https://openreview.net/forum?id=T7YV5UZKBc
Compressor summary: The paper proposes a neural architecture search method to find the best adaptation strategy for few-shot recognition using adapters and freezing layers.
https://openreview.net/forum?id=T5Xb0iGCCv
Compressor summary: Neur2RO is a fast and efficient machine learning algorithm for solving two-stage robust optimization problems, especially on large-scale instances.
https://openreview.net/forum?id=Sy8upuD6Bw
Compressor summary: The study shows that conversational repair improves language performance in signaling games with feedback and noise, challenging the role of compositionality for generalization.
https://openreview.net/forum?id=Sx7BIiPzys
Compressor summary: The paper proposes a deterministic variational approach to train Bayesian last layer neural networks that improves uncertainty estimation and predictive accuracy.
https://openreview.net/forum?id=Sx038qxjek
Compressor summary: CRITIC is a framework that helps large language models improve their outputs by using external tools for validation and revision, similar to how humans use search engines and code interpreters.
https://openreview.net/forum?id=StYc4hQAEi
Compressor summary: The paper proposes efficient control variates to reduce variance in Monte Carlo estimation of sliced Wasserstein distances for comparing probability measures in various applications.
https://openreview.net/forum?id=SsmT8aO45L
Compressor summary: The paper proposes a watermarking method for large language models that ensures high-quality generated text, correct detection, and robustness against text manipulation.
https://openreview.net/forum?id=Spp2i1hKwV
Compressor summary: The paper proposes a selective annotation method for in-context learning that minimizes costs and improves quality by using a directed graph and a greedy algorithm to choose influential unlabeled data subsets.
https://openreview.net/forum?id=SoismgeX7z
Compressor summary: The paper introduces GSBM, a generalized distribution matching algorithm for training diffusion or flow models that can handle nonlinear state costs and preserves feasible transport maps throughout training, leading to improved scalability and convergence.
https://openreview.net/forum?id=Si3YFA641c
Compressor summary: The paper proposes R-EDL, a modified Evidential Deep Learning method that adjusts prior weight and optimizes Dirichlet PDF to improve predictive uncertainty estimation.
https://openreview.net/forum?id=ShjMHfmPs0
Compressor summary: The quality or diversity of future generative models will decline without enough fresh real data in each generation of an autophagous loop, which is called Model Autophagy Disorder (MAD).
https://openreview.net/forum?id=ShQrnAsbPI
Compressor summary: The paper proposes accurate forgetting (AF) and a generative-replay method for federated continual learning, which selectively uses previous knowledge in federated networks based on a normalizing flow model.
https://openreview.net/forum?id=SgjAojPKb3
Compressor summary: OpenNeRF is a method that uses visual-language models to segment arbitrary concepts from 3D scenes in images without relying on predefined training sets or point cloud representations.
https://openreview.net/forum?id=Sf2A2PUXO3
Compressor summary: The text proposes a framework that uses dropout techniques to explore and measure the Rashomon set, which consists of multiple competing models with almost-equally-optimal performance but conflicting outputs, and use this information to mitigate predictive multiplicity in classification tasks.
https://openreview.net/forum?id=SdeAPV1irk
Compressor summary: IRS is a method that reduces computational cost and maintains strong robustness guarantees for certifying modified deep neural networks against adversarial attacks using incremental robustness certification for randomized smoothing.
https://openreview.net/forum?id=SdBApv9iT4
Compressor summary: The paper proposes a new horizon-free regret bound for linear MDPs, where it estimates value functions and confidence sets directly, instead of estimating transition models and inhomogeneous value functions.
https://openreview.net/forum?id=SZzQz8ikwg
Compressor summary: The paper proposes ELLE, a regularization term for single-step adversarial training that reduces catastrophic overfitting and is more efficient than previous methods.
https://openreview.net/forum?id=SZOQ9RKYJu
Compressor summary: The paper introduces OWL, a large language model for IT operations, and proposes HMCE, a method to extend homogeneous Markov context, improving performance on IT tasks.
https://openreview.net/forum?id=SYBdkHcXXK
Compressor summary: The paper analyzes hard pixel errors in semantic segmentation, identifies aliasing as a major cause, and proposes new methods to improve segmentation by removing or adjusting high-frequency components.
https://openreview.net/forum?id=SUUrkC3STJ
Compressor summary: The text proposes a new graph learning method called VCR-Graphormer, which uses personalized PageRank tokenization and virtual connections to enable efficient mini-batch training and improve performance on node classification tasks.
https://openreview.net/forum?id=STUGfUz8ob
Compressor summary: Transformer models struggle with generalizing to new symbols when trained on abstract tasks, requiring large amounts of data or modifications to improve performance.
https://openreview.net/forum?id=SQrHpTllXa
Compressor summary: CABINET improves LLMs' question-answering over tables by filtering out irrelevant information using an unsupervised relevance scorer and a weakly supervised parsing module.
https://openreview.net/forum?id=SQpnEfv9WH
Compressor summary: Social-Transmotion is a transformer-based model that leverages diverse visual cues to enhance human trajectory prediction by using prompts inspired by NLP.
https://openreview.net/forum?id=SQGUDc9tC8
Compressor summary: The authors introduce Social Bias Neurons to identify and suppress harmful information in pre-trained language models, using a novel technique called Integrated Gap Gradients (IG$^2$).
https://openreview.net/forum?id=SNGXbZtK6Q
Compressor summary: The paper proposes neuron activation coverage (NAC) as a simple and effective measure for detecting out-of-distribution inputs in neural networks and evaluating model robustness.
https://openreview.net/forum?id=SLw9fp4yI6
Compressor summary: Model arithmetic is a new way to combine and adjust large language models for better text generation with less retraining and more control.
https://openreview.net/forum?id=SL7djdVpde
Compressor summary: The paper explores how to use a quantum algorithm called Hamming Weight preserving ansatz to incorporate symmetries and constraints in physical systems, which could lead to better performance in problems like ground state energy estimation and feature selection.
https://openreview.net/forum?id=SKulT2VX9p
Compressor summary: The paper proposes a framework for achieving causal fairness in machine learning using partially known causal graphs and interventions, and shows its effectiveness on various datasets.
https://openreview.net/forum?id=SIZWiya7FE
Compressor summary: The text introduces Label-Agnostic Forgetting (LAF), a supervision-free method for removing information from forgotten data in a trained model, which performs well without labels and excels in semi-supervised scenarios.
https://openreview.net/forum?id=SEiuSzlD1d
Compressor summary: The paper proposes a new method called MRVM-NeRF that improves 3D representation learning for NeRF models by using masks to guide ray and view modeling, leading to better generalization and detail capture across diverse scenes.
https://openreview.net/forum?id=SCQfYpdoGE
Compressor summary: The paper proposes a method to check if machine learning models can offer solutions when they make decisions that affect people's access to resources like loans or jobs, and shows how this can help improve fairness in these models.
https://openreview.net/forum?id=SBoRhRCzM3
Compressor summary: Thought Propagation (TP) improves complex reasoning in large language models by solving analogous problems and reusing their solutions to help with multi-step reasoning tasks.
https://openreview.net/forum?id=SBj2Qdhgew
Compressor summary: The paper proposes an information-theoretic approach to analyze fairness trade-offs in federated learning with respect to sensitive attributes, identifying three sources of unfairness and a convex optimization problem that defines the theoretical limits of accuracy and fairness trade-offs.
https://openreview.net/forum?id=SA19ijj44B
Compressor summary: Bayesian neural networks can optimize expensive objective functions and handle non-stationarity better than standard Gaussian process surrogates, but their performance depends on the problem and the inference method used.
https://openreview.net/forum?id=S5yOuNfSA0
Compressor summary: The paper studies the theoretical aspects of CLIP, a multi-modal learning method that uses contrastive pretraining to learn joint image and text representations, and proposes a new approach based on its insights.
https://openreview.net/forum?id=S5aUhpuyap
Compressor summary: The text describes a recurrent network model with dendritic nonlinearities and stochastic somatic integration that can implicitly represent priors over latent variables and encode task-specific posteriors for flexible inference.
https://openreview.net/forum?id=S5EqslEHnz
Compressor summary: Adaptive Inflation (AdaInf) is a data-centric strategy that improves contrastive learning by adjusting the balance between data inflation and augmentation, achieving state-of-the-art results on benchmark datasets without extra computation cost.
https://openreview.net/forum?id=S46Knicu56
Compressor summary: The paper proposes a general variational framework to estimate the average dose-response function with error-contaminated treatment data, using neural networks to handle unobserved treatments and confounding variables.
https://openreview.net/forum?id=S2oTVrlcp3
Compressor summary: SmartPlay is a benchmark and methodology to evaluate large language models' abilities as intelligent agents by testing them on 6 different games that challenge various capabilities of an LLM agent.
https://openreview.net/forum?id=S1RKWSyZ2Y
Compressor summary: MGIE uses multimodal large language models to generate expressive image editing instructions and improve controllability and flexibility of image manipulation.
https://openreview.net/forum?id=RzY9qQHUXy
Compressor summary: The paper proposes an adaptive data augmentation method called DODA that addresses the challenges of long-tailed learning by allowing each class to choose appropriate augmentation methods based on its distribution, reducing the imbalance and improving performance.
https://openreview.net/forum?id=RzNlECeoOB
Compressor summary: The paper introduces $t^3$VAE, a modified VAE framework that uses Student's t-distributions for the prior, encoder, and decoder to better capture rare features in heavy-tailed data.
https://openreview.net/forum?id=RyUvzda8GH
Compressor summary: Incremental predictive coding (iPC) is an improved algorithm for training neuroscience-inspired models that is more efficient, stable, biologically plausible, and performs better on image and language tasks.
https://openreview.net/forum?id=RwI7ZEfR27
Compressor summary: The Brain Language Model is a large-scale model that can predict clinical variables, forecast future brain states, identify functional networks, and generate interpretable latent representations from fMRI data, enabling novel insights into human brain activity dynamics.
https://openreview.net/forum?id=RvfPnOkPV4
Compressor summary: WIMBD is a platform that analyzes large text corpora, revealing their content quality and diversity, and finds issues like duplicate and low-quality data, as well as contamination in benchmark datasets.
https://openreview.net/forum?id=RvUVMjfp8i
Compressor summary: SSL can benefit from unlabeled data but needs to be improved for consistent and diverse data sources in open environments.
https://openreview.net/forum?id=RthOl4jHw5
Compressor summary: Meta-Evolve is a method that uses continuous robot evolution to efficiently transfer policies from a source robot to multiple target robots, achieving significant improvements in simulation cost.
https://openreview.net/forum?id=RtDok9eS3s
Compressor summary: The authors propose a simplified design for deep Transformers by removing some components without losing speed or performance.
https://openreview.net/forum?id=RtAct1E2zS
Compressor summary: The paper develops a theory of error propagation in diffusion models and proposes a regularization method based on cumulative error to improve their performance.
https://openreview.net/forum?id=RsztjXcvUf
Compressor summary: The paper proposes a warm-starting technique for solving variational inequalities and shows its faster convergence than existing methods, while also relaxing smoothness assumptions.
https://openreview.net/forum?id=RsJwmWvE6Q
Compressor summary: The text studies how to estimate residual errors for matrix and vector norms using linear sketches, improving previous bounds and allowing for faster algorithms with sparse matrices.
https://openreview.net/forum?id=Rry1SeSOQL
Compressor summary: The text proposes a new approach to Machine Translation Evaluation that compares translations without references using natural language inference and synthetic data, achieving state-of-the-art results on various benchmarks.
https://openreview.net/forum?id=RjYKTQ0L0W
Compressor summary: The paper proposes a novel method for generating high-quality content-grounded data for tasks like LFQA and summarization, which performs well in human evaluation and outperforms existing models in faithfulness.
https://openreview.net/forum?id=RgELE1dQXx
Compressor summary: The paper proposes a model for human-AI interactions that considers human adherence and selective advice, and presents learning algorithms for optimal policy and timing of advice.
https://openreview.net/forum?id=RemfXx7ebP
Compressor summary: The study presents a data-driven pipeline for RNA sequence design based on tertiary structure using a large curated dataset, structural modeling, and contrastive learning with secondary structures as prior knowledge.
https://openreview.net/forum?id=Rc7dAwVL3v
Compressor summary: NaturalSpeech 2 is a text-to-speech system that uses a neural audio codec and a diffusion model to generate high-quality, diverse speech from large datasets, even with zero-shot learning.
https://openreview.net/forum?id=RaqZX9LSGA
Compressor summary: The paper introduces a new formalism for Markov decision processes with random delays, called stochastic delayed execution MDPs, and proposes a model-based algorithm, Delayed EfficientZero, that achieves optimal performance by searching over the class of Markov policies.
https://openreview.net/forum?id=RZBy8oHTz4
Compressor summary: The paper proposes a simple modification to the spectral contrastive loss that improves performance and robustness in self-supervised learning by introducing zero-mean regularization for negative pairs.
https://openreview.net/forum?id=RXFVcynVe1
Compressor summary: The paper introduces a novel technique to improve graph learning with text by using language model explanations as features, which boost performance and speed on various text-attributed graph datasets.
https://openreview.net/forum?id=RVrINT6MT7
Compressor summary: This study develops a mathematical framework to explain how noisy neural networks can autonomously reactivate the same network states during periods of quiescence, such as sleep, based on their task optimization.
https://openreview.net/forum?id=RRayv1ZPN3
Compressor summary: TAIL is a framework for adapting large pretrained models to new control tasks efficiently using parameter-efficient fine-tuning techniques like LoRA, with minimal data and without forgetting previous knowledge.
https://openreview.net/forum?id=RR8y0WKrFv
Compressor summary: The paper proposes a novel ensemble method for unsupervised constituency parsing using tree averaging and multi-teacher distillation, which outperforms previous methods.
https://openreview.net/forum?id=RR70yWYenC
Compressor summary: The paper proposes a new method, SIOPT-Grad, for minimizing finite-sums with instance-optimal guarantees, which requires fewer gradient evaluations than existing methods.
https://openreview.net/forum?id=RNGUbTYSjk
Compressor summary: The paper presents a new framework of algorithms based on extragradient for min-max optimization in structured problems, with adaptive exploration to avoid limit cycles.
https://openreview.net/forum?id=RMgqvQGTwH
Compressor summary: The paper proposes a new hybrid RL algorithm that combines an on-policy actor-critic method with offline data, achieving both theoretical and empirical benefits.
https://openreview.net/forum?id=RLSWbk9kPw
Compressor summary: The paper proposes a neural sorting network for abstract inputs, improves the swap function's differentiability, and uses a Transformer network with multi-head attention to capture dependencies and leverage model capacity.
https://openreview.net/forum?id=RJDjSXNuAZ
Compressor summary: The paper proposes a weakly supervised object detection method that uses image-level annotations and distills knowledge from a pre-trained model to detect viruses more efficiently than existing methods.
https://openreview.net/forum?id=RIuevDSK5V
Compressor summary: The text introduces ConR, a contrastive regularizer that improves deep imbalanced regression by modeling global and local label similarities in feature space and penalizing incorrect proximities between samples.
https://openreview.net/forum?id=RIu5lyNXjT
Compressor summary: This work studies how different prompt formatting affects large language models' few-shot performance, and proposes FormatSpread to quickly evaluate a range of prompt formats for a task.
https://openreview.net/forum?id=RIcYTbpO38
Compressor summary: Motion Coherent Augmentation is a data augmentation method for video recognition that uses Hue Jittering to introduce appearance variation and improve motion pattern detection.
https://openreview.net/forum?id=RIEW6M9YoV
Compressor summary: The paper proposes a new method for generating graphs using a $K^2$ representation that captures the hierarchical structure of graphs and uses a Transformer-based architecture with a tree positional encoding scheme.
https://openreview.net/forum?id=RDSj6S8WJe
Compressor summary: This paper challenges the common assumption of linear MDPs and proposes a new hierarchical reinforcement learning framework called dynamics aggregation, which improves performance guarantees.
https://openreview.net/forum?id=RDSTjtnqCg
Compressor summary: The paper discusses OOD detection methods, proposing SCALE and ISH to improve OOD/ID accuracy without compromising performance.
https://openreview.net/forum?id=R7rZUSGOPD
Compressor summary: PAE is a framework that teaches AI agents to learn from external knowledge, improving their exploration efficiency and task performance.
https://openreview.net/forum?id=R3Tf7LDdX4
Compressor summary: Memory-consistent neural networks (MCNNs) improve imitation learning by limiting policy errors using memory regions from expert demonstrations.
https://openreview.net/forum?id=R0c2qtalgG
Compressor summary: The paper introduces MetaTool, a benchmark to evaluate if large language models can choose appropriate tools for various tasks, and finds that most LLMs struggle with this task but have room for improvement.
https://openreview.net/forum?id=QzTpTRVtrP
Compressor summary: Large Brain Model (LaBraM) is a unified foundation model for EEG that enables cross-dataset learning and achieves state-of-the-art performance on various downstream tasks.
https://openreview.net/forum?id=QyFm3D3Tzi
Compressor summary: GPDiff is a new method that pre-trains STG models with a diffusion model and a transformer network to adapt to different cities' data and achieve better performance on various smart city tasks.
https://openreview.net/forum?id=QxItoEAVMb
Compressor summary: TorchRL is a PyTorch library that simplifies algorithm development and improves efficiency in decision and control tasks, with TensorDict as a new flexible primitive.
https://openreview.net/forum?id=Qwq4cpLtoX
Compressor summary: This study evaluates fifteen model architectures on synthetic in-context learning tasks and finds that contemporary architectures perform best as task complexity increases, while state-space models are more robust than transformers.
https://openreview.net/forum?id=QuVlUn4T2G
Compressor summary: The paper explores whether a generalized method for synthesizing novel views from dynamic scenes in monocular videos is possible, and finds that it is partially feasible without scene-specific appearance optimization but requires consistent depth estimates.
https://openreview.net/forum?id=QuIiLSktO4
Compressor summary: The text introduces parsimonious algorithms for caching and multi-item single-server settings that use predictions efficiently and maintain good performance with varying prediction error.
https://openreview.net/forum?id=QrEHs9w5UF
Compressor summary: The paper proposes a new method to explain why image classification models fail by using human-understandable tags of images instead of clustering in feature space, resulting in better and simpler descriptions.
https://openreview.net/forum?id=QqjFHyQwtF
Compressor summary: The authors propose a new method, Uncertainty-Aware Dynamic Fusion (UADF), to improve generative error correction by incorporating acoustic information into the language model's output, leading to better speech recognition results.
https://openreview.net/forum?id=Qox9rO0kN0
Compressor summary: The paper proposes CommFormer, a method that learns a communication graph for multiple intelligent agents to improve collaboration in AI applications.
https://openreview.net/forum?id=Qmw9ne6SOQ
Compressor summary: Text-to-image diffusion models use multiple components in the conditional UNet to store knowledge about different visual attributes, while only one causal state in the text-encoder is shared across them, enabling fast data-free model editing with DiffQuickFix.
https://openreview.net/forum?id=QmZKc7UZCy
Compressor summary: LanguageBind extends video-language pretraining to multiple modalities by using language as a bind, leading to improved performance in various tasks.
https://openreview.net/forum?id=QmYNBVukex
Compressor summary: This paper proposes a data selection method that improves pre-fine-tuning of pre-trained language models by aligning their distributions and achieves better performance, faster than existing techniques.
https://openreview.net/forum?id=QlqdXrzzD1
Compressor summary: The paper introduces unified models for 3D object tracking that adapt to different categories by encoding shape and size attributes using a transformer-based network called AdaFormer.
https://openreview.net/forum?id=QibPzdVrRu
Compressor summary: The paper analyzes how a two-layer ReLU network with gradient flow and small initialization learns from well-separated input data, showing an upper bound on the time for neurons to align and convergence rate of the loss.
https://openreview.net/forum?id=QiJuMJl0QS
Compressor summary: The paper proposes a new heterogeneous meta-learning strategy that captures the multi-modality of task distribution using a neural permutation layer with sub-quadratic parameter complexity.
https://openreview.net/forum?id=QgwAYFrh9t
Compressor summary: The paper shows how three-layer neural networks can efficiently learn hierarchical polynomials over Gaussian distribution by feature learning, outperforming kernel methods and two-layer networks.
https://openreview.net/forum?id=QcMdPYBwTu
Compressor summary: SEIGNN is a new efficient implicit GNN that uses mini-batch training and a stochastic solver to handle large graphs, while maintaining prediction accuracy.
https://openreview.net/forum?id=Qbf1hy8b7m
Compressor summary: AugLocal is a local learning method that enhances the synergy between layers and reduces memory usage in deep neural networks by selectively incorporating some layers into auxiliary networks and adjusting their depth.
https://openreview.net/forum?id=Qa0ULgosc9
Compressor summary: OpenTab is a table reasoning framework using LLMs that retrieves relevant tables, generates SQL programs, and conducts grounded inference for accurate responses on structured data.
https://openreview.net/forum?id=QQYpgReSRk
Compressor summary: MOFI is a new vision foundation model that learns from noisy image-text pairs using entity labels and achieves state-of-the-art performance on various tasks.
https://openreview.net/forum?id=QQ6RgKYiQq
Compressor summary: MovingParts is a NeRF-based method for reconstructing dynamic scenes and discovering parts by tracking particle trajectories under both Eulerian and Lagrangian views.
https://openreview.net/forum?id=QLoepRnoue
Compressor summary: HDFE is a vector representation method for continuous objects that works with neural networks and improves performance on tasks like function-to-function mapping and point cloud surface normal estimation.
https://openreview.net/forum?id=QJGj07PD9C
Compressor summary: The paper proposes a method to train Fourier Neural Operators (FNO) in mixed precision, reducing memory usage and increasing training speed while maintaining accuracy.
https://openreview.net/forum?id=QHROe7Mfcb
Compressor summary: The paper proposes a one-shot subgraph reasoning method for efficient and adaptive knowledge graph reasoning using Personalized PageRank heuristics and optimal configuration searching.
https://openreview.net/forum?id=QFYVVwiAM8
Compressor summary: AdaSAP is a method that creates sparse deep learning models with better robustness to unseen input variations by optimizing the loss landscape through weight perturbations during pruning.
https://openreview.net/forum?id=Q53QLftNkA
Compressor summary: The MW-MAE model uses a novel attention module to learn better general-purpose audio representations and improve performance on various downstream tasks, by capturing local and global information independently in each decoder block.
https://openreview.net/forum?id=Q3YaCghZNt
Compressor summary: The text proposes a method to use LLMs with automated reasoners for program verification and shows its effectiveness on benchmarks.
https://openreview.net/forum?id=Q1vkAhdI6j
Compressor summary: MixSup is a 3D object detection method that uses cheap cluster labels and some box labels to learn semantics and geometry, improving performance on various datasets.
https://openreview.net/forum?id=Q1u25ahSuy
Compressor summary: SpQR is a new compression technique that allows near-lossless LLMs on memory-limited devices by isolating outlier weights and compressing others to 3-4 bits, achieving up to 15% speedup with minimal accuracy loss.
https://openreview.net/forum?id=PxoFut3dWW
Compressor summary: Wanda is a novel, straightforward, and effective pruning method for Large Language Models that does not require retraining or weight update.
https://openreview.net/forum?id=PvyOYleymy
Compressor summary: The authors propose a new white-box architecture, CRATE-MAE, for large-scale unsupervised representation learning of images that is interpretable, structured, and efficient.
https://openreview.net/forum?id=PvJnX3dwsD
Compressor summary: Neural Quadratic Models can mimic the "catapult phase" and help analyze neural networks' generalization behavior.
https://openreview.net/forum?id=PudduufFLa
Compressor summary: This paper introduces a new way to represent geographical data in machine learning models using a combination of spherical harmonic functions and sinusoidal networks, which improves performance and avoids artifacts.
https://openreview.net/forum?id=Psl75UCoZM
Compressor summary: The authors propose a novel world modeling approach that combines VQVAE and discrete diffusion to handle complex robotic observations and achieve better predictions than existing methods.
https://openreview.net/forum?id=PsDFgTosqb
Compressor summary: The paper proposes deep learning methods for solving bilevel programs with discrete decision variables, comparing two types of neural networks and introducing an enhanced sampling technique.
https://openreview.net/forum?id=PoDkdFQIu3
Compressor summary: The paper proposes a method to generate synthetic counterfactual data for estimating time-varying causal effects in real-world settings and shows its superior performance over existing methods.
https://openreview.net/forum?id=PnR1MNen7u
Compressor summary: DeepGeoCCA is a novel deep learning framework for multi-view and multi-modal neuroimaging data that enhances the correlation of latent representations while preserving geometric structures.
https://openreview.net/forum?id=PhMrGCMIRL
Compressor summary: The paper proposes a supervised learning method for fusing outputs of different expert models to improve generalization across tasks and domains in various AI applications, including text classification and summarization.
https://openreview.net/forum?id=PfPnugdxup
Compressor summary: This paper introduces Joint Multi-domain Pre-training (JMP), a method that uses data from multiple chemical domains to generate transferable atomic representations for various molecular tasks with high accuracy.
https://openreview.net/forum?id=Pe2lo3QOvo
Compressor summary: Our algorithm for RLHF using preferences over trajectories minimizes regret bound and query complexity by combining randomization, active learning, and Bayesian supervised learning oracles in linear MDPs and general nonlinear function approximation settings.
https://openreview.net/forum?id=PdaPky8MUn
Compressor summary: Pretraining with denoising objectives improves long-range dependency modeling across various architectures and reduces the gap between Transformers and state space models.
https://openreview.net/forum?id=PczQtTsTIX
Compressor summary: CrossQ is a deep reinforcement learning algorithm that achieves high sample efficiency with low computational cost by using Batch Normalization and removing target networks, without relying on advanced bias-reduction schemes.
https://openreview.net/forum?id=PcxQgtHGj2
Compressor summary: This paper shows that pre-training for offline deep reinforcement learning can be done with simple synthetic data, instead of language pre-training, to improve performance.
https://openreview.net/forum?id=Pc8AU1aF5e
Compressor summary: Synapse is an agent that uses trajectory-as-exemplar prompting and memory to improve performance, efficiency, and generalizability in computer control tasks with large language models.
https://openreview.net/forum?id=PXXuLvIH5r
Compressor summary: The paper proposes a graph interpolation method to modify SAT instances by exchanging clause pairs based on node correspondence learned from a matching network, with the option of adding Gumbel noise for hardness preservation.
https://openreview.net/forum?id=PXNrncg2DF
Compressor summary: SOHES is a novel self-supervised approach for open-world entity segmentation that requires no human annotations and captures entities' parts hierarchically.
https://openreview.net/forum?id=PXD3FAVHJT
Compressor summary: RLHF improves OOD generalization but reduces output diversity in LLMs, highlighting a tradeoff between these properties.
https://openreview.net/forum?id=PR6RMsxuW7
Compressor summary: Our framework combines DRL with classical planning using symbolic regression to induce task structures from few demonstrations, achieving better sample efficiency and generalization than baselines.
https://openreview.net/forum?id=PQbFUMKLFp
Compressor summary: The paper introduces a new method for distributed optimization on the Stiefel manifold using Riemannian conjugate gradient descent with reduced computational complexity and global convergence guarantees.
https://openreview.net/forum?id=PQY2v6VtGe
Compressor summary: Confidential-DPproof is a framework that generates a cryptographic certificate of differential privacy during model training, overcoming limitations of post hoc privacy auditing techniques.
https://openreview.net/forum?id=PP1rudnxiW
Compressor summary: The text introduces a new method called Controlled Monte Carlo Diffusion (CMCD) for Bayesian computation, which adapts forward and backward dynamics in a diffusion model and performs well in various experiments.
https://openreview.net/forum?id=POFrdKvpea
Compressor summary: The paper proposes ACRF, a framework for compressing explicit NeRF models by pruning the neural 3D structure, encoding features with importance, and estimating probability distributions of transform coefficients.
https://openreview.net/forum?id=PKICZXVY9M
Compressor summary: The paper proposes a novel approach called OGEN to improve open-domain visual recognition by synthesizing unknown class features and using adaptive self-distillation to prevent overfitting during finetuning.
https://openreview.net/forum?id=PJwAkg0z7h
Compressor summary: EasyTPP is a central repository of event sequence modeling assets, including data, models, evaluation programs, and documentation, to facilitate research and progress in temporal point processes.
https://openreview.net/forum?id=PJVUWpPnZC
Compressor summary: RSRM is a novel machine learning model that uses Monte Carlo tree search, Double Q-learning, and sub-tree discovery to find complex math equations from limited data.
https://openreview.net/forum?id=PHLVmV88Zy
Compressor summary: The paper introduces fast and effective algorithms for linear and deep CCA using stochastic gradient descent, which improve the performance on standard benchmarks and allow big data analysis in biomedical applications. It also establishes a theoretical link between classical CCA and self-supervised learning.
https://openreview.net/forum?id=PHGxChm1l5
Compressor summary: The Compositional VLM model enhances the ability of large language models to reason compositionally about visual entities and their relations, achieving significant improvements on various vision-language benchmarks.
https://openreview.net/forum?id=PEuDO2EiDr
Compressor summary: RTFS-Net is a novel audio-visual speech separation method that uses multi-layered RNNs and attention-based fusion in the time-frequency domain, achieving better performance than existing time-domain methods with fewer parameters and lower computation costs.
https://openreview.net/forum?id=PCm1oT8pZI
Compressor summary: The paper proposes a watermark injection technique for deep neural networks using an out-of-distribution image as a secret key, which is robust against common attacks and does not require training data or samples.
https://openreview.net/forum?id=PAfnMGXief
Compressor summary: The text describes a new algorithm for generating sparse adversarial examples by observing score-based replies to model queries, which helps evaluate model vulnerabilities more efficiently.
https://openreview.net/forum?id=P1aobHnjjj
Compressor summary: The paper studies how stochastic gradient descent (SGD) can escape local minima with different ranks in deep linear networks, and proposes a sequence of sets to capture the minima of lower rank.
https://openreview.net/forum?id=P1ANzoGg3W
Compressor summary: H2O-SDF is a new method for 3D indoor scene reconstruction that uses a two-phase learning approach and the novel Object Surface Field to balance room layout preservation and detailed object surface capture.
https://openreview.net/forum?id=P15CHILQlg
Compressor summary: The paper proposes LED-GFN, a method to improve GFlowNets by learning energy decompositions for state transitions and reparameterizing flow functions using these potentials.
https://openreview.net/forum?id=Oxh5CstDJU
Compressor summary: TD-MPC2 is a new algorithm that improves the performance of model-based reinforcement learning in various tasks and domains by optimizing local trajectories in the latent space of implicit world models.
https://openreview.net/forum?id=OwtMhMSybu
Compressor summary: RECODE is a novelty-based exploration method that uses clustering and masked transformers to achieve state-of-the-art results in challenging 3D and Atari exploration tasks.
https://openreview.net/forum?id=OvlcyABNQT
Compressor summary: ABS is a method that combines BO and policy gradient methods for exploration using deterministic policies, improving performance in locomotion problems.
https://openreview.net/forum?id=Ouj6p4ca60
Compressor summary: The paper proposes using diversity-seeking reinforcement learning to fine-tune large language models for tasks requiring sampling from intractable posterior distributions, such as chain-of-thought reasoning.
https://openreview.net/forum?id=OuV9ZrkQlc
Compressor summary: ImagenHub is a one-stop library that standardizes inference and evaluation for various conditional image generation tasks, providing evaluation datasets, a unified pipeline, human evaluation scores, and guidelines.
https://openreview.net/forum?id=OsGUnYOzii
Compressor summary: SCRaWl is a novel simplicial complex neural network architecture that uses random walks and fast 1D convolutions to achieve better classification performance and higher-order relationships than graph-based methods, while controlling the computational cost.
https://openreview.net/forum?id=OrOd8PxOO2
Compressor summary: The paper presents a universal motion representation for humanoid control that learns from a large unstructured motion dataset, enabling diverse and realistic human behavior in complex tasks.
https://openreview.net/forum?id=OqTMUPuLuC
Compressor summary: The text proposes a new framework, DiLu, that uses large language models to enable autonomous driving systems with knowledge-driven capabilities similar to human drivers, improving decision-making and generalization.
https://openreview.net/forum?id=OkHHJcMroY
Compressor summary: PILOT and PILOT$^+$ are new RL policy evaluation methods that achieve fast convergence, sample efficiency, and apply constant step sizes for nonlinear function approximation.
https://openreview.net/forum?id=Oju2Qu9jvn
Compressor summary: The paper proposes a new approach for dynamic feature selection using mutual information estimation and various improvements to achieve better performance and transparency in prediction processes.
https://openreview.net/forum?id=OfXqQ5TRwp
Compressor summary: ALAM is a novel framework that uses average quantization and a lightweight sensitivity calculation scheme to compress activations in large language models, reducing memory overhead without sacrificing performance.
https://openreview.net/forum?id=Of2nEDc4s7
Compressor summary: The text discusses how gradient-based feature learning in neural networks can improve when dealing with anisotropic input data and suggests using anisotropic weight decay regularization for efficient learning.
https://openreview.net/forum?id=OeQE9zsztS
Compressor summary: The paper proposes a unified theory for learning with unlabeled data and a base kernel, using spectrally transformed kernel regression (STKR), which can handle various smooth target functions and provide scalable and statistically guaranteed implementations.
https://openreview.net/forum?id=OeH6Fdhv7q
Compressor summary: TapMo is a text-driven animation pipeline that generates high-quality motions for various skeleton-free 3D characters without the need for rigging.
https://openreview.net/forum?id=OdpIjS0vkO
Compressor summary: The paper shows theoretically and empirically how larger models, more data, and lower training loss improve performance in random feature regression, which applies to shallow neural networks with enough features and samples.
https://openreview.net/forum?id=Od39h4XQ3Y
Compressor summary: GraphSAM improves efficiency and generalization of graph transformer models by using gradient approximation and loss landscape approximation.
https://openreview.net/forum?id=OZitfSXpdT
Compressor summary: The paper proposes an adaptive method to learn the optimal ratio of using soft and hard supervision in knowledge distillation for training compact networks, improving performance across different tasks and architectures.
https://openreview.net/forum?id=OXv0zQ1umU
Compressor summary: Evoke is a framework that uses two instances of a large language model, one as an author and one as a reviewer, to automatically refine prompts for natural language processing tasks, focusing on hard samples to improve performance.
https://openreview.net/forum?id=OUkZXbbwQr
Compressor summary: The paper proposes a debate-based reward model for reinforcement learning agents to train justifiable policies in high-stakes situations, such as prescribing and justifying treatment decisions for septic patients.
https://openreview.net/forum?id=OUeIBFhyem
Compressor summary: $\infty$-Diff is a generative diffusion model that learns continuous functions in infinite-dimensional Hilbert space and can sample at arbitrary resolutions with high quality and efficiency.
https://openreview.net/forum?id=ORUiqcLpV6
Compressor summary: The paper proposes a novel interpretable 3D visual grounding framework that uses a chain-of-thoughts approach to localize objects in complex scenes, improving performance and data efficiency.
https://openreview.net/forum?id=ONPECq0Rk7
Compressor summary: The study proposes a new pre-training method for language models that reconstructs input embeddings using Contrastive Weight Tying, reducing training time and improving performance on downstream tasks.
https://openreview.net/forum?id=OKf6JtXtoy
Compressor summary: MAP IT is a new method for dimensionality reduction in visualization that captures local information and produces better class structure than existing methods.
https://openreview.net/forum?id=OIsahq1UYC
Compressor summary: The paper proposes a new sampling method (DGFS) for high-dimensional density functions that uses partial trajectories and a flow function to improve learning speed and accuracy.
https://openreview.net/forum?id=OI3RoHoWAN
Compressor summary: The paper proposes GenSim, a method that uses large language models to generate diverse simulation tasks and expert demonstrations, improving task-level generalization and transfer to real-world robotic policies.
https://openreview.net/forum?id=OHpvivXrQr
Compressor summary: PromptMSP is a pre-training and prompt tuning framework that improves multimer structure prediction by extending protein-protein interaction knowledge to different scales of multimers, achieving significant accuracy and efficiency improvements.
https://openreview.net/forum?id=OGtnhKQJms
Compressor summary: The paper presents a unified framework for learning identifiable representations from multiple partial and causally related views using contrastive learning and a set of rules called identifiability algebra.
https://openreview.net/forum?id=OF5x1dzWSS
Compressor summary: The paper proposes a novel adversarial training method that uses distributionally robust optimization to obtain importance weights and improve robustness across the training distribution, especially for vulnerable data points.
https://openreview.net/forum?id=OEL4FJMg1b
Compressor summary: The paper proposes DragonDiffusion, a novel method for editing images using diffusion models that allows precise Drag-style manipulations by leveraging feature correspondence and multi-scale guidance.
https://openreview.net/forum?id=ODSgo2m8aE
Compressor summary: This paper explores using the Lipschitz constant to control output perturbations in graph neural networks (GNNs) caused by biases, improving fairness in GNN training.
https://openreview.net/forum?id=OCqyFVFNeF
Compressor summary: The paper proposes a method to find shared interactions between deep neural networks trained for the same task, which can help explain their knowledge more faithfully.
https://openreview.net/forum?id=O9nZCwdGcG
Compressor summary: The paper proposes a new model, Biased Temporal Convolution Graph Network (BiaTCGNet), that can handle missing data and improve multivariate time series forecasting accuracy by injecting bias into two modules.
https://openreview.net/forum?id=O9PArxKLe1
Compressor summary: The text discusses the problem of untrustworthy users misusing image generators for spam and disinformation, and proposes watermarking as a solution to detect such content, but also shows that adaptive attacks can break existing watermarking methods.
https://openreview.net/forum?id=O8ouVV8PjF
Compressor summary: ShapeConv combines shapelets and CNN to create an interpretable and controllable time series model that performs well.
https://openreview.net/forum?id=O2jyuo89CK
Compressor summary: The paper presents a new generative model for complex vector drawings that uses stroke-clouds and latent codes to capture their distribution on an implicit set space.
https://openreview.net/forum?id=O072Rc8uUy
Compressor summary: The paper introduces Progressive3D, a framework for generating accurate 3D content from complex prompts using multiple editing steps and user-defined regions, as well as a suppression technique to focus on semantic differences.
https://openreview.net/forum?id=NzxCMe88HX
Compressor summary: The text discusses the vulnerabilities of latent diffusion models in generating realistic images and proposes new strategies for effective protection against misuse.
https://openreview.net/forum?id=Ny8NiVfi95
Compressor summary: MAGNeT is a fast text-to-music model that combines masked sequence modeling with rescoring and hybrid decoding, achieving comparable quality to autoregressive models.
https://openreview.net/forum?id=Ny150AblPu
Compressor summary: D-TIML is a new method that uses diffusion models to find and visualize inconsistencies between texts and images in misleading pairs, and it comes with a large dataset to test its effectiveness.
https://openreview.net/forum?id=NxoFmGgWC9
Compressor summary: The paper presents GR-1, a GPT-style model for language-conditioned visual robot manipulation, which improves success rates on the CALVIN benchmark and real robots by using large-scale video generative pre-training.
https://openreview.net/forum?id=NvbeD9Ttkx
Compressor summary: FOSI is a meta-algorithm that enhances first-order optimizers by using second-order information to improve convergence rate and optimization time.
https://openreview.net/forum?id=Nu9mOSq7eH
Compressor summary: This paper proposes a unified language interface for computer vision tasks that uses text-to-image generation to perform various tasks like segmentation, object detection, etc., with the help of natural language instructions.
https://openreview.net/forum?id=Nu7dDaVF5a
Compressor summary: NFP is a method that uses scene priors to train a network that reconstructs 3D scenes from single-view RGB-D images without a fusion module and enables single-image novel-view synthesis.
https://openreview.net/forum?id=NtQqIcSbqv
Compressor summary: The authors study how human hands manipulate objects, collect visual-tactile data, and propose a method to learn a cross-modal latent manifold for manipulation prediction and discovery of latent structure.
https://openreview.net/forum?id=Nshk5YpdWE
Compressor summary: LFlows are a novel method for modeling fluid dynamics using differentiable density transformations and parameterized diffeomorphisms, outperforming existing methods in accuracy and efficiency.
https://openreview.net/forum?id=NsCXDyv2Bn
Compressor summary: VoiceGen is a method to generate diverse and high-quality voice variability from text prompts using a variation network and a prompt generation pipeline, addressing challenges in traditional TTS approaches.
https://openreview.net/forum?id=Nq45xeghcL
Compressor summary: The paper proposes a new algorithm called RISC that trains agents without resets by intelligently switching between two agents based on confidence, and shows its effectiveness on various challenging environments.
https://openreview.net/forum?id=NnyD0Rjx2B
Compressor summary: The paper introduces fairret, a framework for machine learning fairness that integrates easily with automatic differentiation libraries using linear-fractional statistics and shows its effectiveness in experiments.
https://openreview.net/forum?id=NnYaYVODyV
Compressor summary: The paper presents a model called Perceptual Group Tokenizer that uses grouping operations to extract visual features and perform self-supervised representation learning, achieving competitive performance compared to state-of-the-art vision architectures.
https://openreview.net/forum?id=NltzxpG0nz
Compressor summary: Steve-Eye is a multimodal model that combines language and vision to enable embodied agents to better interact with open worlds, using a large dataset and three essential functions for an agent.
https://openreview.net/forum?id=NkmJotfL42
Compressor summary: The text discusses the difficulty of finding tight generalization bounds for neural networks in the overparameterized setting and shows that such bounds are either impossible or come with trade-offs.
https://openreview.net/forum?id=NjU0jtXcYn
Compressor summary: ColaBO is a framework that allows domain experts to customize Bayesian optimization with their own prior beliefs about the function being optimized, improving its sample efficiency and performance.
https://openreview.net/forum?id=NjNfLdxr3A
Compressor summary: ELoRA is a method to reduce storage requirements and trainable parameters in natural language processing by freezing random low-rank matrices and using small scaling vectors.
https://openreview.net/forum?id=NjNGlPh8Wh
Compressor summary: The paper explores how allowing transformers to generate and condition on intermediate tokens (chain of thought) can extend their computational power, depending on the number of steps taken, and characterizes the complexity class of problems they can solve based on this factor.
https://openreview.net/forum?id=NialiwI2V6
Compressor summary: MOTOR is a self-supervised foundation model that uses pretraining on large patient records to improve time-to-event prediction performance and adaptability across medical tasks and databases.
https://openreview.net/forum?id=NgaLU2fP5D
Compressor summary: The paper presents PSI-KT, a hierarchical generative model that combines cognitive traits and knowledge structure to optimize learning materials and improve understanding and retention in intelligent tutoring systems, while maintaining interpretability and efficiency.
https://openreview.net/forum?id=Nf4Lm6fXN8
Compressor summary: RaE is a simple framework that reuses experience from previous experiments to improve RL performance and research iteration times, while being robust to different data quality and amount.
https://openreview.net/forum?id=NddKiWtdUm
Compressor summary: The text introduces a new way to train language models that allows them to learn social values from simulated interactions, improving their generalization and robustness.
https://openreview.net/forum?id=NdcQQ82mfy
Compressor summary: The paper proposes a hybrid RL approach for enhancing branching policies in MIP problems using cost-effective sample augmentation, achieving superior results compared to existing methods.
https://openreview.net/forum?id=NYN1b8GRGS
Compressor summary: GIM is a self-training framework that uses internet videos to improve the generalization of image matching models, and ZEB is a new benchmark for evaluating cross-domain performance.
https://openreview.net/forum?id=NY3wMJuaLf
Compressor summary: FedCOG is a novel method for federated learning that generates complementary data to reduce data heterogeneity and improve model performance, without changing the communication process.
https://openreview.net/forum?id=NW31gAylIm
Compressor summary: FedTPG is a scalable federated learning method for vision-language models that learns context-aware prompts across clients, improving generalization to seen and unseen classes and datasets.
https://openreview.net/forum?id=NW2s5XXwXU
Compressor summary: Our method improves long-tail diffusion model generation by directly transferring knowledge from head data using the score function's multi-objective characteristics, enhancing diversity and robustness.
https://openreview.net/forum?id=NU9AYHJvYe
Compressor summary: The text studies how many labeled examples are needed to learn data representations well using contrastive learning, and gives optimal or near-optimal bounds for different distance functions.
https://openreview.net/forum?id=NSVtmmzeRB
Compressor summary: GeoBFN is a new model that generates high-quality 3D molecule geometry by modeling diverse modalities in the differentiable parameter space of distributions and maintaining SE-(3) invariant density modeling property.
https://openreview.net/forum?id=NSIVHTbZBR
Compressor summary: The paper presents a method to improve diffusion models for inpainting by using probabilistic circuits (PCs) to compute exact conditional probabilities and guide the image generation process with semantic constraints.
https://openreview.net/forum?id=NSDszJ2uIV
Compressor summary: The MARCEL benchmark evaluates how well molecular representation learning models perform with flexible molecules considering different structures and tasks.
https://openreview.net/forum?id=NLevOah0CJ
Compressor summary: PRIOR is a method to learn preferences from feedback faster and better by assigning rewards based on the importance of states in a trajectory, improving performance on locomotion and manipulation tasks.
https://openreview.net/forum?id=NLbRvr840Q
Compressor summary: The paper introduces hypergraph dynamic systems (HDS) that improve the performance and controllability of hypergraph neural networks (HGNNs) by bridging them with dynamic systems and using ordinary differential equations (ODEs).
https://openreview.net/forum?id=NLPzL6HWNl
Compressor summary: FFA-LoRA is a modified version of LoRA that improves efficiency and stability in federated learning by fixing non-zero matrices and only fine-tuning zero-initialized ones.
https://openreview.net/forum?id=NGVljI6HkR
Compressor summary: The paper shows that searching in the programmatic space, induced by a domain-specific language, performs better and is easier to optimize than searching in latent spaces learned by systems like LEAPS and HPRL.
https://openreview.net/forum?id=NG7sS51zVF
Compressor summary: StreamingLLM is a framework that allows large language models to handle long streaming interactions efficiently by using a dedicated attention sink token and caching previous key-value states.
https://openreview.net/forum?id=NDkpxG94sF
Compressor summary: The paper presents a 3D object detector for point clouds using DETR framework with improved performance and efficiency, achieved by introducing 3D Vertex Relative Position Encoding and refining the pipeline.
https://openreview.net/forum?id=N8N0hgNDRt
Compressor summary: MetaMath is a finetuned language model that specializes in mathematical reasoning, achieving significant improvements over existing open-source LLMs on mathematical question answering benchmarks.
https://openreview.net/forum?id=N6o0ZtPzTg
Compressor summary: The study proposes Prompt-OIRL, a method that uses offline inverse reinforcement learning to optimize prompts for improving arithmetic reasoning in Large Language Models without accessing them during evaluation.
https://openreview.net/forum?id=N2WchST43h
Compressor summary: The paper analyzes the convergence of adversarial training on a two-layer neural network and proposes an efficient algorithm with reduced time cost.
https://openreview.net/forum?id=N23A4ybMJr
Compressor summary: The paper proposes a Win-Win method for efficient training and inference of high-resolution vision transformers using random windows, achieving fast and state-of-the-art results in semantic segmentation and optical flow tasks.
https://openreview.net/forum?id=N0nTk5BSvO
Compressor summary: The paper introduces TESTAM, a deep learning model that uses a mixture-of-experts to separately model recurring and non-recurring traffic patterns using different graph representations for better traffic forecasting.
https://openreview.net/forum?id=N0gT4A0jNV
Compressor summary: The paper proposes an efficient and robust alternating minimization framework for low rank matrix completion with incolemntary rows and columns, achieving nearly linear time complexity in the sample complexity.
https://openreview.net/forum?id=N0I2RtD8je
Compressor summary: Vision-language models can be used as reward models for reinforcement learning, enabling sample-efficient learning of complex tasks without manual rewards.
https://openreview.net/forum?id=MzjiMxlWab
Compressor summary: The paper proposes a new method for learning user preferences from behavioral data by aligning them with multi-faceted item characteristics using prototype-based representation learning and a bi-directional binding block.
https://openreview.net/forum?id=My7lkRNnL9
Compressor summary: The text discusses the challenges and performance of forward-only algorithms for training neural networks without backpropagation, comparing different versions and linking them to feedback alignment and backpropagation.
https://openreview.net/forum?id=MukGKGtgnr
Compressor summary: The paper proposes methods for discovering the whole causal structure of linear models with latent variables under mild assumptions and proves their identifiability.
https://openreview.net/forum?id=MrYiwlDRQO
Compressor summary: PeFLL is a new personalized federated learning algorithm that improves accuracy, reduces computation and communication, and has theoretical guarantees by jointly training an embedding network and a hypernetwork.
https://openreview.net/forum?id=MrR3rMxqqv
Compressor summary: This paper studies how many example sequences multi-head attention mechanisms in transformers can memorize and introduces new assumptions about input data independence.
https://openreview.net/forum?id=MpyFAhH9CK
Compressor summary: The authors propose a reinforcement learning approach for controlling reconfigurable soft robots that can change their shape during their lifetime, and introduce Morphological Maze, a benchmark for testing these robots.
https://openreview.net/forum?id=MnMWa94t12
Compressor summary: The Dynamic Scene Transformer (DyST) is a model that captures 3D structure and dynamics of real-world scenes from monocular videos using neural scene representation and co-training on DySO dataset.
https://openreview.net/forum?id=MloaGA6WwX
Compressor summary: The paper introduces TADRED, a data-driven method that optimizes experimental design and machine learning for imaging tasks, improving acquisition time, cost, and deployment.
https://openreview.net/forum?id=MiRPBbQNHv
Compressor summary: The authors study how well deep neural networks (DNNs) can replicate human peripheral vision and find that while they improve with training, they still struggle to match human performance in certain conditions.
https://openreview.net/forum?id=Mhb5fpA1T0
Compressor summary: The paper proposes a method to train robots from videos without action annotations, using images and text as representations, and shows its effectiveness on table-top tasks.
https://openreview.net/forum?id=Mgq6kxl115
Compressor summary: The Diffusion Bridge Network is a new method to predict using multiple neural networks without the high cost of running each one individually, by simulating an equation that connects their outputs.
https://openreview.net/forum?id=MeHmwCDifc
Compressor summary: The paper explores the consistency of reward models in Reinforcement Learning from Human Feedback and proposes methods to measure and improve it, showing its impact on chatbot performance.
https://openreview.net/forum?id=MeB86edZ1P
Compressor summary: The authors propose a new method using lateral connections and Hebbian learning to prevent catastrophic forgetting in spiking neural networks, which could enable energy-efficient continual learning in neuromorphic computing systems.
https://openreview.net/forum?id=McfYbKnpT8
Compressor summary: The paper proposes using supervised learning and classic heuristics to create instance-specific presolving for MIP solvers, which can significantly improve their performance.
https://openreview.net/forum?id=MbfAK4s61A
Compressor summary: The study introduces CipherChat, a framework to test safety alignment of LLMs in non-natural languages like ciphers and finds that some ciphers can bypass LLMs' safety mechanisms, while also discovering a novel human cipher called SelfCipher.
https://openreview.net/forum?id=MY0qlcFcUg
Compressor summary: DTR is a simple add-on strategy for diffusion models that improves performance by selectively activating channels based on task affinity and weights.
https://openreview.net/forum?id=MVmT6uQ3cQ
Compressor summary: The OPTIN framework is a tool to increase the efficiency of pre-trained transformer architectures without re-training, using intermediate feature distillation to capture model dependencies and achieve competitive performance on various tasks.
https://openreview.net/forum?id=MVe2dnWPCu
Compressor summary: PICLE is a modular continual learning framework that uses a probabilistic model to efficiently evaluate module compositions, achieving perceptual, few-shot and latent transfer in large search spaces.
https://openreview.net/forum?id=MSe8YFbhUE
Compressor summary: The paper proposes DrM, a visual RL method that reduces agent's inactivity by minimizing dormant ratio, improving sample efficiency and performance across different continuous control environments.
https://openreview.net/forum?id=MREQ0k6qvD
Compressor summary: The authors propose a new method to model dynamic changes in neural interactions by incorporating prior information about anatomical connectomes into a state-switching generalized linear model.
https://openreview.net/forum?id=MOmqfJovQ6
Compressor summary: This paper proposes an optimization problem to find the best way to group actions in reinforcement learning to balance exploration and exploitation in high-dimensional spaces.
https://openreview.net/forum?id=MO632iPq3I
Compressor summary: DECT is a fast and efficient method for classifying graphs and point clouds using the Euler Characteristic Transform with task-specific learned representations.
https://openreview.net/forum?id=MO5PiKHELW
Compressor summary: The study analyzes how syntax acquisition in masked language models (MLMs) is influenced by the evolution of Syntactic Attention Structure (SAS), which is essential for developing grammatical capabilities, but can also hinder model quality.
https://openreview.net/forum?id=MNyOI3C7YB
Compressor summary: SEABO is a search-based method for offline imitation learning that learns a reward function from expert and unlabeled data, achieving competitive performance to offline RL algorithms with ground-truth rewards.
https://openreview.net/forum?id=MNShbDSxKH
Compressor summary: The paper proposes a generative neuro-symbolic visual reasoning model that grows and reuses modules using large language models, achieving competitive performance and fast generalization to new tasks.
https://openreview.net/forum?id=MN3yH2ovHb
Compressor summary: SyncDreamer is a new diffusion model that creates consistent images from multiple views using a 3D-aware feature attention mechanism.
https://openreview.net/forum?id=MLBdiWu4Fw
Compressor summary: InternVid is a large dataset with millions of videos and their descriptions that enables learning powerful video-text representations for various multimodal tasks, such as action recognition, retrieval, and dialogue systems.
https://openreview.net/forum?id=MJksrOhurE
Compressor summary: The CARD Transformer model improves time series forecasting by capturing channel correlations, generating tokens with different resolutions, and using a robust loss function that considers prediction uncertainties.
https://openreview.net/forum?id=MJJcs3zbmi
Compressor summary: This paper proposes an objective discovery method that enables reinforcement learning algorithms to adapt their learning strategies based on the training horizon, leading to better generalization and performance across different tasks.
https://openreview.net/forum?id=MJ3K7uDGGl
Compressor summary: Temperature scaling in knowledge distillation (KD) can be replaced by transformed teacher matching (TTM), which uses a Rényi entropy term for better generalization, and further improved by weighted TTM (WTTM).
https://openreview.net/forum?id=MIEnYtlGyv
Compressor summary: Symphony is a new generative model for 3D molecular geometries that uses higher-degree E(3)-equivariant features to efficiently represent probability distributions and generate accurate molecules.
https://openreview.net/forum?id=MHjigVnI04
Compressor summary: The study finds that gradient descent in high-dimensional classification tasks aligns with low-rank outlier eigenspaces of Hessian and gradient matrices, leading to suboptimal classifiers in multi-layer settings.
https://openreview.net/forum?id=MFCjgEOLJT
Compressor summary: The authors propose a control theoretic approach to model the zebrafish larva's natural locomotor repertoire using a sparse control signal driving a latent RNN that preserves kinematic features and disentangles different categories of movements.
https://openreview.net/forum?id=MEGQGNUfPx
Compressor summary: FOMO is a novel learning paradigm that alternates between forgetting and relearning phases to mitigate robust overfitting and improve neural network robustness against adversarial attacks.
https://openreview.net/forum?id=MCl0TLboP1
Compressor summary: HUBL is a technique that improves offline RL performance by blending heuristic values with bootstrapped values based on Monte-Carlo returns, reducing complexity and increasing sample efficiency.
https://openreview.net/forum?id=MCNqgUFTHI
Compressor summary: PPDPP is a new dialogue policy planning paradigm that improves the proactivity of large language models in different applications by using a tunable plug-and-play dialogue policy planner.
https://openreview.net/forum?id=M6XWoEdmwf
Compressor summary: AMAGO is an in-context RL agent that uses sequence models, sparse rewards, off-policy data, and hindsight relabeling to solve goal-conditioned problems and extend in-context learning to open-world domains.
https://openreview.net/forum?id=M3QXCOTTk4
Compressor summary: The text describes a negative effect of training diverse data-sharing agents, called the curse of diversity, and proposes a solution named Cross-Ensemble Representation Learning (CERL) to overcome it.
https://openreview.net/forum?id=M0xK8nPGvt
Compressor summary: C-PSRL is a new posterior sampling method for reinforcement learning that uses a causal graph as prior knowledge, improving sample efficiency and Bayesian regret.
https://openreview.net/forum?id=M0MF4t3hE9
Compressor summary: The paper proposes Instruction-oriented Object Detection (IOD), a task that enables detectors to understand user instructions and locate objects, introducing IOD-Bench dataset and Ins-DetCLIP model.
https://openreview.net/forum?id=LzPWWPAdY4
Compressor summary: LoftQ is a novel quantization framework that improves the performance of large language models by simultaneously quantizing them and finding a proper low-rank initialization for LoRA fine-tuning, especially in challenging mixed precision regimes.
https://openreview.net/forum?id=Lvf7GnaLru
Compressor summary: This paper studies diversification methods for handling spurious features and finds that their performance depends on the distribution of unlabeled data, learning algorithm choice, and the number of diverse hypotheses, with no clear improvement from increasing diversity.
https://openreview.net/forum?id=LtuRgL03pI
Compressor summary: InstructScene is a new framework for generating 3D indoor scenes from natural language instructions, improving controllability and fidelity with a semantic graph prior and a layout decoder.
https://openreview.net/forum?id=LsURkIPYR5
Compressor summary: LaneSegNet is a new mapping network that generates lane segments, incorporating geometry and topology information, for autonomous driving systems, and outperforms previous methods on three tasks.
https://openreview.net/forum?id=LqaEEs3UxU
Compressor summary: Sign2GPT is a framework for sign language translation that uses pretrained vision and language models with lightweight adapters to translate gloss-free sign videos without needing annotated gloss orders.
https://openreview.net/forum?id=LqRGsGWOTX
Compressor summary: BO-REP is a new bilevel optimization algorithm that adapts to unbounded smoothness in neural networks by using normalized momentum and periodic updates, achieving state-of-the-art complexity results.
https://openreview.net/forum?id=LokR2TTFMs
Compressor summary: The paper introduces a new method for pretraining 3D point clouds using masked autoencoders that focuses on recovering intrinsic point features instead of location, achieving better results on various point cloud analysis tasks.
https://openreview.net/forum?id=LnLySuf1vp
Compressor summary: SpikeGCL is a novel framework for graph contrastive learning with spiking neural networks that achieves efficient and high-quality 1-bit representations for graphs, outperforming many existing methods.
https://openreview.net/forum?id=LjivA1SLZ6
Compressor summary: EMU is a MARL algorithm that uses memory embeddings and an episodic reward to accelerate learning and avoid local optima in complex tasks.
https://openreview.net/forum?id=LjeqMvQpen
Compressor summary: The paper presents a method for fusing multiple transformer-based neural networks using Optimal Transport to align their components, achieving better performance than vanilla fusion or individual models on image and text tasks.
https://openreview.net/forum?id=LfmZh91tDI
Compressor summary: The paper explores how layer-wise averaging of neural network parameters can improve the performance of independent models, especially when trained on different datasets.
https://openreview.net/forum?id=Let8OMe20n
Compressor summary: Text-Image Alignment Assessment (TIA2) is a benchmark to study reward overoptimization in text-to-image models, and TextNorm helps reduce it by calibrating reward model scores.
https://openreview.net/forum?id=LemSSn8htt
Compressor summary: The paper introduces a new algorithm called $\Delta$-amortized inference that leverages sparsity in probabilistic graphical models to enable faster and more efficient inference using local credit assignment and learned samplers.
https://openreview.net/forum?id=LebzzClHYw
Compressor summary: Instructive Decoding (ID) improves the performance of instruction-tuned language models by generating diverse responses using noisy instructions, especially 'opposite'.
https://openreview.net/forum?id=LbJqRGNYCf
Compressor summary: JoMA is a new framework that analyzes how Transformers learn hierarchies of tokens by integrating the self-attention layer in MLP dynamics.
https://openreview.net/forum?id=LZIOBA2oDU
Compressor summary: The Langevinized Kalman Temporal-Difference (LKTD) algorithm uses Kalman filtering to create a scalable sampling method for deep reinforcement learning, considering the stochastic nature of agent-environment interactions and quantifying uncertainties in policy updates.
https://openreview.net/forum?id=LYG6tBlEX0
Compressor summary: H-GAP is a model that generates humanoid trajectories from human motion-captured data and excels in controlling 56 degrees of freedom humanoids without online learning or individual task training.
https://openreview.net/forum?id=LY3ukUANko
Compressor summary: Convolution-based language models struggle with associative recall, a key task in language modeling, due to fixed filters and variable input distances, but attention can solve this issue.
https://openreview.net/forum?id=LXVswInHOo
Compressor summary: IN-CONTEXT PRETRAINING improves language models' performance by training them on related documents, enabling better contextual reasoning.
https://openreview.net/forum?id=LWuYsSD94h
Compressor summary: The paper proposes a versatile black-box method for multi-agent learning with bandit feedback and non-stationary environments, achieving low regret and testing different types of equilibria.
https://openreview.net/forum?id=LWmuPfEYhH
Compressor summary: The proposed ACORM framework uses attention-guided contrastive role representation learning to improve behavior heterogeneity, knowledge transfer, and skillful coordination in multi-agent reinforcement learning.
https://openreview.net/forum?id=LUpC8KTvdV
Compressor summary: The paper proposes a method called MaskTAS that uses masked image modeling to search for efficient and effective vision transformer architectures without the need for labeled data.
https://openreview.net/forum?id=LSYhE2hLWG
Compressor summary: SineNet is a new deep neural network architecture that improves the modeling of complex, time-evolving dynamics in partial differential equations by using multiple sequentially connected U-shaped blocks to reduce temporally misaligned features and enable multi-scale processing.
https://openreview.net/forum?id=LQ6LQ8f4y8
Compressor summary: The new technique enhances imitation learning robustness by generating corrective data from expert demonstrations, accounting for errors and disturbances in the environment dynamics.
https://openreview.net/forum?id=LNLjU5C5dK
Compressor summary: FIGA is an improved alignment approach that uses fine-grained quality signals from contrasting good and bad responses to train large language models.
https://openreview.net/forum?id=LJWizuuBUy
Compressor summary: The study presents a new neural network pruning technique that effectively reduces noisy gradients using entropic Wasserstein regression, achieving better performance when the network size or target sparsity is large.
https://openreview.net/forum?id=LEYUkvdUhq
Compressor summary: ZipIt! is a method to combine different models trained on separate tasks into one multi-task model without retraining by using a general "zip" operation and partially zipping the models up until a specified layer.
https://openreview.net/forum?id=L9U5MJJleF
Compressor summary: The concept bottleneck layer enhances interpretable generative models by adding a human-understandable concept encoding and loss function, which works across different model families and significantly outperforms baselines.
https://openreview.net/forum?id=L8UNn7Llt4
Compressor summary: The authors propose a modified DICE method for offline reinforcement learning and imitation learning, which uses an orthogonal-gradient update to improve performance and robustness by resolving the conflict between forward and backward gradients.
https://openreview.net/forum?id=L6crLU7MIE
Compressor summary: The text discusses how to assign and estimate exchange values for individual agents in AI systems based on their contribution to the collective desirability score, enabling better imitation of desired behaviors.
https://openreview.net/forum?id=L6L1CJQ2PE
Compressor summary: MALMEN is a method to update large language models with corrected knowledge by using a hyper-network to generate parameter shifts based on least square problem formulation.
https://openreview.net/forum?id=L4nOxziGf9
Compressor summary: RepARe is a framework that improves zero-shot performance of large vision-language models by rephrasing questions using caption and reasoning information extracted from the model itself, selecting the best question candidate based on its confidence score, and increasing syntactic complexity.
https://openreview.net/forum?id=L3FHMoKZcS
Compressor summary: The paper analyzes existing calibration methods for large language models and proposes Batch Calibration, a simple method that addresses contextual bias and improves performance on various tasks.
https://openreview.net/forum?id=L0r0GphlIL
Compressor summary: Teleportation speeds up optimization, improves convergence, and enhances generalization by using loss-invariant transformations in overparametrized models.
https://openreview.net/forum?id=Kz3yckpCN5
Compressor summary: The authors analyze the approach of finetuning language models on outputs from stronger models and find that it only works well for mimicking style but not factuality, suggesting that improving open-source models requires better base LMs instead of imitating proprietary ones.
https://openreview.net/forum?id=Kuj5gVp5GQ
Compressor summary: The Sinkhorn-Newton-Sparse (SNS) algorithm improves the convergence speed of optimal transport calculations by introducing early stopping and a Newton-type subroutine, achieving super-exponential convergence and $O(n^2)$ per-iteration complexity.
https://openreview.net/forum?id=Kuh5qgCGCp
Compressor summary: rainname is a training method that enables multimodal learning in low-resource languages using a strong multilingual language model and English data, achieving state-of-the-art results in Chinese image-to-text and text-to-image generation.
https://openreview.net/forum?id=KuPixIqPiq
Compressor summary: Self-debugging teaches large language models to debug their code by explaining it in natural language, achieving state-of-the-art performance on several code generation benchmarks.
https://openreview.net/forum?id=KsUh8MMFKQ
Compressor summary: This paper introduces ThinShellLab, a simulation platform for robots to learn and evaluate skills in manipulating diverse thin-shell materials, overcoming challenges through an optimization scheme that combines sampling-based and gradient-based methods.
https://openreview.net/forum?id=KrtGfTGaGe
Compressor summary: The Wasserstein Belief Updater is a reinforcement learning algorithm that learns a latent model and an approximate belief update for partially observable environments, providing theoretical guarantees on optimality.
https://openreview.net/forum?id=KrWuDiW4Qm
Compressor summary: The paper proposes a meta-learning method for discovering causal structures in physics-informed machine learning, which improves the robustness of out-of-distribution forecasting tasks.
https://openreview.net/forum?id=KqbCvIFBY7
Compressor summary: Particle guidance is a method to improve diversity and sample efficiency in generative models by using joint-particle time-evolving potentials instead of independent samples.
https://openreview.net/forum?id=KpoQSgxbKH
Compressor summary: SpeechFlow is a pre-trained generative model that can create realistic speech data for various tasks and surpass existing models.
https://openreview.net/forum?id=Koh0i2u8qX
Compressor summary: The paper proposes a graph defense method that is resistant to both intense attacks and heavy computation, using a denoising module and a Mixture-of-Experts approach.
https://openreview.net/forum?id=Kn7tWhuetn
Compressor summary: The paper introduces ForgetNet, a neural network model that reasons about algorithms without relying on historical embeddings, and G-ForgetNet, an improved version that uses a gating mechanism to selectively integrate history during training.
https://openreview.net/forum?id=Kl9CqKf7h6
Compressor summary: FedHyper is a novel algorithm for federated learning that adapts the learning rate and improves efficiency and accuracy in various scenarios.
https://openreview.net/forum?id=KkrDUGIASk
Compressor summary: The paper proposes a novel framework for heterogeneous collaborative perception, which enables agents with different sensors and models to work together effectively and securely.
https://openreview.net/forum?id=KjegfPGRde
Compressor summary: Decomposed Prompt Tuning (DePT) improves parameter-efficient fine-tuning by decomposing soft prompts into shorter tokens and low-rank matrices, reducing memory and time costs without sacrificing performance.
https://openreview.net/forum?id=KjOAHlKMF5
Compressor summary: The text introduces a generalized cascading bandit framework that considers user states and state transitions, and proposes two efficient and near-optimal algorithms for recommendation systems and online advertising.
https://openreview.net/forum?id=KiespDPaRH
Compressor summary: The paper proposes a lightweight, learning-free, and architecture-agnostic regularization method for NeRF models using optimal transport to handle dynamic scenes' dynamics in the probability distribution of light intensity.
https://openreview.net/forum?id=KgaBScZ4VI
Compressor summary: The paper studies how to use different uncertainty measures from generative language models to decide when to defer to a larger model for better quality output while reducing inference costs.
https://openreview.net/forum?id=KdVvOA00Or
Compressor summary: The paper proposes a nonparametric regularized approach called ReTaSA to estimate the importance weight function for addressing the target shift problem in regression settings with different marginal distributions of the continuous target variable in training and testing domains, while the conditional distribution of features remains the same.
https://openreview.net/forum?id=KbetDM33YG
Compressor summary: The paper proposes LeBeD, a score to estimate generalization error of GNNs on unlabeled real-world graphs under distribution shifts, using node prediction and structure reconstruction discrepancies.
https://openreview.net/forum?id=KZSEgJGPxu
Compressor summary: SNIP is a pre-training method for deep learning models that enhances their understanding of symbolic and numeric domains by jointly learning from both, leading to better performance across various tasks.
https://openreview.net/forum?id=KZJehvRKGD
Compressor summary: The paper proposes using $\mu$P parameterized residual networks with a scaled residual branch to transfer optimal hyperparameters across width and depth, and supports the idea theoretically with dynamical mean field theory.
https://openreview.net/forum?id=KY8ZNcljVU
Compressor summary: NetInfoF is a framework for measuring and exploiting network usable information for link prediction and node classification tasks using graph neural networks (GNNs).
https://openreview.net/forum?id=KWO8LSUC5W
Compressor summary: COSMO is a new method for learning the structure of directed graphs that doesn't check for cycles and still works well.
https://openreview.net/forum?id=KUNzEQMWU7
Compressor summary: MathVista is a benchmark to test large language and multimodal models' ability to do mathematical reasoning with visual contexts, showing they are not yet at human level performance.
https://openreview.net/forum?id=KTtEICH4TO
Compressor summary: The paper proposes a contact-based object representation and pretraining pipeline for nonprehensile manipulation using reinforcement learning, which enables efficient learning and zero-shot transfer to novel real-world objects.
https://openreview.net/forum?id=KTf4DGAzus
Compressor summary: The text discusses challenges in model selection for multi-modal agents in complex tasks and proposes a framework called M3 to improve it while maintaining low runtime overhead.
https://openreview.net/forum?id=KSjPaXtxP8
Compressor summary: The paper proposes a framework to define memorization in self-supervised learning and shows that it is essential for high performance of encoders on various tasks.
https://openreview.net/forum?id=KS8mIvetg2
Compressor summary: Our procedure detects test set contamination in language models by comparing the likelihood of canonical orderings versus shuffled orderings, without needing access to pretraining data or model weights.
https://openreview.net/forum?id=KQe9tHd0k8
Compressor summary: The paper proposes a novel algorithm that iteratively pseudo-labels instances based on aggregated labels and bag information, then refines embeddings using these pseudo-labels, and finally trains a classifier with the final pseudo-labels, achieving significant improvements in learning from label proportions problems for both tabular and image datasets.
https://openreview.net/forum?id=KQ2i6jazVK
Compressor summary: The paper presents a method to reconstruct 3D articulated objects from motion cues in videos without using category-specific skeletons and achieves better results than existing methods.
https://openreview.net/forum?id=KPmajBxEaF
Compressor summary: LEAP is a novel pose-free approach for sparse-view 3D modeling that uses a neural volume to encode geometry and texture priors, enabling fast and accurate novel view synthesis without camera poses.
https://openreview.net/forum?id=KOZu91CzbK
Compressor summary: This paper presents a framework for enhancing large language models with autonomous language agents that learn from environment feedback using policy gradient and retrospective summaries.
https://openreview.net/forum?id=KNvubydSB5
Compressor summary: The paper proposes a novel graph generative network that captures the hierarchical structure of graphs, generates communities and cross-edges in parallel, models the output distribution with a multinomial factorization, and achieves high quality results on complex graphs.
https://openreview.net/forum?id=KIPJKST4gw
Compressor summary: The paper investigates how introducing code data in different stages of training large language models can improve their reasoning capabilities for various applications.
https://openreview.net/forum?id=KI9NqjLVDT
Compressor summary: ReMasker is a simple and effective method for imputing missing values in tabular data using masked autoencoding, which learns representations invariant to missingness.
https://openreview.net/forum?id=KC58bVmxyN
Compressor summary: The authors propose a novel cognitive model that learns abstract relational structures across different domains and shows its similarity to human cognitive maps in the hippocampus.
https://openreview.net/forum?id=KBo7Z5aTV0
Compressor summary: The paper introduces PiXL, a framework for semantic segmentation that focuses on tailored learning of pixels, handling pixel-level variance, and enhancing per-pixel recognition capabilities.
https://openreview.net/forum?id=KAk6ngZ09F
Compressor summary: The paper introduces data filtering networks (DFNs) to create high-quality image-text datasets from large uncurated pools, and releases two new state-of-the-art datasets and a method to train DFNs from public data.
https://openreview.net/forum?id=K9sVJ17zvB
Compressor summary: VersVideo is a versatile video generation model that enhances spatial-temporal performance using multi-excitation paths, spatial-temporal attention blocks, and temporal consistency modules, while maintaining controllability across various conditions.
https://openreview.net/forum?id=K9V7ugVuUz
Compressor summary: The paper proposes a new method, DAR, to improve similarity-based representation learning by aligning differences between classes, reducing overfitting and tightening error bounds.
https://openreview.net/forum?id=K6kt50zAiG
Compressor summary: CAMBranch uses contrastive learning and variable shifting to improve MILP branching policies with limited expert data.
https://openreview.net/forum?id=K2c04ulKXn
Compressor summary: The paper proposes DBPM, a method to identify and suppress noisy and faulty positive pairs in time series contrastive learning, improving representation quality.
https://openreview.net/forum?id=K1mcPiDdOJ
Compressor summary: The paper proposes a conditional information bottleneck approach for time series imputation that preserves temporal context, reducing redundancy and improving performance in interpolation and extrapolation, as well as classification tasks.
https://openreview.net/forum?id=JzvIWvC9MG
Compressor summary: The paper presents algorithms to find game parameters that result in expected equilibrium behavior using generative adversarial optimization and shows their effectiveness in predicting electricity market prices.
https://openreview.net/forum?id=JzG7kSpjJk
Compressor summary: Per-IC quantization and Adaptive Dimensions are proposed methods that improve the efficiency of large language models by mitigating outliers and adapting to weight sensitivity patterns.
https://openreview.net/forum?id=JtKGkz9fAe
Compressor summary: ReFusion is a method that fuses retrieval representations into language models for non-knowledge-intensive tasks, improving efficiency and performance.
https://openreview.net/forum?id=JsnR0YO4Fq
Compressor summary: The study analyzes weight balancing, a simple but effective method for long-tailed data, and shows how it can be simplified by understanding its underlying mechanisms.
https://openreview.net/forum?id=Js5PJPHDyY
Compressor summary: The paper proposes a method that improves CLIP's zero-shot classification performance by using Gaussian Discriminant Analysis, which requires less training time and computational resources than existing methods.
https://openreview.net/forum?id=JrmPG9ufKg
Compressor summary: The paper proposes federated SimCLR variants that incorporate user verification and semi-supervised learning, and studies how different non-i.i.d.-ness sources affect their performance.
https://openreview.net/forum?id=JnYaF3vv3G
Compressor summary: The paper proposes a new algorithm for private training of machine learning models that combines gradient projection and stochastic gradient descent, improving both privacy and utility.
https://openreview.net/forum?id=JnRStoIuTe
Compressor summary: RSRS or RS2 is an efficient data pruning method for deep neural networks that reduces time-to-accuracy by learning from repeatedly sampled data subsets throughout training, outperforming many state-of-the-art methods.
https://openreview.net/forum?id=JiTVtCUOpS
Compressor summary: LIFT is a new method for multivariate time series forecasting that leverages local lead-lag relationships to improve predictions by using advance information from leading indicators.
https://openreview.net/forum?id=JgqftqZQZ7
Compressor summary: FLATTEN is a training-free method that improves visual consistency in text-to-video editing by using optical flow in the attention module of diffusion models' U-Nets.
https://openreview.net/forum?id=JfqN3gu0i7
Compressor summary: This paper analyzes the performance of kernel classification methods using statistical theory and proposes a way to estimate interpolation smoothness for practical applications.
https://openreview.net/forum?id=Jf5gplvglq
Compressor summary: The paper introduces Skill-Mix, a new evaluation method to measure how well LLMs can combine different skills they have learned, which reveals differences in model capabilities not captured by existing evaluations.
https://openreview.net/forum?id=JewzobRhay
Compressor summary: Context-based fine-tuning methods are less expressive than full fine-tuning and cannot change the relative attention pattern over the content.
https://openreview.net/forum?id=JePfAI8fah
Compressor summary: The iTransformer model improves Transformer-based forecasters by inverting the roles of attention and feed-forward networks for better representation of multivariate time series.
https://openreview.net/forum?id=Je5SHCKpPa
Compressor summary: MUSE is a graph contrastive learning method for multimodal patient representation learning that handles missing modalities and labels and reduces modality collapse.
https://openreview.net/forum?id=Jc0FssXh2R
Compressor summary: This paper analyzes when two-layer linear neural networks with feature learning outperform normal ridge regression in high-dimensional settings, using a new criterion that acts as an upper bound on the predictive risk and achieves optimal Bayes risk.
https://openreview.net/forum?id=JbcwfmYrob
Compressor summary: SEA is a method that combines sparse and linear attention to achieve better performance and interpretability than previous sparse and linear approaches for natural language understanding tasks while using less memory.
https://openreview.net/forum?id=JYu5Flqm9D
Compressor summary: The paper introduces and evaluates Codable Text Watermarking for LLMs (CTWL), a technique to embed hidden patterns in generated texts, carrying customizable information while preserving text quality.
https://openreview.net/forum?id=JXGph215fL
Compressor summary: The paper proposes a new framework for decision-time planning in imperfect-information games that does not rely on subgames, improves performance in games with large amounts of non-public information, and is based on the notion of update equivalence.
https://openreview.net/forum?id=JWpwDdVbaM
Compressor summary: ARM is a new multivariate temporal-contextual learning method that improves long-term time series forecasting by adapting to individual series patterns and inter-series dependencies, without increasing computational costs.
https://openreview.net/forum?id=JW3jTjaaAB
Compressor summary: AirPhyNet is a novel approach for air quality prediction that uses physics-based neural networks to overcome the limitations of traditional models and achieve better accuracy and interpretability.
https://openreview.net/forum?id=JSS9rKHySk
Compressor summary: The paper analyzes the challenges and applications of general function approximation in offline reinforcement learning, and introduces new lower bounds based on model-realizability.
https://openreview.net/forum?id=JORAfH2xFd
Compressor summary: The paper develops a framework to study the impact of training generative models on mixed datasets of real and synthetic data, proving their stability under certain conditions and quantifying error and radius stability.
https://openreview.net/forum?id=JO7k0SJ5V6
Compressor summary: This paper explores the effects of modifying Rotary Position Embedding in Large Language Models to improve their extrapolation capabilities, and proposes a unified framework to describe the relationship between base value and tuning context length.
https://openreview.net/forum?id=JN7TcCm9LF
Compressor summary: The paper proposes a new generalization bound for neural networks using Koopman operators that works for full-rank weight matrices and is independent of the network width if the weights are orthogonal.
https://openreview.net/forum?id=JDud6zbpFv
Compressor summary: CCQD is a new optimization framework that simplifies the QD problem by dividing a policy network into representation and decision layers, which coevolve together to improve sample efficiency.
https://openreview.net/forum?id=J7ioefqDPw
Compressor summary: The paper studies how graph neural networks (GNNs) can be attacked by manipulating node labels, reveals evaluation issues in the existing literature, and proposes new and stronger attacks.
https://openreview.net/forum?id=J44HfH4JCg
Compressor summary: The paper presents a new dataset and evaluation method for visual instruction tuning, and demonstrates how to reduce hallucination in multi-modal models by using them.
https://openreview.net/forum?id=J2TZgj3Tac
Compressor summary: The paper introduces new algorithms for deep RL in two-agent environments that reduce exploitability and improve convergence, especially Self-Play PSRO which incorporates an optimal stochastic policy into the population.
https://openreview.net/forum?id=J1djqLAa6N
Compressor summary: The study improves score matching methods in deep learning by using deep equilibrium models, which reduce memory costs and enable better modeling.
https://openreview.net/forum?id=IzqZbNMZ0M
Compressor summary: The paper proposes a new algorithm for private optimization that has the same optimal complexity as its non-private counterpart and works on nonconvex and nonsmooth objectives.
https://openreview.net/forum?id=Iyve2ycvGZ
Compressor summary: BOSS is a technique to improve efficiency and image quality of flow-matching generative models by optimizing step sizes and refining the velocity network.
https://openreview.net/forum?id=IxpTsFS7mh
Compressor summary: VQ-TR is a new method for probabilistic time series forecasting that uses discrete latent representations and Attention module to achieve accuracy and efficiency, outperforming current Transformer-based methods.
https://openreview.net/forum?id=Ixi4j6LtdX
Compressor summary: MPDistil is a meta-policy distillation technique that improves knowledge transfer from larger to smaller language models using collaboration, competition, and curriculum learning.
https://openreview.net/forum?id=IuXR1CCrSi
Compressor summary: The paper explores how different ways of writing graphs as text affect the ability of large language models to reason with them, and shows that choosing the right encoding method can significantly improve performance.
https://openreview.net/forum?id=IoKRezZMxF
Compressor summary: The paper presents an efficient method for editing videos using text instructions, which generates paired samples of input and edited videos with consistent long video lengths across batches.
https://openreview.net/forum?id=IkmD3fKBPQ
Compressor summary: The paper explores how well large language models can correct themselves without external feedback and suggests improvements for their reasoning abilities.
https://openreview.net/forum?id=IjMUGuUmBI
Compressor summary: GraphChef is a new GNN model that creates human understandable rules (recipes) to explain the classes in a dataset using decision trees.
https://openreview.net/forum?id=Ifz3IgsEPX
Compressor summary: DP-OPT is a method to create privacy-friendly prompts for large language models without compromising their performance, by tuning them on the client side and applying them to cloud models.
https://openreview.net/forum?id=IcVNBR7qZi
Compressor summary: Reinforcement finetuning (RFT) for language models suffers from vanishing gradients due to small reward standard deviation, which harms optimization and performance compared to supervised finetuning (SFT).
https://openreview.net/forum?id=IcR1OOFzxm
Compressor summary: The paper proposes a conditional generative model (RAISE) that encodes image attributes as latent concepts and decomposes rules into atomic rules, enabling better abstract visual reasoning in machine intelligence for Raven's Progressive Matrix problems.
https://openreview.net/forum?id=IZMPWmcS3H
Compressor summary: The text describes a novel method for generating high-quality 3D models from text using a single-stage optimization, addressing artifacts, inconsistencies, and flickering issues in the previous methods.
https://openreview.net/forum?id=IYxDy2jDFL
Compressor summary: The paper proposes a method called pivotal sampling to improve active learning by reducing sample size and improving coverage, and provides two theoretical results supporting its effectiveness.
https://openreview.net/forum?id=ITq4ZRUT4a
Compressor summary: The text proposes DSG, a graph-based QG/A framework for evaluating text-to-image models that addresses reliability challenges in existing methods.
https://openreview.net/forum?id=IRcv4yFX6z
Compressor summary: CAST is a novel learning framework that combines image segmentation and recognition using adaptive segment tokens inspired by vision transformers.
https://openreview.net/forum?id=IPhm01y9a9
Compressor summary: The paper proposes a simple bug fix called "absolute win" that improves the performance of two computer vision models, Hiera and ViTDet, by correctly handling window attention and position embeddings.
https://openreview.net/forum?id=IOEEDkla96
Compressor summary: The paper proposes a defense strategy for deep neural networks against backdoor attacks by pruning the feature maps that extract the attack information, achieving better performance and coverage than existing methods.
https://openreview.net/forum?id=ILqA09Oeq2
Compressor summary: The paper compares two methods for detecting a hidden signal in multi-view clustering data: one using tensor decomposition and another using matrix approximation, finding that the latter performs worse and has an optimal threshold.
https://openreview.net/forum?id=ILYjDvUM6U
Compressor summary: UAICRL is a robust ICRL algorithm that models and mitigates aleatoric and epistemic uncertainties to infer and adhere to expert constraints more accurately.
https://openreview.net/forum?id=IL71c1z7et
Compressor summary: The paper proposes fleet-merge, a method for distributed learning of policies among robots that can handle symmetries and improve performance on various tasks, including a new tool-use benchmark.
https://openreview.net/forum?id=IGzaH538fz
Compressor summary: GraphGuard is a certified defense that robustly predicts labels for graphs under perturbations, overcoming limitations of existing provable defenses and empirical defenses in graph classification.
https://openreview.net/forum?id=IEduRUO55F
Compressor summary: Eureka is an algorithm that uses large language models to generate human-level rewards for complex low-level manipulation tasks, outperforming expert-designed rewards and enabling in-context learning.
https://openreview.net/forum?id=I5webNFDgQ
Compressor summary: DiffusionSat is a large generative model that produces realistic satellite images using metadata and outperforms previous methods.
https://openreview.net/forum?id=I4wB3HA3dJ
Compressor summary: DISAM is an algorithm that optimizes under domain shifts and improves convergence and generalization by adjusting gradient perturbations based on domain losses.
https://openreview.net/forum?id=I2mIxuXA72
Compressor summary: This paper examines how machine learning algorithms for domain generalization handle label noise and spurious correlations, finding that they have some benefits but not necessarily better performance than classic empirical risk minimization.
https://openreview.net/forum?id=I1quoTXZzc
Compressor summary: ECBMs improve black-box deep learning models by capturing high-order interactions between concepts and quantifying conditional dependencies, leading to higher accuracy and richer interpretations.
https://openreview.net/forum?id=HrRKc9ei7h
Compressor summary: The paper proposes an efficient algorithm for online prediction with group-wise regret guarantees, based on a modification of the sleeping experts technique.
https://openreview.net/forum?id=HobyL1B9CZ
Compressor summary: The paper proposes Chain-of-Experts, a multi-agent framework using large language models to solve complex operation research problems, which also introduces a new benchmark dataset.
https://openreview.net/forum?id=HiYMiZYwkw
Compressor summary: SMA is a domain-agnostic masked modeling method that learns masks to sample data without domain-specific assumptions, achieving state-of-the-art performance in self-supervised learning across protein biology, chemical property prediction, and particle physics.
https://openreview.net/forum?id=HiTg16qhxp
Compressor summary: The paper proposes a novel mechanism called Dynamic Neural Response Tuning (DNRT) for artificial neural networks that mimics biological neurons' dynamic response conditions and improves their performance in various tasks and domains.
https://openreview.net/forum?id=HhfcNgQn6p
Compressor summary: The paper studies how to select a subset of unlabeled data samples to reduce labeling and computational costs, and shows that data selection can improve performance over training on the full sample.
https://openreview.net/forum?id=HgZUcwFhjr
Compressor summary: The paper presents RelatiViT, a new method based on transformers, to recognize spatial relations between objects in images more accurately than current approaches.
https://openreview.net/forum?id=HgOJlxzB16
Compressor summary: This paper shows how to train a 2-layer neural network with ReLU activations and standard SGD on XOR function data using a signal-finding phase and a signal-heavy phase, achieving low population error with polylogarithmic samples.
https://openreview.net/forum?id=HfXDrAzFvG
Compressor summary: The paper extends LipSDP to estimate Lipschitz bounds for neural networks with general activation functions like GroupSort, MaxMin, and Householder by using novel quadratic constraints derived from their properties.
https://openreview.net/forum?id=Hf17y6u9BC
Compressor summary: The text discusses the impact of methodological details in activation patching, a technique to understand machine learning models' internal mechanisms, and provides recommendations for its best practices.
https://openreview.net/forum?id=HdAoLSBYXj
Compressor summary: The paper proposes a novel contrastive learning method for neural topic models that balances the trade-off between reconstruction and generalization objectives using sets of topic vectors.
https://openreview.net/forum?id=HZndRcfyNI
Compressor summary: This paper characterizes how neural network architecture affects the choice of hyperparameters and proposes a method to generalize across different architectures for better AutoML performance.
https://openreview.net/forum?id=HZ3S17EI0o
Compressor summary: OKO learning improves accuracy and calibration of machine learning models by considering sets instead of individual examples.
https://openreview.net/forum?id=HYyRwm367m
Compressor summary: The paper introduces a novel model called SVQ that learns semantic neural discrete representations by quantizing scenes hierarchically and training a prior over them, achieving better generation performance and scene understanding compared to existing methods.
https://openreview.net/forum?id=HXoq9EqR9e
Compressor summary: FairVLM is a method to make vision-language models more fair and robust by jointly debiasing their image and text representations in RKHSs, improving zero-shot prediction accuracy and addressing societal biases and spurious features.
https://openreview.net/forum?id=HXc5aXeoc8
Compressor summary: The paper proposes two techniques to reduce artifacts in image generation using diffusion models, one based on Heavy Ball momentum and another on a variable trade-off between accuracy and artifact suppression.
https://openreview.net/forum?id=HXWTXXtHNl
Compressor summary: TDSM is a new method to train conditional diffusion models with noisy labels by estimating transition probabilities from clean and noisy labels and improving sample quality.
https://openreview.net/forum?id=HX5ujdsSon
Compressor summary: This paper explores how large language models like transformers use a Bayesian perspective to learn and generalize to new functions in different task families, as well as when they deviate from this behavior.
https://openreview.net/forum?id=HT2dAhh4uV
Compressor summary: Our method improves object-centric learning by explicitly encouraging compositionality in representation with a novel objective that ensures valid mixtures of object features from different images.
https://openreview.net/forum?id=HSKaGOi7Ar
Compressor summary: The paper introduces homomorphism expressivity as a novel framework for quantitatively assessing the expressiveness of Graph Neural Networks (GNNs), addressing limitations of existing measures and providing insights, bridging gaps, and settling open questions in the GNN community.
https://openreview.net/forum?id=HRkyLbBRHI
Compressor summary: COCOA is a novel approach to offline reinforcement learning that achieves conservatism by reparameterizing inputs as compositions of anchors and differences, improving the performance of four state-of-the-art algorithms on the D4RL benchmark.
https://openreview.net/forum?id=HMe5CJv9dQ
Compressor summary: The paper proposes a new differentially private data structure that efficiently approximates similarity queries between a large high-dimensional private dataset and any query point, using low-dimensional structures in the similarity function.
https://openreview.net/forum?id=HKgRwNhI9R
Compressor summary: The paper proposes a general method for continuous convolutions using separable basis functions in fluid simulations and shows that Fourier-based networks are more accurate and stable than other architectures, without needing window functions.
https://openreview.net/forum?id=HKGQDDTuvZ
Compressor summary: The FAT block is a novel method that uses multiscale directional analysis to improve learned image compression by capturing anisotropic frequency components and preserving directional details.
https://openreview.net/forum?id=HHbRxoDTxE
Compressor summary: The looped transformer is a new architecture that combines iterative features with transformers, enabling them to solve data-fitting problems more efficiently and with fewer parameters.
https://openreview.net/forum?id=HHWlwxDeRn
Compressor summary: The paper presents a novel method for teaching robots to manipulate diverse objects using sparse 3D features learned from images, feature refinement, and an energy function.
https://openreview.net/forum?id=HFtrXBfNru
Compressor summary: Smart is a method that uses self-supervised graph reconstruction to improve the representation of evolving graphs in Graph Neural Networks and reduce distortion.
https://openreview.net/forum?id=HE9eUQlAvo
Compressor summary: The paper proposes data selection methods based on influence estimation to enhance the utility, fairness, and robustness of convex learning models for a given function of interest, and tests them on diverse datasets and settings.
https://openreview.net/forum?id=HD5Y7M8Xdk
Compressor summary: VIS is a new method that improves latent variable model learning by maximizing the log-likelihood using optimal proposal distributions.
https://openreview.net/forum?id=HC0msxE3sf
Compressor summary: The paper reinterprets a common setting in emergent communication as a different model and shows how choosing the right prior distributions can make the resulting languages more like natural ones in terms of word lengths and segmentation.
https://openreview.net/forum?id=H98CVcX1eh
Compressor summary: The text discusses modular systems' ability to discover compositional structure in tasks, their theoretical and empirical analysis, and application to various problems.
https://openreview.net/forum?id=H4yQefeXhp
Compressor summary: DMV3D is a new method that uses a transformer model to denoise and reconstruct 3D images from multi-view data, achieving high-quality results for single-image reconstruction and text-to-3D generation.
https://openreview.net/forum?id=H4A9e8HvIn
Compressor summary: The paper presents new projection-free algorithms for optimal non-monotone and monotone adversarial continuous DR-submodular optimization with sub-linear regret bounds and various constraints and feedback types.
https://openreview.net/forum?id=H3UayAQWoE
Compressor summary: The paper introduces a framework called PPBench to evaluate psychological aspects of large language models and tests five popular models using it.
https://openreview.net/forum?id=H3IUunLy8s
Compressor summary: CAPABOOST is a strategy that enhances model capacity for fine-tuning large foundation models by using low-rank updates through parallel weight modules with random masks, improving performance on various downstream tasks without extra costs.
https://openreview.net/forum?id=H396R79GiQ
Compressor summary: The paper proposes a new method for micro-expression spotting that uses an unfixed reference frame, a skip-$k$-frame strategy, and a block-wise main directional mean optical flow feature to accurately detect and locate micro-expressions despite their small size and susceptibility to head movements.
https://openreview.net/forum?id=GzNhzX9kVa
Compressor summary: The authors create a dataset to explore model calibration properties in neural architecture search (NAS) using 117,702 unique neural networks across various datasets, robustness, and bin size measurements, addressing several longstanding questions in the field.
https://openreview.net/forum?id=GzNaCp6Vcg
Compressor summary: PINNACLE is a novel algorithm that optimizes the selection of all training points in Physics-Informed Neural Networks, improving their performance in various learning tasks.
https://openreview.net/forum?id=GxCGsxiAaK
Compressor summary: The paper explores the threat of poisoning reinforcement learning data with universal jailbreak backdoors that allow harmful responses without needing adversarial prompts.
https://openreview.net/forum?id=GruDNzQ4ux
Compressor summary: The paper proposes DreamSmooth, a reward smoothing method for model-based reinforcement learning that improves sample efficiency and performance on sparse-reward tasks.
https://openreview.net/forum?id=GnOLWS4Llt
Compressor summary: The text discusses challenges and solutions for offline reinforcement learning in complex applications involving partially observed states, proposing a bisimulation loss to improve sample efficiency.
https://openreview.net/forum?id=GlpawHh80l
Compressor summary: The paper proposes pp-SPA, a new approach to vertex hunting that generates pseudo-points and improves the performance of SPA by reducing noise and outliers in high-dimensional spaces.
https://openreview.net/forum?id=Glcsog6zOe
Compressor summary: The paper introduces Tree-Planner, a method that improves efficiency and scalability of task planning with LLMs by breaking down the process into three phases: plan sampling, action tree construction, and grounded deciding.
https://openreview.net/forum?id=GkJiNn2QDF
Compressor summary: FeatUp is a framework to restore spatial information in deep features, enabling better performance in dense prediction tasks like segmentation and depth prediction without retraining.
https://openreview.net/forum?id=GkJOCga62u
Compressor summary: This paper introduces orbit-equivariance, a relaxed form of equivariance in graph neural networks that enables functions to produce different outputs for similar nodes while preserving important structural biases.
https://openreview.net/forum?id=GgEAdqYPNA
Compressor summary: MMCL achieves robust representations by contrasting intra-class features and sharing inter-class details, preventing spurious features from dominating generalizable ones.
https://openreview.net/forum?id=Gg7cXo3S8l
Compressor summary: DC-FL is a novel contrastive learning approach for forward learning that improves performance and efficiency compared to existing methods and backpropagation.
https://openreview.net/forum?id=Gf15GsnfTy
Compressor summary: REValueD is a new reinforcement learning algorithm that uses value-decomposition to handle high-dimensional discrete action spaces and achieves better performance on challenging tasks.
https://openreview.net/forum?id=GaLCLvJaoF
Compressor summary: The text introduces a control-theoretic technique called $\mathcal{L}_1$-MBRL that improves the robustness of Model-Based Reinforcement Learning algorithms by generating an approximate model of the transition function and adapting the control input accordingly.
https://openreview.net/forum?id=GZ6AcZwA8r
Compressor summary: The paper proposes a novel class of graph kernels called MMD-GK that use MMD to measure graph similarities effectively and adaptively, and demonstrate their advantages over existing methods on various graph tasks.
https://openreview.net/forum?id=GXtmuiVrOM
Compressor summary: DORAEMON is a new method for reinforcement learning that adjusts dynamics parameters during training to maximize entropy and generalization, improving sim-to-real transfer.
https://openreview.net/forum?id=GWSIo2MzuH
Compressor summary: The paper proposes new PAC information-theoretic bounds for generalization analysis that are easier to compute and do not require the bounded loss assumption, by using loss entropies.
https://openreview.net/forum?id=GW4j4n2cjH
Compressor summary: The paper introduces a new task in dance generation called dance accompaniment, where a model generates responsive movements for a follower dancer synchronized with a lead dancer and music.
https://openreview.net/forum?id=GURqUuTebY
Compressor summary: The paper introduces DreamFlow, a fast and high-quality text-to-3D generation method that uses a novel optimization algorithm based on multi-view image-to-image translation and predetermined timesteps.
https://openreview.net/forum?id=GTk0AdOYLq
Compressor summary: The text proposes a diffusion probabilistic end-to-end model for generating raw speech waveforms, which can synthesize unlimited speech duration with high-fidelity, preserving temporal coherence and natural local acoustic behaviors.
https://openreview.net/forum?id=GTUoTJXPBf
Compressor summary: The paper analyzes how overparameterized neural networks generalize well despite interpolating noisy data, finding that they often exhibit "tempered overfitting" but can also suffer from catastrophic overfitting in certain scenarios.
https://openreview.net/forum?id=GSBHKiw19c
Compressor summary: The study introduces a new idea called dynamics reward to improve offline reinforcement learning and proposes MOREC, a method that achieves state-of-the-art performance on benchmarks.
https://openreview.net/forum?id=GQGNLEHmdl
Compressor summary: AutoChunk is an automatic compiler system that reduces activation memory for long sequence inference in deep learning models by applying chunk strategies.
https://openreview.net/forum?id=GPKTIktA0k
Compressor summary: The Reversal Curse is a phenomenon where large language models fail to generalize from one direction of a statement ("A is B") to its reverse ("B is A").
https://openreview.net/forum?id=GN921JHCRw
Compressor summary: RAPTOR is a novel approach that recursively embeds, clusters, and summarizes text chunks into a tree, enabling better integration and understanding of information across long documents for various tasks, especially complex question-answering.
https://openreview.net/forum?id=GIUjLsDP4Z
Compressor summary: The paper proposes a new structural encoding method for graph neural networks based on discrete Ricci curvature, which outperforms existing methods and improves downstream tasks more than rewiring techniques.
https://openreview.net/forum?id=GH2LYb9XV0
Compressor summary: Grokking is when a linear model learns to generalize well after training on simple tasks, and it can be explained by the dynamics of the training and generalization data covariance matrix.
https://openreview.net/forum?id=GEcwtMk1uA
Compressor summary: ToolEmu is a framework that uses an LM to emulate tool execution and enables scalable testing of LM agents against diverse tools and scenarios, helping identify risks associated with language model agents.
https://openreview.net/forum?id=GACjMj1MS1
Compressor summary: The paper proposes a method to assess and impose fairness constraints on machine learning algorithms using uncertainty and covariance measures, achieving accurate and fair results in decision-making systems.
https://openreview.net/forum?id=G7UtIGQmjm
Compressor summary: The authors propose a method to improve large language models' inductive reasoning ability by generating hypotheses in natural language and converting them into Python programs, which outperforms the direct prompting approach and requires minimal human input.
https://openreview.net/forum?id=G32oY4Vnm8
Compressor summary: PTaRL is a prototype-based representation learning framework that improves prediction performance and consistency for tabular tasks by constructing disentangled representations around global data prototypes.
https://openreview.net/forum?id=G2cG3mQqop
Compressor summary: IC$|$TC is a new method for image clustering based on text criteria, which gives users more control and better results than traditional methods.
https://openreview.net/forum?id=G1Hlubz1fR
Compressor summary: This paper proposes a novel modular skill learning approach that categorizes and efficiently organizes skills for neural networks, improving knowledge transfer and sample efficiency in multi-task learning.
https://openreview.net/forum?id=G0vdDSt9XM
Compressor summary: CRAFT is a framework that enhances LLMs by generating and retrieving task-specific code snippets, improving their performance on various complex tasks without fine-tuning.
https://openreview.net/forum?id=Fx2SbBgcte
Compressor summary: AnimateDiff is a framework for animating personalized text-to-image models by adding a plug-and-play motion module, which can be fine-tuned with MotionLoRA to adapt to new motion patterns.
https://openreview.net/forum?id=FvK2noilxT
Compressor summary: The paper presents GeneOH Diffusion, a novel approach to refine hand-object interactions from noisy sequences by using a contact-centric representation and a denoising via diffusion strategy.
https://openreview.net/forum?id=FsVxd9CIlb
Compressor summary: AttEXplore is a novel attribution method for DNN models that uses decision boundary exploration and frequency manipulation to provide enhanced explainability and outperforms other interpretability methods.
https://openreview.net/forum?id=FoMZ4ljhVw
Compressor summary: The paper introduces "Direct Inversion," a technique that separates source and target diffusion branches for better image editing, and presents PIE-Bench, a benchmark to evaluate it.
https://openreview.net/forum?id=Fn655mJ4bv
Compressor summary: The paper presents a new way to explain how structured models make decisions by finding important features in each part of the input space, considering output variable correlations, and training an energy-based interpreter function.
https://openreview.net/forum?id=FlvtjAB0gl
Compressor summary: The paper proposes LaVIT, a model that uses a visual tokenizer to translate images into tokens for LLM to understand and generate multi-modal content simultaneously, improving performance on vision-language tasks.
https://openreview.net/forum?id=FlhjUkC7vH
Compressor summary: DisenBooth is a framework that uses disentangled embeddings to generate images based on text descriptions while preserving the subject identity and controlling the irrelevant information.
https://openreview.net/forum?id=FlY7WQ2hWS
Compressor summary: Our method incentivizes clients in federated learning by providing differentiated training-time rewards, avoiding post-training challenges and ensuring optimal models.
https://openreview.net/forum?id=Fk5IzauJ7F
Compressor summary: The paper proposes a candidate label set pruning method that filters out false labels in a training-free manner, improving deep partial-label learning performance.
https://openreview.net/forum?id=Fj7Fzm5lWL
Compressor summary: The paper proposes a deep-learning framework to classify dynamical regimes and characterize bifurcation boundaries in various systems using topological invariants extracted from data.
https://openreview.net/forum?id=FhQSGhBlqv
Compressor summary: The paper presents a new framework for causal discovery that can handle hidden variables in complex networks, using rank information from covariance matrices.
https://openreview.net/forum?id=Ffjc8ApSbt
Compressor summary: The paper analyzes how to balance function classes in collaborative filtering to reduce biases, and proposes a universal kernel-based method and an adaptive causal balancing method for this purpose.
https://openreview.net/forum?id=Feiz5HtCD0
Compressor summary: The study finds that using InstructGPT for collaborative writing reduces diversity in essays, while GPT3 does not have this effect.
https://openreview.net/forum?id=FddFxi08J3
Compressor summary: The paper explores the connection between graph neural networks and the Weisfeiler-Leman test for distinguishing graphs based on their patterns, and provides a method to compute the least dimension needed for this distinction using local information.
https://openreview.net/forum?id=FdVXgSJhvz
Compressor summary: The paper introduces Alpagasus, a data selection strategy that uses a large language model to filter out low-quality instances from IFT datasets, resulting in faster training and improved performance on instruction-following tasks.
https://openreview.net/forum?id=FbuyDzZTPt
Compressor summary: The paper proposes a regularization method and a simplified prompt-based approach for class-incremental learning, improving accuracy and reducing computation cost compared to previous rehearsal-free methods.
https://openreview.net/forum?id=FYKVPOHCpE
Compressor summary: H-NTL is a novel method that uses causal modeling to disentangle content and style factors in learning non-transferable representations for better domain adaptation.
https://openreview.net/forum?id=FWJAmwE0xH
Compressor summary: NSR is a Neural-Symbolic model with a Grounded Symbol System core representation that learns perception, parsing, and reasoning modules and achieves better systematic generalization on various tasks.
https://openreview.net/forum?id=FVhmnvqnsI
Compressor summary: MDC is a method that compresses multiple dataset condensation processes into one with adaptive subset loss to address challenges in on-device scenarios and improve training efficiency.
https://openreview.net/forum?id=FUgrjq2pbB
Compressor summary: MVDream is a multi-view diffusion model that generates consistent 3D images from text prompts using both 2D and 3D data, improving on existing methods for 3D generation and concept learning.
https://openreview.net/forum?id=FRCHDhbxZF
Compressor summary: Scene Flow via Distillation is a method that uses unlabeled data to train a fast feedforward model for scene flow estimation without human labels, achieving state-of-the-art performance and enabling real-time applications.
https://openreview.net/forum?id=FQepisCUWu
Compressor summary: The paper introduces ChatEval, a multi-agent debating system using large language models to evaluate texts more accurately and effectively than single-agent approaches, and shows that diverse role prompts are crucial for its performance.
https://openreview.net/forum?id=FOSBQuXgAq
Compressor summary: The paper explores the complex posterior distribution of deep Bayesian Neural Networks, studying its approximation, quality, impact on uncertainty, visualization, and weight-space symmetries in various vision tasks and architectures.
https://openreview.net/forum?id=FNq3nIvP4F
Compressor summary: The paper introduces SEINE, a short-to-long video diffusion model that generates coherent and creative transitions between different scenes based on textual descriptions.
https://openreview.net/forum?id=FMMF1a9ifL
Compressor summary: GOLF is a new framework that uses neural networks and gradients to optimize molecular conformations faster and more accurately for drug discovery and materials design.
https://openreview.net/forum?id=FM5xfcaR2Y
Compressor summary: The paper proposes a simple modification rule for binary classifiers to satisfy different group fairness constraints, which can be applied to multiple datasets without needing sensitive attribute information during inference.
https://openreview.net/forum?id=FKksTayvGo
Compressor summary: Denoising Diffusion Bridge Models (DDBMs) are a new type of generative model that uses diffusion bridges, interpolating processes between two distributions, to map data without relying on random noise or cumbersome methods like guidance or projected sampling.
https://openreview.net/forum?id=FJiUyzOF1m
Compressor summary: Laplace-LoRA is a Bayesian method that improves calibration and reduces overconfidence in fine-tuned large language models using the Laplace approximation.
https://openreview.net/forum?id=FJWT0692hw
Compressor summary: SequenceMatch is a method that improves autoregressive modeling by using imitation learning to minimize divergences between generated sequences and the dataset, including out-of-distribution errors.
https://openreview.net/forum?id=FIplmUWdm3
Compressor summary: QLLM is an efficient and accurate low-bitwidth post-training quantization method for large language models that uses adaptive channel reassembly and tuning to mitigate outliers and improve performance.
https://openreview.net/forum?id=FIGXAxr9E4
Compressor summary: The text investigates how data-balancing can help reduce biases in CLIP models and discusses its strengths, limitations, and impact on different tasks.
https://openreview.net/forum?id=FHqAzWl2wE
Compressor summary: The paper proposes a method to learn joint distributions from multiple probability densities using dynamical transport of measure, which can extract correspondences among the marginals and has applications in various tasks.
https://openreview.net/forum?id=FE2e8664Sl
Compressor summary: The paper introduces Few-shot Hybrid Domain Adaptation (HDA), where a pre-trained generator adapts to multiple target domains with different attributes while preserving the source domain's characteristics using a novel directional subspace loss.
https://openreview.net/forum?id=FDb2JQZsFH
Compressor summary: The paper introduces an AID module that improves the performance of TPR-based models on systematic generalization tasks by enhancing the binding of structured representations to input features using attention.
https://openreview.net/forum?id=FDQF6A1s6M
Compressor summary: LOQA is a reinforcement learning algorithm that helps agents optimize their utility while cooperating in partially competitive environments and performs well in benchmark scenarios.
https://openreview.net/forum?id=FAGtjl7HOw
Compressor summary: The paper explores interpretable kernel clustering using decision trees to approximate nonlinear extensions of k-means, while maintaining flexibility and interpretability in practice.
https://openreview.net/forum?id=F7QnIKlC1N
Compressor summary: The paper introduces GTMGC, a novel deep learning network that predicts molecular conformations from their 2D topology using a new self-attention mechanism called MSRSA, and shows it outperforms existing methods on two benchmark datasets.
https://openreview.net/forum?id=F76bwRSLeK
Compressor summary: The paper proposes a method to identify interpretable features in neural networks that can help understand their internal workings and causal effects on tasks.
https://openreview.net/forum?id=F5dhGCdyYh
Compressor summary: The paper introduces a new type of adversarial attack on reinforcement learning agents that is hard to detect by both automated methods and humans, aiming to improve robustness against attacks on sensory inputs.
https://openreview.net/forum?id=F4bmOrmUwc
Compressor summary: This paper investigates zero-mean neural collapse for fixed non-negative orthogonal classifiers, which improve performance in classification problems by enhancing softmax masking and tackling mixup limitations.
https://openreview.net/forum?id=F1TKzG8LJO
Compressor summary: The RT-Trajectory method enables robots to generalize to new tasks by representing them through rough trajectory sketches that balance low-level guidance and situational context, and can be specified through simple human inputs or automated methods.
https://openreview.net/forum?id=EvyYFSxdgB
Compressor summary: The paper introduces a novel difficulty-aware task sampler (DATS) for meta-learning physics-informed neural networks (PINNs) that can improve the accuracy and reduce disparity across different partial differential equations (PDEs).
https://openreview.net/forum?id=EvDeiLv7qc
Compressor summary: The paper introduces a new Mixture of Experts (MoE) architecture that is extremely parameter-efficient, generalizes to unseen tasks, and performs well with limited resources.
https://openreview.net/forum?id=Ev10F9TWML
Compressor summary: The paper presents new methods to make the proofs of deep neural network robustness more understandable, showing that different training methods affect how the proofs use or remove input features.
https://openreview.net/forum?id=EriR6Ec69a
Compressor summary: The authors study how to make recurrent neural networks more robust to changes in the environment by adjusting their connectivity, and find that low-rank, sparse connectivity works best for closed-form continuous-time neural networks.
https://openreview.net/forum?id=EpYnZpDpsQ
Compressor summary: The paper proposes a self-supervised representation learning method that does not rely on data augmentations, but uses random projections for reconstruction. It shows superior performance on various tasks and modalities.
https://openreview.net/forum?id=EpVe8jAjdx
Compressor summary: Sensory scaffolding helps novice learners by providing additional observation streams, and _Scaffolder_ is a reinforcement learning approach that uses privileged sensing to improve artificial agents' performance in various simulated robotic tasks.
https://openreview.net/forum?id=Ep0TtjVoap
Compressor summary: ToRA agents use external tools to solve complex math problems, outperforming open-source models and achieving competitive results with GPT-4.
https://openreview.net/forum?id=Eo7kv0sllr
Compressor summary: This paper introduces emulated fine-tuning (EFT), a method to test the impact of scaling pre-training and fine-tuning stages in language models, and shows that it improves helpfulness and factuality in various scenarios.
https://openreview.net/forum?id=EnXJfQqy0K
Compressor summary: The text describes a novel framework, CoELA, that leverages large language models to enhance multi-agent cooperation with natural language communication and commonsense reasoning in embodied environments.
https://openreview.net/forum?id=EmQSOi1X2f
Compressor summary: This paper investigates self-contradictions in large language models, proposes a detection and mitigation framework using prompts, and releases a tool to help users avoid such issues.
https://openreview.net/forum?id=ElykcDu5YK
Compressor summary: The paper proposes a new meta-Bayesian optimization algorithm for risk measures that uses previous tasks' results, has scaling and shifting invariance, and robustness to harmful tasks.
https://openreview.net/forum?id=EhrzQwsV4K
Compressor summary: L2MAC is a novel memory-augmented language model that enables long and consistent code generation by managing context and interacting with a file store, outperforming other methods on system design tasks.
https://openreview.net/forum?id=EhmEwfavOW
Compressor summary: The paper presents a method to extend spectral convolutions to directed graphs, enabling new state of the art results in node classification and stability to topological perturbations.
https://openreview.net/forum?id=Eh0Od2BJIM
Compressor summary: HyperAttention is a new attention mechanism that addresses the computational challenges of large language models by using two parameters to measure the difficulty of the problem and achieving linear time sampling even with unbounded entries or large stable rank, leading to significant speed improvements.
https://openreview.net/forum?id=EcetCr4trp
Compressor summary: This paper proposes a feature learning theory to understand convergence and generalization properties in Federated Learning, showing how communication helps achieve low test error and addressing data heterogeneity issues with a weighted method.
https://openreview.net/forum?id=Ebt7JgMHv1
Compressor summary: This paper warns about interpretability illusions from naïve subspace interventions in neural models, showing examples and suggesting more careful analysis for valid interpretations.
https://openreview.net/forum?id=EanCFCwAjM
Compressor summary: The paper introduces a novel camera pose estimation method that uses rays, image features, and set-level transformers, and learns from uncertainties with a denoising diffusion model.
https://openreview.net/forum?id=EXitynZhYn
Compressor summary: The text proposes a novel VQA benchmark using visual classification datasets, follow-up questions, and LLM-based metrics to evaluate text-generative vision-language models more accurately.
https://openreview.net/forum?id=EWTFMkTdkT
Compressor summary: The paper introduces a new model class that efficiently learns complicated dynamics from high-dimensional data by combining variational autoencoders, spatio-temporal attention, and scientifically-motivated invariances.
https://openreview.net/forum?id=EW8ZExRZkJ
Compressor summary: The paper analyzes dilated CNNs' performance in estimating maps between infinite-dimensional input and output spaces and shows they achieve minimax optimal error rates, outperforming linear estimators.
https://openreview.net/forum?id=EPNEazJoAg
Compressor summary: The feature interaction score (FIS) measures how much feature interactions vary among equally accurate predictive models, helping to understand their behavior.
https://openreview.net/forum?id=EJPIzl7mgc
Compressor summary: The paper proposes an adversarial training method to improve layout-to-image generation, enabling better alignment and editability.
https://openreview.net/forum?id=EIPLdFy3vp
Compressor summary: The text discusses contrastive learning for time series data, proposing AutoTCL, an adaptive augmentation method that improves representation learning and outperforms existing approaches on forecasting and classification tasks.
https://openreview.net/forum?id=EHrvRNs2Y0
Compressor summary: ResFields are neural fields with temporal residual layers that can effectively model complex temporal signals and improve performance on various tasks involving 2D video approximation, dynamic shape modeling, and dynamic NeRF reconstruction.
https://openreview.net/forum?id=EHg5GDnyq1
Compressor summary: AgentVerse is a framework that helps AI agents cooperate and perform better on various tasks than individually.
https://openreview.net/forum?id=EH2O3h7sBI
Compressor summary: The paper proposes a new approach (PGP) that combines prompt-tuning and gradient projection for continual learning, which reduces forgetting by selecting relevant prompts and using self-attention and SVD in vision-transformers.
https://openreview.net/forum?id=EGQBpkIEuu
Compressor summary: The text analyzes existing data augmentation techniques for image-based deep reinforcement learning, compares their effects, and proposes a novel regularization term called tangent prop that improves performance and sample efficiency.
https://openreview.net/forum?id=EG68RSznLT
Compressor summary: FTB is a method that uses generative models and preference augmentation to improve trajectories in offline PbRL without learning inaccurate rewards from preferences.
https://openreview.net/forum?id=EDXkkUAIFW
Compressor summary: The paper proposes a one-shot active learning method for training multiple models using different representations without repeated model training, which samples and reweights instances based on their maximum Lewis weights.
https://openreview.net/forum?id=EDPxCjXzSb
Compressor summary: CIReVL is a training-free method for Compositional Image Retrieval using large vision and language models to caption images and apply textual modifications, achieving competitive performance and human-understandability.
https://openreview.net/forum?id=EArTDUmILF
Compressor summary: The paper proposes a new method using multi-modal physiological signals and a Variational Bayesian Heterogeneous Graph Neural Network with Relationship Distribution Adaptation to improve cross-subject emotion recognition accuracy without aligning signal features.
https://openreview.net/forum?id=E78OaH2s3f
Compressor summary: The paper introduces a universal condition alignment score for diverse conditional generation tasks that works across all conditions and requires no additional models.
https://openreview.net/forum?id=E60SIDItyT
Compressor summary: The paper compares two loss functions for learning from aggregate data, introduces a new interpolating estimator, and analyzes the effect of bag size on risk-privacy trade-offs in linear regression.
https://openreview.net/forum?id=E34AlVLN0v
Compressor summary: The paper presents a parallel algorithm that greatly speeds up GPU evaluation and training of sequential models without sacrificing accuracy, enabling their use in long sequence problems.
https://openreview.net/forum?id=E1NxN5QMOE
Compressor summary: Aranyani is an ensemble of oblique decision trees for making fair decisions in online settings by using aggregate statistics of previous decisions, reducing storage and computation requirements.
https://openreview.net/forum?id=DuQkqSe9en
Compressor summary: AILBoost is a novel and principled off-policy adversarial imitation learning algorithm that uses an ensemble of weighted weak learners and a discriminator to improve sample efficiency and scalability in various environments.
https://openreview.net/forum?id=DsEhqQtfAG
Compressor summary: The paper proposes a diffusion sampling strategy that combines Krylov subspace methods with modern diffusion models for efficient and high-quality solutions to inverse problems, such as MRI and CT reconstruction.
https://openreview.net/forum?id=DrhZneqz4n
Compressor summary: SinMDM is a single motion diffusion model that learns from a single input motion and synthesizes various motions of different lengths while maintaining the original motion's essence, applicable to multiple contexts like spatial and temporal in-betweening, motion expansion, style transfer, and crowd animation.
https://openreview.net/forum?id=DqziS8DG4M
Compressor summary: Point2SSM is a novel unsupervised learning approach to construct statistical shape models of anatomy from noisy point clouds using attention-based correspondence mappings.
https://openreview.net/forum?id=DqD59dQP37
Compressor summary: This paper analyzes how unobserved confounding affects causal fairness in machine learning predictions and proposes a neural framework to ensure fairness even under observed confounding.
https://openreview.net/forum?id=DpFeMH4l8Q
Compressor summary: GPO is a framework to align large language models with the preferences of different groups using few-shot learning and meta-learning.
https://openreview.net/forum?id=Dnc3paMqDE
Compressor summary: The paper introduces Spherical Patch Fields, a technique for SO(3)-equivariant 3D point clouds, and presents DeepSPF, which improves Scan-to-CAD performance in registration, retrieval, and completion tasks.
https://openreview.net/forum?id=DmD1wboID9
Compressor summary: BayesPrompt is a novel approach that learns prompts with domain discriminative information to improve few-shot learning using pre-trained language models.
https://openreview.net/forum?id=DjzvJCRsVf
Compressor summary: CLIPSelf is a method that improves CLIP ViTs for local image region tasks like object detection and segmentation by aligning their representations without needing extra data.
https://openreview.net/forum?id=Djw0XhjHZb
Compressor summary: The paper proposes a new geometric deep learning method using differential forms to represent simplices and embed them in $\mathbb{R}^n$, achieving universal approximation, interpretability, and efficiency without message passing.
https://openreview.net/forum?id=DjeQ39QoLQ
Compressor summary: The paper proposes a perturb-then-diagonalize method for initializing state-space models, improving their robustness and performance compared to previous diagonal-based approaches.
https://openreview.net/forum?id=DjIsNDEOYX
Compressor summary: The research proposes SMNN, a partially connected network with a scalable monotonic hidden layer, which preserves monotonicity and is easy to implement and train.
https://openreview.net/forum?id=Diq6urt3lS
Compressor summary: Cleanba is a new platform for distributed reinforcement learning that improves reproducibility and efficiency over existing methods.
https://openreview.net/forum?id=DiWRG9JTWZ
Compressor summary: The text discusses OOD problems in few-shot classification, focusing on two types (CD-FSC and SC-FSC), introduces a new benchmark (MetaCoCo) for SC-FSC, and evaluates existing methods on it.
https://openreview.net/forum?id=DfPtC8uSot
Compressor summary: The paper defines and analyzes expected robustness in GNNs, proposes a new robust variant called GCORNs, and evaluates it on real-world datasets.
https://openreview.net/forum?id=DesYwmUG00
Compressor summary: The text presents a new probabilistic method for image editing using stochastic differential equations (SDEs) that improves upon existing ordinary differential equation (ODE) methods and shows promising results in various tasks.
https://openreview.net/forum?id=Dc4rXq3HIA
Compressor summary: The paper proposes DG, a new approach to handle domain shifts in machine learning, which learns domain-specific models and reweights them based on domain metadata for better out-of-domain generalization.
https://openreview.net/forum?id=DayPQKXaQk
Compressor summary: The authors propose a constrained decoding method for label projection in zero-shot cross-lingual transfer, which improves translation quality and can be applied to both training and test data strategies.
https://openreview.net/forum?id=DZqic2sPTY
Compressor summary: GraphPulse is a framework that combines Mapper and RNN techniques to analyze temporal graphs and improve prediction accuracy.
https://openreview.net/forum?id=DZUzOKE6og
Compressor summary: The paper introduces hyperedge filling, a novel generative self-supervised learning task for hypergraphs, and HYPEBOY, a hypergraph neural network method that learns effective hypergraph representations using this task.
https://openreview.net/forum?id=DYIIRgwg2i
Compressor summary: The paper presents a data-driven compression method for large Transformer models that can reduce their size by 20%-30% with minimal performance loss and achieve state-of-the-art results in different types of pruning.
https://openreview.net/forum?id=DRu8PMHgCh
Compressor summary: FedTrans is a method to improve federated learning by selecting clients based on their noise level and utility, using Bayesian inference with auxiliary data.
https://openreview.net/forum?id=DO2WFXU1Be
Compressor summary: PINNsFormer is a new framework that uses Transformers to improve the accuracy of numerical solutions to PDEs by capturing temporal dependencies and incorporating a wavelet activation function.
https://openreview.net/forum?id=DLJznSp6X3
Compressor summary: ReLoRA is a new method that trains large neural networks more efficiently, saving memory and speeding up training.
https://openreview.net/forum?id=DJZDgMOLXQ
Compressor summary: PEC is a novel approach for class-incremental learning that uses prediction error to estimate class scores, reducing forgetting and improving performance compared to existing methods.
https://openreview.net/forum?id=DGez4B2a6Y
Compressor summary: PIRATE is a novel deep learning method for biomedical image registration that uses a pre-trained CNN denoiser and an explicit data-fidelity penalty to improve registration accuracy.
https://openreview.net/forum?id=DFTHW0MyiW
Compressor summary: The authors propose a method to improve reinforcement learning policies' robustness against state-adversarial attacks by refining the policy class before test time and using an adversarial bandit subroutine.
https://openreview.net/forum?id=DEJIDCmWOz
Compressor summary: The study evaluates the reliability of watermarking as a way to identify machine-generated text in different scenarios, finding that watermarks remain detectable even after human and machine paraphrasing.
https://openreview.net/forum?id=DDX1u29Gqr
Compressor summary: DreamCraft3D is a method that creates high-quality 3D objects by using a 2D reference image and a hierarchical process involving geometry sculpting and texture boosting guided by view-dependent diffusion models.
https://openreview.net/forum?id=DCDT918ZkI
Compressor summary: The paper proposes a new adversarial training paradigm for Graph Neural Networks that uses out-of-distribution detection to enhance robustness without relying on prior knowledge, but shows a trade-off between graph attack efficacy and defensibility.
https://openreview.net/forum?id=DASh78rJ7g
Compressor summary: The paper proposes new plugin estimators for selective classification with out-of-distribution detection, which are theoretically grounded and improve over existing heuristic methods.
https://openreview.net/forum?id=D9rJdtmIG6
Compressor summary: SpaCE is a toolkit that helps scientists evaluate causal inference methods for dealing with spatial confounding in studies involving spatial data.
https://openreview.net/forum?id=D7KJmfEDQP
Compressor summary: The paper proposes a new uncertainty-based method to improve weighted-averaging of merged models by addressing gradient mismatches and outperforms other schemes on language and vision tasks.
https://openreview.net/forum?id=D5mJSNtUtv
Compressor summary: The paper proposes a new system for improving learned lossless data compression by using an efficient entropy coder and a discrete latent space optimization scheme, achieving better compression ratios with minimal extra computational time.
https://openreview.net/forum?id=D4NJFfrqoq
Compressor summary: The paper proposes a new algorithm for constrained Bayesian optimization that balances exploration and exploitation in the decoupled setting, with a theoretical performance guarantee and connections to active learning.
https://openreview.net/forum?id=D2eOVqPX9g
Compressor summary: FedSARSA is a novel on-policy FRL scheme that converges to near-optimal policies and achieves linear speedup with agent collaboration.
https://openreview.net/forum?id=Cy5v64DqEF
Compressor summary: Idempotence in image codecs is related to perceptual compression, and a new method based on unconditional generative models with idempotence constraints improves image quality without training new models.
https://openreview.net/forum?id=CvYBvgEUK9
Compressor summary: The paper studies first-order algorithms for solving Bilevel Optimization problems with smooth but possibly nonconvex objectives, using penalty methods to connect the objective functions and hyper-objective, and proposes efficient algorithms with provable convergence guarantees.
https://openreview.net/forum?id=Cu5wJa5LGO
Compressor summary: LLCP is a causal framework for video reasoning that uses self-supervised local auto-regression to answer accident attribution and counterfactual prediction questions without needing annotated data.
https://openreview.net/forum?id=CtOA9aN8fr
Compressor summary: The paper proposes a method to prune large-scale multimodal datasets for training CLIP-style models, improving data and training efficiency while achieving better results on ImageNet zero-shot accuracy and DataComp Medium benchmark.
https://openreview.net/forum?id=CpnKq3UJwp
Compressor summary: MAZero is a model-based multi-agent reinforcement learning algorithm with improved efficiency, which combines a centralized model with Monte Carlo Tree Search and uses novel techniques for action spaces and policy optimization.
https://openreview.net/forum?id=ChHx5ORqF0
Compressor summary: The paper introduces a method called Label-Guided Pseudo-Labeling (LGPL) to translate bounding boxes and class labels across different object detection datasets, improving downstream detection performance.
https://openreview.net/forum?id=Ch7WqGcGmb
Compressor summary: This paper explores a modern error feedback method called EF21, which has better theoretical guarantees and communication complexity than previous methods, and shows its effectiveness in practice with experiments.
https://openreview.net/forum?id=CgPs04l9TO
Compressor summary: This paper investigates oscillations in behavior cloning with neural networks caused by minibatch SGD updates, which lead to large errors over long horizons, and shows that exponential moving average (EMA) can effectively reduce this issue.
https://openreview.net/forum?id=CfXh93NDgH
Compressor summary: The paper proposes a method to create complex instruction data for training large language models using an evolved set of instructions, which improves their performance compared to baselines.
https://openreview.net/forum?id=Cf4FJGmHRQ
Compressor summary: The text proposes a new neural network model, PAC-FNO, that can handle varying image resolutions and natural variations using the frequency domain approach and improve image recognition performance.
https://openreview.net/forum?id=CdjnzWsQax
Compressor summary: The authors propose FTS-Diffusion, a framework that generates realistic financial time series by extracting and synthesizing recurring patterns and modeling their temporal transitions, improving prediction accuracy.
https://openreview.net/forum?id=Cc0qk6r4Nd
Compressor summary: The paper proposes InCo Aggregation, a method that uses internal cross-layer gradients to improve model performance in federated learning with system heterogeneity.
https://openreview.net/forum?id=CanomFZssu
Compressor summary: GADAM is a novel framework for unsupervised graph anomaly detection that combines efficient MLP-based local inconsistency mining with hybrid attention based adaptive message passing to capture global anomaly signals, achieving superior performance and efficiency on nine benchmark datasets.
https://openreview.net/forum?id=CZiY6OLktd
Compressor summary: The MG-TSD model improves time series forecasting by using multiple granularity levels within the data to guide the learning process of diffusion models, without relying on external data.
https://openreview.net/forum?id=CYmF38ysDa
Compressor summary: FLASK is a fine-grained evaluation protocol for assessing large language models based on skill sets, which improves interpretability and reliability of model performance.
https://openreview.net/forum?id=CX2RgsS29V
Compressor summary: The paper proposes a new method for efficiently updating truncated SVD of large, sparse, and evolving data matrices by exploiting sparsity and using an extended decomposition.
https://openreview.net/forum?id=CUfSCwcgqm
Compressor summary: The paper proposes Neural Atoms, a method that abstracts atomic groups in molecules and improves the ability of Graph Neural Networks (GNNs) to capture long-range interactions for drug discovery.
https://openreview.net/forum?id=CTlUHIKF71
Compressor summary: RAPL is a new method that helps robots learn visual rewards that match human preferences by using feedback and optimal transport, resulting in better and more efficient robot behaviors.
https://openreview.net/forum?id=COYDmKkQH4
Compressor summary: AutoCast++ is a system that uses news articles to predict real-world events and improves forecasting accuracy by ranking and summarizing relevant passages.
https://openreview.net/forum?id=COO51g41Q4
Compressor summary: The paper proposes STAR, a pruning method for Vision Transformers that combines intra-layer and inter-layer patch importance scoring to reduce computational costs without much performance degradation.
https://openreview.net/forum?id=CNL2bku4ra
Compressor summary: The authors propose a method to fine-tune language models on query neighbors retrieved from a text embedding index at test time, improving performance across various language modeling tasks.
https://openreview.net/forum?id=CMzF2aOfqp
Compressor summary: Label Wave is a novel early stopping method that works without validation data by tracking model predictions on training data and halting training before mislabeled data is excessively fitted.
https://openreview.net/forum?id=CK5Hfb5hBG
Compressor summary: The paper proposes ChannelViT, a modified ViT architecture for multi-channel images, which improves reasoning across input channels and uses Hierarchical Channel Sampling as a regularization technique for robustness.
https://openreview.net/forum?id=CIqjp9yTDq
Compressor summary: This paper provides a rigorous theoretical analysis of heavy-ball momentum with decaying learning rates for quadratic regression problems under anisotropic gradient noise, showing accelerated convergence and usefulness in large-batch settings.
https://openreview.net/forum?id=CIj1CVbkpr
Compressor summary: The paper proposes Online Spiking Renormalization (OSR) and Online Threshold Stabilizer (OTS) to improve online training of spiking neural networks, achieving better performance with less memory consumption.
https://openreview.net/forum?id=CHGcP6lVWd
Compressor summary: The paper introduces Meta-Distribution Energy (MDE), a novel measure for automated model evaluation that improves efficiency and effectiveness by using energy-based learning and providing better predictions without ground-truth labels.
https://openreview.net/forum?id=CGlczSBBSj
Compressor summary: SEAL is a framework for systematically evaluating super-resolution methods on diverse real-world images using clustered degradation cases and coarse-to-fine metrics.
https://openreview.net/forum?id=CF8H8MS5P8
Compressor summary: The text discusses a paradox in generative AI, where models can produce outputs beyond human expertise but still make basic errors, suggesting that their generation capability is not based on understanding like humans.
https://openreview.net/forum?id=CEkIyshNbC
Compressor summary: Double descent occurs when imperfect deep learning models learn noisy data and then use over-parameterization to separate signal from noise, but not in well-regularized models.
https://openreview.net/forum?id=CAqdG2dy5s
Compressor summary: The paper proposes a novel graph deep learning framework called Graph-graph Network (GgNet) that can infer signals at unmonitored locations by exploiting correlations between variables and locations using spatio-temporal measurements from physical sensors.
https://openreview.net/forum?id=C61sk5LsK6
Compressor summary: InfoBatch is a framework that prunes less informative samples from the data and rescales gradients to achieve lossless training acceleration without gradient bias.
https://openreview.net/forum?id=C4CxQmp9wc
Compressor summary: Jumanji is a fast, flexible, and scalable open-source RL environment suite designed for combinatorial problems and general decision-making tasks, with customizable initial states and problem complexities.
https://openreview.net/forum?id=C4BikKsgmK
Compressor summary: Str2Str is a new method to sample protein conformations using structure-to-structure translation, which is fast and does not need simulation data for training or inference.
https://openreview.net/forum?id=C36v8541Ns
Compressor summary: The paper proposes a method to improve the robustness of deep neural networks using randomized smoothing and shows significant improvement in certified accuracy compared to existing methods.
https://openreview.net/forum?id=C1sQBG6Sqp
Compressor summary: Dual Randomized Smoothing (DRS) improves the robustness of image classifiers by down-sampling and smoothing images in lower dimensions, overcoming the curse of dimensionality issue in Randomized Smoothing.
https://openreview.net/forum?id=ByR3NdDSZB
Compressor summary: The text proposes a new framework, PARL, that addresses policy alignment issues in reinforcement learning using preference-based feedback, leading to significant improvements in sample complexity.
https://openreview.net/forum?id=BxHgpC6FNv
Compressor summary: The paper shows that two-layer ReLU networks trained by gradient descent on noisy XOR data can overfit to training labels and then suddenly generalize well, demonstrating benign overfitting and the grokking phenomenon.
https://openreview.net/forum?id=BuFNoKBiMs
Compressor summary: The Decoupled MTPP framework uses Neural ODEs to learn continuous dynamics of event influences and disentangle inter-event dependencies in complex temporal data.
https://openreview.net/forum?id=BteuUysuXX
Compressor summary: The paper proposes verbose images that manipulate VLMs to generate long and inefficient sentences by using three losses and a temporal weight algorithm.
https://openreview.net/forum?id=BtZ7vCt5QY
Compressor summary: The paper introduces a new stochastic deep learning method for causal inference with complex high-dimensional data that uses sparse deep learning theory and stochastic neural networks to overcome challenges of high dimensionality, unknown data generation process, and missing values.
https://openreview.net/forum?id=BtT6o5tfHu
Compressor summary: The paper proposes a new sampling method for diffusion-based image super-resolution models that leverages and optimizes the randomness of the reverse process, improving the quality of results with fewer steps.
https://openreview.net/forum?id=BrjLHbqiYs
Compressor summary: The paper proposes methods to quantify multimodal interactions in semi-supervised learning using information theory and shows their usefulness for task performance and data selection.
https://openreview.net/forum?id=BqHaLnans2
Compressor summary: This paper proposes a method to train an LLM for vision-language capabilities on medical images by instructing it to answer questions about image inputs and generate both text and image responses.
https://openreview.net/forum?id=BqEvdOS1Hs
Compressor summary: The paper proposes a "human-centered" approach for improving game AI agents' collaboration with humans by enhancing their goal achievement, and evaluates it in a MOBA game.
https://openreview.net/forum?id=Bpkhu2ExxU
Compressor summary: The paper explores dropout's mechanism in neural networks training by deriving its stochastic equations theoretically and analyzing its impact on loss landscape through empirical studies.
https://openreview.net/forum?id=Bpcgcr8E8Z
Compressor summary: The paper introduces Fast-DetectGPT, an efficient zero-shot detector that distinguishes between human and machine-generated text using conditional probability curvature, which improves on DetectGPT's performance and speed.
https://openreview.net/forum?id=Bo6GpQ3B9a
Compressor summary: The paper proposes a new method to use unlabeled data in semi-supervised classification, combining robust optimization with self-supervision, and shows improved generalization error bounds in simulations and real data.
https://openreview.net/forum?id=BnQY9XiRAS
Compressor summary: This paper presents a new method to represent crystals as graphs using their periodic patterns and introduces ComFormer, a transformer model for crystalline materials with two variants that outperform existing methods.
https://openreview.net/forum?id=BllUWdpIOA
Compressor summary: The paper proposes a continual momentum filtering (CMF) framework that uses the Kalman filter to improve online test-time adaptation of deep neural networks by balancing adaptability and information retention, achieving better results than existing methods in various scenarios.
https://openreview.net/forum?id=BlkxbI6vzl
Compressor summary: The paper proposes a new second-order algorithm for sparse phase retrieval that converges faster than existing first-order methods, given enough measurements.
https://openreview.net/forum?id=Bl8u7ZRlbM
Compressor summary: WildChat is a large and diverse dataset of user-ChatGPT interactions that reveals how people use the chatbot in practice, including potentially unsafe scenarios, and can be used to improve instruction following models.
https://openreview.net/forum?id=BifeBRhikU
Compressor summary: This paper introduces a novel approach, Partially-Binarized LLM (PB-LLM), that compresses Large Language Models to single bit using salient weights and improves their linguistic reasoning capacity through post-training quantization and quantization-aware training.
https://openreview.net/forum?id=BdPvGRvoBC
Compressor summary: This paper analyzes the effects of two gradient clipping methods on the convergence of FedAvg, a federated learning algorithm, and provides convergence guarantees under different assumptions.
https://openreview.net/forum?id=Bb4VGOWELI
Compressor summary: OPRO is a method that uses large language models to generate solutions for optimization problems described in natural language, achieving better performance than human-designed prompts.
https://openreview.net/forum?id=Bb21JPnhhr
Compressor summary: The text proposes AntGPT, which uses large language models to predict an actor's future behavior from video observations by inferring goals and modeling temporal dynamics.
https://openreview.net/forum?id=BZtEthuXRF
Compressor summary: MDF is a method for learning diffusion models on non-euclidean data using an intrinsic coordinate system, which improves diversity and accuracy on various datasets and problems.
https://openreview.net/forum?id=BXY6fe7q31
Compressor summary: The paper introduces a new module, VPG-C, to improve multimodal language models' understanding of demonstrative instructions by completing missing visual details using synthetic discriminative training.
https://openreview.net/forum?id=BWAhEjXjeG
Compressor summary: The paper proposes RSCP+, a framework to improve conformal prediction methods' robustness against adversarial examples, and introduces two new methods, PTT and RCT, that reduce uncertainty sets while maintaining efficiency.
https://openreview.net/forum?id=BV1PHbTJzd
Compressor summary: The paper proposes a new distributed stochastic optimization algorithm using a non-linear Markov chain, the Self-Repellent Random Walk (SRRW), which improves convergence and reduces variance compared to existing methods.
https://openreview.net/forum?id=BTKAeLqLMw
Compressor summary: The authors study how to automatically select data for instruction tuning of language models, proposing a simple strategy and a new model (Deita) that performs well with minimal data.
https://openreview.net/forum?id=BRdEBlwUW6
Compressor summary: The DAFA method addresses the robust fairness problem by adjusting adversarial margins and loss weights for different class similarities, improving the worst class's performance.
https://openreview.net/forum?id=BPb5AhT2Vf
Compressor summary: The paper proposes a method to register images with point clouds by matching diffusion features from pretrained models and geometric features from monocular depth estimators, achieving significant improvements over existing methods on indoor and outdoor benchmarks.
https://openreview.net/forum?id=BPHcEpGvF8
Compressor summary: This paper analyzes backdoor attacks in machine learning, which are stealthy ways to manipulate model behavior with hidden triggers, and proposes a theoretical framework to measure their effectiveness on clean and compromised data.
https://openreview.net/forum?id=BOfDKxfwt0
Compressor summary: RealChat-1M is a large-scale dataset of real-world conversations with language models for various applications, including content moderation, safety benchmarking, instruction-following, and challenging questions.
https://openreview.net/forum?id=BLGQ3oqldb
Compressor summary: The paper introduces LogicMP, a new neural layer that enables efficient and modular integration of first-order logic constraints with neural networks using mean-field variational inference over an MLN.
https://openreview.net/forum?id=BKXvPDekud
Compressor summary: The authors propose a new pre-trained model, CellPLM, for single-cell data that better captures cell-cell relationships and accounts for noisy data compared to existing models inspired by natural language processing.
https://openreview.net/forum?id=BIveOmD1Nh
Compressor summary: Machine learning can speed up molecular docking by learning a fast scoring function from cross-correlation of ligand and protein fields using graph neural networks and Fourier transforms.
https://openreview.net/forum?id=BI1N3lTWtn
Compressor summary: The paper proposes a multi-level framework for efficient training of large-scale deep learning models, using operators like coalescing, de-coalescing and interpolation, which reduces computational cost while maintaining performance.
https://openreview.net/forum?id=BGkqypmGvm
Compressor summary: Our approach enhances computer vision models by adding a multidimensional State Space Model layer that improves 2-D inductive bias and performance, especially for Vision Transformers.
https://openreview.net/forum?id=BEyEziZ4R6
Compressor summary: The paper proposes using Lipschitz constrained networks to train DP-DNNs with tighter sensitivity bounds and without gradient clipping, and provides a Python package for implementing them.
https://openreview.net/forum?id=BEH4mGo7zP
Compressor summary: The paper proposes a pre-training strategy called ProteinINR that incorporates information from protein sequences, 3D structures, and surfaces to improve protein representation learning and performance in downstream tasks.
https://openreview.net/forum?id=BE5aK0ETbp
Compressor summary: The paper introduces a unified framework for continual learning methods that reconciles different approaches and reveals their shared mathematical structures, as well as a new concept called refresh learning that enhances performance by unlearning and relearning data.
https://openreview.net/forum?id=BBD6KXIGJL
Compressor summary: The Hybrid Directional Graph Neural Network (HDGNN) combines equivariant operations with learnable modules to predict chemical properties of molecules, achieving state-of-the-art performance on several tasks.
https://openreview.net/forum?id=B9klVS7Ddk
Compressor summary: LLM-KICK is a benchmark to evaluate compressed large language models based on various tasks, revealing strengths and weaknesses of existing compression methods.
https://openreview.net/forum?id=B6pQxqUcT8
Compressor summary: ToolChain* is a tree search-based algorithm that helps LLMs efficiently navigate large action spaces to solve complex problems using APIs.
https://openreview.net/forum?id=B4vzu2aokv
Compressor summary: This paper proposes a Mutual Distillation Module for point-level supervised instance segmentation that leverages both instance position and semantic information to enhance object perception and achieve better results than existing methods.
https://openreview.net/forum?id=AyzkDpuqcl
Compressor summary: CDRL is an approach that improves energy-based model training by learning from noisy versions of a dataset and refining samples with an initializer model, resulting in better sample quality and faster sampling time.
https://openreview.net/forum?id=AyXIDfvYg8
Compressor summary: The paper shows that adding a ReLU activation to the final layer of a model using supervised contrastive loss can restore symmetry in learned representations under class imbalances, without affecting test accuracy.
https://openreview.net/forum?id=Ax2yRhCQr1
Compressor summary: This paper analyzes the role of augmentation in self-supervised representation learning, providing generalization bounds and a quantitative measure of augmentation complexity.
https://openreview.net/forum?id=AwyxtyMwaG
Compressor summary: The study reveals how transformer models use function vectors (FV) to represent and execute tasks across different contexts.
https://openreview.net/forum?id=AssIuHnmHX
Compressor summary: The paper proposes a framework to understand when and how Transformers can length generalize, using RASP programming language and showing a correlation between RASP-simplicity and generalization.
https://openreview.net/forum?id=ArpwmicoYW
Compressor summary: FairTune is a framework to optimize fairness in AI models by adjusting fine-tuning parameters based on validation data.
https://openreview.net/forum?id=AqN23oqraW
Compressor summary: KoLA is a benchmark that evaluates large language models based on their knowledge abilities using carefully designed tasks, data, and metrics.
https://openreview.net/forum?id=Aq35gl2c1k
Compressor summary: Critical learning periods are observed in both biological and artificial networks, and depend on network depth and data structure, as well as feature competition and task relationship in deep linear models.
https://openreview.net/forum?id=AhizIPytk4
Compressor summary: The text introduces ImageNetCT-9K, a large 3D CT dataset for pre-training models, which improves transfer learning for 3D image segmentation tasks.
https://openreview.net/forum?id=Agyicd577r
Compressor summary: The paper proposes BPE + SEAS to improve BatchPrompt, a prompting strategy that batches data points into one input for LLMs. This increases token utilization and efficiency, while enhancing quality through permutation, voting, and early stopping. The approach outperforms or matches SinglePrompt on several NLP tasks with less LLM resources.
https://openreview.net/forum?id=AgM3MzT99c
Compressor summary: The text proposes using large language models to prioritize tasks for open-ended algorithms based on human notions of interestingness, improving the selection of learnable and novel tasks.
https://openreview.net/forum?id=AgDICX1h50
Compressor summary: Analogical Prompting is a new approach that guides language models to generate relevant knowledge from past experiences and apply it to solve reasoning tasks, improving their performance.
https://openreview.net/forum?id=AfnsTnYphT
Compressor summary: The Dynamic Signal Distribution model shows that CNNs have statistical advantages over LCNs and FCNs in capturing locality and translation invariance in real-world images.
https://openreview.net/forum?id=AfhNyr73Ma
Compressor summary: The paper presents a general proof framework for stability analysis of zeroth-order optimization algorithms, which improves generalization bounds for SGD, GD, and SVRG using coordinate estimation.
https://openreview.net/forum?id=Ad87VjRqUw
Compressor summary: G-Shell is a parametrization method that enables fast reconstruction and generative modeling of both watertight and non-watertight 3D meshes from multiview images.
https://openreview.net/forum?id=Ad81awoBVS
Compressor summary: The study finds a strong linear relationship between rotation prediction accuracy and one-class classification (OCC) performance, suggesting that estimating the transformation distribution in the dataset is crucial for self-supervised representation learning.
https://openreview.net/forum?id=AcoXPIPh4A
Compressor summary: ASGD can perform better generalization than SGD in overparameterized linear regression by adapting to different eigenvalue subspaces of the data covariance matrix.
https://openreview.net/forum?id=AcSChDWL6V
Compressor summary: Graph Transformers and Message Passing Graph Neural Networks with virtual nodes have different universal expressivity, but both can effectively learn from global information on graphs.
https://openreview.net/forum?id=AcRfzLS6se
Compressor summary: BLOOD is a method for detecting out-of-distribution (OOD) data in deep neural networks based on the smoothness of transformation between intermediate layers, which works well on pre-trained models without training data access and outperforms other methods.
https://openreview.net/forum?id=AcJrSoArlh
Compressor summary: The paper proposes a new method to improve the transferability of adversarial examples by targeting the common weaknesses of model ensemble, which are flatness and closeness to local optimum. The method works well on various tasks and models.
https://openreview.net/forum?id=Abr7dU98ME
Compressor summary: ForwardGNN is a new forward learning procedure for graph neural networks that avoids the constraints of backpropagation by performing layer-wise local forward training on graph data.
https://openreview.net/forum?id=AbXGwqb5Ht
Compressor summary: The article establishes a regularization method for deep residual networks trained with gradient flow, showing that they converge to neural ODEs when initialized as discretizations of them.
https://openreview.net/forum?id=AZW3qlCGTe
Compressor summary: The paper proposes a new method to improve image classification by using set-level labels that capture relationships among instances and provides theoretical analysis and experimental results showing its effectiveness.
https://openreview.net/forum?id=AZGIwqCyYY
Compressor summary: The authors propose a meta learning method to train a graph neural network for Hamiltonian dynamics, enabling it to adapt to new physical systems and capture a generalized representation across domains.
https://openreview.net/forum?id=AY9KyTGcnk
Compressor summary: The paper presents StABL, an optimal online learning algorithm that adapts quickly to changing environments with minimal queries per round.
https://openreview.net/forum?id=AY6aM13gGF
Compressor summary: The paper presents LaMo, a framework that uses Decision Transformers and pre-trained Language Models to achieve state-of-the-art performance in offline reinforcement learning tasks, especially with limited data samples.
https://openreview.net/forum?id=AXbN2qMNiW
Compressor summary: The paper proposes a self-supervised learning method for protein-ligand binding representation using Transformer interactions and two pre-training tasks, achieving better generalization across various binding tasks.
https://openreview.net/forum?id=AXC9KydyZq
Compressor summary: M3C is a learning-free algorithm for graph matching and clustering that guarantees convergence, offers flexibility, and outperforms existing methods.
https://openreview.net/forum?id=AU2gS9ut61
Compressor summary: BrainPy is a differentiable brain simulator that bridges the gap between brain simulation and brain-inspired computing by providing efficient and scalable brain modeling capabilities.
https://openreview.net/forum?id=ATEawsFUj4
Compressor summary: GAIA is a data-driven approach that generates diverse, natural, and high-quality talking avatars using speech-to-motion and appearance disentanglement.
https://openreview.net/forum?id=ARPrtuzAnQ
Compressor summary: The paper investigates if incorporating symmetries into neural networks can reduce the difficulty of learning them with gradient descent, and finds that it does not for several types of networks.
https://openreview.net/forum?id=AOJyfhWYHf
Compressor summary: OpenChat is a novel framework that improves open-source language models with mixed-quality data using a class-conditioned policy in reinforcement learning, achieving the highest performance among 13b models.
https://openreview.net/forum?id=ANvmVS2Yr0
Compressor summary: DNNs trained for denoising can learn high-dimensional densities and perform shrinkage operations in adapted bases, despite memorization concerns.
https://openreview.net/forum?id=ALVwQjZRS8
Compressor summary: The authors propose Coeditor, a fine-tuned language model that predicts multiple edits to a code region based on recent changes within the same codebase, outperforming existing models in single and multi-round code completion tasks.
https://openreview.net/forum?id=ALGFFPXWSi
Compressor summary: The unified likelihood ratio (ULR) method is a flexible and efficient gradient estimation technique for neural network training that reduces computation, improves device adaptation, and enhances robustness.
https://openreview.net/forum?id=AJBkfwXh3u
Compressor summary: The paper proposes a causality-inspired generative model to improve the interpretability of dynamic graph neural networks by identifying static, dynamic, and causal relationships in complex spatial-temporal correlations.
https://openreview.net/forum?id=AJBGSVSTT2
Compressor summary: The paper proposes a new backdoor attack method for federated learning that targets crucial layers of the model, improving stealthiness and effectiveness under various defense strategies.
https://openreview.net/forum?id=AHgc5SMdtd
Compressor summary: The paper proposes a novel zero-shot anomaly classification and segmentation method for industrial vision that uses mutual scoring of unlabeled test images and achieves superior performance on MVTec AD and VisA datasets.
https://openreview.net/forum?id=AF9Q8Vip84
Compressor summary: The paper introduces SpeechTokenizer, a unified speech tokenizer for speech large language models that uses Encoder-Decoder architecture with residual vector quantization (RVQ) and proposes Unified Speech Language Model (USLM), which performs well on benchmarks and text-to-speech tasks.
https://openreview.net/forum?id=ADSxCpCu9s
Compressor summary: The text presents a benchmark system that evaluates how well language models can plan tasks for home-service robots using different datasets, simulators, and prompts.
https://openreview.net/forum?id=ADDCErFzev
Compressor summary: The text discusses how varying dropout proportion in object recognition models can reveal an optimal balance between high-dimensional efficient codes and low-dimensional robust codes in hierarchical vision systems, as observed in human occipitotemporal cortex.
https://openreview.net/forum?id=A7t7z6g6tM
Compressor summary: The paper proposes a novel framework called HENN that models predictive uncertainty caused by composite set labels in DNNs using Subjective Logic, and shows improved performance over existing methods.
https://openreview.net/forum?id=A4mJuFRMN8
Compressor summary: The paper proposes RENT, a resampling method for learning with noisy labels using the transition matrix, and shows it performs better than existing methods.
https://openreview.net/forum?id=A2mRcRyGdl
Compressor summary: Semantic Flow learns 3D motion information from monocular videos using neural networks and semantic labels to enable various tasks like editing, completion, tracking, and adaptation.
https://openreview.net/forum?id=A18gWgc5mi
Compressor summary: The paper explores autoencoders for modeling nonlinear dynamical systems and proposes a new method called Periodic Reencoding to improve future state prediction.
https://openreview.net/forum?id=A0HKeKl4Nl
Compressor summary: Fine-tuning changes machine learning models' capabilities in subtle ways, which could affect their safety and performance.
https://openreview.net/forum?id=9zhHVyLY4K
Compressor summary: The authors propose an unsupervised method to align pre-trained generative models with new neural datasets for neuroscience applications.
https://openreview.net/forum?id=9x6yrFAPnx
Compressor summary: The paper presents a new CVaR RL algorithm that combines UCB bonus, exploration, and representation learning in low-rank MDPs with nonlinear functions. It shows the sample complexity and efficiency of the algorithm.
https://openreview.net/forum?id=9w3iw8wDuE
Compressor summary: DeYO is a novel test-time adaptation method that uses a new confidence metric, Pseudo-Label Probability Difference (PLPD), to select and weight samples based on the influence of object shape on predictions.
https://openreview.net/forum?id=9pKtcJcMP3
Compressor summary: Video language planning (VLP) is a novel algorithm that uses generative models to create multimodal plans for complex tasks, which can then be executed by real robots with higher success rates.
https://openreview.net/forum?id=9nsNyN0vox
Compressor summary: The text proposes a method called "Neural Comprehension" that incorporates compiled neural networks into language models to improve their rule-based reasoning and symbolic comprehension abilities.
https://openreview.net/forum?id=9m02ib92Wz
Compressor summary: DataInf is a fast and accurate method for approximating influence scores in large generative AI models, enabling efficient data attribution and improving model transparency.
https://openreview.net/forum?id=9kG7TwgLYu
Compressor summary: The paper proposes fair and efficient algorithms for the online knapsack problem with applications to cloud resource allocation.
https://openreview.net/forum?id=9k0krNzvlV
Compressor summary: Watermark distillation trains a student model to generate watermarked text like a teacher model that uses decoding-based watermarking, with varying results depending on the watermarking strategy and conditions.
https://openreview.net/forum?id=9j1RD9LlWH
Compressor summary: The paper proposes a new method to efficiently optimize expensive functions using Gaussian Cox processes and shows its effectiveness in various datasets.
https://openreview.net/forum?id=9hjVoPWPnh
Compressor summary: The paper introduces a framework and algorithm for machine unlearning image-to-image generative models without compromising performance or violating data retention policies.
https://openreview.net/forum?id=9cQtXpRshE
Compressor summary: AGILE3D is an attention-based model that enables simultaneous and more accurate 3D object segmentation with fewer user clicks and faster inference by leveraging spatial-temporal queries and click attention.
https://openreview.net/forum?id=9bmTbVaA2A
Compressor summary: Concept-QA+V-IP is a method that combines a language model (GPT) to propose semantic concepts as queries for image classification, a question-answering network (Concept-QA) to answer the queries, and Variational Information Pursuit (V-IP) to select informative query chains that serve as explanations for predictions.
https://openreview.net/forum?id=9ZUYJpvIys
Compressor summary: TOSS uses text to guide the generation of novel views from a single image, improving quality and control over existing methods like Zero123.
https://openreview.net/forum?id=9WD9KwssyT
Compressor summary: Zipformer is a faster, more memory-efficient, and better-performing transformer model for automatic speech recognition that uses a U-Net-like encoder structure, reorganized block structure, BiasNorm, new activation functions, and ScaledAdam optimizer.
https://openreview.net/forum?id=9W6KaAcYlr
Compressor summary: The authors propose an optimization-based clustering approach using deep predictive models to identify functional cell types in retina and visual cortex using Maximally Discriminative Stimuli (MDS), which are interpretable and can be used for on-the-fly assignment of cell types during data acquisition.
https://openreview.net/forum?id=9UIGyJJpay
Compressor summary: The Vector Field Network (VFN) is a new encoder for protein design that can model both frames and atoms using learnable vector computations, improving the performance of de novo protein design methods.
https://openreview.net/forum?id=9ROuKblmi7
Compressor summary: NECO is a new method that uses neural collapse and principal component spaces to detect out-of-distribution data in machine learning models.
https://openreview.net/forum?id=9RNfX0ah0K
Compressor summary: The paper introduces LOOD, a framework to measure data memorization and information leakage in machine learning algorithms by analyzing the changes in output distribution when leaving out one data point at a time.
https://openreview.net/forum?id=9RIbNmx984
Compressor summary: The paper analyzes how network size, regularization, and the ratio between parameters and visited states affect TD algorithms in deep RL, finding a double-descent pattern in performance.
https://openreview.net/forum?id=9OevMUdods
Compressor summary: Pinocchio is a benchmark to evaluate the factual knowledge in large language models, revealing their limitations and challenges for trustworthy AI.
https://openreview.net/forum?id=9NqC72m31m
Compressor summary: The authors propose a Neural Field Classifier framework that transforms regression tasks in neural fields into multi-label classification tasks, improving performance and robustness with minimal additional cost.
https://openreview.net/forum?id=9JQtrumvg8
Compressor summary: The paper introduces WebAgent, a large language model-driven agent that learns from self-experience to complete tasks on real websites using natural language instructions, planning, summarization, and code generation.
https://openreview.net/forum?id=9DvDRTTdlu
Compressor summary: ED-NeRF is a novel approach for fast and effective 3D NeRF editing using a refinement layer in the latent space of a diffusion model and an improved loss function tailored for editing.
https://openreview.net/forum?id=9DXXMXnIGm
Compressor summary: The paper proposes a training-free method to improve diffusion models by using off-the-shelf classifiers for guidance, achieving significant performance improvements with minimal extra cost.
https://openreview.net/forum?id=9Cu8MRmhq2
Compressor summary: Norton is a method that uses optimal transport to learn correspondence between video clips and captions, addressing misalignment issues in temporal learning for video-language tasks.
https://openreview.net/forum?id=99tKiMVJhY
Compressor summary: The paper proposes Dec-POMFC, a method for multi-agent reinforcement learning that considers decentralization and partial observability, and shows its effectiveness on swarming tasks.
https://openreview.net/forum?id=97Dl82avFs
Compressor summary: The paper proposes an alt-text generation method for Twitter images that uses both image and text inputs and shows superior performance over previous methods.
https://openreview.net/forum?id=93LoCyww8o
Compressor summary: The paper proposes a Hybrid Internal Model method for quadruped robots' locomotion control using joint encoders and an IMU, which improves agility, robustness, and adaptability on various terrains.
https://openreview.net/forum?id=92btneN9Wm
Compressor summary: The proposed SPDER neural network architecture learns a positional embedding and overcomes spectral bias, resulting in faster training, lower losses, and better performance on various tasks.
https://openreview.net/forum?id=92KV9xAMhF
Compressor summary: The text discusses how diffusion models use generative processes and vector fields to estimate densities and sample data in high dimensions, and explores the importance of constraining these vector fields to be conservative for accurate inference and dimensionality estimation.
https://openreview.net/forum?id=90yw2uM6J5
Compressor summary: The paper presents Hierarchical Contact Mesh Transformer (HCMT), a GNN model that uses hierarchical mesh structures to learn long-range dependencies and quickly propagate collision effects in flexible body dynamics, achieving better performance than existing methods.
https://openreview.net/forum?id=8xliOUg9EW
Compressor summary: MUSTARD is a data generation framework that creates high-quality theorem and proof datasets for training large language models in mathematical reasoning and theorem proving tasks.
https://openreview.net/forum?id=8vKknbgXxf
Compressor summary: AD can compute a generalized derivative called a Clarke subderivative for neural networks with non-differentiable functions, under certain conditions and minibatch size.
https://openreview.net/forum?id=8utTlmhw8v
Compressor summary: The paper shows how to efficiently learn Nash equilibria in rank-1 bimatrix games using optimistic gradient descent/ascent, without requiring the Minty property.
https://openreview.net/forum?id=8sKcAWOf2D
Compressor summary: Fine-tuning language models on generalized tasks improves their entity tracking ability by enhancing positional information handling.
https://openreview.net/forum?id=8p3fu56lKc
Compressor summary: The paper studies how one-layer transformers with linear self-attention learn to implement gradient descent on synthetic noisy linear regression data and finds that the learned algorithm depends on the distribution of covariates, weight vector, and responses.
https://openreview.net/forum?id=8nxy1bQWTG
Compressor summary: Diffusion models are improved hierarchical variational autoencoders with flexible mean functions and weighted noise variances that achieve state of the art results on CIFAR-10.
https://openreview.net/forum?id=8j9hz8DVi8
Compressor summary: CASPR is a technique that improves the speed and performance of training deep neural networks by using different preconditioners for each parameter axis and combining them with Kronecker-sum approximation.
https://openreview.net/forum?id=8ishA3LxN8
Compressor summary: The authors propose a simpler method called finite scalar quantization (FSQ) for representing latent variables in variational autoencoders (VAE), which can perform well on various computer vision tasks without the drawbacks of vector quantization (VQ).
https://openreview.net/forum?id=8iTpB4RNvP
Compressor summary: This paper introduces a new backdoor attack on face forgery detection methods that uses specific trigger patterns to deceive detectors and embeds them with backdoors, outperforming existing attacks.
https://openreview.net/forum?id=8htNAnMSyP
Compressor summary: The study proposes a data-driven approach to design effective quantum feature maps for real-world datasets using feature-selection techniques and a deep neural predictor.
https://openreview.net/forum?id=8g26Yv1EOu
Compressor summary: The text presents a model-based reinforcement learning approach with neural ODEs and GD-MPC to intervene in large-scale networked excitatory point processes, using an ANI framework for knowledge transfer and sharing across diverse contexts.
https://openreview.net/forum?id=8euJaTveKw
Compressor summary: PROMETHEUS is an open-source language model that can evaluate text as well as GPT-4 when given appropriate reference materials, trained on a new dataset called FEEDBACK COLLECTION that includes score rubrics and feedback generated by GPT-4.
https://openreview.net/forum?id=8dN7gApKm3
Compressor summary: The paper proposes uncertainty quantification models for hyperspectral image classification, addressing epistemic and aleatoric uncertainties, and improving out-of-distribution and misclassification detection with regularization terms.
https://openreview.net/forum?id=8Wuvhh0LYW
Compressor summary: OmniQuant is a technique that optimizes quantization parameters for large language models to improve efficiency and performance across various settings.
https://openreview.net/forum?id=8VPWfqtQMX
Compressor summary: The paper proposes In-Context Risk Minimization (ICRM), an algorithm that uses contextual information to improve domain generalization and out-of-distribution performance.
https://openreview.net/forum?id=8Ur2xmuw7w
Compressor summary: The paper explores link prediction principles across diverse datasets, identifying three critical factors and their relationships, and providing instructions for GNN4LP model design and dataset selection.
https://openreview.net/forum?id=8QfK9Dq4q0
Compressor summary: The paper proposes TPLR, a novel method for predicting task IDs in class incremental learning using likelihood ratios and replay data, which significantly improves performance over existing approaches.
https://openreview.net/forum?id=8Q4uVOJ5bX
Compressor summary: The paper presents a novel zero-shot grounded T2I generation method with cross-attention guidance, discrete sampling, and region/boundary-aware losses that improves image quality and layout fidelity.
https://openreview.net/forum?id=8OxL034uEr
Compressor summary: The paper presents a new neural operator architecture using multigrid structures for learning nonlinear operators, which improves training ease and reduces overfitting compared to existing methods.
https://openreview.net/forum?id=8OBuqbLb8h
Compressor summary: Fast-ELECTRA is a more efficient ELECTRA-style pre-training method that uses an existing language model as the auxiliary model, smooths the main model's output distribution with temperature scaling, and reduces training cost.
https://openreview.net/forum?id=8Ju0VmvMCW
Compressor summary: The authors propose a new kernel that considers label information to measure interactions between samples in artificial neural networks and demonstrate its usefulness for understanding learning phenomena and detecting poisoning samples.
https://openreview.net/forum?id=8ISRqgtjPc
Compressor summary: CoBIT is a novel framework that unifies three pre-training objectives for vision-and-language tasks using a unicoder-decoder structure with image and text encoders and a cross-modal decoder, achieving improved performance in various tasks.
https://openreview.net/forum?id=8HCARN2hhw
Compressor summary: The text proposes learning an actionable representation of 3D scenes for agents without explicit scene reconstruction, using a blind auxiliary agent to optimize navigation and improve robustness to distribution changes.
https://openreview.net/forum?id=8FHWkY0SwF
Compressor summary: The paper proposes FedSDR, a method for personalized federated learning that discovers and removes shortcut features to improve out-of-distribution generalization using structural causal models.
https://openreview.net/forum?id=8F6bws5JBy
Compressor summary: The paper proposes a new oversampling framework using language models to generate synthetic minority samples for balancing class distribution and improving machine learning performance.
https://openreview.net/forum?id=8EyRkd3Qj2
Compressor summary: The paper proposes checkerboard learning, a novel imputation framework for fragmented multimodal data in distributed clients, which adapts transformations between modalities using a Pareto min-max framework.
https://openreview.net/forum?id=8BAkNCqpGW
Compressor summary: The paper introduces a new method to estimate policy gradients for confounded POMDPs with continuous state and observation spaces in the offline setting using non-parametric identification, moment restrictions, and min-max learning.
https://openreview.net/forum?id=89A5c6enfc
Compressor summary: The paper proposes using noisy node labels as a proxy for additional information to improve fast local graph clustering methods.
https://openreview.net/forum?id=86zAUE80pP
Compressor summary: CPPO is a method for continually aligning language models with human preferences using reinforcement learning, without needing complete retraining.
https://openreview.net/forum?id=86NGO8qeWs
Compressor summary: The paper proposes CompA, a collection of benchmarks to evaluate compositional reasoning in audio-language models (ALMs), and presents CompA-CLAP, an improved version of CLAP that performs better on these benchmarks.
https://openreview.net/forum?id=84n3UwkH7b
Compressor summary: The authors present a method to detect and prevent image generation models from memorizing and replicating training data, which can help avoid legal issues and improve creativity.
https://openreview.net/forum?id=82Mc5ilInM
Compressor summary: FreeDyG is a novel method for link prediction in dynamic graphs that uses frequency domain analysis to capture node interaction patterns, improving performance over existing methods.
https://openreview.net/forum?id=80wh3jjCZf
Compressor summary: The paper proposes Dynamic Neighborhood Construction (DNC), a novel exploitation paradigm for handling large and structured discrete action spaces in reinforcement learning, which is more efficient and scalable than existing approaches.
https://openreview.net/forum?id=7zY781bMDO
Compressor summary: The paper proposes return-conditioned supervised learning (RCSL) as an improved off-policy technique for sequential decision-making, addressing convergence issues and enabling learning from sub-optimal datasets using trajectory stitching.
https://openreview.net/forum?id=7wY67ZDQTE
Compressor summary: The paper proposes a new method for improving deep neural networks using the information bottleneck principle, which leverages the Cauchy-Schwarz divergence to avoid MSE-based regression and achieve better generalization, robustness, and trade-off between prediction accuracy and compression.
https://openreview.net/forum?id=7vVWiCrFnd
Compressor summary: This paper explores how graph neural networks (GNNs) learn from probabilistic graphical models, introduces a hierarchical framework to analyze their performance, and proposes novel methods to improve their expressive power.
https://openreview.net/forum?id=7pWRLDBAtc
Compressor summary: LG-Mix uses NTK-based convergence to mix local and global updates based on their importance, improving personalized federated learning performance across various datasets.
https://openreview.net/forum?id=7oLshfEIC2
Compressor summary: TimeMixer uses a novel multiscale-mixing approach to disentangle complex temporal variations in time series, achieving state-of-the-art forecasting performance with efficient run-time.
https://openreview.net/forum?id=7hxoYxKDTV
Compressor summary: The paper proposes PQDiff, a novel image outpainting method that can generate content beyond the input boundaries with arbitrary and continuous multiples in a single step, without relying on a pre-trained backbone network.
https://openreview.net/forum?id=7gUrYE50Rb
Compressor summary: The paper introduces new embodied question answering tasks and a multimodal representation learning model that improves performance on them.
https://openreview.net/forum?id=7gLfQT52Nn
Compressor summary: The paper introduces a new objective and algorithm for approximating the Laplacian representation in reinforcement learning, which improves exploration, generalization, and transfer by accurately recovering true eigenvectors and eigenvalues.
https://openreview.net/forum?id=7gDENzTzw1
Compressor summary: The text introduces a new robust reinforcement learning algorithm that protects against state perturbations by deriving a pessimistic policy, inferring belief states, and purifying state distributions.
https://openreview.net/forum?id=7etoNfU9uF
Compressor summary: SpikePoint is a novel point-based SNN architecture that effectively processes sparse event cloud data, achieving state-of-the-art performance in event-based action recognition with low power consumption and few parameters.
https://openreview.net/forum?id=7erlRDoaV8
Compressor summary: The authors propose an attack-and-defense framework for deleting sensitive information directly from pretrained language model weights, but find that existing methods are vulnerable to extraction attacks.
https://openreview.net/forum?id=7avlrpzWqo
Compressor summary: The paper presents a new aggregation system for distributed deep learning that uses Beta densities and optimization techniques to handle failures and data augmentation, and shows improved performance in various tasks.
https://openreview.net/forum?id=7WsivwyHrS
Compressor summary: YOQO is a new label-only attack method for membership inference that can reduce the query budget to just one and achieve high accuracy while being resistant to common defenses.
https://openreview.net/forum?id=7W3GLNImfS
Compressor summary: The text discusses the limitations and biases of using human feedback for evaluating and training large language models, especially regarding factuality errors and assertiveness.
https://openreview.net/forum?id=7VPTUWkiDQ
Compressor summary: The paper studies when object-centric representations, learned by specific autoencoders with consistent encoder-decoder structures, can generalize to novel compositions of known concepts, using identifiability theory and experiments.
https://openreview.net/forum?id=7UhxsmbdaQ
Compressor summary: The paper proposes Beam Enumeration, which enhances generative molecular design by explaining substructures and improving sample efficiency using reinforcement learning.
https://openreview.net/forum?id=7Ttk3RzDeu
Compressor summary: The paper investigates how to improve book summarization by LLMs using human and automatic evaluation methods, introducing BooookScore as a metric for coherence errors, and comparing different prompting workflows and LLM models.
https://openreview.net/forum?id=7TOs9gjAg1
Compressor summary: InfoCORE is a method to remove batch effects and improve drug screening data analysis by maximizing information and adaptively reweighting samples.
https://openreview.net/forum?id=7QI7tVrh2c
Compressor summary: The paper proposes a new method to improve the neural network approximation of partial differential equations by optimizing both the model and the random samples in the training set using a deep generative model.
https://openreview.net/forum?id=7Ol6foUi1G
Compressor summary: DIMAP is a pruning method for hierarchical vision transformers that considers local self-attention and weight distribution, achieving efficient compression without sacrificing performance.
https://openreview.net/forum?id=7NzgkEdGyr
Compressor summary: The paper introduces Orthogonal Butterfly, a parameter-efficient way to adapt large models to downstream tasks using orthogonal matrices and butterfly structures.
https://openreview.net/forum?id=7M0EzjugaN
Compressor summary: The text proposes realistic continuous learning scenarios for embodied agents and introduces CAMA, a new method that updates information based on confidence scores without task boundary information, which performs better than previous methods in the proposed setups.
https://openreview.net/forum?id=7KDuQPrAF3
Compressor summary: The paper introduces a foundation model for Error Correction Codes that uses a modified Transformer architecture to learn from multiple codes and adapt to unseen ones, achieving similar or better results than specialized code-specific models.
https://openreview.net/forum?id=7Jwpw4qKkb
Compressor summary: AutoDAN is a novel jailbreak attack against aligned LLMs that generates stealthy prompts using a hierarchical genetic algorithm, overcoming scalability and stealthiness issues in existing techniques.
https://openreview.net/forum?id=7JfKCZQPxJ
Compressor summary: The paper introduces a new video evaluation metric called STREAM, which independently assesses the spatial and temporal aspects of videos, addressing the limitations of existing metrics like Frechet Video Distance (FVD) and providing insights for improving video generative models.
https://openreview.net/forum?id=7FeIRqCedv
Compressor summary: SLiMe is a method to segment images using vision-language models like Stable Diffusion, requiring only one annotated sample and achieving better performance than other methods.
https://openreview.net/forum?id=7ERQPyR2eb
Compressor summary: Real3D-Portrait is a framework for generating realistic talking portrait videos from unseen images, using 3D reconstruction, motion-conditioned animation, and natural torso and background synthesis.
https://openreview.net/forum?id=7D9X2cFnt1
Compressor summary: Elastic Feature Consolidation (EFC) is a method for learning from sequential tasks without previous data that reduces feature drift and improves performance on image classification tasks.
https://openreview.net/forum?id=79rfgv3jw4
Compressor summary: The paper presents a framework for creating and evaluating AI agents that can collaborate with less-skilled partners in complex settings, using chess variants as models. It shows that skill-compatible AI agents outperform conventional chess AI despite being weaker individually.
https://openreview.net/forum?id=79FVDdfoSR
Compressor summary: Equivariant neural networks can be improved by using specific point-wise activations and coordinate systems, which this paper characterizes for general symmetries and applications to graph and image analysis.
https://openreview.net/forum?id=78iGZdqxYY
Compressor summary: MIRAGE is a novel graph distillation algorithm that compresses computation data in message-passing GNNs to create a smaller synthetic training set without compromising model performance or requiring full dataset training.
https://openreview.net/forum?id=776lhoaulC
Compressor summary: The paper proposes a novel method for nighttime optical flow that adapts appearance and boundary features from daytime and event domains using a common latent space.
https://openreview.net/forum?id=770DetV8He
Compressor summary: The Markov Bridge Model is a new generative framework for single-step retrosynthesis planning in chemistry that uses Markov processes to approximate intractable discrete distributions and does not require tractable noise distributions as sampling proxies.
https://openreview.net/forum?id=70IgE3tRbu
Compressor summary: The paper introduces CIL, a new method for learning invariant features in continuous domain problems, which improves generalization and performance over existing techniques.
https://openreview.net/forum?id=6yv8UHVJn4
Compressor summary: The paper introduces two algorithms for online reinforcement learning in adversarial settings that outperform existing methods in terms of regret performance.
https://openreview.net/forum?id=6xfe4IVcOu
Compressor summary: The paper proposes a new technique, Chain of Hindsight, that enables language models to learn from any feedback by converting it into sentences and fine-tuning the model on sequences of model generations paired with feedback, improving alignment with human preferences and performance on summarization and dialogue tasks.
https://openreview.net/forum?id=6vF0ZJGor4
Compressor summary: ImplicitSLIM is an unsupervised learning method for sparse high-dimensional data that enhances embedding-based models by extracting embeddings from SLIM-like models, improving performance and speeding up convergence in collaborative filtering applications.
https://openreview.net/forum?id=6tqgL8VluV
Compressor summary: The paper studies the necessary conditions for providing error guarantees in machine learning-based database operations, such as indexing, cardinality estimation, and range-sum estimation, by presenting lower bounds on model size.
https://openreview.net/forum?id=6pPYRXKPpw
Compressor summary: The text introduces simulation environments and datasets to evaluate imitation learning models' ability to learn diverse multi-modal human behaviors using tractable metrics and benchmarks existing methods.
https://openreview.net/forum?id=6p8lpe4MNf
Compressor summary: The paper proposes a semantic invariant watermarking method for large language models that can resist attacks based on synonym substitution and text paraphrasing while maintaining security robustness.
https://openreview.net/forum?id=6okaSfANzh
Compressor summary: The paper presents a cost-saving cascade pipeline for reasoning tasks using LLMs, which switches between cheaper and more expensive models based on answer consistency.
https://openreview.net/forum?id=6mLjDwYte5
Compressor summary: The paper shows that combining sparse mixture-of-experts and instruction tuning improves the performance of large language models with fewer parameters and computational cost.
https://openreview.net/forum?id=6iwg437CZs
Compressor summary: STanHop-Net is a novel neural network that uses sparse temporal and cross-series representations learned by tandem Hopfield layers and enhanced with external memory modules for fast response to events in multivariate time series prediction.
https://openreview.net/forum?id=6hvtSLkKeZ
Compressor summary: The paper introduces CCBPP, a vector Bin Packing Problem variant that considers item classes and sizes, and proposes an Encoder-Decoder Model with GCN and Cluster Decode to solve it efficiently.
https://openreview.net/forum?id=6cFcw1Rxww
Compressor summary: This paper introduces local search for GFlowNets, a method that improves diverse sample generation by exploiting high rewarded solutions using destruction and reconstruction guided by backward and forward policies.
https://openreview.net/forum?id=6bcAD6g688
Compressor summary: The study introduces a framework to evaluate and fix label errors in datasets used for training harmless language models, improving their performance and safety.
https://openreview.net/forum?id=6RR3wU4mSZ
Compressor summary: Our method speeds up self-attention in pretrained transformer models for long sequences without retraining or hardware changes.
https://openreview.net/forum?id=6PmJoRfdaK
Compressor summary: LongLoRA is a method to efficiently extend the context sizes of pre-trained language models by using sparse local attention during fine-tuning and parameter-efficient context expansion.
https://openreview.net/forum?id=6PbvbLyqT6
Compressor summary: DDCFR is a new algorithm for solving imperfect-information games that learns how to weight each iteration dynamically, improving convergence rate and performance.
https://openreview.net/forum?id=6O3Q6AFUTu
Compressor summary: The paper proposes a new interpolation method that combines noise addition and diffusion models to create better image results and handle various types of images.
https://openreview.net/forum?id=6N8TW504aa
Compressor summary: The paper introduces GMOGP, a novel framework for multi-output regression that improves flexibility, optimality, and scalability over existing MOGP methods.
https://openreview.net/forum?id=6MRm3G4NiU
Compressor summary: The authors introduce SaProt, a protein language model that incorporates 3D structure information, improving performance on various downstream tasks.
https://openreview.net/forum?id=6LLho5X6xV
Compressor summary: UniTabE is a new method that pretrains models on tables and uses Transformer encoders to improve their performance in data science tasks.
https://openreview.net/forum?id=6J3ehSUrMU
Compressor summary: The text discusses Federated Domain Adaptation, a challenging problem in federated learning where source and target clients have different domains and limited data, and proposes a new aggregation rule ($ exttt{FedGP}$) and an auto-weighting scheme to improve performance.
https://openreview.net/forum?id=6IjN7oxjXt
Compressor summary: Adversarial training improves neural network robustness but can cause trade-offs; CURE is a new framework that selectively updates layers to balance robustness and generalization while mitigating "robust overfitting".
https://openreview.net/forum?id=6Gzkhoc6YS
Compressor summary: PerSAM is a method to customize SAM, a powerful segmentation framework, for specific visual concepts using one-shot data and without any training, improving performance on various benchmarks and enabling personalized text-to-image synthesis.
https://openreview.net/forum?id=6CZ50WgfCG
Compressor summary: DrS is a method for learning dense rewards from stages that can be reused in unseen tasks, improving RL performance and sample efficiency on physical robot manipulation tasks.
https://openreview.net/forum?id=6CIGhcJYJH
Compressor summary: Two-timescale extragradient efficiently optimizes minimax problems without assuming nondegenerate Hessians.
https://openreview.net/forum?id=6ARlSgun7J
Compressor summary: This paper proposes a method to reduce label variance in extreme classification using a tail-robust teacher model and knowledge distillation, improving tail performance on challenging label sets.
https://openreview.net/forum?id=6ALuy19mPa
Compressor summary: DreamClean is a training-free method that restores high-quality images from diverse degradations by embedding the degraded image in the latent of pre-trained diffusion models and re-sampling it through a novel Variance Preservation Sampling technique.
https://openreview.net/forum?id=66arKkGiFy
Compressor summary: The paper analyzes the relationship between PnP-ULA, a Bayesian method for imaging inverse problems, and its data-fidelity and denoiser, providing an error bound and validating it numerically on several examples.
https://openreview.net/forum?id=64kSvC4iPg
Compressor summary: The paper proposes a compression method for online language models that uses a lightweight adapter to reduce memory and attention needs, enabling similar performance to full context models with much smaller memory spaces.
https://openreview.net/forum?id=62K7mALO2q
Compressor summary: The study analyzes how large language models learn new tasks through in-context learning, revealing emergent abilities and underlying concepts using random binary sequences as context.
https://openreview.net/forum?id=60lNoatp7u
Compressor summary: NeurRev is a framework for dynamic sparse training that awakens dormant neurons by pruning and reduces memory consumption on resource-constrained platforms.
https://openreview.net/forum?id=5xadJmgwix
Compressor summary: The paper presents a new way to use diffusion models for generating realistic and complex sketches from pixels, using guidance from classifiers and a special sampling strategy.
https://openreview.net/forum?id=5tGGWOijvq
Compressor summary: The paper proposes Prompt Risk Control, a lightweight framework for selecting language model prompts based on rigorous risk bounds, to avoid high loss in deployment.
https://openreview.net/forum?id=5t57omGVMw
Compressor summary: The paper proposes a bandit algorithm that can learn optimal parameters for Successive Over-Relaxation (SOR), a linear system solver, without extra matrix computations and shows its effectiveness in speeding up numerical simulations.
https://openreview.net/forum?id=5t44vPlv9x
Compressor summary: The paper proposes a two-branch neural network that adapts and encodes body part correlations and global frequencies to reconstruct dynamic human motion and shape with high detail and generalization.
https://openreview.net/forum?id=5sjxMwWmk8
Compressor summary: The paper proposes GNNSync, a framework using neural networks to estimate unknown angles from noisy measurements in applications like sensor network localization and clock synchronization.
https://openreview.net/forum?id=5sixirvG0I
Compressor summary: The paper proposes a deep reinforcement learning algorithm based on the Whittle index for multi-agent inventory management, which performs better than existing methods on large-scale problems.
https://openreview.net/forum?id=5o9G4XF1LI
Compressor summary: The paper discusses how imperfect reward functions in reinforcement learning can lead to suboptimal outcomes and proposes methods to address this issue, including optimal early stopping and worst-case reward maximization.
https://openreview.net/forum?id=5nM2AHzqUj
Compressor summary: The authors propose and test a new self-attention mechanism called Linear Log-Normal Attention that improves the scalability of transformer models by emulating the original attention's distribution and concentration properties.
https://openreview.net/forum?id=5liV2xUdJL
Compressor summary: The paper proposes a new method for training stateful policies in reinforcement learning using a stochastic internal state kernel and a stateless policy, which improves speed, simplicity, and scalability over Backpropagation Through Time.
https://openreview.net/forum?id=5jcav5RcKw
Compressor summary: The JAM framework combines text and image models to produce seamless multimodal outputs using a data-efficient instruction-tuning strategy.
https://openreview.net/forum?id=5jWsW08zUh
Compressor summary: This paper investigates how different factors affect the Lipschitz continuity of neural networks, which is important for their robustness, generalization, and adversarial resistance.
https://openreview.net/forum?id=5iENGLEJKG
Compressor summary: The authors propose a method to interpret and edit vision transformers using natural language explanations of their latent tokens.
https://openreview.net/forum?id=5hAMmCU0bK
Compressor summary: Robust IQL is a new offline reinforcement learning approach that uses Huber loss and quantile estimators to handle various types of data corruption, improving upon the resilience of implicit Q-learning.
https://openreview.net/forum?id=5h0qf7IBZZ
Compressor summary: The paper proposes MiniLLM, a technique that distills smaller language models from generative larger ones using reverse KLD objective and improves their performance in instruction-following tasks.
https://openreview.net/forum?id=5ep85sakT3
Compressor summary: The paper proposes a neural network approach for Neural Contextual Bandits and shows improved regret bounds for online regression with almost convex losses that satisfy the Quadratic Growth condition.
https://openreview.net/forum?id=5dlfiJIXoh
Compressor summary: S-ViLM is a video-language modeling framework that enhances fine-grained information in videos and captions, improving performance on various downstream tasks.
https://openreview.net/forum?id=5bNYf0CqxY
Compressor summary: The text introduces a new technique to improve the robustness of spiking neural networks against adversarial perturbation of inputs by using rate-encoded models and randomized smoothing.
https://openreview.net/forum?id=5a79AqFr0c
Compressor summary: ControlVideo is a text-to-video model that uses a pre-trained image generator and frame smoothing techniques to produce high-quality, controllable videos from text prompts and motion sequences, without the need for extensive training.
https://openreview.net/forum?id=5RielfrDkP
Compressor summary: The Multiresolution Meta-Framelet-based Graph Convolutional Network (MM-FGCN) is a new method to adaptively learn multiresolution representations of graphs for different tasks, using meta-learning and framelets.
https://openreview.net/forum?id=5RUf9nEdyC
Compressor summary: TEDDY is a framework for finding sparse graph lottery tickets by exploiting edge-degree information and achieving better performance than iterative approaches.
https://openreview.net/forum?id=5Nn2BLV7SB
Compressor summary: PandaLM is a large language model that can judge other LLMs based on various factors like conciseness and clarity, ensuring fairer and cost-effective hyperparameter tuning without privacy risks.
https://openreview.net/forum?id=5Lp6qU9hzV
Compressor summary: The paper proposes a Multiscale Positive-Unlabeled (MPU) training framework to improve the detection of AI-generated texts, especially on short texts like SMSes and Tweets.
https://openreview.net/forum?id=5KojubHBr8
Compressor summary: MMICL is a new approach that enhances vision-language models' ability to understand complex multi-modal prompts using in-context learning and a large dataset, improving their performance on various tasks.
https://openreview.net/forum?id=5KUiMKRebi
Compressor summary: Our approach combines deterministic and probabilistic components in deep learning models to achieve accurate predictions with well-calibrated uncertainty estimates, outperforming existing methods.
https://openreview.net/forum?id=5JWAOLBxwp
Compressor summary: The authors propose a new feature representation for 3D vision algorithms that improves the performance of Vector Neuron by capturing more details in 3D data.
https://openreview.net/forum?id=5HCnKDeTws
Compressor summary: The paper studies how different factors like model size, data size, and tuning methods affect the performance of large language models in various tasks, finding that they follow a power-based scaling law and have task- and data-dependent optimal methods.
https://openreview.net/forum?id=5EniAcsO7f
Compressor summary: The paper proposes improving visual localization by expanding the training set with synthetic images generated from text-to-image models and using a tailored training approach.
https://openreview.net/forum?id=5ES5Hdlbxw
Compressor summary: The paper explains deep RL's success using a new algorithm (SQIRL) that separates exploration and learning, showing that random exploration works well and deep function approximators generalize effectively in many stochastic MDPs.
https://openreview.net/forum?id=5Dwqu5urzs
Compressor summary: The paper introduces a physics-regulated deep reinforcement learning framework for safe autonomous systems that integrates data-driven and model-based methods with invariant principles.
https://openreview.net/forum?id=5Ca9sSzuDp
Compressor summary: The paper analyzes how CLIP's image encoder works by breaking down its components, finding text representations for attention heads, and using this knowledge to improve a zero-shot segmentation model.
https://openreview.net/forum?id=5BXAXOpaWu
Compressor summary: ISA is a method to retrieve images based on a query image and text, using an adaptive token learner to map the image to a sentence, and an asymmetric structure for flexible deployment of large and small models.
https://openreview.net/forum?id=5BCFlnfE1g
Compressor summary: MetaCLIP is a method to create balanced datasets for language-image pre-training by using metadata derived from CLIP's concepts, outperforming CLIP's data on multiple benchmarks.
https://openreview.net/forum?id=567BjxgaTp
Compressor summary: The paper presents a simple and accurate lie detector for large language models that works without access to their internal workings or factual knowledge.
https://openreview.net/forum?id=55uj7mU7Cv
Compressor summary: The text discusses a study that introduces a theory to eliminate multiple translation functions and improve unsupervised domain translation by aligning conditional distributions over auxiliary variable-induced subsets of domains.
https://openreview.net/forum?id=52fz5sUAy2
Compressor summary: The paper proposes a method to address selection bias and neighborhood effects in recommender systems by formulating the problem as an interference issue and introducing a treatment representation for causal inference.
https://openreview.net/forum?id=4znwzG92CE
Compressor summary: Habitat 3.0 is a simulation platform that studies human-robot collaboration in home environments using accurate humanoid simulation, human-in-the-loop infrastructure, and collaborative tasks such as Social Navigation and Social Rearrangement.
https://openreview.net/forum?id=4zZFGliCl9
Compressor summary: The paper investigates the posterior collapse phenomenon in variational autoencoders (VAEs) and advances the theory for two less studied VAE variants, conditional VAEs and hierarchical VAEs, revealing their causes of posterior collapse.
https://openreview.net/forum?id=4yaFQ7181M
Compressor summary: The authors propose a novel data-driven method for computational fluid dynamics that can predict solutions at arbitrary spatial and temporal locations using recurrent GNNs and spatio-temporal attention.
https://openreview.net/forum?id=4vPVBh3fhz
Compressor summary: The paper proposes a new method for constructing prediction sets with guaranteed uncertainty quantification in the presence of label shift by estimating importance weights and propagating them through Gaussian elimination.
https://openreview.net/forum?id=4rBEgZCubP
Compressor summary: VPD is a method that learns a particle-based representation of 3D scenes, a neural simulator, and a renderer from RGB-D videos without privileged information, enabling scene editing and long-term predictions for various applications.
https://openreview.net/forum?id=4r2ybzJnmN
Compressor summary: The text proposes a new algorithm to learn delays in spiking neural networks, which improves their performance for temporal data processing tasks such as speech recognition.
https://openreview.net/forum?id=4olqbTBt1Y
Compressor summary: DREAM is a novel method that uses graph-level and subgraph-enhanced branches to explore both global and local structures in graphs for open-set domain adaptation, incorporating posterior regularization, mixup, and k nearest neighbor-based graph-of-graphs.
https://openreview.net/forum?id=4kLVvIh8cp
Compressor summary: The paper proposes PNLSVI, an oracle-efficient algorithm for offline RL with non-linear function approximation that has minimax optimal instance-dependent regret and a wide range of applicability.
https://openreview.net/forum?id=4iPw1klFWa
Compressor summary: SNNKs are scalable neural network kernels that can replace regular feedforward layers, leading to more efficient and expressive models with reduced parameters and computational cost.
https://openreview.net/forum?id=4h1apFjO99
Compressor summary: Diffusion-TS is a new DDPM-based framework that generates high-quality multivariate time series using an encoder-decoder Transformer with disentangled temporal representations, enabling conditional generation tasks and achieving state-of-the-art results.
https://openreview.net/forum?id=4g02l2N2Nx
Compressor summary: Hedgehog is a learnable linear attention that mimics softmax attention's properties and achieves near-linear complexity, improving Transformer efficiency and performance in various settings.
https://openreview.net/forum?id=4eJDMjYZZG
Compressor summary: The paper shows how to use reinforcement learning to fine-tune language models to evade existing detectors that try to identify if a text was written by a model or not.
https://openreview.net/forum?id=4bSQ3lsfEV
Compressor summary: The paper presents a new concept of functionally equivalent features in neural networks, which are features that output the same value under certain transformations, and a metric for measuring their complexity. It also proposes an algorithm to prune networks based on FEFs, which can improve performance and reduce parameters. The experiments confirm the existence and effects of FEFs across different networks and tasks.
https://openreview.net/forum?id=4bLXfRd0CX
Compressor summary: Earth Mover Distance Optimization (EMO) is a new method for improving language models by addressing issues with existing cross-entropy based approaches, leading to better performance and easier fine-tuning.
https://openreview.net/forum?id=4aywmeb97I
Compressor summary: Asynchronous federated learning can be slowed down by data heterogeneity, but a new method that caches and reuses client updates can improve its convergence under non-i.i.d. data.
https://openreview.net/forum?id=4Zz5UELkIt
Compressor summary: The paper proposes a method to adaptively design data collection policies for indirect experiments, using influence functions to minimize mean-squared error and improve sample efficiency.
https://openreview.net/forum?id=4YESQqIys7
Compressor summary: The paper introduces NfgTransformer, a neural network architecture for representing normal-form games and enabling deep learning systems to reason about strategic interaction concepts.
https://openreview.net/forum?id=4WnqRR915j
Compressor summary: Llemma is a powerful pretrained language model for mathematics that surpasses existing models on the MATH benchmark and can perform theorem proving without finetuning.
https://openreview.net/forum?id=4WM0OogPTx
Compressor summary: The paper proposes Conservative Density Estimation (CDE), a novel algorithm for offline reinforcement learning that handles out-of-distribution errors by constraining state-action occupancy distributions, achieving state-of-the-art performance on the D4RL benchmark.
https://openreview.net/forum?id=4VgBjsOC8k
Compressor summary: This paper shows that depthwise-separable convolutional neural networks have trained filters that cluster into a few main patterns resembling biological vision systems, suggesting potential for more interpretable and biologically-inspired designs.
https://openreview.net/forum?id=4VIgNuQ1pY
Compressor summary: The paper introduces stable Neural SDEs classes for handling irregular and incomplete time series data, and evaluates their performance on various tasks and datasets.
https://openreview.net/forum?id=4VGEeER6W9
Compressor summary: This paper develops non-asymptotic theory to analyze the data generation process of diffusion models in discrete time, with or without score function estimates.
https://openreview.net/forum?id=4UiLqimGm5
Compressor summary: The paper proposes Coordinate-Aware Modulation (CAM), which improves neural field representations by combining MLPs and grid methods, enhancing performance and stability across various signals.
https://openreview.net/forum?id=4Ua4hKiAJX
Compressor summary: The paper proposes a novel graph rewiring framework for GNNs that balances over-squashing, locality, and sparsity, and shows its effectiveness on various benchmarks.
https://openreview.net/forum?id=4UIBysXjVq
Compressor summary: The paper proposes a novel spectral graph neural network (RQGNN) for detecting anomalies in graphs by using Rayleigh Quotient and Chebyshev Wavelet GNN with RQ-pooling, achieving better performance than existing methods.
https://openreview.net/forum?id=4NhMhElWqP
Compressor summary: The DAM is a neural model that uses randomly sampled histories to produce an adjustable basis composition for universal forecasting of time series across domains and datasets, outperforming existing models on various metrics.
https://openreview.net/forum?id=4N97bz1sP6
Compressor summary: The paper proposes a method to separate single-source sounds in audio using language descriptions, improving unsupervised and supervised learning for audio separation.
https://openreview.net/forum?id=4MsfQ2H0lP
Compressor summary: GAPN is a novel method that uses a generative adversarial policy network to predict the structure of large protein complexes efficiently and accurately.
https://openreview.net/forum?id=4L0xnS4GQM
Compressor summary: The Chain-of-Table framework uses tabular data as a proxy for intermediate thoughts in the reasoning chain to improve table understanding tasks with large language models.
https://openreview.net/forum?id=4KqkizXgXU
Compressor summary: The paper introduces curiosity-driven exploration to train red team models, which generates diverse and effective test cases for large language models without relying solely on human testers.
https://openreview.net/forum?id=4KZpDGD4Nh
Compressor summary: Cosmos is a framework that uses neurosymbolic grounding to create object-centric world models for compositional generalization tasks, achieving state-of-the-art results.
https://openreview.net/forum?id=4JbrdrHxYy
Compressor summary: Zip combines CLIP's boundary prior with SAM to enable annotation-free object detection and instance segmentation, improving performance significantly in various settings.
https://openreview.net/forum?id=4IT2pgc9v6
Compressor summary: One for All (OFA) is a framework that uses text-attributed graphs and a single model to handle various tasks on graphs, such as node, link, and graph tasks, without fine-tuning.
https://openreview.net/forum?id=4DoSULcfG6
Compressor summary: The paper proposes a new attack called Chameleon that can effectively infer membership in machine learning models using only the predicted labels by using data poisoning and query selection.
https://openreview.net/forum?id=4Ay23yeuz0
Compressor summary: TabSYN is a method that uses diffusion models within a VAE to synthesize high-quality tabular data with generality, quality, and speed advantages over existing methods.
https://openreview.net/forum?id=49z97Y9lMq
Compressor summary: The paper introduces a new efficient metric (LCOT) for circular probability measures that allows applying ML algorithms and improves representation learning.
https://openreview.net/forum?id=488A64eOf6
Compressor summary: The paper proposes a decoding method for language models that optimizes multiple aspects of text quality and shows its effectiveness in aligning with human texts.
https://openreview.net/forum?id=47hDbAMLbc
Compressor summary: The paper studies the challenges and solutions of robust memorization in neural networks, showing that it is NP-hard, providing upper and lower bounds for network width, and constructing optimal networks for different data sets.
https://openreview.net/forum?id=43cYe4oogi
Compressor summary: The paper analyzes the expressivity of Graph Neural Networks (GNNs) for Knowledge Graph (KG) reasoning, formalizes the rule structures they can learn, and proposes a novel labeling strategy to improve their performance.
https://openreview.net/forum?id=42lcaojZug
Compressor summary: The paper proposes a neural network-based rate control system for learning-based video compression, which assigns optimal bitrates to frames and predicts coding parameters for a given rate, enhancing the compression performance.
https://openreview.net/forum?id=3zvB14IF6D
Compressor summary: The paper proposes DORSal, a video diffusion model for high-fidelity 3D scene generation from object-centric representations, enabling scalable neural rendering with object-level editing.
https://openreview.net/forum?id=3zQo5oUvia
Compressor summary: REBAR is a novel approach for self-supervised contrastive learning in time-series, which uses cross-attention to reconstruct subsequences and labels them with REBAR error, achieving state-of-the-art results on downstream tasks.
https://openreview.net/forum?id=3zKtaqxLhW
Compressor summary: Generalized Knowledge Distillation (GKD) is a new method that trains smaller models on their own outputs with feedback from larger models to overcome distribution mismatch issues in compressing sequence models.
https://openreview.net/forum?id=3z60EWfh1p
Compressor summary: The paper introduces ${\it GATE}$, a novel transfer learning technique for regression tasks using differential geometry to align latent vectors between tasks and enable knowledge transfer.
https://openreview.net/forum?id=3y2TfP966N
Compressor summary: T-Rep is a self-supervised method that learns vector embeddings of time and uses them to extract temporal features for robust time series representation learning, outperforming existing algorithms on classification, forecasting, and anomaly detection tasks.
https://openreview.net/forum?id=3y1K6buO8c
Compressor summary: The authors propose a new MEG decoding model that improves image retrieval and reveals high-level visual features of brain activity with high temporal resolution.
https://openreview.net/forum?id=3xHDeA8Noi
Compressor summary: Sophia is a scalable second-order optimizer that speeds up language model training with GPT models by using a simple estimate of the diagonal Hessian as the pre-conditioner.
https://openreview.net/forum?id=3xDaj4pRna
Compressor summary: SAM, a neural network training method, can improve feature quality and diversity by adaptively suppressing well-learned features in datasets with redundant or spurious features.
https://openreview.net/forum?id=3w6xuXDOdY
Compressor summary: The paper studies the generalization abilities of offline RL algorithms and introduces a benchmark with diverse datasets to evaluate them, finding that existing methods struggle to generalize.
https://openreview.net/forum?id=3tM1l5tSbv
Compressor summary: The paper proposes a new generative learning approach using a rectified flow model to tackle multi-valued input-solution mappings in non-convex problems, outperforming existing methods in stability, complexity, and solution optimality.
https://openreview.net/forum?id=3rmpixOjPS
Compressor summary: This paper proposes a novel module for lightweight Vision Transformers that improves their performance by adapting re-parameterization technology, achieving state-of-the-art results on various vision tasks.
https://openreview.net/forum?id=3qo1pJHabg
Compressor summary: The paper proposes a new method to defend against adversarial perturbations in visual object tracking using a spatial-temporal continuous representation and semantic text guidance.
https://openreview.net/forum?id=3pf2hEdu8B
Compressor summary: The paper proposes a new uniformity metric for self-supervised learning that is sensitive to dimensional collapse and improves the performance of downstream tasks.
https://openreview.net/forum?id=3pWSL8My6B
Compressor summary: The study shows how AI models learn sparse, symbolic concepts from occluded input samples and proves conditions for their emergence.
https://openreview.net/forum?id=3oTPsORaDH
Compressor summary: SEGNO is a new GNN model that incorporates second-order continuity and velocity information to improve generalization for complex physical systems.
https://openreview.net/forum?id=3mnWvUZIXt
Compressor summary: The paper investigates how to learn good representations for decision-making using video data, and shows that some common methods work well with iid noise but struggle with exogenous, non-iid noise.
https://openreview.net/forum?id=3fEKavFsnv
Compressor summary: The paper proposes a novel optimization method (MMD-MP) to detect machine-generated texts using maximum mean discrepancy, addressing the challenges posed by multiple text populations from different large language models.
https://openreview.net/forum?id=3f5PALef5B
Compressor summary: LEGO-Prover uses a growing skill library with verified lemmas to enhance LLMs for theorem proving, enabling modular proofs, skill creation, and evolution, improving pass rates on math problems.
https://openreview.net/forum?id=3eFMnZ3N4J
Compressor summary: Efficient-3DiM is a framework that simplifies and speeds up the process of generating novel views from single images using diffusion models.
https://openreview.net/forum?id=3d0OmYTNui
Compressor summary: The authors propose a differential privacy framework for aligning large language models through reinforcement learning, with applications in text generation and summarization.
https://openreview.net/forum?id=3cuJwmPxXj
Compressor summary: The paper proposes Rep4Ex, a method to learn identifiable representations that enable non-linear extrapolation for intervention prediction tasks using autoencoders with linearity constraints on latent features.
https://openreview.net/forum?id=3bq3jsvcQ1
Compressor summary: Step-Back Prompting is a technique that helps LLMs do abstractions and improve their reasoning skills, leading to significant performance gains on various tasks.
https://openreview.net/forum?id=3aZCPl3ZvR
Compressor summary: Sharpness-Aware Minimization (SAM) enhances performance, especially under label noise, by altering the gradient contribution from clean examples and modifying the network Jacobian, which can be replicated with cheaper alternatives.
https://openreview.net/forum?id=3ZqKxMHcAg
Compressor summary: The authors propose evaluating language models' performance and alignment in negotiation games, which reveal their ability to plan and negotiate, and show that open-source models are weak, cooperation is challenging, and the best models do not always prevail.
https://openreview.net/forum?id=3Z1gxuAQrA
Compressor summary: PoSE is a training method that simulates long inputs using a fixed context window, allowing large language models to adapt to different input lengths efficiently and with minimal performance impact.
https://openreview.net/forum?id=3Y7r6xueJJ
Compressor summary: The text discusses how dataset bias in continual learning algorithms can lead to undesirable transfer and accumulation of bias, and proposes a baseline method to address this issue.
https://openreview.net/forum?id=3Vw7DQqq7U
Compressor summary: LEMON is a method to initialize and train scaled Transformers using the weights of their smaller pre-trained counterparts, reducing computational costs by up to 56.7%.
https://openreview.net/forum?id=3VD4PNEt5q
Compressor summary: The paper proposes an adversarial attack on camera-LiDAR fusion models for 3D object detection by exploiting their weakness in the camera modality.
https://openreview.net/forum?id=3UWuFoksGb
Compressor summary: The paper introduces PARL, a framework for learning abstract actions and states from language-annotated demonstrations, which can generalize to new scenarios and tasks.
https://openreview.net/forum?id=3TO3TtnOFl
Compressor summary: The text introduces binary token representations (BTR) for retrieval augmentation in language models, which significantly reduce computation, storage, and improve inference speed on knowledge-intensive NLP tasks.
https://openreview.net/forum?id=3SJE1WLB4M
Compressor summary: The paper studies how kernel methods like neural networks are trained with different algorithms, and shows how their generalization error depends on the learning profile of these algorithms for various data models.
https://openreview.net/forum?id=3ROGsTX3IR
Compressor summary: The paper explores how deep neural networks learn new features and experience Grokking phenomena by applying the adaptive kernel approach to teacher-student models and comparing it to phase transitions in physics.
https://openreview.net/forum?id=3QkzYBSWqL
Compressor summary: The authors propose a universal data poisoning attack that exploits inter-class transferability to control misclassifications in deep image classifiers with minimal poison samples.
https://openreview.net/forum?id=3QLkwU40EE
Compressor summary: SPTNet uses a two-stage adaptation approach and spatial prompt tuning method to improve generalized category discovery, achieving state-of-the-art results with minimal additional parameters.
https://openreview.net/forum?id=3NnfJnbJT2
Compressor summary: GIO is a scalable method to select informative data subsets for model training, achieving high performance with few examples, across different tasks and domains.
https://openreview.net/forum?id=3NmO9lY4Jn
Compressor summary: The paper proposes a novel framework to generate high-quality minority samples using diffusion models by introducing a uniqueness metric and a sampling technique called minority guidance.
https://openreview.net/forum?id=3M0GXoUEzP
Compressor summary: CrIBo is a new method for improving visual representation learning by bootstrapping objects at the image level, enhancing in-context learning and achieving state-of-the-art results on downstream tasks.
https://openreview.net/forum?id=3KDbIWT26J
Compressor summary: The study explores why large language models can translate without parallel data and finds that unintentional bilingualism in pre-training corpus, especially word alignment data, helps them learn translation skills.
https://openreview.net/forum?id=3K3s9qxSn7
Compressor summary: The paper investigates how easily circuits can represent various aspects of reinforcement learning problems and shows that model-based algorithms often have better performance due to simpler environment rules.
https://openreview.net/forum?id=3JjJezzVkT
Compressor summary: Momentum may help reduce variance in stochastic optimization, but its benefits are limited for practical deep learning training when the optimal learning rate is not very large.
https://openreview.net/forum?id=3GurO0kRue
Compressor summary: The paper proposes a method called SHE to address class-level imbalance by discovering hidden subpopulations and adjusting predictions for better generalization.
https://openreview.net/forum?id=3FJOKjooIj
Compressor summary: The paper proposes a novel self-supervised learning method for heterogeneous graphs that captures both homogeneity and heterogeneity without pre-defined meta-paths, using a self-expressive matrix and two losses to improve task performance.
https://openreview.net/forum?id=3EWTEy9MTM
Compressor summary: The text explains how a sequence of intermediate steps (chain of thought) helps large language models solve complex arithmetic and symbolic reasoning tasks by enabling serial computation that they otherwise lack.
https://openreview.net/forum?id=39cPKijBed
Compressor summary: The paper introduces a new technique called time-dependent importance reweighting to reduce dataset bias in diffusion models and improve sample quality.
https://openreview.net/forum?id=38E4yUbrgr
Compressor summary: RLC is a novel LMSI method that uses evaluation scores from language models to improve their performance on various tasks without external supervision.
https://openreview.net/forum?id=36L7W3ri4U
Compressor summary: Q-Replicator Dynamics (QRD) is a class of continuous-time no-regret learning dynamics in potential games, which includes gradient descent, log-barrier and replicator dynamics as special cases. The paper studies QRD's ability to select socially optimal equilibria and shows that QRD dynamics have unbounded gaps between average case performance and Price of Anarchy analysis even in larger settings.
https://openreview.net/forum?id=34STseLBrQ
Compressor summary: The paper studies how the dimension of the latent space affects the performance of DeepSets, a neural network architecture for set representation, and proposes new embeddings that require less dimension.
https://openreview.net/forum?id=33XGfHLtZg
Compressor summary: The paper presents an improved conformal prediction method that can control the expected value of any monotone loss function and is applicable to various tasks and settings.
https://openreview.net/forum?id=327tbF3S65
Compressor summary: The paper proposes a new model that generates adaptive positional embeddings for synthesizing implicit neural representations (INRs) across various domains, improving the quality of generation over existing methods.
https://openreview.net/forum?id=31IOmrnoP4
Compressor summary: Repelling random walks is a new method that helps explore graphs more efficiently by inducing correlations between walker trajectories, improving statistical estimators without bias.
https://openreview.net/forum?id=30aSE3FB3L
Compressor summary: The paper presents new neural network architectures on Riemannian manifolds, such as SPD and Grassmann manifolds, for computer vision tasks like human action recognition and node classification.
https://openreview.net/forum?id=30N3bNAiw3
Compressor summary: The authors propose a new method for contrastive analysis using mutual information terms to learn semantically expressive representations and assess its performance on various datasets.
https://openreview.net/forum?id=2tVHNRZuCs
Compressor summary: The paper proposes an implicit self-improvement framework for large language models that learns from human preference data to improve response quality without relying on explicit rubrics.
https://openreview.net/forum?id=2pvECsmld3
Compressor summary: SparseFormer is a new method that mimics human's sparse visual recognition by using few tokens to represent images, achieving performance similar to existing models with lower computational costs.
https://openreview.net/forum?id=2oWRumm67L
Compressor summary: Light-MILPopt is a lightweight optimization framework that uses a small-scale optimizer and dataset to solve large-scale MILPs more efficiently than existing methods.
https://openreview.net/forum?id=2msbbX3ydD
Compressor summary: Ferret is a new language model that can understand and describe different shapes and regions in images, using a novel hybrid representation and a large dataset called GRIT.
https://openreview.net/forum?id=2lDQLiH1W4
Compressor summary: Instant3D is a novel text-to-3D method that uses a diffusion model for fast and high-quality 3D asset generation from text prompts in two stages: text-to-image and image-to-NeRF.
https://openreview.net/forum?id=2inBuwTyL2
Compressor summary: The paper proposes a method for precise relative pose prediction in robot manipulation tasks that is provably SE(3)-equivariant, learned from few demonstrations, and generalizes across object variations.
https://openreview.net/forum?id=2iGiSHmeAN
Compressor summary: BroGNet combines SDEs and GNNs to learn Brownian dynamics from trajectories, preserving linear momentum conservation and outperforming baselines on various systems.
https://openreview.net/forum?id=2dnO3LLiJ1
Compressor summary: The paper identifies and solves artifacts in ViT networks' feature maps, improving their performance for self-supervised learning and visual processing.
https://openreview.net/forum?id=2dhxxIKhqz
Compressor summary: The paper proposes a technique to convert neural networks into function space, enabling scalability, retention of prior knowledge, and efficient incorporation of new data in sequential learning paradigms.
https://openreview.net/forum?id=2cRzmWXK9N
Compressor summary: This paper introduces $f$-DPO, a generalized method to optimize large language models for human preference alignment using various divergence constraints, which improves alignment performance and generation diversity.
https://openreview.net/forum?id=2Y5kBPtU0o
Compressor summary: MEND is a method that learns to condense demonstrations without retraining, improving in-context learning performance and reducing computational overhead for large language models.
https://openreview.net/forum?id=2XkTz7gdpc
Compressor summary: The paper proposes a method for generating large graphs by progressively expanding a single node, using denoising diffusion to add nodes and edges in a localized manner while achieving high expressivity and computational efficiency.
https://openreview.net/forum?id=2XBBumBGeP
Compressor summary: This paper proposes a new normalizing flows framework for generating realistic noise images using estimated camera settings, which improves noise modeling and enhances image denoising performance.
https://openreview.net/forum?id=2V1Z0Jdmss
Compressor summary: The text explores over-memorization, a type of overfitting that affects deep neural networks' generalization capacity, and proposes a framework to prevent it by distracting the network from high-confidence natural patterns.
https://openreview.net/forum?id=2UnCj3jeao
Compressor summary: The text introduces a new method to improve neural Monge map estimators by incorporating unbalanced optimal transport, enhancing their performance in unpaired domain translation tasks like image and cellular data.
https://openreview.net/forum?id=2SwHngthig
Compressor summary: TAO is a general approach that uses pre-trained Transformers to learn offline opponent models and adapt to unknown fixed policies in competitive environments.
https://openreview.net/forum?id=2Rwq6c3tvr
Compressor summary: The paper proposes a method to identify data contamination in large language models using guided instruction and classification techniques.
https://openreview.net/forum?id=2Q8TZWAHv4
Compressor summary: The paper proposes Graph Output Attribution (GOAt), a method to explain Graph Neural Networks (GNNs) by attributing graph outputs to input features using scalar products, achieving faithful, discriminative, and stable explanations.
https://openreview.net/forum?id=2Oiee202rd
Compressor summary: The paper proposes a two-step zero-shot image classification method called PerceptionCLIP that infers contextual attributes from an image and uses them for object classification, improving performance and reducing reliance on spurious features.
https://openreview.net/forum?id=2NpAw2QJBY
Compressor summary: The paper presents a novel neural architecture that enhances large neighborhood search for multi-agent path finding by using convolution and attention to represent subproblems.
https://openreview.net/forum?id=2Mo7v69otj
Compressor summary: The paper proposes a Category theory approach to handle shifts and imbalances in covariates for overparameterized DNN models in data pooling tasks.
https://openreview.net/forum?id=2JF8mJRJ7M
Compressor summary: The paper proposes a new method, Lipsum-FT, for fine-tuning pre-trained vision-language models that preserves their robustness against distribution shifts while improving performance on image classification tasks.
https://openreview.net/forum?id=2HJRwwbV3G
Compressor summary: The authors question the Knowledge Neuron Thesis, which claims that large language models recall facts from a training corpus through MLP weights resembling key-value memory, and suggest looking into other aspects of the models to better understand knowledge representation.
https://openreview.net/forum?id=2DbVeuoa6a
Compressor summary: Neural Spectral Methods is a technique for solving PDEs using neural networks that learns mappings between spectral coefficients and achieves faster and more accurate results than previous methods.
https://openreview.net/forum?id=28L2FCtMWq
Compressor summary: Ground-A-Video is a training-free video editing method that uses cross-frame gated attention and other techniques to accurately edit multiple attributes of videos in a consistent way.
https://openreview.net/forum?id=26XphugOcS
Compressor summary: The paper proposes a zero-shot method to transfer continuous prompts between NLP models by encoding them into relative space and searching for target prompts.
https://openreview.net/forum?id=23b9KSNQTX
Compressor summary: RETSim is a robust deep learning model for finding similar texts, outperforming existing methods on several tasks and introducing a new benchmark for evaluation.
https://openreview.net/forum?id=22pyNMuIoa
Compressor summary: PromptAgent is an automated system that uses strategic planning and error feedback to discover expert-level prompts for large language models, outperforming existing methods on 12 tasks across three domains.
https://openreview.net/forum?id=22OTbutug9
Compressor summary: RA-DIT is a method that fine-tunes language models to use external data for better performance on knowledge-intensive tasks, improving over existing approaches by a significant margin.
https://openreview.net/forum?id=1vrS1zwekw
Compressor summary: MUFFIN is a new instruction-following dataset curation scheme for large language models that diversifies tasks with input facets, improving their performance across various benchmarks.
https://openreview.net/forum?id=1vmSEVL19f
Compressor summary: DRaFT is a method to improve diffusion models using reward gradients that works well for various reward functions and enhances image quality.
https://openreview.net/forum?id=1vDArHJ68h
Compressor summary: The Recall to Imagine (R2I) method improves the performance and speed of model-based reinforcement learning agents by integrating state space models for better long-term memory and credit assignment.
https://openreview.net/forum?id=1vCnDyQkjg
Compressor summary: UniHSI is a framework for human-scene interaction using language commands, which translates tasks into contact sequences and executes them uniformly.
https://openreview.net/forum?id=1tZbq88f27
Compressor summary: MiniGPT-4 shows that aligning visual features with an advanced language model can achieve multi-modal abilities similar to GPT-4, like image description and website creation from sketches.
https://openreview.net/forum?id=1pSL2cXWoz
Compressor summary: The paper proposes a new OOD detection method based on Bregman divergence that uses an optimal norm coefficient to measure the distance between data distributions and achieves state-of-the-art performance.
https://openreview.net/forum?id=1oqedRt6Z7
Compressor summary: The paper introduces convolutional deep kernel machines with novel techniques, achieving state-of-the-art performance for kernel methods on image datasets.
https://openreview.net/forum?id=1op5YGZu8X
Compressor summary: The paper explains why adversarial training can harm the robustness of deep neural networks and proposes a new method to improve it for infinite-width DNNs.
https://openreview.net/forum?id=1oijHJBRsT
Compressor summary: The paper introduces instruction backtranslation, a method to improve language models by generating and selecting instruction prompts for web documents and finetuning them.
https://openreview.net/forum?id=1ndDmZdT4g
Compressor summary: Dynamic Sparse No Training (DSNT) is a training-free approach for fine-tuning large language models that updates sparse models without backpropagation or weight updates, improving performance at high sparsity levels.
https://openreview.net/forum?id=1mjsP8RYAw
Compressor summary: The paper proposes SFAVEL, an unsupervised framework for fact verification using pre-trained language models and a novel contrastive loss function that improves alignment quality and semantic relationships.
https://openreview.net/forum?id=1mNFsbvo2P
Compressor summary: The paper proposes a Bayesian model to estimate risk for patients who were tested or not tested for disease, incorporating domain constraints from health settings to improve parameter inference and cancer risk prediction.
https://openreview.net/forum?id=1k4yZbbDqX
Compressor summary: The paper presents InstaFlow, a one-step diffusion-based text-to-image generator using Rectified Flow and reflow to achieve high image quality at ultra-fast inference speed with low computational cost.
https://openreview.net/forum?id=1jbh2e0b2K
Compressor summary: This paper analyzes why multitask finetuning improves the performance of foundation models on target tasks with limited labeled samples, and proposes a practical task selection algorithm.
https://openreview.net/forum?id=1ii8idH4tH
Compressor summary: The study proposes a new Byzantine robust algorithm called Momentum Screening for nonconvex federated learning problems with adaptive hyperparameters and improved optimization error rate in certain conditions.
https://openreview.net/forum?id=1hsVvgW0rU
Compressor summary: The paper proposes a new feedback model for learning in hard reinforcement learning problems, which allows collecting additional observations from latent states without observing them directly.
https://openreview.net/forum?id=1d2cLKeNgY
Compressor summary: The paper presents a ReSimAD method that converts domain knowledge into invariant representations and simulates target domains using 3D meshes, enabling zero-shot perception in autonomous driving across different domains.
https://openreview.net/forum?id=1bbPQShCT2
Compressor summary: I-PHYRE is a framework that tests agents' abilities to reason about physics, plan multiple steps, and intervene in real time during dynamic events.
https://openreview.net/forum?id=1bAUywYJTU
Compressor summary: The paper proposes a method to improve 3D content creation by aligning the 3D optimization process with the sampling process of diffusion models using prioritized timestep sampling.
https://openreview.net/forum?id=1YPfmglNRU
Compressor summary: This paper argues that considering the type of expertise in decision-makers can help design and select methods for estimating treatment effects, as it influences their performance and can be predicted from data.
https://openreview.net/forum?id=1YO4EE3SPB
Compressor summary: The paper proposes a variational approach to approximate the posterior distribution in diffusion models for inverse problems, using denoisers with different structural constraints and SNR-based weights, resulting in improved image restoration tasks.
https://openreview.net/forum?id=1Wi0Ys33Nm
Compressor summary: The paper extends a proof about neural network convergence to different weight distributions and shows how to find the Edge of Chaos for better training.
https://openreview.net/forum?id=1VeQ6VBbev
Compressor summary: Dynamical policy gradient combines dynamic programming and policy gradient methods to improve convergence bounds and exploit structure in finite time horizon problems, such as optimal stopping or language model training.
https://openreview.net/forum?id=1SbkubNdbW
Compressor summary: Label smoothing can increase or decrease model privacy depending on the method used; negative factor smoothing enhances resilience against information extraction attacks.
https://openreview.net/forum?id=1SIBN5Xyw7
Compressor summary: The paper introduces Meta-SpikeFormer, a versatile and energy-efficient Transformer-based spiking neural network that outperforms existing CNN-based spiking neural networks on various vision tasks.
https://openreview.net/forum?id=1RrOtCmuKr
Compressor summary: The paper proposes a method to reduce the memory footprint of deep neural networks by jointly learning codebooks and weight mappings, enabling efficient compression and loading on commodity devices.
https://openreview.net/forum?id=1RE0H6mU7M
Compressor summary: The paper proposes a new model-based meta-RL method that achieves better sample efficiency and returns on common and challenging high-dimensional domains.
https://openreview.net/forum?id=1PXEY7ofFX
Compressor summary: Bespoke solvers improve sampling quality for diffusion models by optimizing custom ODE solvers tailored to the model's equations.
https://openreview.net/forum?id=1OfAO2mes1
Compressor summary: The paper proposes a novel method to automatically identify and separate backdoored data in machine learning systems without needing clean data or manual thresholds.
https://openreview.net/forum?id=1NHgmKqOzZ
Compressor summary: Progressive dataset distillation creates multiple small sets of synthetic images at different phases of training to reduce the time and memory requirement of training deep networks on large datasets.
https://openreview.net/forum?id=1JtTPYBKqt
Compressor summary: The authors propose a new problem called Neural Architecture Retrieval that uses motif-based graph representation learning to find similar neural architectures efficiently and automatically.
https://openreview.net/forum?id=1JbsdayvhO
Compressor summary: The paper presents a diffusion model that can reconstruct and generate realistic 3D scenes from 2D images using a novel neural scene representation, without needing additional supervision or regularizers.
https://openreview.net/forum?id=1CPta0bfN2
Compressor summary: The text proposes a method that uses sparse matrix factorization to approximate cross-encoder similarity scores and perform $k$-nearest neighbor search, achieving better recall than existing methods while being more efficient.
https://openreview.net/forum?id=1CK45cqkEh
Compressor summary: The paper proposes UOL, a novel clustering algorithm for ordered data, which groups objects into ordered clusters, constructs an embedding space, and estimates ranks via nearest neighbor search.
https://openreview.net/forum?id=1BuWv9poWz
Compressor summary: The paper proposes a gradient normalization scaling method to improve adversarial attacks on Vision Transformers by exploiting their attention points in the frequency domain.
https://openreview.net/forum?id=1BmveEMNbG
Compressor summary: The paper discusses challenges in reasoning on knowledge graphs, proposes a new dataset and method using fuzzy logic to improve reasoning capabilities for complex queries.
https://openreview.net/forum?id=17pVDnpwwl
Compressor summary: Depth-muP is a method for scaling deep neural networks that improves feature learning and diversity while preventing performance degradation.
https://openreview.net/forum?id=14rn7HpKVk
Compressor summary: SALMONN is a multimodal AI model that combines text and audio processing to achieve competitive performance on various speech and audio tasks and discover new emergent abilities.
https://openreview.net/forum?id=0w42S2Gp70
Compressor summary: This paper proposes a robust perceptual similarity metric called LipSim that uses 1-Lipschitz neural networks to protect against adversarial attacks and provide guarantees.
https://openreview.net/forum?id=0uI5415ry7
Compressor summary: The paper studies a simple linearized shallow Transformer model to better understand and improve Transformer training.
https://openreview.net/forum?id=0tWTxYYPnW
Compressor summary: The text discusses how human feedback in preference learning contains hidden context, which can lead to counter-intuitive results and vulnerabilities in methods like reinforcement learning from human feedback (RLHF). It introduces distributional preference learning (DPL) as a solution to address these issues.
https://openreview.net/forum?id=0t1O8ziRZp
Compressor summary: This study introduces RGLS, a parameter that adapts pre-trained agents' recommendations for logic synthesis, improving circuit quality and reducing runtime significantly.
https://openreview.net/forum?id=0kWd8SJq8d
Compressor summary: The paper proposes a new method to estimate Mutual Information using score-based diffusion models that outperforms existing methods in accuracy and passes self-consistency tests.
https://openreview.net/forum?id=0jsfesDZDq
Compressor summary: The paper proposes a method for creating sparse recurrent spiking neural networks (RSNNs) from untrained models, using graph sparsification and Lyapunov exponents, which improves computational efficiency and task performance compared to traditional pruning methods.
https://openreview.net/forum?id=0jHkUDyEO9
Compressor summary: Magic123 is a method that generates high-quality 3D meshes from single images using neural networks and 2D and 3D priors.
https://openreview.net/forum?id=0j9ZDzMPqr
Compressor summary: The paper proposes a method to generate counterfactual explanations for unsupervised node representation learning, helping to understand the underlying subgraphs that affect the node's behavior in downstream tasks.
https://openreview.net/forum?id=0i6Z9N5MLY
Compressor summary: The paper proposes variance reduction methods for solving finite-sum monotone inclusion problems with improved oracle complexity and provides a near-optimal complexity bound.
https://openreview.net/forum?id=0gTW5JUFTW
Compressor summary: The paper introduces a new method for understanding road scenes and planning driving routes, using improved detectors for lanes and traffic elements and a simple neural network pipeline for reasoning their topology relationships.
https://openreview.net/forum?id=0d1gQI114C
Compressor summary: LiDAR-PTQ is a novel post-training quantization method for 3D lidar detection that achieves state-of-the-art performance and speedup while being cost-effective compared to other methods.
https://openreview.net/forum?id=0bMmZ3fkCk
Compressor summary: NEFTune is a simple augmentation that improves language model finetuning on various tasks and datasets.
https://openreview.net/forum?id=0akLDTFR9x
Compressor summary: The paper proposes a new time series representation learning method based on temporal difference contrastive predictive coding, which improves goal-conditioned reinforcement learning performance in less data and more noisy settings.
https://openreview.net/forum?id=0aR1s9YxoL
Compressor summary: This paper explores how to maintain plasticity in visual reinforcement learning and proposes a new method called Adaptive RR that adjusts reuse frequency based on the critic's plasticity level.
https://openreview.net/forum?id=0VBsoluxR2
Compressor summary: The text introduces MOFDiff, a model that generates new metal-organic frameworks (MOFs) with unique structures and improved performance for carbon capture using diffusion and graph neural networks.
https://openreview.net/forum?id=0V5TVt9bk0
Compressor summary: Q-Bench is a benchmark to assess MLLMs' low-level visual perception, description, and quality assessment abilities using LLVisionQA, LLDescribe, and quantifiable quality scores datasets.
https://openreview.net/forum?id=0QAzIMq32X
Compressor summary: Inner classifier-free guidance (ICFG) improves the balance between fidelity and diversity in conditional diffusion models when the condition is continuous and has a specific structure, by providing an alternative perspective on classifier-free guidance.
https://openreview.net/forum?id=0Nui91LBQS
Compressor summary: SEED is an image tokenizer that enables text and images to be processed interchangeably within a unified autoregressive Transformer, improving multimodal comprehension and generation in large language models.
https://openreview.net/forum?id=0JsRZEGZ7L
Compressor summary: Latent Topology Inference (LTI) with Differentiable Cell Complex Module (DCM) improves Graph Neural Networks (GNNs) by learning higher-order cell complexes for multi-way data interactions, achieving better results especially when no input graph is given.
https://openreview.net/forum?id=0H6DFoZZXZ
Compressor summary: The paper proposes a Language to Control Diffusion model (LCD) that uses natural language instructions to plan and control generalist agents efficiently and effectively across high-dimensional inputs, long horizons, and novel tasks.
https://openreview.net/forum?id=0BqyZSWfzo
Compressor summary: The paper proposes a "one-shot" approach to audit or estimate differentially private algorithms' privacy loss efficiently during model training without additional assumptions or knowledge.
https://openreview.net/forum?id=09xFexjhqE
Compressor summary: This paper proposes AutoLoRa, a method that disentangles and automates Robust Fine-Tuning to improve adversarial robustness in downstream applications using a low-rank branch and heuristic strategies for learning rate scheduling.
https://openreview.net/forum?id=09iOdaeOzp
Compressor summary: Structured pruning helps create smaller yet powerful large language models (LLMs) from pre-trained, larger models by removing layers and dynamically adjusting data composition, outperforming state-of-the-art open-source models with much less compute cost.
https://openreview.net/forum?id=06lrITXVAx
Compressor summary: The paper proposes a dropout-based bilevel optimization model to address overfitting in machine learning tasks with limited datasets and analyzes its optimization properties.
https://openreview.net/forum?id=02f3mUtqnM
Compressor summary: The paper proposes a hybrid inference approach using a router that assigns queries to small or large NLP models based on difficulty and desired quality, saving cost without sacrificing quality.