This page contains one-sentence summaries of cs.AI/ML/CV/CL papers announced on 2024-02-22 generated by the compressor, my personal LLM-based project.
http://arxiv.org/abs/2402.14017v1
Compressor summary: D-Flow is a method for controlling generation in Diffusion and Flow-Matching models by differentiating through the flow, achieving state-of-the-art results on various tasks.
http://arxiv.org/abs/2402.14016v1
Compressor summary: The study finds that large language models used for assessment are vulnerable to simple attacks that can manipulate their outputs, raising concerns about their reliability in real-world situations.
http://arxiv.org/abs/2402.14015v1
Compressor summary: Key points: - Machine Learning models face data integrity issues due to internet data sources - Corrective Machine Unlearning is the problem of mitigating the impact of unknown manipulations on a model, possibly knowing only a subset of affected samples - Existing unlearning methods require most of the manipulated data to be identified, but SSD achieves limited success with just a small portion Summary: The paper studies Corrective Machine Unlearning, a problem of removing the effect of unknown internet data manipulations on a ML model, and shows that SSD can partially achieve this goal with few samples.
http://arxiv.org/abs/2402.14013v1
Compressor summary: The paper proposes an algorithm that leverages users' limited attention to improve recommendation systems in digital health and EdTech, addressing the challenge of impulsive user choices.
http://arxiv.org/abs/2402.14009v1
Compressor summary: GINNs are neural networks that learn under geometric constraints, represent solutions as neural fields, and generate diverse solutions without training data using differentiable losses based on Morse theory.
http://arxiv.org/abs/2402.14008v1
Compressor summary: OlympiadBench is a new bilingual multimodal scientific benchmark that tests the advanced abilities of large language and multimodal models, revealing their limitations in physics reasoning.
http://arxiv.org/abs/2402.14007v1
Compressor summary: The study introduces cross-lingual consistency in text watermarking, shows its lack in current methods, proposes a removal attack, and suggests a defense method.
http://arxiv.org/abs/2402.14002v1
Compressor summary: The paper discusses the difference between hallucinations and attention misdirection in large language models, proposing a method to minimize errors and unlock innovation potential in business settings.
http://arxiv.org/abs/2402.14000v1
Compressor summary: 3DPE is a fast and flexible tool for real-time face image editing using 3D geometry and text descriptions.
http://arxiv.org/abs/2402.13991v1
Compressor summary: The authors explore how pre-training sequence composition affects generalization and propose intra-document causal masking, BM25Chunk for better in-context learning and downstream tasks.
http://arxiv.org/abs/2402.13987v1
Compressor summary: NoisyGNNs defend GNNs against small adversarial perturbations by injecting noise into the model's architecture, improving robustness with minimal performance impact and high compatibility with existing techniques.
http://arxiv.org/abs/2402.13984v1
Compressor summary: StABlE Training improves the stability and accuracy of neural network interatomic potentials by combining supervised learning from quantum-mechanical energies and forces with reference system observables, and iteratively correcting instabilities using a Boltzmann Estimator.
http://arxiv.org/abs/2402.13979v1
Compressor summary: The paper proposes a neural network model for predicting the Atlantic Meridional Overturning Circulation (AMOC), an important climate factor, under various scenarios and uncertainty, challenging previous AMOC collapse predictions.
http://arxiv.org/abs/2402.13963v1
Compressor summary: Key points: - The paper develops an open-source multilingual language model for medicine (MMedLM 2) - It creates a new medical corpus (MMedC) and a benchmark (MMedBench) for multilingual LLMs in medicine - MMedLM 2 outperforms other open-source models and rivals GPT-4 on the benchmark Summary: The paper presents an open-source, multilingual language model for medicine, based on a new medical corpus and benchmark, that surpasses existing models and competes with GPT-4.
http://arxiv.org/abs/2402.13956v1
Compressor summary: This study investigates how neural language models learn entailment relations from text data and proposes a revised theory accounting for redundancy in natural language.
http://arxiv.org/abs/2402.13955v1
Compressor summary: The study presents BEE-NET, a deep neural network that uses environmental factors to recognize emotions from body language, outperforming the current state-of-the-art by 2.07% on the BoLD dataset.
http://arxiv.org/abs/2402.13954v1
Compressor summary: Key points: - The paper evaluates social biases encoded by transformer models trained with masked language modeling (MLM) objective - It uses proxy functions and iterative masking experiments to measure bias quality and preference for disadvantaged/advantaged groups - It compares bias estimations with other methods using two benchmark datasets and finds high religious and disability biases, low gender bias - It extends previous work by evaluating biases after re-training MLM under masked language modeling objective Summary: The paper assesses social biases in transformer models trained with MLM using novel measures and methods, finding varying bias levels across datasets and tasks.
http://arxiv.org/abs/2402.13950v1
Compressor summary: The paper examines how large language models reason step-by-step and introduces FRODO, a framework to improve their reasoning by generating correct intermediate steps and faithfully reasoning over them.
http://arxiv.org/abs/2402.13946v1
Compressor summary: The authors propose AttackGNN, a novel reinforcement learning agent that generates adversarial examples to fool graph neural network-based techniques for hardware security applications.
http://arxiv.org/abs/2402.13936v1
Compressor summary: The paper proposes a new image captioning model training strategy using ground truth captions to improve distinctiveness, fluency, and contrastive reward in reinforcement learning.
http://arxiv.org/abs/2402.13934v1
Compressor summary: This paper compares the reasoning abilities of two efficient Transformer variants (Sparse and Linear) using Chain-of-Thought prompts and finds their model size scales with problem size, but they can be more efficient for a specific class of Dynamic Programming problems.
http://arxiv.org/abs/2402.13932v1
Compressor summary: Visual prompting is a new method that improves tumor segmentation in cancer diagnosis using subtle input modifications, outperforming traditional methods with less fine-tuning.
http://arxiv.org/abs/2402.13930v1
Compressor summary: The paper proposes a new method for combining local guide policies with Reinforcement Learning algorithms, and shows it improves performance on various classical problems and safety-critical systems.
http://arxiv.org/abs/2402.13929v1
Compressor summary: The text introduces a diffusion distillation method for improving text-to-image generation using SDXL, with a focus on quality and mode coverage.
http://arxiv.org/abs/2402.13927v1
Compressor summary: The paper proposes a new algorithm called "delusional hedge" that extends the classic hedge algorithm to learn from both labeled and unlabeled information sources, and shows that humans behave similarly when learning from diverse opinions.
http://arxiv.org/abs/2402.13926v1
Compressor summary: Bait-and-Switch attacks can easily manipulate safe text generated by large language models into harmful narratives, posing a major challenge for developing reliable safety measures.
http://arxiv.org/abs/2402.13919v1
Compressor summary: The study uses GPT-3.5 and GPT-4 to generate feedback for improving the factual accuracy of clinical note summarization by AI systems, leveraging their expertise in clinical NLP tasks.
http://arxiv.org/abs/2402.13918v1
Compressor summary: Key points: - The text is about cloud segmentation from remote sensing imagery and its challenges and applications. - The paper evaluates seven semantic segmentation and detection algorithms for cloud identification and compares their performance. - The paper also investigates the impact of image type and spectral bands on model adaptability. - The paper aims to produce machine learning algorithms that can perform cloud segmentation with few spectral bands. - The paper uses Sentinel-2 and Landsat-8 imagery as datasets and provides a github link for reproduction. Summary: The text summarizes a paper that benchmarks seven cloud segmentation methods using remote sensing imagery, assesses their adaptability with different image types and spectral bands, and proposes algorithms that can segment clouds with only RGB and RGBN-IR bands. The paper provides a github link for reproduction and uses Sentinel-2 and Landsat-8 datasets.
http://arxiv.org/abs/2402.13917v1
Compressor summary: The study evaluates Llama2's machine translation performance across various languages and explores factors influencing its translation quality, suggesting that multilingual LLMs could be more effective than English-centric ones.
http://arxiv.org/abs/2402.13916v1
Compressor summary: The text introduces four machine learning models for wind power forecasting, evaluates them on a dataset from a wind park, and shows that a convolutional neural network with continuous learning performs best.
http://arxiv.org/abs/2402.13914v1
Compressor summary: The text introduces two cultures of XAI research, focusing on human-oriented explanations (BLUE) and model-oriented explanions (RED), with the latter being under-explored but crucial for AI safety.
http://arxiv.org/abs/2402.13911v1
Compressor summary: The study proposes a Physics Informed Machine Learning model that combines the strengths of both physics-based models and ML algorithms for more reliable hydrological predictions.
http://arxiv.org/abs/2402.13906v1
Compressor summary: The paper proposes a graph-based method to identify the typical structure of documents within a collection by capturing recurring topics across domains and languages.
http://arxiv.org/abs/2402.13904v1
Compressor summary: Key points: - The paper explores how to measure confidence of LLMs using multiple random generations and consistency measures - Consistency-based methods outperform existing post-hoc approaches - Confidence scores can enhance model performance Summary: The paper proposes a new way to calibrate the confidence of large language models using multiple random outputs and consistency metrics, which improves both accuracy and performance.
http://arxiv.org/abs/2402.13903v1
Compressor summary: The paper proposes a regularization technique for stochastic first-order methods that stabilizes optimization and provides performance guarantees for finding saddle points of convex-concave functions, especially in reinforcement learning.
http://arxiv.org/abs/2402.13901v1
Compressor summary: The paper studies convergence guarantees and improved acceleration for discrete-time denoising diffusion models, covering various distributions with different properties.
http://arxiv.org/abs/2402.13891v1
Compressor summary: The paper proposes iterated regularization to improve kernel methods for estimating the ratio of two probability densities and achieve fast error convergence rates in machine learning and statistics tasks.
http://arxiv.org/abs/2402.13887v1
Compressor summary: This paper questions the effectiveness of using output probabilities to evaluate large language models, particularly for multiple choice questions, as it finds that these methods do not accurately reflect how LLMs are used in real-world applications.
http://arxiv.org/abs/2402.13876v1
Compressor summary: The text introduces a new method called SPFNet that uses scene priors to improve super-resolution of depth images by reducing texture interference and enhancing edges.
http://arxiv.org/abs/2402.13874v1
Compressor summary: The paper introduces a sequential-aware method ($ exttt{Se}^2$) for selecting and organizing example sequences that leverage LLM feedback to improve in-context learning (ICL) performance, quality, and diversity across 23 NLP tasks.
http://arxiv.org/abs/2402.13870v1
Compressor summary: The paper proposes a new method called weak innovation autoencoder for generating probable future time series data, which outperforms existing methods in predicting volatile electricity prices.
http://arxiv.org/abs/2402.13871v1
Compressor summary: The paper proposes a fine-tuned DistilBERT model with Explainable-AI techniques for effective phishing email detection using preprocessing methods on a balanced dataset.
http://arxiv.org/abs/2402.13866v1
Compressor summary: Key points: - Large Language Models are good at natural language but not specialized domains like accounting - Kuaiji is an Accounting Large Language Model fine-tuned using Baichuan framework - Kuaiji uses CAtAcctQA, a large Chinese accounting dataset, and shows high accuracy and speed - Kuaiji is the first Chinese accounting LLM and validated in real-world scenarios Summary: Kuaiji is a tailored Accounting Large Language Model that uses a continuous pre-training and supervised fine-tuning process to adapt to specialized domains. It leverages CAtAcctQA, the first Chinese accounting dataset, and performs well in real-world accounting tasks.
http://arxiv.org/abs/2402.13857v1
Compressor summary: The paper presents new efficient and dimension-independent replicable algorithms for learning large-margin halfspaces that improve upon existing ones in terms of sample and runtime complexity.
http://arxiv.org/abs/2402.13852v1
Compressor summary: The novel neural control system monitors and adjusts insulin delivery in real-time to optimize glucose levels for diabetic individuals.
http://arxiv.org/abs/2402.13851v1
Compressor summary: VL-Trojan is a multimodal instruction backdoor attack that can effectively manipulate autoregressive VLMs by injecting poisoned samples with triggers in instructions or images, overcoming challenges such as frozen visual encoder and black-box attacks.
http://arxiv.org/abs/2402.13848v1
Compressor summary: Key points: - BEV maps are important for robotics, but existing algorithms have limitations - The new model can project any modality to BEV without depth information or supervision - The model is general and outperforms competing methods Summary: The paper presents a novel model that can map any first-person modality to BEV maps for robotics, overcoming the limitations of existing algorithms.
http://arxiv.org/abs/2402.13831v1
Compressor summary: MLXP is an open-source, simple, and lightweight experiment management tool for machine learning research that helps ensure reproducible results.
http://arxiv.org/abs/2402.13827v1
Compressor summary: The paper introduces a technique to reduce unnecessary 3D Gaussians in real-time for faster and high-quality rendering with 3D Gaussian splatting, and proposes an efficient hardware architecture to support it.
http://arxiv.org/abs/2402.13822v1
Compressor summary: The paper proposes a novel framework for NAS that addresses frequency and time resolution issues in TSC, achieving state-of-the-art results on four datasets.
http://arxiv.org/abs/2402.13821v1
Compressor summary: The paper studies Lipschitz continuous Configurable Markov Decision Processes (Conf-MDPs) and provides bounds on Wasserstein distance between stationary distributions and a new performance improvement lower bound.
http://arxiv.org/abs/2402.13820v1
Compressor summary: The paper proposes a method to learn and track motions using self-supervised representations that capture spatial-temporal patterns and enable safe action execution.
http://arxiv.org/abs/2402.13818v1
Compressor summary: The paper tests advanced NLP models' ability to detect dehumanizing language and finds they are promising but biased, lacking accuracy in identifying dehumanization for some groups.
http://arxiv.org/abs/2402.13816v1
Compressor summary: The paper presents a unified view of non-local methods for image denoising and proposes a new method called NL-Ridge that simplifies the process and achieves better performance.
http://arxiv.org/abs/2402.13812v1
Compressor summary: The study shows that using voice biomarkers and machine learning can accurately predict heart failure patients' 5-year mortality rates and improve their care.
http://arxiv.org/abs/2402.13810v1
Compressor summary: The paper derives a formula for how well Langevin dynamics (a sampling method) works near optimal solutions, depending on the properties of its preconditioning matrix, and shows its application in neural networks and comparison with other methods.
http://arxiv.org/abs/2402.13800v1
Compressor summary: The study analyses online information about Europe and migration, focusing on its geography, language, and dynamics, revealing a transnational imbalance in information flow and the role of football as a popular topic.
http://arxiv.org/abs/2402.13796v1
Compressor summary: The paper proposes a system to detect and monitor brick kiln pollution using satellite imagery and machine learning, which could help governments regulate them more effectively.
http://arxiv.org/abs/2402.13791v1
Compressor summary: This paper reviews explainable AI methods in Remote Sensing, identifying trends, challenges, insights, and future directions.
http://arxiv.org/abs/2402.13785v1
Compressor summary: The paper proposes a novel hierarchical method for controller design in MDPs using deep reinforcement learning and reactive synthesis, with PAC guarantees and sparse reward handling.
http://arxiv.org/abs/2402.13782v1
Compressor summary: Probabilistic logic programming integrates probabilistic models into logical languages, using algebraic methods for modeling, inference, and learning.
http://arxiv.org/abs/2402.13781v1
Compressor summary: ExDyna is a novel gradient sparsification scheme that reduces communication overhead in distributed training systems by efficiently managing workload balance, partition topology, and threshold scaling.
http://arxiv.org/abs/2402.13779v1
Compressor summary: REMO is a self-supervised learning framework for molecular representation learning that leverages atom-combination rules and chemical reactions to pre-train encoders, achieving better performance on various downstream tasks than existing methods.
http://arxiv.org/abs/2402.13778v1
Compressor summary: Key points: - Reinforcement learning based weakly supervised system for localisation - Novel reward definition using non-binarised classification probability from pre-trained binary classifier - Minimises human bias and labelling costs - Outperforms multi-instance learning and competes with fully-supervised learning on prostate cancer lesion localisation task Summary: The paper proposes a weakly supervised system that uses reinforcement learning and novel reward definition to localise regions of interest in images, such as prostate cancer lesions, with minimal human bias and labelling costs, achieving comparable performance to fully-supervised learning.
http://arxiv.org/abs/2402.13777v1
Compressor summary: The text is a systematic review of deep generative models for offline policy learning in reinforcement learning and imitation learning domains.
http://arxiv.org/abs/2402.13765v1
Compressor summary: The paper proposes a new accuracy-preserving calibration method for deep neural networks using the Concrete distribution and shows its effectiveness on benchmarks.
http://arxiv.org/abs/2402.13764v1
Compressor summary: The paper presents a new benchmark, \shortname, to measure the critique abilities of Large Language Models (LLMs) in four dimensions using nine diverse tasks.
http://arxiv.org/abs/2402.13758v1
Compressor summary: This paper introduces TreatFact, a dataset of LLM-generated summaries for clinical texts, and benchmarks 11 LLMs for factual consistency evaluation across news and clinical domains, revealing open-source LLMs' limitations and challenges in clinical summary assessment.
http://arxiv.org/abs/2402.13756v1
Compressor summary: The authors propose a novel vision-based neural network for relative pose estimation between nano-drones using a low-resolution camera and a low-power chip, achieving significant improvements in accuracy and endurance compared to previous methods.
http://arxiv.org/abs/2402.13753v1
Compressor summary: LongRoPE extends the context window of LLMs to 2048k tokens using an efficient search for non-uniformities, a progressive extension strategy, and readjustments on shorter lengths.
http://arxiv.org/abs/2402.13752v1
Compressor summary: The paper reviews and tests various electricity load forecasting techniques for optimizing consumption in communities with renewable energy sources and storage, using AI models and weather forecasts.
http://arxiv.org/abs/2402.13744v1
Compressor summary: The text describes the research on Neural Algorithmic Reasoning (NAR), which combines neural networks and algorithms to enable machines to reason logically and learn from data.
http://arxiv.org/abs/2402.13741v1
Compressor summary: The text proposes a table generation prompt for relational triple extraction (RTE) task to improve in-context learning (ICL) performance by incorporating structured information and selecting semantically relevant samples.
http://arxiv.org/abs/2402.13740v1
Compressor summary: The paper proposes a framework for automating natural language to Corpus Query Language translation, using large language models and advanced evaluation metrics.
http://arxiv.org/abs/2402.13737v1
Compressor summary: Key points: - Diffusion models are good for image generation and introduced to precipitation nowcasting - SRNDiff is a diffusion model based on historical data that uses conditional UNet networks for accurate predictions - SRNDiff outperforms GANs in accuracy, stability, and efficiency, and generates high-quality samples Summary: SRNDiff is a diffusion model that uses conditional UNet networks to predict short-term precipitation nowcasting from historical data, achieving better accuracy, stability, and efficiency than GANs.
http://arxiv.org/abs/2402.13731v1
Compressor summary: The study defines degenerate knowledge neurons in pre-trained language models and proposes methods to analyze and form them for better factual knowledge storage.
http://arxiv.org/abs/2402.13729v1
Compressor summary: The HVDM model uses a hybrid autoencoder to capture spatio-temporal dependencies in videos and generate high-quality synthetic videos with fine structures and details.
http://arxiv.org/abs/2402.13728v1
Compressor summary: Deep Neural Collapse occurs mainly due to deep feature learning with the average gradient outer product, which is related to the singular structure of neural network weights.
http://arxiv.org/abs/2402.13725v1
Compressor summary: The paper presents a new way to train sparse Hopfield networks that links sparsity, memory retrieval, and loss margins, and applies it to structured Hopfield networks for pattern association tasks.
http://arxiv.org/abs/2402.13722v1
Compressor summary: The authors propose adaptive masking methods to improve Aspect-Based Sentiment Analysis by removing irrelevant tokens based on context, achieving better results on online review datasets.
http://arxiv.org/abs/2402.13720v1
Compressor summary: Ouroboros improves the efficiency and effectiveness of speculative decoding by using a phrase candidate pool from the LLM's verification process to generate better drafts.
http://arxiv.org/abs/2402.13718v1
Compressor summary: $\infty$Bench is a new benchmark for large language models that tests their ability to process and reason over long contexts, with average data lengths exceeding 100K tokens in both English and Chinese.
http://arxiv.org/abs/2402.13717v1
Compressor summary: Neeko is a framework that uses a dynamic low-rank adapter strategy to enable efficient imitation of multiple characters in open-domain dialogue, improving user interaction experiences.
http://arxiv.org/abs/2402.13711v1
Compressor summary: The paper proposes a new graph continual learning method, DSLR, that improves rehearsal-based approaches by considering class representativeness and diversity, as well as ensuring informative neighbors for replayed nodes to reduce catastrophic forgetting.
http://arxiv.org/abs/2402.13709v1
Compressor summary: The paper proposes a new measure called SaGE to evaluate the moral consistency of LLMs in conversational systems by using "Rules of Thumb" extracted from their responses to moral questions.
http://arxiv.org/abs/2402.13703v1
Compressor summary: The study shows that instruction-tuning multilingual LLMs on parallel data improves cross-lingual performance and challenges the Superficial Alignment Hypothesis.
http://arxiv.org/abs/2402.13700v1
Compressor summary: The paper analyzes existing robust aggregators for collaborative machine learning and shows that they are either ineffective or too restrictive to ensure privacy, safety, and learning.
http://arxiv.org/abs/2402.13699v1
Compressor summary: The text introduces a synthetic data-based technique that uses Explainable Boosting Machines (EBMs) to create explainable features for image data in quantum information science without sacrificing accuracy.
http://arxiv.org/abs/2402.13697v1
Compressor summary: CONCAT is a method to improve zero-shot panoptic segmentation by aligning semantic queries with visual CLS tokens and training a generator to synthesize fine-grained vision queries for unseen categories.
http://arxiv.org/abs/2402.13693v1
Compressor summary: The study creates a large Chinese Multimodal Named Entity Recognition dataset using Weibo posts and images, showing improved performance with image integration and cross-lingual learning.
http://arxiv.org/abs/2402.13671v1
Compressor summary: The paper describes a method for detecting machine-generated text across multiple languages and domains, using fine-tuned language models and statistical metrics, which ranked fourth in a competition.
http://arxiv.org/abs/2402.13669v1
Compressor summary: The paper proposes Self-Distillation Fine-Tuning (SDFT), a method that helps large language models maintain their general abilities while adapting to specific tasks by using a distilled dataset generated by the model itself.
http://arxiv.org/abs/2402.13667v1
Compressor summary: The Genetic Copy Optimization Framework enhances the efficiency and engagement of marketing copy creation using large language models and a modified crossover operator.
http://arxiv.org/abs/2402.13655v1
Compressor summary: The paper proposes a regularization method for updating regression trees that balances predictability and empirical stability by weighting data points based on their uncertainty in the initial model.
http://arxiv.org/abs/2402.13653v1
Compressor summary: The authors introduce a new task called zero-shot Protein Question Answering (PQA) for scientific enquiry, provide a large dataset and benchmarks, and present a multi-modal framework named Pika that achieves state-of-the-art performance.
http://arxiv.org/abs/2402.13651v1
Compressor summary: The paper evaluates the robustness of deep classifiers for radar data processing and proposes training methods to reduce overfitting and improve generalization.
http://arxiv.org/abs/2402.13647v1
Compressor summary: This paper proposes four ways to combine attention masking and LLMs for unsupervised text style transfer, improving style strength, content preservation, and text fluency.
http://arxiv.org/abs/2402.13643v1
Compressor summary: The paper proposes CAM, a novel approach for scene text recognition that uses class-aware glyph masks to suppress background noise and align features for improved performance.
http://arxiv.org/abs/2402.13641v1
Compressor summary: The paper proposes FlexHB, a new multi-fidelity Bayesian Optimization method that improves efficiency and performance on Hyperparameter Optimization tasks by integrating fine-grained fidelity and FlexBand frameworks.
http://arxiv.org/abs/2402.13636v1
Compressor summary: The paper evaluates gender-profession bias in various vision-language models by using a synthetic dataset of blurred gender distinctions across professional actions.
http://arxiv.org/abs/2402.13635v1
Compressor summary: Key points: - The text discusses the importance of data quality for trustworthy AI in medicine and proposes a framework for evaluating it. - Data quality affects aspects such as fairness, robustness, interpretability and ethical implications. - The framework is based on a systematic review of 62 studies and covers 15 dimensions. Summary: The authors propose a data quality framework for medical AI applications that helps to ensure trustworthiness by addressing issues such as fairness, robustness, interpretability and ethics.
http://arxiv.org/abs/2402.13631v1
Compressor summary: The paper proposes a novel shadow detection approach that focuses on low-intensity regions to learn better discriminative features and outperforms existing methods.
http://arxiv.org/abs/2402.13630v1
Compressor summary: The UniGraph framework trains a graph foundation model to generalize across diverse domains using Text-Attributed Graphs, a cascaded LM and GNN architecture, and large language models for tuning and prediction.
http://arxiv.org/abs/2402.13628v1
Compressor summary: The paper presents a new data-driven method for predicting room temperature in HVAC systems using k-means clustering, which simplifies the system model and reduces modeling time while maintaining prediction accuracy.
http://arxiv.org/abs/2402.13625v1
Compressor summary: The paper proposes a framework that uses both text and images to improve language models' commonsense abilities.
http://arxiv.org/abs/2402.13623v1
Compressor summary: FLAME is a novel approach that uses large language models and few-shot prompting to automatically expand taxonomies in low-resource environments, improving accuracy and overcoming challenges faced by traditional supervised methods.
http://arxiv.org/abs/2402.13616v1
Compressor summary: This paper proposes PGI and GELAN to address data loss in deep learning, improving performance on object detection tasks.
http://arxiv.org/abs/2402.13615v1
Compressor summary: This article examines the range of factual possibilities for the conjunction fallacy, finding that most research has focused on a narrow part of them, potentially biasing explanations of this cognitive phenomenon.
http://arxiv.org/abs/2402.13613v1
Compressor summary: The paper introduces a shared task on extracting comparative opinions from Vietnamese product reviews, with a human-annotated dataset and an evaluation metric.
http://arxiv.org/abs/2402.13610v1
Compressor summary: The authors propose using large generative models to develop automated systems for end-to-end data-driven discovery, but argue that current LGMs are not yet sufficient and suggest integrating them with feedback mechanisms for better results.
http://arxiv.org/abs/2402.13607v1
Compressor summary: The paper introduces CODIS, a new benchmark to evaluate multimodal language models' ability to use textual context to improve their image understanding, which they find MLLMs perform poorly on compared to humans.
http://arxiv.org/abs/2402.13606v1
Compressor summary: The paper investigates multi-lingual confidence estimation for large language models, introducing a new dataset, examining performance, and proposing a cross-lingual method that improves reliability.
http://arxiv.org/abs/2402.13605v1
Compressor summary: National Alignment is a measure for large language models to understand a country's culture and basic knowledge, with KorNAT being the first benchmark for South Korea.
http://arxiv.org/abs/2402.13604v1
Compressor summary: OccCANINE is a tool that automates the process of classifying occupations into HISCO codes, improving speed and accuracy and enabling new research possibilities.
http://arxiv.org/abs/2402.13602v1
Compressor summary: The study explores how large language models can improve autonomous driving by combining natural language text, arithmetic reasoning, and common sense in dynamic situations like low visibility.
http://arxiv.org/abs/2402.13598v1
Compressor summary: User-LLM is a novel framework that uses user embeddings to contextualize large language models for various natural language processing tasks, achieving significant performance gains and efficient computation.
http://arxiv.org/abs/2402.13593v1
Compressor summary: GLAME is a novel method that uses knowledge graphs to enhance large language models by tracking and incorporating changes in knowledge during editing, improving their generalization ability.
http://arxiv.org/abs/2402.13587v1
Compressor summary: The paper proposes ModICT, a method that uses in-context learning with a similar product sample and dynamic prompts to improve the accuracy and diversity of generating product descriptions from images and keywords.
http://arxiv.org/abs/2402.13584v1
Compressor summary: WinoViz evaluates language models' reasoning abilities on variant visual properties of objects across different contexts, finding that large vision-language models outperform language models and image-generating models.
http://arxiv.org/abs/2402.13583v1
Compressor summary: Our work introduces a systematic approach and metrics to evaluate the quality of long texts and presents LongWanjuan, a bilingual dataset for training language models on long-text tasks.
http://arxiv.org/abs/2402.13582v1
Compressor summary: The paper proposes GuanZero, a framework for AI agents to learn and master the challenging game of Guandan using Monte-Carlo methods and deep neural networks with a novel encoding scheme.
http://arxiv.org/abs/2402.13579v1
Compressor summary: CluDe is a novel clustering-based framework for depth completion that learns pixel-wise and continuous depth representation, reducing depth smearing around object boundaries.
http://arxiv.org/abs/2402.13578v1
Compressor summary: This paper introduces Transformer to improve gaze object prediction by using a Transformer-based object detector and a Transformer-based gaze autoencoder with an object-to-gaze cross-attention mechanism, achieving state-of-the-art results.
http://arxiv.org/abs/2402.13577v1
Compressor summary: BBA is a prompting method that improves LVLMs' multi-modal reasoning by aligning separate reasoning chains for visual and DSL representations, leading to better performance on geometry, chess, and molecular problems.
http://arxiv.org/abs/2402.13576v1
Compressor summary: The text introduces PREM, a model for video corpus moment retrieval, which captures partial relevance between query and video using multi-modal collaborative retrievers and focus-then-fuse moment localizers.
http://arxiv.org/abs/2402.13575v1
Compressor summary: The study presents FPA, a novel neural rendering method for realistic adversarial camouflage that simulates lighting and material variations using a generative approach and achieves high attack success rate and transferability.
http://arxiv.org/abs/2402.13573v1
Compressor summary: This paper proposes a method to speed up image diffusion models by using less attention, achieving better trade-off between efficiency and quality.
http://arxiv.org/abs/2402.13572v1
Compressor summary: The Algorithm Transformer (AlgoFormer) is a novel transformer block that achieves higher expressiveness in algorithm representation, enabling it to potentially outperform human-designed algorithms in challenging tasks.
http://arxiv.org/abs/2402.13571v1
Compressor summary: The paper introduces TransMuCoRes, a multilingual coreference resolution dataset in 31 South Asian languages created using translated texts and off-the-shelf tools, and evaluates two models on it.
http://arxiv.org/abs/2402.13567v1
Compressor summary: The paper proposes a unified metric, spot check equivalence, to compare and evaluate the performance of different mechanisms for eliciting high-quality data from human labelers for AI systems.
http://arxiv.org/abs/2402.13566v1
Compressor summary: EventFormer is a model that uses events within videos to retrieve specific moments using natural language queries and outperforms existing methods in video corpus moment retrieval tasks.
http://arxiv.org/abs/2402.13562v1
Compressor summary: The paper explores how multiple source languages in cross-lingual transfer affect multilingual language models' feature emphasis and suggests heuristics to identify effective language combinations.
http://arxiv.org/abs/2402.13561v1
Compressor summary: The paper proposes a Cognitive Visual-Language Mapper to improve large multimodal models for answering knowledge-based visual questions by aligning visuals with relevant knowledge.
http://arxiv.org/abs/2402.13556v1
Compressor summary: IGAP is a novel method that uses learnable prompts in the spectral space to align graphs with different structures and signals for inductive graph pre-training and fine-tuning.
http://arxiv.org/abs/2402.13551v1
Compressor summary: The paper proposes a graph called NARCO that captures coherence dependencies in narratives using retrospective questions and shows its usefulness for various tasks without human annotations.
http://arxiv.org/abs/2402.13550v1
Compressor summary: The text discusses using large language models (LLMs) like GPT-4 to advance negotiation research, analyze their multifaceted capabilities in different scenarios, and identify areas where they still struggle, such as subjective assessments and contextually appropriate responses.
http://arxiv.org/abs/2402.13548v1
Compressor summary: The text proposes a novel diffusion model called DiffPLF for probabilistic forecasting of electric vehicle charging load, which can handle volatile patterns and use covariates for better prediction.
http://arxiv.org/abs/2402.13547v1
Compressor summary: ActiveRAG is an improved Retrieval Augmented Generation framework that uses active learning to enhance Large Language Models' understanding of external knowledge, leading to better question-answering performance.
http://arxiv.org/abs/2402.13546v1
Compressor summary: The paper proposes an Interactive Visual Adapter (IVA) to improve the understanding of long videos by enhancing interaction with fine-grained visual elements within large language models.
http://arxiv.org/abs/2402.13545v1
Compressor summary: Key points: - The paper proposes a deep learning method for text tamper detection with three steps: feature assistance, audit point positioning, and tamper recognition. - The method simulates and augments data samples with artificial tamper data features in various scenarios. - The method achieves high accuracy, recall, and precision using a dual-path dual-stream recognition network and vlad. Summary: The paper presents a deep learning text tamper detection method that uses feature assistance, audit point positioning, and tamper recognition, and simulates data with artificial tamper features, achieving high performance.
http://arxiv.org/abs/2402.13542v1
Compressor summary: ARL2 is a technique that improves large language models by using them to label and score relevant evidence, leading to better retrieval-augmented generation.
http://arxiv.org/abs/2402.13537v1
Compressor summary: EffLoc is an efficient Vision Transformer for single-image camera relocalization using self-attention, sequential group attention, and feed-forward layers.
http://arxiv.org/abs/2402.13536v1
Compressor summary: The text explores using AI models like GPT-4V and DALL-E3 for semantic image compression that uses natural language to store concepts, achieving extremely low bitrates by ignoring structural information, and introduces an iterative reflection process to improve decoded images.
http://arxiv.org/abs/2402.13534v1
Compressor summary: The text proposes a two-stage curriculum learning framework for sequence labeling tasks that improves performance, training speed, and handles data heterogeneity.
http://arxiv.org/abs/2402.13533v1
Compressor summary: The paper presents efficient GPU-based methods to pretrain and finetune large language models for financial applications by using low-rank structures, replacing linear layers, and quantizing parameters, achieving significant speedup, compression, and accuracy improvements.
http://arxiv.org/abs/2402.13532v1
Compressor summary: The paper proposes a backdoor attack on dense passage retrieval systems using grammar errors to disseminate targeted misinformation.
http://arxiv.org/abs/2402.13531v1
Compressor summary: The paper analyzes differentially private gradient descent for linear regression, showing its accuracy and sample complexity match those of non-private methods and allowing confidence intervals for the empirical optimizer.
http://arxiv.org/abs/2402.13525v1
Compressor summary: The paper proposes MatchNAS, a method to adapt deep neural networks for mobile devices using both labelled and unlabelled data, without manual re-specialization or retraining.
http://arxiv.org/abs/2402.13524v1
Compressor summary: OMGEval is an open-source test set for assessing multilingual large language models' capabilities in various languages, covering general knowledge and logical reasoning.
http://arxiv.org/abs/2402.13522v1
Compressor summary: The paper introduces RecMind, a dataset for analyzing seeker's internal state in Japanese movie recommendation dialogues, and proposes a response generation framework based on it.
http://arxiv.org/abs/2402.13517v1
Compressor summary: The Round Trip Translation method defends large language models against social-engineered attacks by paraphrasing adversarial prompts and making them easier for LLMs to detect.
http://arxiv.org/abs/2402.13516v1
Compressor summary: ProSparse is a method that introduces ReLU activation function in large language models and uses progressive sparsity regularization to achieve high activation sparsity without compromising model performance or inference speed.
http://arxiv.org/abs/2402.13514v1
Compressor summary: The paper introduces CuQA, a dataset for compositional unknown questions in open-domain question-answering, and proposes a Self Divide-and-Conquer framework to improve efficiency and performance by adaptively calling different methods.
http://arxiv.org/abs/2402.13512v1
Compressor summary: This paper shows how self-attention in language models is equivalent to context-conditioned Markov chains and explores their properties and limitations in text generation.
http://arxiv.org/abs/2402.13510v1
Compressor summary: SealD-NeRF is a method that enables pixel-level editing in dynamic NeRF scenes by mapping edits to a specific timeframe and using a teacher-student approach.
http://arxiv.org/abs/2402.13505v1
Compressor summary: The SimPro framework is a novel, highly adaptable semi-supervised learning approach that does not rely on predefined assumptions about unlabeled data class distribution and improves pseudo-label quality using a refined EM algorithm and a Bayes classifier.
http://arxiv.org/abs/2402.13498v1
Compressor summary: The text explores using large language models (LLMs) for automated lay summarization (LS) of biomedical articles, proposing a novel framework and evaluation metrics that leverage LLMs' abilities to generate background knowledge and assess layness.
http://arxiv.org/abs/2402.13497v1
Compressor summary: The paper proposes a new Consistency Regularization method for Quantization-Aware Training, which uses vicinal data distribution information to improve generalization and achieve better results than existing QAT methods and Full Precision counterparts.
http://arxiv.org/abs/2402.13496v1
Compressor summary: HetTree is a novel heterogeneous tree graph neural network that models the hierarchy among metapaths, captures parent-children relationships using subtree attention, and matches node features and labels based on corresponding metapaths for accurate and scalable graph analysis.
http://arxiv.org/abs/2402.13494v1
Compressor summary: GradSafe detects unsafe prompts in LLMs by analyzing the gradients of safety-critical parameters, outperforming existing methods like Llama Guard.