This page contains one-sentence summaries of cs.AI/ML/CV/CL papers announced on 2024-07-25 generated by the compressor, my personal LLM-based project.
http://arxiv.org/abs/2407.17470v1
Compressor summary: SV4D is a latent video diffusion model that generates consistent novel views for dynamic 3D objects from a reference video and optimizes an implicit 4D representation without SDS optimization.
http://arxiv.org/abs/2407.17469v1
Compressor summary: The paper introduces CouldAsk, a benchmark for evaluating question reformulation, and shows that current language models struggle to improve unanswerable questions.
http://arxiv.org/abs/2407.17468v1
Compressor summary: WildHallucinations is a benchmark that tests factuality of LLMs by generating and fact-checking information about real-world entities from user-chatbot conversations, revealing Hallucination patterns across domains.
http://arxiv.org/abs/2407.17467v1
Compressor summary: The paper proposes a critical mixture ratio (CMR) to balance the general and domain-specific knowledge of large language models during continual pre-training, improving their efficiency and effectiveness in specialized domains.
http://arxiv.org/abs/2407.17466v1
Compressor summary: The paper studies multi-objective reinforcement learning (MORL) with multiple reward functions, analyzes different optimization targets for finding Pareto optimal policies, and proposes efficient algorithms using Tchebycheff scalarization.
http://arxiv.org/abs/2407.17465v1
Compressor summary: u-$\mu$P combines $\mu$P and Unit Scaling to simplify and improve hyperparameter optimization for efficient training and low-precision computing.
http://arxiv.org/abs/2407.17459v1
Compressor summary: The paper investigates how demographic inference errors affect the fairness performance of various learning-to-rank models and suggests that fair re-ranking strategies are more robust to these errors than LTR-based methods.
http://arxiv.org/abs/2407.17458v1
Compressor summary: EuroCropsML is a new dataset for crop type classification in Europe using Sentinel-2 satellite images.
http://arxiv.org/abs/2407.17457v1
Compressor summary: CSCPR is a new algorithm for RGB-D indoor place recognition that uses global retrieval, reranking, and context clusters to handle noisy data and achieve better performance than existing models.
http://arxiv.org/abs/2407.17454v1
Compressor summary: The paper proposes a machine learning-based cycle for generating and choosing explanations in Explainable AI, using a taxonomy of explanation selection criteria derived from sociology and cognitive science.
http://arxiv.org/abs/2407.17453v1
Compressor summary: The authors propose VILA^2, a visual language model family that uses self-augmentation and specialist-augmentation to improve data quality and performance on various tasks.
http://arxiv.org/abs/2407.17449v1
Compressor summary: The paper proposes a new method to identify and reduce unwanted correlations in deep neural networks using anomaly detection and data augmentation.
http://arxiv.org/abs/2407.17447v1
Compressor summary: The authors improve existing algorithms for jailbreaking safety-tuned language models by using a new distillation-based approach that makes the attacks more powerful and fluent, achieving high success rates on various models.
http://arxiv.org/abs/2407.17442v1
Compressor summary: The paper proposes an Adaptive Hybrid-Memory-Fusion (AHMF) model that incorporates working memory and long-term memory to predict driver attention more human-like, using domain adaptation techniques for better performance.
http://arxiv.org/abs/2407.17438v1
Compressor summary: Key points: - Human image animation generates videos from a photo of a character - Existing approaches face challenges with training data, 2D motion, and camera motions - HumanVid is a new dataset that combines real-world and synthetic data with rich annotations - CamAnimate is a baseline model that shows state-of-the-art performance on HumanVid Summary: HumanVid introduces a large-scale dataset for human image animation, addressing the limitations of existing approaches by providing diverse and precise real-world and synthetic data with human and camera motion annotations.
http://arxiv.org/abs/2407.17437v1
Compressor summary: Nerva is a fast C++ neural network library that uses sparse matrix operations to reduce training time and memory usage while maintaining accuracy.
http://arxiv.org/abs/2407.17418v1
Compressor summary: 3D Gaussian Splatting is a technique that converts multi-view images into 3D representations for real-time rendering, and this survey reviews existing works, challenges, and opportunities in the field.
http://arxiv.org/abs/2407.17417v1
Compressor summary: This paper investigates how watermarking large language models can deter copyright violations, while also considering the impact on membership inference attacks and proposing an adaptive technique to improve detection.
http://arxiv.org/abs/2407.17412v1
Compressor summary: The paper proposes a novel algorithm, PASS, that leverages visual prompts to determine channel importance and achieve structural sparsity in neural networks, resulting in improved efficiency and performance.
http://arxiv.org/abs/2407.17409v1
Compressor summary: The paper introduces lanelet2_ml_converter, an extension to the HD map framework Lanelet2 that supports machine learning tasks and training using map data.
http://arxiv.org/abs/2407.17406v1
Compressor summary: The paper introduces Dependency Transformer Grammars (DTGs), a new type of Transformer model that uses dependency structures to improve generalization, and shows that it outperforms previous models based on constituency trees.
http://arxiv.org/abs/2407.17404v1
Compressor summary: The paper explores using large language models with game description grammar to improve automated game design with limited data.
http://arxiv.org/abs/2407.17399v1
Compressor summary: The paper proposes a method called Noise2VST that uses an off-the-shelf Gaussian denoiser and a variance-stabilizing transform to efficiently remove noise from real-world images without additional training data.
http://arxiv.org/abs/2407.17396v1
Compressor summary: The paper proposes a GNN architecture that treats node embeddings as epistemic states and shows it can achieve state-of-the-art results in reasoning with relational domains.
http://arxiv.org/abs/2407.17395v1
Compressor summary: The authors challenge the common assumption of data-generating probability distributions in machine learning and propose an alternative framework that focuses on finite populations for better modeling and theory.
http://arxiv.org/abs/2407.17390v1
Compressor summary: CovScore is a new method for evaluating extracted thematic title sets from documents using five metrics, which simplifies and speeds up the evaluation process and can be applied to relevant datasets like Holocaust survivor testimonies.
http://arxiv.org/abs/2407.17387v1
Compressor summary: PERSONA is a test bed that evaluates and improves language models' ability to align with diverse user values by generating synthetic personas and feedback.
http://arxiv.org/abs/2407.17383v1
Compressor summary: The research uses neural networks and BERT models to correct spelling errors in written text, especially in the Persian language.
http://arxiv.org/abs/2407.17379v1
Compressor summary: The text introduces a new multi-image relation association task (MMRA) to evaluate large visual language models' ability to perceive associative relations between multiple images and their details, finding that current models struggle with fine-grained tasks and spatial perception.
http://arxiv.org/abs/2407.17378v1
Compressor summary: The paper introduces PrevPredMap, a framework that uses previous predictions to create online vectorized HD maps, with two essential modules and a dual-mode strategy for robust performance.
http://arxiv.org/abs/2407.17377v1
Compressor summary:
http://arxiv.org/abs/2407.17365v1
Compressor summary:
http://arxiv.org/abs/2407.17361v1
Compressor summary:
http://arxiv.org/abs/2407.17356v1
Compressor summary:
http://arxiv.org/abs/2407.17353v1
Compressor summary:
http://arxiv.org/abs/2407.17349v1
Compressor summary:
http://arxiv.org/abs/2407.17344v1
Compressor summary:
http://arxiv.org/abs/2407.17339v1
Compressor summary:
http://arxiv.org/abs/2407.17336v1
Compressor summary:
http://arxiv.org/abs/2407.17333v1
Compressor summary:
http://arxiv.org/abs/2407.17331v1
Compressor summary:
http://arxiv.org/abs/2407.17328v1
Compressor summary:
http://arxiv.org/abs/2407.17310v1
Compressor summary:
http://arxiv.org/abs/2407.17303v1
Compressor summary:
http://arxiv.org/abs/2407.17284v1
Compressor summary:
http://arxiv.org/abs/2407.17272v1
Compressor summary:
http://arxiv.org/abs/2407.17267v1
Compressor summary:
http://arxiv.org/abs/2407.17246v1
Compressor summary:
http://arxiv.org/abs/2407.17230v1
Compressor summary:
http://arxiv.org/abs/2407.17229v1
Compressor summary:
http://arxiv.org/abs/2407.17227v1
Compressor summary:
http://arxiv.org/abs/2407.17226v1
Compressor summary:
http://arxiv.org/abs/2407.17219v1
Compressor summary:
http://arxiv.org/abs/2407.17213v1
Compressor summary:
http://arxiv.org/abs/2407.17209v1
Compressor summary:
http://arxiv.org/abs/2407.17206v1
Compressor summary:
http://arxiv.org/abs/2407.17197v1
Compressor summary:
http://arxiv.org/abs/2407.17193v1
Compressor summary:
http://arxiv.org/abs/2407.17182v1
Compressor summary:
http://arxiv.org/abs/2407.17174v1
Compressor summary:
http://arxiv.org/abs/2407.17170v1
Compressor summary:
http://arxiv.org/abs/2407.17164v1
Compressor summary:
http://arxiv.org/abs/2407.17165v1
Compressor summary:
http://arxiv.org/abs/2407.17163v1
Compressor summary:
http://arxiv.org/abs/2407.17162v1
Compressor summary:
http://arxiv.org/abs/2407.17160v1
Compressor summary:
http://arxiv.org/abs/2407.17157v1
Compressor summary:
http://arxiv.org/abs/2407.17156v1
Compressor summary: