PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.
Live Feed

AI Research Paper Feed

Discover cutting-edge research papers in AI and machine learning. Stay ahead with the latest breakthroughs, insights, and discoveries from top researchers worldwide.

22,178 Research Papers
Multi-Domain Coverage
Real-time Updates
arXivFeb 5, 2026

Parity, Sensitivity, and Transformers

Alexander Kozachinskiy, Tomasz Steifer et al.

TLDR: This paper presents a new construction of a transformer that can solve the PARITY problem using a single layer with practical features, and establishes a lower bound proving that a single-layer, single-head transformer cannot solve PARITY.

0764
ArXivFeb 5, 2026

A Systematic Evaluation of Large Language Models for PTSD Severity Estimation: The Role of Contextual Knowledge and Modeling Strategies

Panagiotis Kaliosis, Adithya V Ganesan et al.

TLDR: The study evaluates large language models for PTSD severity estimation, finding that detailed contextual knowledge and strategic modeling significantly enhance accuracy.

0166
ArXivFeb 5, 2026

Path Sampling for Rare Events Boosted by Machine Learning

Porhouy Minh, Sapna Sarupria

TLDR: AIMMD is a new algorithm that uses machine learning to improve the efficiency of transition path sampling for studying molecular processes.

093
ArXivFeb 5, 2026

PACE: Defying the Scaling Hypothesis of Exploration in Iterative Alignment for Mathematical Reasoning

Jun Rao, Zixiong Yu et al.

TLDR: PACE introduces a more efficient method for mathematical reasoning in language models by using minimal exploration, outperforming traditional methods with less computational cost.

0199
ArXivFeb 5, 2026

Private Prediction via Shrinkage

Chao Yan

TLDR: The paper presents a method to achieve differentially private prediction with reduced dependence on the number of queries, improving efficiency in streaming settings.

086
ArXivFeb 5, 2026

Fairness Under Group-Conditional Prior Probability Shift: Invariance, Drift, and Target-Aware Post-Processing

Amir Asiaee, Kaveh Aryan

TLDR: The paper addresses fairness in machine learning under group-conditional prior probability shift and introduces a method to maintain fairness when label prevalences change across demographic groups between training and deployment.

0153
ArXivFeb 5, 2026

Detecting Misbehaviors of Large Vision-Language Models by Evidential Uncertainty Quantification

Tao Huang, Rui Wang et al.

TLDR: The paper introduces Evidential Uncertainty Quantification (EUQ), a method to detect misbehaviors in large vision-language models by assessing internal conflicts and knowledge gaps, outperforming existing methods in identifying issues like hallucinations and adversarial vulnerabilities.

01,046
ArXivFeb 5, 2026

Faithful Bi-Directional Model Steering via Distribution Matching and Distributed Interchange Interventions

Yuntai Bao, Xuhong Zhang et al.

TLDR: The paper introduces Concept DAS (CDAS), a novel intervention-based model steering method that uses distribution matching to achieve more faithful and stable control compared to traditional preference-optimization methods.

074
ArXivFeb 5, 2026

Grammatical Error Correction Evaluation by Optimally Transporting Edit Representation

Takumi Goto, Yusuke Sakai et al.

TLDR: The paper introduces UOT-ERRANT, a new metric for evaluating grammatical error correction systems by optimally transporting edit vectors, showing improved performance and interpretability.

0247
ArXivFeb 5, 2026

Consistency-Preserving Concept Erasure via Unsafe-Safe Pairing and Directional Fisher-weighted Adaptation

Yongwoo Kim, Sungmin Cha et al.

TLDR: The paper introduces PAIR, a framework for concept erasure in text-to-image models that maintains semantic consistency by aligning unsafe concepts with safe alternatives.

070
arXivFeb 5, 2026

Clifford Kolmogorov-Arnold Networks

Matthias Wolff, Francesco Alesiani et al.

TLDR: The Clifford Kolmogorov-Arnold Network (ClKAN) is a new architecture for approximating functions in Clifford algebra spaces, utilizing Randomized Quasi Monte Carlo methods and novel batch normalization strategies for improved scalability and efficiency.

095
ArXivFeb 5, 2026

Accelerated Sequential Flow Matching: A Bayesian Filtering Perspective

Yinan Huang, Hans Hao-Hsun Hsu et al.

TLDR: This paper introduces Sequential Flow Matching, a Bayesian filtering approach that accelerates real-time sequential prediction by efficiently updating predictive distributions, reducing inference latency compared to traditional methods.

0173
ArXivFeb 5, 2026

Hinge Regression Tree: A Newton Method for Oblique Regression Tree Splitting

Hongyi Li, Han Lin et al.

TLDR: The Hinge Regression Tree (HRT) is a new method for creating oblique decision trees using a Newton method that improves split quality and convergence speed, outperforming traditional tree models.

0184
ArXivFeb 5, 2026

How Do Language Models Acquire Character-Level Information?

Soma Sato, Ryohei Sasano

TLDR: This paper investigates how language models acquire character-level information, identifying key factors related to tokenization and semantic associations.

0361
ArXivFeb 5, 2026

Logarithmic-time Schedules for Scaling Language Models with Momentum

Damien Ferbach, Courtney Paquette et al.

TLDR: ADANA, an optimizer with time-varying schedules for hyperparameters, improves large-scale language model training efficiency by up to 40% compared to AdamW.

02,859
ArXivFeb 5, 2026

A Short and Unified Convergence Analysis of the SAG, SAGA, and IAG Algorithms

Feng Zhu, Robert W. Heath et al.

TLDR: This paper presents a unified convergence analysis for the SAG, SAGA, and IAG algorithms, providing a simpler and more comprehensive understanding of their performance.

0925
ArXivFeb 5, 2026

Radon--Wasserstein Gradient Flows for Interacting-Particle Sampling in High Dimensions

Elias Hess-Childs, Dejan Slepčev et al.

TLDR: The paper introduces new Radon--Wasserstein gradient flows for efficient high-dimensional sampling using interacting particles with linear scaling costs.

02,496
ArXivFeb 5, 2026

VRIQ: Benchmarking and Analyzing Visual-Reasoning IQ of VLMs

Tina Khezresmaeilzadeh, Jike Zhong et al.

TLDR: VRIQ benchmark reveals that current Vision Language Models struggle with visual reasoning, primarily due to perception limitations.

077
ArXivFeb 5, 2026

OpenMAG: A Comprehensive Benchmark for Multimodal-Attributed Graph

Chenxi Wan, Xunkai Li et al.

TLDR: OpenMAG is a comprehensive benchmark designed to evaluate Multimodal-Attributed Graph (MAG) models across various domains, encoders, and tasks, addressing existing limitations in fair evaluation standards.

0335
ArXivFeb 5, 2026

OPUS: Towards Efficient and Principled Data Selection in Large Language Model Pre-training in Every Iteration

Shaobo Wang, Xuan Ouyang et al.

TLDR: OPUS is a dynamic data selection framework that improves the efficiency of large language model pre-training by selecting better tokens based on optimizer-induced updates, achieving significant performance gains with less data.

0160
Showing 1-20 of 22178 papers
per page
…
1 / 1109