PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.
Live Feed

AI Research Paper Feed

Discover cutting-edge research papers in AI and machine learning. Stay ahead with the latest breakthroughs, insights, and discoveries from top researchers worldwide.

22,578 Research Papers
Multi-Domain Coverage
Real-time Updates
ArXivFeb 19, 2026

ABCD: All Biases Come Disguised

Mateusz Nowak, Xavier Cadet et al.

TLDR: The paper introduces a bias-reduced evaluation protocol for LLMs in multiple-choice questions that improves robustness to answer permutations with minimal performance loss.

01
ArXivFeb 19, 2026

Mine and Refine: Optimizing Graded Relevance in E-commerce Search Retrieval

Jiaqi Xi, Raghav Saboo et al.

TLDR: The 'Mine and Refine' framework enhances e-commerce search retrieval by optimizing semantic text embeddings through a two-stage process involving contrastive training and policy-aligned fine-tuning.

01
ArXivFeb 19, 2026

Instructor-Aligned Knowledge Graphs for Personalized Learning

Abdulrahman AlRabah, Priyanka Kargupta et al.

TLDR: InstructKG is a framework that automatically constructs knowledge graphs from course materials to capture learning dependencies and aid personalized learning.

03,773
ArXivFeb 19, 2026

A Unified Framework for Locality in Scalable MARL

Sourav Chakraborty, Amit Kiran Rege et al.

TLDR: The paper presents a unified framework for addressing locality in scalable multi-agent reinforcement learning (MARL) by introducing a policy-dependent approach to the exponential decay property (EDP) of value functions.

03
ArXivFeb 19, 2026

Variational Grey-Box Dynamics Matching

Gurjeet Sangra Singh, Frantzeska Lavda et al.

TLDR: The paper introduces a novel grey-box method that integrates incomplete physics models into generative models for learning dynamics from observational data without relying on ground-truth physics parameters.

02
ArXivFeb 19, 2026

Phase-Aware Mixture of Experts for Agentic Reinforcement Learning

Shengtian Yang, Yu Li et al.

TLDR: The paper introduces Phase-Aware Mixture of Experts (PA-MoE) to enhance reinforcement learning by allowing expert specialization for complex tasks without being dominated by simpler tasks.

01,905
ArXivFeb 19, 2026

A Privacy by Design Framework for Large Language Model-Based Applications for Children

Diana Addae, Diana Rogachova et al.

TLDR: This paper proposes a Privacy-by-Design framework for developing AI applications for children that integrates privacy regulations to ensure data protection and legal compliance.

05,480
ArXivFeb 19, 2026

KLong: Training LLM Agent for Extremely Long-horizon Tasks

Yue Liu, Zhiyuan Hu et al.

TLDR: KLong is a new LLM agent designed to tackle long-horizon tasks using a novel training method combining trajectory-splitting SFT and progressive RL, outperforming existing models on various benchmarks.

05,462
ArXivFeb 19, 2026

Deeper detection limits in astronomical imaging using self-supervised spatiotemporal denoising

Yuduo Guo, Hao Zhang et al.

TLDR: ASTERIS, a self-supervised denoising algorithm, enhances astronomical imaging detection limits by leveraging spatiotemporal data, improving detection by 1 magnitude and identifying previously undetectable features in deep space images.

03,661
ArXivFeb 19, 2026

Universal Fine-Grained Symmetry Inference and Enforcement for Rigorous Crystal Structure Prediction

Shi Yin, Jinming Mu et al.

TLDR: This paper presents a novel approach to crystal structure prediction using large language models and constrained optimization to improve symmetry inference and enforce physical validity, achieving state-of-the-art results without relying on existing databases.

02,926
ArXivFeb 19, 2026

Deep Reinforcement Learning for Optimal Portfolio Allocation: A Comparative Study with Mean-Variance Optimization

Srijan Sood, Kassiani Papasotiriou et al.

TLDR: This study compares Deep Reinforcement Learning (DRL) and Mean-Variance Optimization (MVO) for portfolio allocation, showing DRL's strong performance across various financial metrics.

02,282
ArXivFeb 19, 2026

Transforming Behavioral Neuroscience Discovery with In-Context Learning and AI-Enhanced Tensor Methods

Paimon Goulart, Jordan Steinhauser et al.

TLDR: This paper presents an AI-enhanced pipeline using In-Context Learning and tensor methods to improve data analysis in behavioral neuroscience, particularly for studying fear generalization in mice, which can help understand PTSD.

01,486
ArXivFeb 19, 2026

Sign Lock-In: Randomly Initialized Weight Signs Persist and Bottleneck Sub-Bit Model Compression

Akira Sakai, Yuma Ichikawa

TLDR: The paper identifies that weight sign persistence is a bottleneck in sub-bit model compression and proposes methods to reduce sign flips while maintaining performance.

02,159
ArXivFeb 19, 2026

Retaining Suboptimal Actions to Follow Shifting Optima in Multi-Agent Reinforcement Learning

Yonghyeon Jo, Sunwoo Lee et al.

TLDR: The paper introduces Successive Sub-value Q-learning (S2Q), a method that improves adaptability in multi-agent reinforcement learning by retaining multiple high-value actions, outperforming existing algorithms.

01,880
ArXivFeb 19, 2026

Evaluating Chain-of-Thought Reasoning through Reusability and Verifiability

Shashank Aggarwal, Ram Vikas Mishra et al.

TLDR: This paper introduces reusability and verifiability as new metrics to evaluate the quality of Chain-of-Thought reasoning in multi-agent IR pipelines, revealing that these metrics are not correlated with traditional accuracy measures.

04,766
ArXivFeb 19, 2026

The Anxiety of Influence: Bloom Filters in Transformer Attention Heads

Peter Balogh

TLDR: Certain transformer attention heads in language models act as membership testers, identifying repeated tokens with high precision, similar to Bloom filters.

04,703
ArXivFeb 19, 2026

LORA-CRAFT: Cross-layer Rank Adaptation via Frozen Tucker Decomposition of Pre-trained Attention Weights

Kasun Dewage, Marianna Pensky et al.

TLDR: CRAFT is a parameter-efficient fine-tuning method using Tucker decomposition on pre-trained attention weights, achieving competitive performance with minimal adaptation parameters.

04,611
ArXivFeb 19, 2026

Fine-Grained Uncertainty Quantification for Long-Form Language Model Outputs: A Comparative Study

Dylan Bouchard, Mohit Singh Chauhan et al.

TLDR: This study introduces a taxonomy for fine-grained uncertainty quantification in long-form language model outputs, revealing that claim-level scoring and uncertainty-aware decoding improve factuality in generated content.

04,753
ArXivFeb 19, 2026

From Labor to Collaboration: A Methodological Experiment Using AI Agents to Augment Research Perspectives in Taiwan's Humanities and Social Sciences

Yi-Chih Huang

TLDR: This study proposes a collaborative AI workflow for humanities and social sciences research, using Taiwan's Claude.ai data to validate its feasibility and effectiveness.

03,286
ArXivFeb 19, 2026

Catastrophic Forgetting Resilient One-Shot Incremental Federated Learning

Obaidullah Zaland, Zulfiqar Ahmad Khan et al.

TLDR: This paper introduces One-Shot Incremental Federated Learning (OSI-FL), a framework that addresses communication overhead and catastrophic forgetting in federated learning by using category-specific embeddings and selective sample retention.

00
Showing 1-20 of 22578 papers
per page
…
1 / 1129