Paper Archive

Browse and export your curated research paper collection

238
Archived Days
2368
Total Papers
7.9
Avg Score
9
Categories

Export Archive Data

Download your archived papers in various formats

JSON: Complete data with analysis • CSV: Tabular data for analysis • Markdown: Human-readable reports • BibTeX: Academic citations
Browse by Date

Papers for May 4, 2026

10 papers found

Jinpai Zhao, Nishant Panda, Yen Ting Lin, Eirik Valseth, Diane Oyen, Clint Dawson 5/1/2026 arxiv

machine learning

We introduce HyCOP, a modular framework that learns parametric PDE solution operators by composing simple modules (advection, diffusion, learned closures, boundary handling) in a query-conditioned way. Rather than learning a monolithic map, HyCOP learns a policy over short programs - which module to...

Keywords: HyCOP, hybrid composition, neural operators, PDEs, modularity, interpretability, OOD generalization, scientific machine learning

Sailesh Panda, Pritam Kadasi, Abhishek Upperwal, Mayank Singh 5/1/2026 arxiv

machine learning

Large language models (LLMs) often achieve strong performance on reasoning benchmarks, but final-answer accuracy alone does not show whether they faithfully execute the procedure specified in a prompt. We study this question through a controlled diagnostic benchmark for procedural execution, where m...

Keywords: LLMs, procedural execution, diagnostic benchmark, arithmetic algorithms, faithfulness, generation errors, model evaluation, hallucination

Siyuan Huang, Xiaoye Qu, Yafu Li, Tong Zhu, Zefeng He, Muxin Fu, Daizong Liu, Wei-Long Zheng, Yu Cheng 5/1/2026 arxiv

machine learning

While autoregressive Large Vision-Language Models (LVLMs) demonstrate remarkable proficiency in multimodal tasks, they face a "Visual Signal Dilution" phenomenon, where the accumulation of textual history expands the attention partition function, causing visual attention to decay inversely with gene...

Keywords: Persistent Visual Memory, PVM, Visual Signal Dilution, LVLMs, Qwen3-VL, attention decay, feed-forward network

Yan Fang, Mengcheng Lan, Zilong Huang, Weixian Lei, Yunqing Zhao, Yujie Zhong, Yingchen Yu, Qi She, Yao Zhao, Yunchao Wei 5/1/2026 arxiv

machine learning

In this paper, we present \textbf{Gen}erative \textbf{L}anguage-\textbf{I}mage \textbf{P}re-training (GenLIP), a minimalist generative pretraining framework for Vision Transformers (ViTs) designed for multimodal large language models (MLLMs). To better align vision encoders with the autoregressive n...

Keywords: ViT, GenLIP, generative pretraining, vision encoder, MLLM, language modeling objective, Recap-DataComp-1B, multi-resolution

Ziyang Huang, Yi Cao, Ali K. Shargh, Jing Luo, Ruidong Mei, Mohd Zaki, Zhan Liu, Wyatt Bunstine, William Jurayj, Somdatta Goswami, Tyrel McQueen, Michael Shields, Jaafar El-Awady, Paulette Clancy, Benjamin Van Durme, Nicholas Andrews, William Walden, Daniel Khashabi 5/1/2026 arxiv

machine learning

Large language models are increasingly deployed as autonomous coding agents and have achieved remarkably strong performance on software engineering benchmarks. However, it is unclear whether such success transfers to computational scientific workflows, where tasks require not only strong coding abil...

Keywords: AutoMat, LLM-based agents, computational materials science, reproducibility, benchmark, coding agents, toolchains, execution fragility

Pavlin G. Poličar, Andraž Pevcin, Blaž Zupan 5/1/2026 arxiv

machine learning

Generating diverse, readable statistical charts from tabular data remains challenging for LLMs, as many failures become apparent after rendering and are not detectable from data or code alone. Existing chart datasets also rarely provide fully aligned artifacts, such as executable code, dataset conte...

Keywords: LLM, chart generation, rendered-output validation, data visualization, multimodal LLMs, UCI datasets, question-answer pairs

Arunabh Srivastava, Mohammad A., Khojastepour, Srimat Chakradhar, Sennur Ulukus 5/1/2026 arxiv

natural language processing

Humans solve problems by executing targeted plans, yet large language models (LLMs) remain unreliable for structured workflow execution. We propose RunAgent, a multi-agent plan execution platform that interprets natural-language plans while enforcing stepwise execution through constraints and rubric...

Keywords: RunAgent, LLM, plan execution, constraint-guided, agentic language, code generation, tool use, rubrics

Alfredo Madrid-García, Miguel Rujas 5/1/2026 arxiv

machine learning

Background: Patient-facing medical chatbots based on retrieval-augmented generation (RAG) are increasingly promoted to deliver accessible, grounded health information. AI-assisted development lowers the barrier to building them, but they still demand rigorous security, privacy, and governance contro...

Keywords: RAG, retrieval-augmented generation, medical chatbot, privacy, security, system prompt, LLM-assisted audit, Chrome DevTools

Xihao Chen, Yangyang Guo, Roger Zimmermann 5/1/2026 arxiv

machine learning

Key-Value (KV) cache has become a de facto component of modern Large Vision-Language Models (LVLMs) for inference. While it enhances decoding efficiency in Large Language Models (LLMs), its direct adoption in LVLMs introduces substantial GPU memory overhead due to the large number of vision tokens p...

Keywords: LightKV, KV cache, LVLM, vision-token compression, cross-modality message passing, prompt-aware, prefill, GPU memory

Yinhao Xiao, Rongbo Xiao, Yihan Zhang 5/1/2026 arxiv

machine learning

Reliable spatial analysis in GIScience requires preserving coordinate semantics, topology, units, and geographic plausibility. Current LLM-based GIS systems generate fluent scripts but rarely enforce these geographic rules at scale. We present GeoContra, a verification and repair framework for LLM-d...

Keywords: GeoContra, geospatial contract, GIS, LLM, verification, repair, CRS, topology
Loading...

Preparing your export...