Paper Archive

Browse and export your curated research paper collection

33
Archived Days
330
Total Papers
7.8
Avg Score
7
Categories

Export Archive Data

Download your archived papers in various formats

JSON: Complete data with analysis • CSV: Tabular data for analysis • Markdown: Human-readable reports • BibTeX: Academic citations
Browse by Date

Papers for October 14, 2025

10 papers found

Takuya Nakabayashi, Navami Kairanda, Hideo Saito, Vladislav Golyanik 10/13/2025 arxiv

computer vision

Event cameras offer various advantages for novel view rendering compared to synchronously operating RGB cameras, and efficient event-based techniques supporting rigid scenes have been recently demonstrated in the literature. In the case of non-rigid objects, however, existing approaches additionally...

Keywords: event cameras, novel view synthesis, non-rigid, deformable 3D, Gaussian Splatting, monocular, image rendering

Chengqi Duan, Kaiyue Sun, Rongyao Fang, Manyuan Zhang, Yan Feng, Ying Luo, Yufang Liu, Ke Wang, Peng Pei, Xunliang Cai, Hongsheng Li, Yi Ma, Xihui Liu 10/13/2025 arxiv

multimodal learning

Recent advances in Large Language Models (LLMs) and Vision Language Models (VLMs) have shown significant progress in mathematical reasoning, yet they still face a critical bottleneck with problems requiring visual assistance, such as drawing auxiliary lines or plotting functions to solve the problem...

Keywords: CodePlot-CoT, Math-VR, image-to-code, visual reasoning, VLM, chain-of-thought, multimodal, mathematical reasoning

Tsung-Han Wu, Mihran Miroyan, David M. Chan, Trevor Darrell, Narges Norouzi, Joseph E. Gonzalez 10/13/2025 arxiv

machine learning

Large Reasoning Models (LRMs) excel at complex reasoning but are traditionally evaluated in static, "frozen world" settings: model responses are assumed to be instantaneous, and the context of a request is presumed to be immutable over the duration of the response. While generally true for short-ter...

Keywords: Large Reasoning Models, interruptibility, dynamic context, robustness, reasoning leakage, assistive programming, long-form reasoning

Edward Stevinson, Lucas Prieto, Melih Barsbey, Tolga Birdal 10/13/2025 arxiv

machine learning

Fundamental questions remain about when and why adversarial examples arise in neural networks, with competing views characterising them either as artifacts of the irregularities in the decision landscape or as products of sensitivity to non-robust input features. In this paper, we instead argue that...

Keywords: adversarial_examples, superposition, feature_interference, representational_compression, transferability, Vision_Transformer, CIFAR-10, robustness

Zhaochen Yu, Ling Yang, Jiaru Zou, Shuicheng Yan, Mengdi Wang 10/13/2025 arxiv

reinforcement learning

Recently, the emergence of agentic RL has showcased that RL could also effectively improve the agentic reasoning ability of LLMs, yet the key design principles and optimal practices remain unclear. In this work, we conduct a comprehensive and systematic investigation to demystify reinforcement learn...

Keywords: agentic RL, LLMs, tool-use, SFT, reward shaping, policy entropy, deliberative strategy, dataset

Wei Huang, Yi Ge, Shuai Yang, Yicheng Xiao, Huizi Mao, Yujun Lin, Hanrong Ye, Sifei Liu, Ka Chun Cheung, Hongxu Yin, Yao Lu, Xiaojuan Qi, Song Han, Yukang Chen 10/13/2025 arxiv

reinforcement learning

We propose QeRL, a Quantization-enhanced Reinforcement Learning framework for large language models (LLMs). While RL is essential for LLMs' reasoning capabilities, it is resource-intensive, requiring substantial GPU memory and long rollout durations. QeRL addresses these issues by combining NVFP4 qu...

Keywords: quantization, reinforcement_learning, LLMs, NVFP4, LoRA, Adaptive Quantization Noise, AQN, QLoRA

Lingfei Qian, Xueqing Peng, Yan Wang, Vincent Jim Zhang, Huan He, Hanley Smith, Yi Han, Yueru He, Haohang Li, Yupeng Cao, Yangyang Yu, Alejandro Lopez-Lira, Peng Lu, Jian-Yun Nie, Guojun Xiong, Jimin Huang, Sophia Ananiadou 10/13/2025 arxiv

machine learning

Although Large Language Model (LLM)-based agents are increasingly used in financial trading, it remains unclear whether they can reason and adapt in live markets, as most studies test models instead of agents, cover limited periods and assets, and rely on unverified data. To address these gaps, we i...

Keywords: Agent Market Arena, AMA, LLM agents, live trading, benchmark, verified data, expert-checked news, cryptocurrency

Arjun Sahney, Ram Gorthi, Cezary Łastowski, Javier Vega 10/13/2025 arxiv

machine learning

We present Operand Quant, a single-agent, IDE-based architecture for autonomous machine learning engineering (MLE). Operand Quant departs from conventional multi-agent orchestration frameworks by consolidating all MLE lifecycle stages -- exploration, modeling, experimentation, and deployment -- with...

Keywords: Operand Quant, autonomous MLE, single-agent, IDE-based, MLE-Benchmark 2025, linear non-blocking agent, AutoML, MLE lifecycle

Chenghao Xiao, Hou Pong Chan, Hao Zhang, Weiwen Xu, Mahani Aljunied, Yu Rong 10/13/2025 arxiv

machine learning

Recent multimodal embedding approaches leveraging multimodal large language models (MLLMs) fine-tuned with contrastive learning (CL) have shown promising results, yet the underlying reasons behind their superiority remain underexplored. This work argues that a crucial advantage of MLLM-based approac...

Keywords: LCO-Emb, MLLM, contrastive learning, cross-modal alignment, generative pretraining, GRSL, anisotropy, kernel similarity

Taira Tsuchiya 10/13/2025 arxiv

machine learning

In two-player zero-sum games, the learning dynamic based on optimistic Hedge achieves one of the best-known regret upper bounds among strongly-uncoupled learning dynamics. With an appropriately chosen learning rate, the social and individual regrets can be bounded by $O(\log(mn))$ in terms of the nu...

Keywords: optimistic Hedge, regret bounds, two-player zero-sum games, social regret, individual regret, last-iterate convergence, dynamic regret, lower bounds
Loading...

Preparing your export...