YuyaoGe's Website
YuyaoGe's Website
About
Highlight Publications
Other Publications
Projects
Posts
Light
Dark
Automatic
Home
Tags
LLM
LLM
Kimi K2.5: Visual Agentic Intelligence
We introduce Kimi K2.5, an open-source multimodal agentic model designed to advance general agentic intelligence. The model focuses on …
Kimi Team
,
Yuyao Ge 葛钰峣
PDF
Cite
arXiv
Blog
Model
GitHub
Gated Differentiable Working Memory for Long-Context Language Modeling
Abstract: Long contexts challenge transformers as attention scores dilute across thousands of tokens and critical information is often lost in the middle. We reframe test-time adaptation as a budget-constrained memory consolidation problem and propose Gdwm (Gated Differentiable Working Memory), which introduces a write controller that estimates Contextual Utility, an information-theoretic measure of long-range contextual dependence, to allocate gradient steps efficiently.
Lingrui Mei
,
Shenghua Liu
,
Yiwei Wang
,
Yuyao Ge 葛钰峣
,
Baolong Bi
,
Jiayu Yao
,
Jun Wan
,
Ziling Yin
,
Jiafeng Guo
,
Xueqi Cheng
Jan 19, 2026
PDF
Cite
arXiv
Reward and Guidance through Rubrics: Promoting Exploration to Improve Multi-Domain Reasoning
Abstract: Reinforcement learning (RL) has shown great promise in enhancing LLM reasoning, but current approaches mainly focus on single domains with verifiable rewards. We propose RGR-GRPO, a rubric-driven RL framework for multi-domain reasoning that uses rubrics to provide fine-grained reward signals and offline guidance.
Baolong Bi
,
Shenghua Liu
,
Yiwei Wang
,
Siqian Tong
,
Lingrui Mei
,
Yuyao Ge 葛钰峣
,
Yilong Xu
,
Jiafeng Guo
,
Xueqi Cheng
Nov 15, 2025
PDF
Cite
arXiv
Awesome Vibe Coding
We introduce Vibe Coding as a formal discipline, formalizing it through a Constrained Markov Decision Process that captures the dynamic triadic relationship among human developers, software projects, and coding agents.
Yuyao Ge 葛钰峣
GitHub
Paper
Can Graph Descriptive Order Affect Solving Graph Problems with LLMs?
In this work, we propose a Dynamically Adaptive Density Control Strategy based on the degree of reconstruction of the background of the scene, which adaptive the spatial sample point generation strategy dynamically according to the training results and prevents the generation of redundant data in the model.
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Baolong Bi
,
Yiwei Wang
,
Lingrui Mei
,
Wenjie Feng
,
Lizhe Chen
,
Xueqi Cheng
PDF
Cite
Slides
Video
DOI
Paper
Poster
GitHub
Kimi K2: Open Agentic Intelligence
We introduce Kimi K2, a Mixture-of-Experts large language model with 32 billion activated parameters and 1 trillion total parameters. …
Kimi Team
,
Yuyao Ge 葛钰峣
PDF
Cite
arXiv
Blog
Model
GitHub
PIS:Linking Importance Sampling and Attention Mechanisms for Efficient Prompt Compression
Abstract: Large language models (LLMs) have achieved remarkable progress, demonstrating unprecedented capabilities across various natural language processing tasks. However, the high costs associated with such exceptional performance limit the widespread adoption of LLMs, highlighting the need for prompt compression.
Lizhe Chen
,
Binjia Zhou
,
Yuyao Ge 葛钰峣
,
Jiayi Chen
,
Shiguang Ni
Apr 23, 2025
PDF
Cite
DOI
arXiv
a1: Steep Test-time Scaling Law via Environment Augmented Generation
Large Language Models (LLMs) have made remarkable breakthroughs in reasoning, yet continue to struggle with hallucinations, logical …
Lingrui Mei
,
Shenghua Liu
,
Yiwei Wang
,
Baolong Bi
,
Yuyao Ge 葛钰峣
,
Jun Wan
,
Yurong Wu
,
Xueqi Cheng
Apr 20, 2025
PDF
Cite
DOI
Innate Reasoning is Not Enough : In-Context Learning Enhances Reasoning Large Language Models with Less Overthinking
TBD
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yiwei Wang
,
Lingrui Mei
,
Lizhe Chen
,
Baolong Bi
,
Xueqi Cheng
PDF
Cite
DOI
arXiv
Hugging Face
EMNLP2024论文分享 | Fewer is More:CoT示例要少而精
作者提出CoT-Influx方法,一种对CoT的示例和内容进行优化从而提高LLMs推理能力的方法,其核心思想是通过剪枝最大化有效信息的输入。
Yuyao Ge 葛钰峣
Oct 24, 2024
2 min read
论文分享
»
Cite
×