YuyaoGe's Website
YuyaoGe's Website
About
Highlight Papers
Experience
Gallery
Projects
Posts
Light
Dark
Automatic
Home
Publications
Publications
Type
Conference paper
Journal article
Preprint
Date
2026
2025
2024
2023
Kimi K2.6: Advancing Open-Source Coding
We are open sourcing our latest model, Kimi K2.6, featuring state-of-the-art coding, long-horizon execution, and agent swarm …
Kimi Team
,
Yuyao Ge 葛钰峣
Cite
Blog
Hugging Face
Ask KIMI
SkillForge: Co-Evolving Skills and Agents via Dynamic Skill Lifecycles
An agentic RL method that evolves the skill library through a fitness-driven lifecycle, enabling skills and the model to co-evolve throughout training.
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yiwei Wang
,
Tianyu Liu
,
Yuchen He
,
Baolong Bi
,
Lingrui Mei
,
Jiayu Yao
,
Lizhe Chen
,
Jiafeng Guo
,
Xueqi Cheng
PDF
Prism-Δ: Differential Subspace Steering for Prompt Highlighting in Large Language Models
We propose PRISM-Δ, a differential subspace steering method for prompt highlighting that matches or exceeds the best existing method on 19 of 20 configurations with relative gains up to +10.6%, while halving the fluency cost.
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yiwei Wang
,
Tianyu Liu
,
Baolong Bi
,
Lingrui Mei
,
Jiayu Yao
,
Jiafeng Guo
,
Xueqi Cheng
Cite
PDF
arXiv
Hugging Face
GitHub
Project
Ask KIMI
Do Large Language Models Already Know the Answer Before They Finish Thinking?
Probing hidden states during reasoning reveals that LLMs already know the answer before finishing thinking. We detect overthinking via ‘jumps’ and intervene during inference to improve reasoning.
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yiwei Wang
,
Tianyu Liu
,
Lingrui Mei
,
Baolong Bi
,
Jiayuan Guo
,
Jiayu Yao
,
Jiafeng Guo
,
Xueqi Cheng
Code
PDF
PromptCD: Test-Time Behavior Enhancement via Polarity-Prompt Contrastive Decoding
Abstract: We present PromptCD, a test-time method for controlling LLM behavior without additional training. The approach creates paired positive and negative guiding prompts for a target behavior and contrasts model responses at the token-probability level for LLMs and through visual attention patterns for VLMs.
Baolong Bi
,
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yuchen He
,
Siqian Tong
,
Lizhe Chen
,
Lingrui Mei
,
Zehao Li
,
Yiwei Wang
,
Yujun Cai
,
Ming-Hsuan Yang
,
Xueqi Cheng
Cite
PDF
arXiv
Ask KIMI
Kimi K2.5: Visual Agentic Intelligence
We introduce Kimi K2.5, an open-source multimodal agentic model designed to advance general agentic intelligence. The model focuses on …
Kimi Team
,
Yuyao Ge 葛钰峣
Cite
PDF
arXiv
Blog
Hugging Face
GitHub
Ask KIMI
Modality-Grounded Contrastive Decoding for Cross-Modal Hallucination Mitigation
A training-free framework that softly anchors fused predictions toward the target modality’s own judgment when overshooting, mitigating cross-modal hallucination in multimodal LLMs.
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yiwei Wang
,
Tianyu Liu
,
Baolong Bi
,
Lingrui Mei
,
Jiayu Yao
,
Jiafeng Guo
,
Xueqi Cheng
PDF
Gated Differentiable Working Memory for Long-Context Language Modeling
Abstract: Long contexts challenge transformers as attention scores dilute across thousands of tokens and critical information is often lost in the middle. We reframe test-time adaptation as a budget-constrained memory consolidation problem and propose Gdwm (Gated Differentiable Working Memory), which introduces a write controller that estimates Contextual Utility, an information-theoretic measure of long-range contextual dependence, to allocate gradient steps efficiently.
Lingrui Mei
,
Shenghua Liu
,
Yiwei Wang
,
Yuyao Ge 葛钰峣
,
Baolong Bi
,
Jiayu Yao
,
Jun Wan
,
Ziling Yin
,
Jiafeng Guo
,
Xueqi Cheng
Cite
PDF
arXiv
Ask KIMI
Reward and Guidance through Rubrics: Promoting Exploration to Improve Multi-Domain Reasoning
Abstract: Reinforcement learning (RL) has shown great promise in enhancing LLM reasoning, but current approaches mainly focus on single domains with verifiable rewards. We propose RGR-GRPO, a rubric-driven RL framework for multi-domain reasoning that uses rubrics to provide fine-grained reward signals and offline guidance.
Baolong Bi
,
Shenghua Liu
,
Yiwei Wang
,
Siqian Tong
,
Lingrui Mei
,
Yuyao Ge 葛钰峣
,
Yilong Xu
,
Jiafeng Guo
,
Xueqi Cheng
Cite
PDF
arXiv
Ask KIMI
A Survey of Vibe Coding with Large Language Models
The advancement of large language models (LLMs) has catalyzed a paradigm shift from code generation assistance to autonomous coding …
Yuyao Ge 葛钰峣
,
Lingrui Mei
,
Zenghao Duan
,
Tianhao Li
,
Yujia Zheng
,
Yiwei Wang
,
Lexin Wang
,
Jiayu Yao
,
Tianyu Liu
,
Yujun Cai
,
Baolong Bi
,
Fangda Guo
,
Jiafeng Guo
,
Shenghua Liu
,
Xueqi Cheng
Cite
PDF
arXiv
Hugging Face
Ask KIMI
Focusing by Contrastive Attention: Enhancing VLMs' Visual Reasoning
Vision-Language Models (VLMs) have demonstrated remarkable success across diverse visual tasks, yet their performance degrades in …
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yiwei Wang
,
Lingrui Mei
,
Baolong Bi
,
Xuanshan Zhou
,
Jiayu Yao
,
Jiafeng Guo
,
Xueqi Cheng
Cite
PDF
arXiv
Ask KIMI
PaperWeekly
Are All Prompt Components Value-Neutral? Understanding the Heterogeneous Adversarial Robustness of Dissected Prompt in Large Language Models
Abstract: Prompt-based adversarial attacks have become an effective means to assess the robustness of large language models (LLMs). However, existing approaches often treat prompts as monolithic text, overlooking their structural heterogeneity-different prompt components contribute unequally to adversarial robustness.
Yujia Zheng
,
Tianhao Li
,
Haotian Huang
,
Tianyu Zeng
,
Jingyu Lu
,
Chuangxin Chu
,
Yuekai Huang
,
Ziyou Jiang
,
Qian Xiong
,
Yuyao Ge 葛钰峣
,
Mingyang Li
Cite
PDF
arXiv
x
Ask KIMI
Can Graph Descriptive Order Affect Solving Graph Problems with LLMs?
We present the first comprehensive analysis of how the order of graph descriptions impacts LLM performance, evaluating four graph description orders across six graph problems using six mainstream LLMs.
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Baolong Bi
,
Yiwei Wang
,
Lingrui Mei
,
Wenjie Feng
,
Lizhe Chen
,
Xueqi Cheng
Cite
Slides
Video
PDF
ACL Anthology
Ask KIMI
Poster
GitHub
Kimi K2: Open Agentic Intelligence
We introduce Kimi K2, a Mixture-of-Experts large language model with 32 billion activated parameters and 1 trillion total parameters. …
Kimi Team
,
Yuyao Ge 葛钰峣
Cite
PDF
arXiv
Blog
Hugging Face
GitHub
Ask KIMI
A Survey of Context Engineering for Large Language Models
The performance of Large Language Models (LLMs) is fundamentally determined by the contextual information provided during inference. …
Lingrui Mei
,
Jiayu Yao
,
Yuyao Ge 葛钰峣
,
Yiwei Wang
,
Baolong Bi
,
Yujun Cai
,
Jiazhi Liu
,
Mingyu Li
,
Zhong-Zhi Li
,
Duzhen Zhang
,
Chenlin Zhou
,
Jiayi Mao
,
Tianze Xia
,
Jiafeng Guo
,
Shenghua Liu
Cite
PDF
arXiv
Hugging Face
GitHub
Ask KIMI
Who is in the Spotlight: The Hidden Bias Undermining Multimodal Retrieval-Augmented Generation
Abstract: Multimodal Retrieval-Augmented Generation (RAG) systems have become essential in knowledge-intensive and open-domain tasks. As retrieval complexity increases, ensuring the robustness of these systems is critical. However, current RAG models are highly sensitive to the order in which evidence is presented, often resulting in unstable performance and biased reasoning, particularly as the number of retrieved items or modality diversity grows.
Jiayu Yao
,
Shenghua Liu
,
Yiwei Wang
,
Lingrui Mei
,
Baolong Bi
,
Yuyao Ge 葛钰峣
,
Zhecheng Li
,
Xueqi Cheng
Cite
PDF
arXiv
Ask KIMI
PIS:Linking Importance Sampling and Attention Mechanisms for Efficient Prompt Compression
Abstract: Large language models (LLMs) have achieved remarkable progress, demonstrating unprecedented capabilities across various natural language processing tasks. However, the high costs associated with such exceptional performance limit the widespread adoption of LLMs, highlighting the need for prompt compression.
Lizhe Chen
,
Binjia Zhou
,
Yuyao Ge 葛钰峣
,
Jiayi Chen
,
Shiguang Ni
Cite
PDF
arXiv
Ask KIMI
a1: Steep Test-time Scaling Law via Environment Augmented Generation
Large Language Models (LLMs) have made remarkable breakthroughs in reasoning, yet continue to struggle with hallucinations, logical …
Lingrui Mei
,
Shenghua Liu
,
Yiwei Wang
,
Baolong Bi
,
Yuyao Ge 葛钰峣
,
Jun Wan
,
Yurong Wu
,
Xueqi Cheng
Cite
PDF
Ask KIMI
Innate Reasoning is Not Enough : In-Context Learning Enhances Reasoning Large Language Models with Less Overthinking
We present the first comprehensive analysis of the impacts of CoT prompting on Reasoning LLMs, finding that one-shot CoT consistently enhances performance and reduces excessive reflections by approximately 90%.
Yuyao Ge 葛钰峣
,
Shenghua Liu
,
Yiwei Wang
,
Lingrui Mei
,
Lizhe Chen
,
Baolong Bi
,
Xueqi Cheng
Cite
PDF
arXiv
Hugging Face
Ask KIMI
Translating Words to Worlds: Zero-Shot Synthesis of 3D Terrain from Textual Descriptions Using Large Language Models
Abstract: The current research on text-guided 3D synthesis predominantly utilizes complex diffusion models, posing significant challenges in tasks like terrain generation. This study ventures into the direct synthesis of text-to-3D terrain in a zero-shot fashion, circumventing the need for diffusion models.
Guangzi Zhang
,
Lizhe Chen
,
Yu Zhang
,
Yan Liu
,
Yuyao Ge 葛钰峣
,
Xingquan Cai
Cite
PDF
Frequency-Importance Gaussian Splatting for Real-Time Lightweight Radiance Field Rendering
In this work, we propose a Dynamically Adaptive Density Control Strategy based on the degree of reconstruction of the background of the scene, which adaptive the spatial sample point generation strategy dynamically according to the training results and prevents the generation of redundant data in the model.
Lizhe Chen
,
Yan Hu
,
Yu Zhang
,
Yuyao Ge 葛钰峣
,
Haoyu Zhang
,
Xingquan Cai
Cite
PDF
Attack based on data : A novel perspective to attack sensitive points directly
Adversarial attack for time-series classification model is widely explored and many attack methods are proposed. But there is not a …
Yuyao Ge 葛钰峣
,
Zhongguo Yang
,
Lizhe Chen
,
Yiming Wang
,
Chengyang Li
Cite
Dataset
PDF
Page
Ask KIMI
Cite
×