I am a final-year Ph.D. student in Machine Learning at the Georgia Institute of Technology, advised by Prof. Faramarz Fekri.

My research focuses on improving long-horizon reasoning and decision-making in large language models. I study post-training and system design for LLM agents, with a focus on deliberate reasoning and planning, as well as scalable agent systems that support adaptive retrieval and long-context interaction.

πŸš€ Featured Projects

Long-Horizon LLM Agents

ICLR 2026 @ SPOT
sym

Scaling Search-Augmented Reasoning Agents via Adaptive Information Control
Siheng Xiong, Oguzhan Gungordu, Blair Johnson, James C. Kerce, Faramarz Fekri

Project

DeepControl is a post-training framework for search-augmented LLM agents that adaptively controls retrieval and expansion based on the agent’s reasoning state, improving long-horizon reasoning performance across diverse benchmarks.

Project 2025
sym

Evidence-Based Expert-Level Scientific Claim Verification
Siheng Xiong, Oguzhan Gungordu, Blair Johnson, Mika Okamoto, James C. Kerce, Faramarz Fekri

Project

DeepVerify is an agentic framework for expert-level scientific claim verification, combining search, tool use, and structured reasoning to ground decisions in retrieved evidence.

ICLR 2026
sym

Enhancing Language Model Reasoning with Structured Multi-Level Modeling
Siheng Xiong, Ali Payani, Faramarz Fekri

Project

MLR is a lightweight planner-executor framework for long-horizon reasoning, combining multi-level decomposition with scalable step-level supervision to improve both accuracy and training stability.

ACL 2025
sym

Deliberate Reasoning in Language Models as Structure-Aware Planning with an Accurate World Model
Siheng Xiong, Ali Payani, Yuan Yang, Faramarz Fekri

Project

SWAP frames multi-step reasoning as structure-aware planning, where a world model predicts structured future states to guide action selection.

Long-Context and Memory Systems

NeurIPS 2025 @ ER
sym

Long-Context Modeling with Dynamic Hierarchical Sparse Attention for On-Device LLMs
Siheng Xiong, Joe Zou, Faramarz Fekri, Yae Jee Cho

Project

DHSA is an efficient sparse attention framework for long-context LLM inference that reduces prefill cost and memory usage while preserving accuracy under tight device constraints.

Preprint 2024
sym

The Compressor-Retriever Architecture for Language Model OS
Yuan Yang, Siheng Xiong, Ehsan Shareghi, Faramarz Fekri

Project

Compressor-Retriever is a model-agnostic architecture for lifelong context management in LLM-based systems, using the base model’s forward pass to compress and retrieve memory while remaining fully differentiable.

πŸ“ Selected Publications

πŸ“ Conference Papers

πŸ“ Workshop Papers

πŸ“ Preprints

πŸŽ– Honors and Awards

  • China National Scholarship (Top 1% Rank)
  • China UHV Scholarship (Top 1% Rank)
  • Kaggle Santander Value Prediction Challenge, Silver Medal (Top 3.4% Rank)

πŸ“– Education

  • Georgia Institute of Technology, Ph.D. in Machine Learning
  • Shanghai Jiao Tong University, M.S. in Electrical and Computer Engineering
  • Xi’an Jiaotong University, B.S. in Electrical and Computer Engineering

πŸ’» Experience

  • 2025.05 - 2025.08, Applied Research Intern, Google, Sunnyvale, California
  • 2023.09 - 2024.04, Research Intern, Cisco Research, San Jose, California
  • 2020.05 - 2021.01, Research Student Assistant, Rutgers University (Mentor: Dimitris N. Metaxas), New Brunswick, New Jersey

πŸ“„ Services

  • Program Committee for AAAI
  • Conference Reviewer for NeurIPS, ICML, ICLR, ACL, EMNLP, NAACL, EACL, KDD, AAAI
  • Journal Reviewer for ACM TIST