Junyou Su 苏军又

I'm currently a master’s student at Peking University Shenzhen Graduate School. My research focuses on domain-specific large language models (LLMs), data synthesis, and post-training methods.

Previously, I completed my Bachelor’s degree in Computer Science and Technology at the Southern University of Science and Technology (SUSTech), where I was recognized as an Outstanding Graduate (Top 5%). During my undergraduate studies, I conducted research under the supervision of Prof. Xuan Song at SUSTech, and collaborated with researchers from Tsinghua University and The University of Tokyo.

I am currently seeking research or internship opportunities

Last Update: November 2025


Education

Peking University

Shenzhen Graduate School, Academic Master’s Program
Research Interests: Data Synthesis, Domain-specific LLMs, Post-training Techniques
Sep 2025 – Jul 2028 (Expected)

Southern University of Science and Technology

B.Eng. in Computer Science and Technology
University Outstanding Graduate; GPA: 3.87/4.00 (Top 5%)
Sep 2021 – Jul 2025

Research Experience

PlanGPT Series (Large Models for Urban Planning)

Collaborative project at Peking University. PlanGPT 1.5 accepted by ACL Industry Track (Oral), and PlanGPT-VL accepted by EMNLP.

  • PlanGPT 1.0: Developed a customized large language model and an efficient retrieval system to assist urban planning.
  • PlanGPT 1.5: Extended the planning capability boundary based on version 1.0.
  • PlanGPT-VL: Proposed the PlanAnno-V framework for generating high-quality visual QA data in planning maps; introduced the Critical Point Thinking mechanism to reduce hallucination via structured verification.
  • PlanBench: Built benchmarks evaluating contextual reasoning, policy understanding, spatial logic, and value assessment in planning tasks.
  • PlanBench-V: Constructed benchmarks for national land-use planning map understanding.
Feb 2024 – Present

ASFT: Anchored Supervised Fine-Tuning (Model Post-Training)

Proposed a stable and generalizable post-training framework; submitted to ICLR.

  • Proposed the ASFT framework, integrating dynamic weighting with lightweight KL anchoring to mitigate training drift in Dynamic Fine-Tuning (DFT).
  • Unified DFT under the Reward-Weighted Regression theoretical framework and proved ASFT has a tighter lower bound and more stable convergence than standard SFT.
  • Demonstrated significant improvements across math reasoning, medical QA, and code generation with ~3% of full RL compute.
  • Introduced ASFT-LoRA, reducing VRAM usage by 50% while retaining performance gains.
Jul 2025 – Present

TAG-INSTRUCT (Data Synthesis)

Accepted by Findings of ACL 2025. Proposed TAG-INSTRUCT to enhance controllability of instruction complexity via structured semantic compression and difficulty-guided expansion.

  • Compresses instructions into compact tag spaces; expands via RL-guided complexity augmentation.
  • Shows superior controllability and stability across multiple benchmarks.
Jul 2024 – Dec 2024

FANNO (Data Synthesis)

Published in Findings of ACL 2025. FANNO is an open-source, end-to-end framework for automated LLM instruction data synthesis.

  • Leverages pre-filtering, seed generation, and UCB-based dynamic enhancement to synthesize high-quality data without proprietary models.
  • Introduced the “Think Different” prompting strategy and UCB-based dynamic weighting mechanism for diversity and complexity.
  • Generated 10K instruction data outperforming Alpaca-GPT4, WizardLM, etc., on AlpacaEval 2.0 and Arena-Hard.
Mar 2024 – Sep 2024

Other Works

  • IRSC (Benchmark)Paper Link: Accepted by NLPCC. Built multilingual RAG evaluation benchmark; proposed SSCI and RCCI; revealed cross-lingual embedding limitations.
  • Ship ETA Trajectory Prediction with LLMs: Applied fine-tuned LLMs for ETA prediction; accepted by SpatialDI.
  • Ship ETA Forecasting Enhanced by Maritime News: Designed an LLM-based maritime news agent; Outstanding Thesis.
  • Natural-Language Spatiotemporal Analysis Agent: With University of Tokyo, developed a Mobility Agent via LangChain with trajectory analysis tools.

Internship Experience

LocationMind Inc.

Tokyo, Japan (On-site + Remote)
Role: Research Fellow
Supervisors: Ryosuke Shibasaki, Zi Pei Fan

Core Work: Research on spatiotemporal analysis integrated with large language models.

Jul 2024 – Present

Behavioral and Spatial Intelligence Lab, Peking University

Shenzhen, China
Role: Research Assistant
Supervisor: Wenjia Zhang

Core Work: Assisted in developing the first large language model for urban planning in China.

Feb 2024 – Jul 2024

Project Experience

Beijing Urban Planning and Design Institute

Served as technical lead, developing an RAG-based planning knowledge QA model.

  • Built a large-scale knowledge base using OCR pipelines for unstructured planning documents.
  • Designed a hierarchical retrieval architecture for precise and efficient RAG generation.
  • Integrated speech input for improved user interaction and accessibility.
Mar 2025 – Present

China Academy of Urban Planning and Design (Shenzhen Branch)

Served as technical lead, developing a planning knowledge QA model based on RAG.

  • Constructed an institutional knowledge base from internal planning documents using LLM-driven sub-knowledge segmentation.
  • Implemented a hierarchical retrieval system for high-precision RAG generation.
  • Built a knowledge-based QA and intelligent review system integrating institutional and open-domain information.
Sep 2024 – Dec 2024

Other Projects

  • Shenzhen Marine Development Research Center: Participated in developing an RAG-based intelligent review agent for planning documents.

Skills

Technical Skills
  • Programming Languages: Java, Python > C++, C, SQL > Go, Vue, JavaScript > Verilog, MATLAB
  • LLM-related: Prompt Engineering, Fine-tuning, Agent Development, Retrieval-Augmented Generation, Multi-Agent Simulation, Reinforcement Learning, Deep-Research
  • Full-Stack: Experienced in frontend and backend. Representative projects: Othello Game, Course Management System, SUSTech Activity Center, StackOverflow Visualization.

Honors & Awards

  • 2nd Prize, SUSTech Freshman Scholarship — 2021
  • 3rd Prize, SUSTech Outstanding Student Scholarship — 2022, 2023, 2024
  • Provincial 3rd Prize, National College Mathematical Modeling Contest — 2023
  • University-level, SUSTech Outstanding Student Award — 2023, 2024
  • University-level, SUSTech Outstanding Graduate — 2025
  • University-level, SUSTech Outstanding Thesis Award — 2025