Parameter-Efficient LLM Adaptation
Developing low-rank and progressive adaptation strategies for large language models.
Xiequn Wang
Master's candidate, Computer Science, SUSTech
Machine Learning + Large Language Models
I am a Master's candidate in Computer Science at the Southern University of Science and Technology (SUSTech) in Shenzhen, China. My research spans machine learning and large language models, with a focus on parameter-efficient adaptation, continual learning, rigorous evaluation, and reproducible research.
My work explores efficient adaptation for large language models, continual learning, evaluation and benchmarks, and vision-language model alignment.
Developing low-rank and progressive adaptation strategies for large language models.
Proactive allocation methods that preserve performance across evolving tasks.
Normalizing soft prompts and aligning multimodal representations.
Probing model priors and reasoning with targeted diagnostics and benchmarks.
Corresponding author is marked with a dagger in the CV.
Notes on experiments, benchmarks, and the research process.
A short guide to staging low-rank adapters across a growing task list without leaking evaluation signal.
How I structure small, scriptable evaluations for one-token verification and chain-of-thought ablations.
Lessons learned when moving from static ranks to a proactive allocator across a task stream.
A simple workflow for capturing experiments without drowning in logs, branches, and reruns.
No posts match that search.
Current research themes connected to ongoing papers.
Mosaic shared adaptation strategies that improve efficiency without sacrificing quality.
Proactive low-rank allocation to sustain performance over long task sequences.
Normalizing soft prompts to stabilize and align vision-language models.
Recent teaching assistant roles at SUSTech.
Teaching Assistant, Spring 2024.
Teaching Assistant, Spring 2021 and 2024; Fall 2021 and 2023.
Teaching Assistant, Spring 2022.
I am open to research collaborations and internships in machine learning and large language models. Email me with a short note about your interests.
Based in Shenzhen, China.
Progressive adaptation sounds simple: start small, grow ranks as tasks arrive, then freeze. In practice, the schedule decides how much capacity each task gets and when you stop updating older adapters.
My default heuristic is to allocate a base rank for each new task, then increase only if validation plateaus across a rolling window. That prevents early tasks from hogging capacity and keeps evaluation clean.
I prefer a small harness over a heavy framework. A few scripts that standardize prompts, capture outputs, and score results are enough to keep reasoning experiments repeatable.
The key is to separate prompt templates, evaluation configs, and scoring logic so you can swap them independently. That is where most reproducibility breaks down.
Static ranks struggle when the task stream changes. The proactive allocator I use predicts which layers need capacity before performance collapses, then expands only those layers.
The best signal has been a combination of forgetting trends and gradient overlap. It is lightweight enough to run alongside training without slowing throughput.
I keep notebooks small and opinionated. Each notebook answers one question and logs the exact config, dataset version, and git commit used for a run.
The habit that helped most is to write a short summary after every session. It turns scraps of output into a narrative that is easy to revisit.