Jiachen Hu
Algorithm Developer at ByteDance. PhD of Peking University.
My name is Jiachen Hu (胡家琛), now working as an algorithm developer at ByteDance. I graduated from Peking University and received the PhD degree in 2025, where I am fortunate to be advised by Professor Liwei Wang, and spend wonderful times working with Chi Jin and Lihong Li remotely for past years. Before becoming a PhD candidate, I obtained my B.S. from Turing Class, Peking University.
I have broad interests in sample efficient reinforcement learning and online learning, especially the application-driven problems. In the past few years, my researches focused on statistically efficient bandits (e.g., multi-armed bandits, linear bandits), online exploration in structured MDPs/POMDPs, and understanding sim-to-real transfer. Please feel free to contact me if you are interested in my researches or having a chat with me!
Contact: nickh at pku.edu.cn
news
| Jul 2025 | Joining ByteDance as an algorithm developer! |
|---|---|
| May 2025 | One paper accepted at ICML 2025! |
| May 2024 | One paper accepted at ICML 2024! |
| May 2024 | One paper accepted at TQC 2024! |
| Jun 2023 | Will visit Princeton University for the next 6 months! |
| Jan 2023 | One paper accepted at ICLR 2023! |
selected publications
- PreprintNew Sphere Packings from the Antipode ConstructionIn arXiv preprint, 2025
We construct non-lattice sphere packings in dimensions 19, 20, 21, 23, 44, 45, and 47, demonstrating record densities that surpass all previously documented results in these dimensions. The construction applies the antipode method to suboptimal cross-sections of \(\Lambda_{24}\) and \(P_{48p}\).
- ICLRProvable Sim-to-real Transfer in Continuous Domain with Partial ObservationsIn International Conference on Learning Representations, 2023
We study sim-to-real transfer in continuous domains with partial observations, modeled by linear quadratic Gaussian (LQG) systems. We show that a popular robust adversarial training algorithm can learn a policy from simulation that is competitive to the optimal real-world policy, providing the first provable guarantee in this setting.
- ICLRUnderstanding Domain Randomization for Sim-to-real TransferIn International Conference on Learning Representations(Spotlight, top 6%) , 2022
We provide a theoretical framework for domain randomization, modeling the simulator as a set of MDPs with tunable parameters. We prove sharp bounds on the sim-to-real gap and show that successful transfer is achievable without any real-world training samples, highlighting the importance of history-dependent policies.
- ICMLNear-Optimal Representation Learning for Linear Bandits and Linear RLIn Proceedings of the 38th International Conference on Machine Learning, 2021
We study multi-task representation learning for linear bandits and episodic RL with linear value function approximation. Our algorithm MTLR-OFUL achieves \(\tilde{O}(M\sqrt{dkT} + d\sqrt{kMT})\) regret, significantly improving over the \(\tilde{O}(Md\sqrt{T})\) baseline, yielding the first theoretical characterization of multi-task representation learning benefits in RL exploration.
- ICLRDistributed Bandit Learning: Near-Optimal Regret with Efficient CommunicationIn International Conference on Learning Representations, 2020
We design communication protocols for distributed bandit learning with M agents under central coordination. For multi-armed bandits, we achieve near-optimal regret with only \(O(M\log(MK))\) communication cost — independent of the time horizon T and matching the lower bound up to a log factor.