Jiachen Hu

Algorithm Developer at ByteDance. PhD of Peking University.

myimg_formal.jpg

My name is Jiachen Hu (胡家琛), now working as an algorithm developer at ByteDance. I graduated from Peking University and received the PhD degree in 2025, where I am fortunate to be advised by Professor Liwei Wang, and spend wonderful times working with Chi Jin and Lihong Li remotely for past years. Before becoming a PhD candidate, I obtained my B.S. from Turing Class, Peking University.

I have broad interests in sample efficient reinforcement learning and online learning, especially the application-driven problems. In the past few years, my researches focused on statistically efficient bandits (e.g., multi-armed bandits, linear bandits), online exploration in structured MDPs/POMDPs, and understanding sim-to-real transfer. Please feel free to contact me if you are interested in my researches or having a chat with me!

Contact: nickh at pku.edu.cn

news

Jul 2025 Joining ByteDance as an algorithm developer!
May 2025 One paper accepted at ICML 2025!
May 2024 One paper accepted at ICML 2024!
May 2024 One paper accepted at TQC 2024!
Jun 2023 Will visit Princeton University for the next 6 months!
Jan 2023 One paper accepted at ICLR 2023!

selected publications

  1. Preprint
    New Sphere Packings from the Antipode Construction
    Ruitao Chen, Jiachen Hu, Binghui Li, and 2 more authors
    In arXiv preprint, 2025

    We construct non-lattice sphere packings in dimensions 19, 20, 21, 23, 44, 45, and 47, demonstrating record densities that surpass all previously documented results in these dimensions. The construction applies the antipode method to suboptimal cross-sections of \(\Lambda_{24}\) and \(P_{48p}\).

  2. ICLR
    Provable Sim-to-real Transfer in Continuous Domain with Partial Observations
    Jiachen Hu*, Han Zhong*, Chi Jin, and 1 more author
    In International Conference on Learning Representations, 2023

    We study sim-to-real transfer in continuous domains with partial observations, modeled by linear quadratic Gaussian (LQG) systems. We show that a popular robust adversarial training algorithm can learn a policy from simulation that is competitive to the optimal real-world policy, providing the first provable guarantee in this setting.

  3. ICLR
    Understanding Domain Randomization for Sim-to-real Transfer
    Xiaoyu Chen*, Jiachen Hu*, Chi Jin, and 2 more authors
    In International Conference on Learning Representations(Spotlight, top 6%) , 2022

    We provide a theoretical framework for domain randomization, modeling the simulator as a set of MDPs with tunable parameters. We prove sharp bounds on the sim-to-real gap and show that successful transfer is achievable without any real-world training samples, highlighting the importance of history-dependent policies.

  4. ICML
    Near-Optimal Representation Learning for Linear Bandits and Linear RL
    Jiachen Hu*, Xiaoyu Chen*, Chi Jin, and 2 more authors
    In Proceedings of the 38th International Conference on Machine Learning, 2021

    We study multi-task representation learning for linear bandits and episodic RL with linear value function approximation. Our algorithm MTLR-OFUL achieves \(\tilde{O}(M\sqrt{dkT} + d\sqrt{kMT})\) regret, significantly improving over the \(\tilde{O}(Md\sqrt{T})\) baseline, yielding the first theoretical characterization of multi-task representation learning benefits in RL exploration.

  5. ICLR
    Distributed Bandit Learning: Near-Optimal Regret with Efficient Communication
    Yuanhao Wang*, Jiachen Hu*, Xiaoyu Chen, and 1 more author
    In International Conference on Learning Representations, 2020

    We design communication protocols for distributed bandit learning with M agents under central coordination. For multi-armed bandits, we achieve near-optimal regret with only \(O(M\log(MK))\) communication cost — independent of the time horizon T and matching the lower bound up to a log factor.