Songchen Tan
i@tansongchen.com +1 (857) 298-9702 tansongchen songchentan

Education

Doctor of Science in Computational Science and Engineering (Mathematics track)

Cambridge, MA
Massachusetts Institute of Technology09/2023 – Current

GPA: 4.75 / 5.00

Relevant Coursework: Eigenvalue of Random Matrices (A), Nonlinear Dynamics and Chaos (A), Fast Methods for Partial Differential Equations (A)


Master of Science in Computational Science and Engineering

Cambridge, MA
Massachusetts Institute of Technology09/2021 – 06/2023

GPA: 5.00 / 5.00

Relevant Coursework: Parallel Computing & Scientific Machine Learning (A), Optimization Methods (A), Numerical Methods for Partial Differential Equations (A+), Introduction to Numerical Methods (A)


Bachelor of Science in Chemistry & Bachelor of Science in Physics

Beijing
Peking University09/2017 – 07/2021

GPA: 3.89 / 4.00, rank 1 / 137, honored as Weiming Bachelor (top 1%)

Relevant Coursework: Introduction to Computation (97), Data Structure and Algorithms (96), Computational Physics (92), Ordinary Differential Equations (92), Mathematical Method in Physics (95), Advanced Mathematics I, II (92, 96), Advanced Algebra I, II (92, 94)


Exchange Student

Los Angeles, CA
University of California, Los Angeles09/2019 – 12/2019

GPA: 4.00 / 4.00

Relevant Coursework: Introduction to Probability (A+), Applied Numerical Methods (A)

Publication

S. Tan, K. Miao, A. Edelman, C. Rackauckas; Scalable Higher-order Nonlinear Solvers via Higher-order Automatic Differentiation Proceedings of 16th International Modelica and FMI Conference 2025
S. Tan, J. Zhu, A. Edelman, C. Rackauckas; TaylorDiff.jl: Efficient and Versatile Higher-Order Derivatives in Julia ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming 2025, Differentiable Programming Workshop 2025
S. Tan, A. Edelman, C. Rackauckas; Fast Higher-order Automatic Differentiation for Physical Models Proceedings of the JuliaCon Conferences 2023
S. Tan; Higher-Order Automatic Differentiation and Its Applications Master's Thesis at MIT 2023
S. Tan; Data-Driven Density Functional Models Bachelor's Thesis at Peking University 2021
I. Leven, H. Hao, S. Tan, ..., T. Head-Gordon; Recent advances for improving the accuracy, transferability, and efficiency of reactive force fields J. Chem. Theory Comput. 2021
S. Tan, I. Leven, ..., T. Head-Gordon; Stochastic constrained extended system dynamics for solving charge equilibration models J. Chem. Theory Comput. 2020

Professional Experience

Optimization Engineering Intern

Cupertino, CA
Apple05/2025 – 08/2025

Advisor: Qilin He, Jiaqi Jiang (Platform Architecture)

  • Worked on machine learning models and differentiable solvers.

Deep Learning Compiler Engineering Intern

Santa Clara, CA
NVIDIA05/2022 – 08/2022

Advisor: Yuan Lin (Deep Learning Compiler Team)

  • Investigated the heuristics of choosing kernels and compilation parameters during GEMM + epilogue fusion, and improved the optimal tactic hit by 70%
  • Accelerated the layer fusion compilation by 40%, by a combination of caching, multi-threading and reducing trial compilations

Research Experience

Efficient Higher-order Automatic Differentiation for Differential Models

Cambridge, MA
Massachusetts Institute of Technology09/2022 – Current

Advisor: Christopher Rackauckas & Alan Edelman

  • Developing higher-order forward-mode automatic differentiation (AD) algorithms that scale linearly with the order, suitable for differential models like ODEs and PDEs
  • Synthesizing code that is compiler-friendly and compatible with reverse-mode AD libraries like Zygote.jl, by aggressively specializing with compile-time type information
  • Deriving higher-order differentiation rules automatically from first-order chain rules with symbolic computation and metaprogramming in Julia

Low-level Automatic Differentiation for Linear Algebra Routines

Cambridge, MA
Massachusetts Institute of Technology09/2021 – 06/2022

Advisor: Christopher Rackauckas & Alan Edelman

  • Joined the Enzyme project (enzyme.mit.edu), an automatic differentiation framework based on source code transformation at LLVM intermediate representation (IR) level, which can differentiate through all languages with a LLVM backend (e.g. Julia, C++, Fortran)
  • Synthesized derivatives of BLAS/LAPACK kernels with generated kernels and calls to other BLAS/LAPACK kernels, and performed extensive optimizations based on linear algebra relations
  • Outperformed other high-level AD frameworks in Julia with 1.3× speed on a linear algebra benchmark set

Optimization Methods for Self-Consistent Field Functional Models

Beijing
Peking University12/2020 – 06/2021

Advisor: Weinan E & Linfeng Zhang

  • Modeled the exchange-correlation density functional in generalized Kohn-Sham theory with deep neural networks and descriptors from density matrices
  • Established a comprehensive theory for using physical quantities data that depends on the functional minimization result to train the functional model, in other words, addressed the “differentiate through argmin” problem
  • Implemented the training process with multiple types of physical quantity data, such as energy band structure and dipole moment
  • Improved the accuracy and generalization performance of the model, obtained an average energy error of 0.06 kcal/mol on a test set that includes 1200 water molecule configurations labeled with SCAN0 functional (48% less than previous methods)

Extended Lagrangian Scheme for Simulating Reactive Chemical Systems

Berkeley, CA
University of California, Berkeley12/2019 – 04/2020

Advisor: Teresa Head-Gordon & Lin Lin

  • Investigated reactive chemial systems with fluctuating charges described with differential-algebraic equations
  • Developed an extended Lagrangian scheme to replace the algebraic part with differential dynamics of an extended system, thereby eliminating the expensive charge-equilibration step (i.e. algebraic equation solving) in simulation
  • Proved the correctness of this scheme both theoretically and practically (the modified simulation still reproduced statistic and dynamic properties of that system)

Awards

MathWorks Prize for Outstanding Masters Research, MIT Center for Computational Science and Engineering
03/2023
Weiming Bachelor, Peking University (top 1%)
07/2021
Academic Award, College of Chemistry and Molecular Engineering, Peking University (top 2%)
07/2021
2020 Wusi Scholarship & Merit Student, Peking University (top 1%)
11/2020
National Second Prize in Contemporary Undergraduate Mathematical Contest in Modeling, China Society for Industrial and Applied Mathematics
12/2019
2019 National Scholarship & Merit Student, Peking University (top 1%)
11/2019
Education Aboard Program Scholarhsip, Peking University
05/2019
2018 National Scholarship & Merit Student, Peking University (top 1%)
11/2018

Skills

Programming Languages: C/C++, Julia, Python, Rust, Fortran, JavaScript/TypeScript
High-Performance Computing: CUDA, MPI & OpenMP, compiler optimizations with LLVM
Scientific Computing & Machine Learning Software: PyTorch, Flux.jl, PySCF, OpenMM, LAMMPS