The performance of lithium-ion batteries is fundamentally limited by ionic transport within solid materials. In cathodes and solid electrolytes, lithium moves through a crystalline lattice by hopping between energetically stable sites. The rate of this hopping process determines:
Ionic conductivity
Rate capability
Power density
Temperature stability
From transition state theory, the hopping rate k follows:
k∝exp(−Ea/kBT) ,
where Ea is the migration (activation) barrier, kB is Boltzmann’s constant and T is the temperature. This exponential dependence means that even small errors in Ea lead to orders-of-magnitude differences in predicted conductivity. Thus, accurately predicting migration barriers is critical for screening next-generation battery materials.
Lithium diffusion occurs via hopping between two equilibrium lattice sites.
Consider:
An initial site with energy Einitial
A final site with similar equilibrium energy
A saddle point along the minimum energy path
The migration barrier is defined as:
Ea =Esaddle − Einitial
Graphically, this corresponds to the energy difference between the lowest energy configuration (stable site) and the highest energy configuration along the minimum energy path.
Physically, the saddle point represents the most constrained configuration during hopping, where:
Li ion passes through a bottleneck
Electrostatic repulsion is maximal
Local coordination is distorted
The magnitude of this barrier determines how easily lithium can move through the material.
To compute Ea, one must determine the minimum energy path (MEP) between initial and final configurations. The standard method is the Nudged Elastic Band (NEB) method:
Generate a sequence of intermediate images between initial and final states.
Relax these images while constraining them along the path. The force perpendicular to the path is 0.
Identify the highest-energy image — the saddle point.
When forces and energies are computed with Density Functional Theory (DFT), this becomes DFT-NEB.
However, DFT-NEB is computationally expensive because:
Each image requires a full DFT calculation.
A single hop may require 5–9 images.
Supercells often contain 100+ atoms.
Screening thousands of materials becomes prohibitive.
To overcome DFT cost, approximate methods like Bond Valence Site Energy (BVSE) can estimate migration barriers cheaply.
BVSE:
Uses empirical electrostatic models.
Estimates site energies from bond valence parameters.
Provides rapid barrier estimates.
Advantages:
Fast
Scalable to 100k+ structures
Limitations:
Approximate physics
Less accurate than DFT
Systematic biases
Thus, BVSE provides large-scale but imperfect data.
Can large-scale approximate physics be used to learn transferable representations that improve high-fidelity predictions?
Formally:
Given:
Large dataset DBVSE (approximate physics)
Small dataset DDFT (ground truth)
We ask:
Does learning on DBVSE improve generalization on DDFT?
Migration barriers are not simple functions of composition alone. They depend on:
Local atomic coordination
Geometry of diffusion bottleneck
Bond lengths and distortions
Electrostatic environment
These are inherently structural and geometric properties.
We therefore represent each migration hop as a graph:
Nodes: Atoms in the supercell
Special centroid atom (“X”) marks bottleneck location
Edges: Neighbor interactions within cutoff
Node features: Atomic number + distance to centroid
Edge features: Interatomic distances
This allows the model to learn local geometric environments directly.
We hypothesize:
Pretraining on 122k BVSE barriers allows the model to learn structural migration physics.
Fine-tuning on 1.6k DFT barriers adjusts energy scale and corrects systematic bias.
The resulting model generalizes better than training on DFT alone.
This approach leverages:
Abundant approximate physics
Scarce high-fidelity quantum data
If successful, this framework enables:
Rapid screening of thousands of battery materials.
Reduced reliance on expensive DFT-NEB.
Scalable discovery of solid electrolytes.
Data-efficient learning in materials science.
More broadly, this project demonstrates large-scale approximate simulations can serve as effective pretraining corpora for high-fidelity quantum predictions.