Evaluation Setup
Pretrain on a source PDE/distribution, then fine-tune on a target PDE/distribution to evaluate transfer efficiency and generalization
Benchmarks:
- Navier-Stokes (NS) (Fluid dynamics prediction (viscosity var))
- Shallow Water Equations (SWE) (Geophysical fluid dynamics)
- Compressible Euler (Gas dynamics)
Metrics:
- Relative L2 Error (Test Error)
- Training Efficiency (Convergence speed)
- Statistical methodology: Not explicitly reported in the paper
Key Results
| Benchmark |
Metric |
Baseline |
This Paper |
Δ |
| Comparison of different pretraining strategies on the FNO backbone, fine-tuned on Navier-Stokes (Re=500). Lower L2 error is better. |
| Navier-Stokes (Re=500) |
Relative L2 Error |
0.0156 |
0.0134 |
-0.0022
|
| Navier-Stokes (Re=500) |
Relative L2 Error |
0.0156 |
0.0142 |
-0.0014
|
| Navier-Stokes |
Relative L2 Error |
0.0156 |
0.0138 |
-0.0018
|
| Results demonstrating performance in low-data regimes (few-shot learning). |
| Shallow Water Equations (10% Data) |
Relative L2 Error |
0.045 |
0.038 |
-0.007
|
Main Takeaways
- Physics-based strategies (Derivative, Coefficient) and dense vision tasks (Masked) generally outperform simple sorting tasks (Binary, TimeSort).
- Data augmentation (Scale, Noise) is a highly effective, low-cost way to improve neural operator performance, often matching more complex pretraining.
- Transformer-based backbones (OFormer) benefit more from large-scale pretraining than CNN-based or FNO backbones, aligning with trends in NLP/Vision.
- Pretraining is most beneficial when the downstream task has limited data or is distributionally similar to the pretraining data.