Understanding exoplanet atmospheric spectra from the NeurIPS Ariel Data Challenge 2024
Each spectrum tells us what molecules are in the exoplanet's atmosphere. Water shows absorption at certain wavelengths, COโ at others, etc.
We also experimented with various ML models on a simplified spectral regression task. Note: This is not the same as solving the full competition problem.
| Rank | Model | Architecture | Paper Date | Test Rยฒ |
|---|---|---|---|---|
| ๐ฅ | FourierKAN | Fourier-based KAN (Statistically Verified) | Jun 2024 | 0.9474 |
| ๐ฅ | Griffin (RG-LRU) | Gated Linear Recurrence (Google DeepMind) | Feb 2024 | 0.9415 |
| ๐ฅ | Liquid NN | Liquid Time-Constant Networks (MIT) | Jun 2020 | 0.9457 |
| 4th | Smooth Fourier | Physics-Informed Fourier Basis | Novel | 0.9441 |
| 5th | Griffin++ | Optimized RG-LRU Variant | Novel | 0.9434 |
| 6th | MEGA | EMA + Gated Attention (Meta) | Sep 2022 | 0.9415 |
| 7th | Neural ODE | ODE-based Dynamics (NeurIPS 2018 Best) | Jun 2018 | 0.9410 |
| 8th | JacobiKAN | Jacobi Polynomial KAN | Jun 2024 | 0.9391 |
| 9th | KAN | Polynomial KAN (MIT) | Apr 2024 | 0.9387 |
| 10th | TTT-Linear | Test-Time Training (Stanford/Meta) | Jul 2024 | 0.9382 |
| 11th | Hyena | Long Conv + Gating (Stanford) | Mar 2023 | 0.9379 |
| 12th | 1D-CNN | Spectral Convolutions | Classic | 0.9366 |
| 13th | Bayesian | MC Dropout | ICML 2016 | 0.9359 |
| 14th | Transformer | ViT-style Spectral | Oct 2020 | 0.9331 |
| 15th | xLSTM | Exponential Gating | May 2024 | 0.9262 |
New to exoplanet spectroscopy? Start here! These visualizations explain everything from scratch.
Ready to go deeper? These visualizations explore the full dataset in detail.
Fourier series as learnable activation functions. Ideal for spectral data with periodic wavelength patterns. arXiv:2406.01034
Kolmogorov-Arnold Networks use learnable polynomial/spline activations on edges. Published by MIT & Northeastern. arXiv:2404.19756
State Space Model with selective mechanism by Gu & Dao (CMU/Princeton). O(n) complexity. arXiv:2312.00752
Extended LSTM by Hochreiter et al. (JKU Linz). Exponential gating and matrix memory. arXiv:2405.04517
Classic but effective for spectral data. Treats input as 1D signal and extracts local patterns through convolution. Fast and proven.
arXiv:2402.19427 (Feb 2024)
De, Smith et al. Google DeepMind. Gated Linear Recurrence - OUR CHAMPION!
arXiv:2406.01034 (June 2024)
Fourier Kolmogorov-Arnold Network. Rยฒ = 0.9463
arXiv:2404.19756 (April 2024)
Liu et al., MIT & Northeastern. Learnable polynomial activations.
arXiv:2406.09798 (June 2024)
Fractional KAN with Jacobi orthogonal polynomials.
arXiv:2405.12832 (May 2024)
Bozorgasl & Chen, Boise State. Morlet wavelet multi-resolution.
arXiv:2006.04439 (June 2020)
Hasani, Lechner et al. MIT & TU Wien. Liquid Time-Constant Networks.
arXiv:1806.07366 (NeurIPS 2018 Best Paper)
Chen et al., U of Toronto. Continuous ODE dynamics.
arXiv:2302.10866 (March 2023)
Poli et al., Stanford. Long convolutions + data-controlled gating.
arXiv:2209.10655 (Sep 2022)
Ma et al., Meta AI (FAIR). EMA + Gated Attention.
arXiv:2407.04620 (July 2024)
Sun et al., Stanford & Meta. Test-Time Training with hidden state learning.
arXiv:2312.00752 (Dec 2023)
Gu & Dao, CMU & Princeton. Selective state spaces, O(n) complexity.
arXiv:2405.21060 (ICML 2024)
Dao & Gu. Structured State Space Duality. 2-8ร faster.
arXiv:2405.04517 (May 2024)
Beck, Hochreiter et al., JKU Linz. Exponential gating & matrix memory.
arXiv:1506.02142 (ICML 2016)
Gal & Ghahramani, Cambridge. Dropout as Bayesian approximation.
arXiv:2010.11929 (Oct 2020)
Dosovitskiy et al., Google Research. Patch-based transformer.