Joint Optimization of Neural Autoregressors via Scoring rules
1 authors
arXiv:2601.05683v1
Authors
Jonas Landsgesell
Abstract
Non-parametric distributional regression has achieved significant milestones in recent years. Among these, the Tabular Prior-Data Fitted Network (TabPFN) has demonstrated state-of-the-art performance on various benchmarks. However, a challenge remains in extending these grid-based approaches to a truly multivariate setting. In a naive non-parametric discretization with $N$ bins per dimension, the complexity of an explicit joint grid scales exponentially and the paramer count of the neural networks rise sharply. This scaling is particularly detrimental in low-data regimes, as the final projection layer would require many parameters, leading to severe overfitting and intractability.
Paper Information
- arXiv ID:
- 2601.05683v1
- Published:
- Categories:
- cond-mat.soft, cs.AI