Let's connect! ๐Ÿฅณ

Nima
Find Nima on:
Ben
Find Ben on:
Bishwajit
Find Bishwajit on:
Leo
Find Leo on:
Jean-Jacques
Find Jean-Jacques on:
Dima
Find Dima on:

Barebones page to important links to this work

NRGPT: An Energy-based Alternative for GPT

Nima Dehmamy
Benjamin Hoover
Bishwajit Saha
Leo Kozachkov
Jean-Jacques Slotine
Dmitry Krotov

Generative Pre-trained Transformer (GPT) architectures are the most popular design for language modeling. Energy-based modeling is a different paradigm that views inference as a dynamical process operating on an energy landscape. We propose a minimal modification of the GPT setting to unify it with the EBM framework. The inference step of our model, which we call eNeRgy-GPT (NRGPT), is conceptualized as an exploration of the tokens on the energy landscape. We prove, and verify empirically, that under certain circumstances this exploration becomes gradient descent, although they don't necessarily lead to the best performing models. We demonstrate that our model performs well for simple language (Shakespeare dataset), algebraic ListOPS tasks, and richer settings such as OpenWebText language modeling. We also observe that our models may be more resistant to overfitting, doing so only during very long training.

Cite this work

@inproceedings{
dehmamy2026nrgpt,
title={{NRGPT}: An Energy-based Alternative for {GPT}},
author={Nima Dehmamy and Benjamin Hoover and Bishwajit Saha and Leo Kozachkov and Jean-Jacques Slotine and Dmitry Krotov},
booktitle={The Fourteenth International Conference on Learning Representations},
year={2026},
url={https://openreview.net/forum?id=B3Muyi2zgo}
}

Links