Bitcoin

Bitcoin

$108,741.08

BTC 0.20%

Ethereum

Ethereum

$4,451.82

ETH 1.11%

  • Login
  • Register
Metaverse Media Group
  • Home
  • Crypto
  • NFTs
  • Artificial Intelligence
  • More
    • Technology
    • Business
    • Newsletter
No Result
View All Result
  • Home
  • Crypto
  • NFTs
  • Artificial Intelligence
  • More
    • Technology
    • Business
    • Newsletter
No Result
View All Result
Metaverse Media Group

New Energy-Based Transformer architecture aims to bring better “System 2 thinking” to AI models

New Energy-Based Transformer architecture aims to bring better “System 2 thinking” to AI models

The Decoderby The Decoder
11 July 2025
A new architecture called Energy-Based Transformer is designed to teach AI models to solve problems analytically and step by step. The article New Energy-Based Transformer architecture aims to bring better “System 2 thinking” to AI models appeared first on THE DECODER….

summary
Summary

A new architecture called the Energy-Based Transformer aims to teach AI models how to solve problems analytically and step by step.

Most current AI models operate much like what Daniel Kahneman described as “System 1 thinking”: they’re fast, intuitive, and excel at pattern recognition. But according to a study from researchers at UVA, UIUC, Stanford, Harvard, and Amazon GenAI, these models often fail at tasks that require the slower, more analytical “System 2 thinking” – such as complex logical reasoning or advanced mathematics.

The paper, “Energy-Based Transformers are Scalable Learners and Thinkers,” asks whether these kinds of reasoning skills can emerge purely from unsupervised learning. The researchers’ answer is a new architecture: the Energy-Based Transformer (EBT).

How Energy-Based Transformers work

The EBT approach treats thinking as an iterative optimization process. Instead of generating an answer in a single step, the model starts with a random solution. It then evaluates this solution by calculating an “energy” value.

THE DECODER Newsletter
The most important AI news straight to your inbox.
✓ Weekly
✓ Cancel at any time

The lower the energy, the better the prediction fits the context. Through repeated adjustments using gradient descent, the answer is gradually refined until the energy reaches a minimum. This lets the model spend more computation on harder problems.

Image: Gladstone et al.
Recommend our article

The idea of framing this process in terms of energy isn’t new – Meta’s chief AI scientist Yann LeCun and others have discussed “energy-based models” for years.

More efficient learning and generalization

In experiments, the researchers compared EBTs with an advanced Transformer variant (Transformer++). Their results suggest EBTs scale more efficiently: the paper reports up to a 35 percent higher scaling rate in terms of data, parameter count, and compute. This points to improved data and computational efficiency.

The real strength, however, shows up in what the authors call “thinking scalability” – the ability to boost performance by allocating extra compute at runtime. On language tasks, EBTs improved performance by up to 29 percent, especially on problems that differed significantly from their training data.

Image: Gladstone et al.

In image denoising tests, EBTs outperformed Diffusion Transformers (DiTs) while requiring 99 percent fewer computation steps. The study also found that EBTs learned image representations that delivered roughly ten times better classification accuracy on ImageNet-1k, suggesting a deeper understanding of content.

Recommendation

Significant hurdles remain

Despite these promising results, open questions remain. The main issue is compute: according to the paper, training EBTs requires 3.3 to 6.6 times more computing power (FLOPs) than standard Transformers. This extra overhead could be a barrier for many real-world applications. The study also measures “System 2 thinking” mainly through perplexity improvements, rather than actual reasoning tasks, and comparisons to state-of-the-art reasoning models are missing due to limited compute budgets.

All scaling predictions are based on experiments with models up to just 800 million parameters – much smaller than today’s largest AI systems. Whether EBTs’ advantages hold at larger scales remains to be seen.

Join our community
Join the DECODER community on Discord, Reddit or Twitter – we can’t wait to meet you.

Read the full article on The-Decoder.com
in AI
Reading Time: 3 mins read
0
0
24
VIEWS
Share on TwitterShare on Facebook

Subscribe to our newsletter

For the latest news & monthly prize giveaways
Join Now

Subscribe to our newsletter

For the latest news & monthly prize giveaways
Join Now
ADVERTISEMENT

Related Posts

OpenAI prepares to launch GPT-5, but big leaps are unlikely
AI

OpenAI prepares to launch GPT-5, but big leaps are unlikely

4 weeks ago
29
Psychologist says ChatGPT helps understanding even if it may not understand
AI

Psychologist says ChatGPT helps understanding even if it may not understand

4 weeks ago
26
Under mounting pressure, Apple plans to increase its spending on artificial intelligence projects
AI

Under mounting pressure, Apple plans to increase its spending on artificial intelligence projects

4 weeks ago
26

Comments

Please login to join discussion
ADVERTISEMENT

Latest News

  • All
  • Crypto
  • NFTs
  • Technology
  • Business
XRP Positioned to Play Key Role in Bridging Tokenized Assets Across Jurisdictions, Says Ripple CTO
Crypto

XRP Positioned to Play Key Role in Bridging Tokenized Assets Across Jurisdictions, Says Ripple CTO

Bitcoin.com News
by Bitcoin.com News
4 weeks ago
29
SEC v Ripple: SEC Has Just Days Left to Respond on XRP Case, Says Legal Expert
Crypto

SEC v Ripple: SEC Has Just Days Left to Respond on XRP Case, Says Legal Expert

Bitcoin.com News
by Bitcoin.com News
4 weeks ago
30
XRP ETF From Teucrium Sees Enormous Interest, Gains Massive Traction With Extraordinary Inflows
Crypto

XRP ETF From Teucrium Sees Enormous Interest, Gains Massive Traction With Extraordinary Inflows

Bitcoin.com News
by Bitcoin.com News
4 weeks ago
29
Latam Insights: El Salvador’s Bitcoin ‘Shuffling,’ Brazil Gets 50% Tariffs
Crypto

Latam Insights: El Salvador’s Bitcoin ‘Shuffling,’ Brazil Gets 50% Tariffs

Bitcoin.com News
by Bitcoin.com News
4 weeks ago
28
Mining Crunch? Bitcoin Hashrate Slides Below 900 EH/s
Crypto

Mining Crunch? Bitcoin Hashrate Slides Below 900 EH/s

Bitcoin.com News
by Bitcoin.com News
4 weeks ago
32
Justin Sun Makes History as Youngest Chinese Commercial Astronaut With Blue Origin’s NS-34 Spaceflight
Crypto

Justin Sun Makes History as Youngest Chinese Commercial Astronaut With Blue Origin’s NS-34 Spaceflight

Bitcoin.com News
by Bitcoin.com News
4 weeks ago
26
Load More
Next Post
10 AI Chatbots Predict Bitcoin’s Year-End Price — One Eyes a $225K Finish

10 AI Chatbots Predict Bitcoin’s Year-End Price — One Eyes a $225K Finish

ADVERTISEMENT

Follow Us

Categories

  • Crypto
  • NFTs
  • AI
  • Technology
  • Business
  • Crypto
  • NFTs
  • AI
  • Technology
  • Business
Subscribe to our Newsletter

© 2022 Metaverse Media Group – The Metaverse Mecca

Privacy and Cookie Policy | Sitemap

Welcome Back!

Sign In with Google
OR

Login to your account below

Forgotten Password? Sign Up

Create New Account!

Sign Up with Google
OR

Fill the forms below to register

*By registering into our website, you agree to the Terms & Conditions and Privacy Policy.
All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • Home
  • Crypto
  • NFTs
  • Artificial Intelligence
  • More
    • Technology
    • Business
    • Newsletter
Bitcoin

Bitcoin

$108,741.08

BTC 0.20%

Ethereum

Ethereum

$4,451.82

ETH 1.11%

  • Login
  • Sign Up
This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy and Cookie Policy.

Subscribe to our newsletter

Get the latest news & win monthly prizes

Subscribe to our newsletter

For the Latest News and Monthly Prize Giveaways

Join Now
Join Now