BPOI Banner
This New Technique Slashes AI Energy Use by 95% This New Technique Slashes AI Energy Use by 95%

This New Technique Slashes AI Energy Use by 95%

A new technique could put AI models on a strict energy diet, potentially cutting power consumption by up to 95% without compromising quality.

Researchers at BitEnergy AI, Inc. have developed Linear-Complexity Multiplication (L-Mul), a method that replaces energy-intensive floating-point multiplications with simpler integer additions in AI computations.

For those unfamiliar with the term, floating-point is a mathematical shorthand that allows computers to handle very large and very small numbers efficiently by adjusting the placement of the decimal point. You can think of it like scientific notation, in binary. They are essential for many calculations in AI models, but they require a lot of energy and computing power. The bigger the number, the better the model is—and the more computing power it requires. Fp32 is generally a full precision model, with developers reducing precision to fp16, fp8, and even fp4, so their models can run on local hardware.

Image: Wikipedia

AI’s voracious appetite for electricity has become a growing concern. ChatGPT alone gobbles up 564 MWh dailyenough to power 18,000 American homes. The overall AI industry is expected to consume 85-134 TWh annually by 2027, roughly the same as Bitcoin mining operations, according to estimations shared by the Cambridge Centre for Alternative Finance.

L-Mul tackles the AI energy problem head-on by reimagining how AI models handle calculations. Instead of complex floating-point multiplications, L-Mul approximates these operations using integer additions. So, for example, instead of multiplying 123.45 by 67.89, L-Mul breaks it down into smaller, easier steps using addition. This makes the calculations faster and uses less energy, while still maintaining accuracy.

The results seem promising. “Applying the L-Mul operation in tensor processing hardware can potentially reduce 95% energy cost by element wise floating point tensor multiplications and 80% energy cost of dot products,” the researchers claim. Without getting overly complicated, what that means is simply this: If a model used this technique, it would require 95% less energy to think, and 80% less energy to come up with new ideas, according to this research.

The algorithm’s impact extends beyond energy savings. L-Mul outperforms current 8-bit standards in some cases, achieving higher precision while using significantly less bit-level computation. Tests across natural language processing, vision tasks, and symbolic reasoning showed an average performance drop of just 0.07%a negligible tradeoff for the potential energy savings.

Transformer-based models, the backbone of large language models like GPT, could benefit greatly from L-Mul. The algorithm seamlessly integrates into the attention mechanism, a computationally intensive part of these models. Tests on popular models such as Llama, Mistral, and Gemma even revealed some accuracy gain on certain vision tasks.

Image: Bitenergy.ai via ArXiv

At an operational level, L-Mul’s advantages become even clearer. The research shows that multiplying two float8 numbers (the way AI models would operate today) requires 325 operations, while L-Mul uses only 157—less than half. “To summarize the error and complexity analysis, L-Mul is both more efficient and more accurate than fp8 multiplication,” the study concludes.

But nothing is perfect and this technique has a major achilles heel: It requires a special type of hardware, so the current hardware isn’t optimized to take full advantage of it.

Plans for specialized hardware that natively supports L-Mul calculations may be already in motion. “To unlock the full potential of our proposed method, we will implement the L-Mul and L-Matmul kernel algorithms on hardware level and develop programming APIs for high-level model design,” the researchers say. This could potentially lead to a new generation of AI models that are fast, accurate, and super cheap—making energy-efficient AI a real possibility.

Generally Intelligent Newsletter

A weekly AI journey narrated by Gen, a generative AI model.

Source link

Jose Antonio Lanz

https://decrypt.co/285154/new-technique-slashes-ai-energy

2024-10-08 18:56:57

bitcoin
Bitcoin (BTC) $ 91,287.46 3.49%
ethereum
Ethereum (ETH) $ 3,099.82 0.95%
tether
Tether (USDT) $ 1.00 0.10%
solana
Solana (SOL) $ 218.70 4.01%
bnb
BNB (BNB) $ 620.66 0.24%
dogecoin
Dogecoin (DOGE) $ 0.378596 1.84%
xrp
XRP (XRP) $ 0.886291 7.52%
usd-coin
USDC (USDC) $ 1.00 0.13%
staked-ether
Lido Staked Ether (STETH) $ 3,095.95 1.01%
cardano
Cardano (ADA) $ 0.724619 22.61%
tron
TRON (TRX) $ 0.191136 7.65%
shiba-inu
Shiba Inu (SHIB) $ 0.000025 6.59%
the-open-network
Toncoin (TON) $ 5.39 2.15%
avalanche-2
Avalanche (AVAX) $ 33.35 5.81%
wrapped-bitcoin
Wrapped Bitcoin (WBTC) $ 91,102.41 3.47%
wrapped-steth
Wrapped stETH (WSTETH) $ 3,665.59 0.84%
sui
Sui (SUI) $ 3.72 13.96%
pepe
Pepe (PEPE) $ 0.000023 9.22%
weth
WETH (WETH) $ 3,099.99 0.96%
chainlink
Chainlink (LINK) $ 13.86 5.97%
bitcoin-cash
Bitcoin Cash (BCH) $ 432.08 2.84%
polkadot
Polkadot (DOT) $ 5.19 7.64%
leo-token
LEO Token (LEO) $ 7.67 2.92%
near
NEAR Protocol (NEAR) $ 5.61 2.67%
aptos
Aptos (APT) $ 12.13 7.56%
litecoin
Litecoin (LTC) $ 84.52 3.48%
wrapped-eeth
Wrapped eETH (WEETH) $ 3,262.34 0.83%
usds
USDS (USDS) $ 1.00 0.42%
uniswap
Uniswap (UNI) $ 8.63 5.13%
crypto-com-chain
Cronos (CRO) $ 0.1707 18.29%
stellar
Stellar (XLM) $ 0.142082 5.80%
internet-computer
Internet Computer (ICP) $ 8.68 7.71%
dogwifcoin
dogwifhat (WIF) $ 3.90 10.02%
bittensor
Bittensor (TAO) $ 519.36 3.23%
ethereum-classic
Ethereum Classic (ETC) $ 23.09 4.16%
kaspa
Kaspa (KAS) $ 0.137127 2.70%
fetch-ai
Artificial Superintelligence Alliance (FET) $ 1.29 4.38%
dai
Dai (DAI) $ 1.00 0.12%
whitebit
WhiteBIT Coin (WBT) $ 22.32 0.62%
ethena-usde
Ethena USDe (USDE) $ 1.00 0.19%
bonk
Bonk (BONK) $ 0.000045 25.34%
polygon-ecosystem-token
POL (ex-MATIC) (POL) $ 0.373765 4.28%
hedera-hashgraph
Hedera (HBAR) $ 0.076536 16.62%
render-token
Render (RENDER) $ 7.18 8.43%
blockstack
Stacks (STX) $ 1.87 2.78%
monero
Monero (XMR) $ 145.07 1.50%
okb
OKB (OKB) $ 44.08 1.91%
floki
FLOKI (FLOKI) $ 0.000271 27.18%
first-digital-usd
First Digital USD (FDUSD) $ 1.00 0.03%
filecoin
Filecoin (FIL) $ 4.19 7.58%