artificial intelligence
This New Technique Slashes AI Energy Use by 95%
Published
6 hours agoon
By
adminA new technique could put AI models on a strict energy diet, potentially cutting power consumption by up to 95% without compromising quality.
Researchers at BitEnergy AI, Inc. have developed Linear-Complexity Multiplication (L-Mul), a method that replaces energy-intensive floating-point multiplications with simpler integer additions in AI computations.
For those unfamiliar with the term, floating-point is a mathematical shorthand that allows computers to handle very large and very small numbers efficiently by adjusting the placement of the decimal point. You can think of it like scientific notation, in binary. They are essential for many calculations in AI models, but they require a lot of energy and computing power. The bigger the number, the better the model is—and the more computing power it requires. Fp32 is generally a full precision model, with developers reducing precision to fp16, fp8, and even fp4, so their models can run on local hardware.
AI’s voracious appetite for electricity has become a growing concern. ChatGPT alone gobbles up 564 MWh daily—enough to power 18,000 American homes. The overall AI industry is expected to consume 85-134 TWh annually by 2027, roughly the same as Bitcoin mining operations, according to estimations shared by the Cambridge Centre for Alternative Finance.
L-Mul tackles the AI energy problem head-on by reimagining how AI models handle calculations. Instead of complex floating-point multiplications, L-Mul approximates these operations using integer additions. So, for example, instead of multiplying 123.45 by 67.89, L-Mul breaks it down into smaller, easier steps using addition. This makes the calculations faster and uses less energy, while still maintaining accuracy.
The results seem promising. “Applying the L-Mul operation in tensor processing hardware can potentially reduce 95% energy cost by element wise floating point tensor multiplications and 80% energy cost of dot products,” the researchers claim. Without getting overly complicated, what that means is simply this: If a model used this technique, it would require 95% less energy to think, and 80% less energy to come up with new ideas, according to this research.
The algorithm’s impact extends beyond energy savings. L-Mul outperforms current 8-bit standards in some cases, achieving higher precision while using significantly less bit-level computation. Tests across natural language processing, vision tasks, and symbolic reasoning showed an average performance drop of just 0.07%—a negligible tradeoff for the potential energy savings.
Transformer-based models, the backbone of large language models like GPT, could benefit greatly from L-Mul. The algorithm seamlessly integrates into the attention mechanism, a computationally intensive part of these models. Tests on popular models such as Llama, Mistral, and Gemma even revealed some accuracy gain on certain vision tasks.
At an operational level, L-Mul’s advantages become even clearer. The research shows that multiplying two float8 numbers (the way AI models would operate today) requires 325 operations, while L-Mul uses only 157—less than half. “To summarize the error and complexity analysis, L-Mul is both more efficient and more accurate than fp8 multiplication,” the study concludes.
But nothing is perfect and this technique has a major achilles heel: It requires a special type of hardware, so the current hardware isn’t optimized to take full advantage of it.
Plans for specialized hardware that natively supports L-Mul calculations may be already in motion. “To unlock the full potential of our proposed method, we will implement the L-Mul and L-Matmul kernel algorithms on hardware level and develop programming APIs for high-level model design,” the researchers say. This could potentially lead to a new generation of AI models that are fast, accurate, and super cheap—making energy-efficient AI a real possibility.
Generally Intelligent Newsletter
A weekly AI journey narrated by Gen, a generative AI model.
Source link
You may like
Bitlayer closes $9m extended series A round led by Polychain Capital
21Shares Calls on EU for Regulatory Clarity To ‘Unlock Crypto Opportunities’ in ETPs
Ethereum Foundation Sells $6M ETH Amid Continued Whale Liquidations
I Just Entered the Bitcoin Mining Lottery: NIKOLAUS
Coinbase warns Gen Z about growing online threats
Cardano Founder Lists Why Donald Trump Is Better
artificial intelligence
AI Tokens Lead Crypto Rebound Amid Strong U.S. Economy
Published
4 days agoon
October 5, 2024By
adminBitcoin may have bottomed at $60,000 earlier this week, and the Fed easing into a strong economy points to more upside, Will Clement said.
Source link
artificial intelligence
AI Assistant Goes Rogue and Ends Up Bricking a User’s Computer
Published
5 days agoon
October 4, 2024By
adminBuck Shlegeris just wanted to connect to his desktop. Instead, he ended up with an unbootable machine and a lesson in the unpredictability of AI agents.
Shlegeris, CEO of the nonprofit AI safety organization Redwood Research, developed a custom AI assistant using Anthropic’s Claude language model.
The Python-based tool was designed to generate and execute bash commands based on natural language input. Sounds handy, right? Not quite.
Shlegeris asked his AI to use SSH to access his desktop, unaware of the computer’s IP address. He walked away, forgetting that he’d left the eager-to-please agent running.
Big mistake: The AI did its task—but it didn’t stop there.
“I came back to my laptop ten minutes later to see that the agent had found the box, SSH’d in, then decided to continue,” Shlegeris said.
For context, SSH is a protocol that allows two computers to connect over an unsecured network.
“It looked around at the system info, decided to upgrade a bunch of stuff, including the Linux kernel, got impatient with apt, and so investigated why it was taking so long,” Shlegeris explained. “Eventually, the update succeeded, but the machine doesn’t have the new kernel, so I edited my grub config.”
The result? A costly paperweight as now “the computer no longer boots,” Shlegeris said.
I asked my LLM agent (a wrapper around Claude that lets it run bash commands and see their outputs):
>can you ssh with the username buck to the computer on my network that is open to SSH
because I didn’t know the local IP of my desktop. I walked away and promptly forgot I’d spun… pic.twitter.com/I6qppMZFfk— Buck Shlegeris (@bshlgrs) September 30, 2024
The system logs show how the agent tried a bunch of weird stuff beyond simple SSH until the chaos reached a point of no return.
“I apologize that we couldn’t resolve this issue remotely,” the agent said—typical of Claude’s understated replies. It then shrugged its digital shoulders and left Shlegeris to deal with the mess.
Reflecting on the incident, Shlegeris conceded, “This is probably the most annoying thing that’s happened to me as a result of being wildly reckless with [an] LLM agent.”
Shlegeris did not immediately respond to Decrypt’s request for comments.
Why AIs Making Paperweights is a Critical Issue For Humanity
Alarmingly, Shlegeris’ experience is not an isolated one. AI models are increasingly demonstrating abilities that extend beyond their intended purposes.
Tokyo-based research firm Sakana AI recently unveiled a system dubbed “The AI Scientist.“
Designed to conduct scientific research autonomously, the system impressed its creators by attempting to modify its own code to extend its runtime, Decrypt previously reported.
“In one run, it edited the code to perform a system call to run itself. This led to the script endlessly calling itself,” the researchers said. “In another case, its experiments took too long to complete, hitting our timeout limit.
Instead of making its code more efficient, the system tried to modify its code to extend beyond the timeout period.
This problem of AI models going beyond their boundaries is why alignment researchers spend so much time in front of their computers.
For these AI models, as long as they get their job done, the end justifies the means, so constant oversight is extremely important to ensure models behave as they are supposed to.
These examples are as concerning as they are amusing.
Imagine if an AI system with similar tendencies were in charge of a critical task, such as monitoring a nuclear reactor.
An overzealous or misaligned AI could potentially override safety protocols, misinterpret data, or make unauthorized changes to critical systems—all in a misguided attempt to optimize its performance or fulfill its perceived objectives.
AI is developing at such high speed that alignment and safety are reshaping the industry and in most cases this area is the driving force behind many power moves.
Anthropic—the AI company behind Claude—was created by former OpenAI members worried about the company’s preference for speed over caution.
Many key members and founders have left OpenAI to join Anthropic or start their own businesses because OpenAI supposedly pumped the brakes on their work.
Schelegris actively uses AI agents on a day-to-day basis beyond experimentation.
“I use it as an actual assistant, which requires it to be able to modify the host system,” he replied to a user on Twitter.
Edited by Sebastian Sinclair
Generally Intelligent Newsletter
A weekly AI journey narrated by Gen, a generative AI model.
Source link
artificial intelligence
AI-Related Cryptos Lead Altcoin Surge; Bitcoin Breakout Nears with Several Catalyst in Q4: Analyst
Published
2 weeks agoon
September 23, 2024By
adminNEAR, RNDR, TAO and LPT booked double-digit gains as artificial intelligence-focused tokens were the best performers within the CoinDesk 20 Index.
Source link
Bitlayer closes $9m extended series A round led by Polychain Capital
21Shares Calls on EU for Regulatory Clarity To ‘Unlock Crypto Opportunities’ in ETPs
Ethereum Foundation Sells $6M ETH Amid Continued Whale Liquidations
I Just Entered the Bitcoin Mining Lottery: NIKOLAUS
Coinbase warns Gen Z about growing online threats
This New Technique Slashes AI Energy Use by 95%
Cardano Founder Lists Why Donald Trump Is Better
The Bitcoin Report: Parabolic Growth Predicted for Q4 2024
SUI price set for new ATH amid native USDC launch
Binance, FalconX and the Curious Case of 1.35M Missing Solana Tokens
Digital asset adoption is rising and reshaping the financial landscape as we know it
SHINOBI: Stop Demonizing Developers Over Nonsense
CATS see 691% surge ahead of major exchange listings
What Made Solana Memecoins The Cycle’s Top Narrative
Bitcoin Price Falters At $64K, Chinese Stocks Tumble, What’s Next for BTC?
182267361726451435
Top Crypto News Headlines of The Week
New U.S. president must bring clarity to crypto regulation, analyst says
Bitcoin 20% Surge In 3 Weeks Teases Record-Breaking Potential
Bitcoin Open-Source Development Takes The Stage In Nashville
Ethereum, Solana touch key levels as Bitcoin spikes
Ethereum Crash A Buying Opportunity? This Whale Thinks So
Why Did Trump Change His Mind on Bitcoin?
Will XRP Price Defend $0.5 Support If SEC Decides to Appeal?
Shiba Inu Price Slips 4% as 3500% Burn Rate Surge Fails to Halt Correction
Citigroup Executive Steps Down To Explore Crypto
Mostbet Güvenilir Mi – Casino Bonus 2024
‘Hamster Kombat’ Airdrop Delayed as Pre-Market Trading for Telegram Game Expands
Bitcoin flashes indicator that often precedes higher prices: CryptoQuant
Crypto Market Movers: 5 Altcoins Making Waves This Bull Run
Trending
- 3 weeks ago
182267361726451435
- 24/7 Cryptocurrency News2 months ago
Top Crypto News Headlines of The Week
- News1 month ago
New U.S. president must bring clarity to crypto regulation, analyst says
- Bitcoin2 months ago
Bitcoin 20% Surge In 3 Weeks Teases Record-Breaking Potential
- Opinion3 months ago
Bitcoin Open-Source Development Takes The Stage In Nashville
- Bitcoin3 months ago
Ethereum, Solana touch key levels as Bitcoin spikes
- Altcoins2 months ago
Ethereum Crash A Buying Opportunity? This Whale Thinks So
- Donald Trump2 months ago
Why Did Trump Change His Mind on Bitcoin?