Over the last year, the emergence of accessible generative artificial intelligence (AI)Â tools has unleashed a frenzy of usage among consumers. From answering modest questions to using technology to perform excessive work tasks, technology is becoming more of a mainstay in everyday life.
The data backs this up, with the most popular AI platforms seeing a massive increase in traffic. OpenAIâs popular AI chatbot, ChatGPT, has around 180.5 million monthly users as of January 2024.
In this rapidly evolving landscape of generative AI, the demand for computing power to run the technology is reaching unprecedented levels, however. As businesses grapple with the complexities of managing this surge in computational requirements, industry experts are trying to find practical solutions.
In an interview with Cointelegraph, Doug Petkanics, co-founder and CEO of video infrastructure network Livepeer, shared insights into the escalating demand for AI computing power in 2024 and shed light on strategies to manage the burgeoning requirements to run the technology.
Related: Michigan University to let AI participate in classes, choose major, earn degree
Understanding the significance of computing power
Computing power is the backbone of AI development and deployment, as it governs the speed and efficiency of AI models.
According to the Livepeer co-founder, the three stages in the Al lifecycle that typically need substantial computing power come during its training, fine-tuning and inference, which is when the trained and tuned model produces outputs or predictions based on a set of inputs (also known as prompts).
The urgency for faster responses, however, collides with the economic reality of steep costs.
Petkanics explained that âmore computing power generally correlates with faster responses. But thereâs always a balance between user experience (speed) and the projectâs economic viability (cost).â
Some estimates have OpenAIâs daily operating costs at $700,000. The incredible costs of computing are exacerbated by a shortage of suitable GPUs for AI computing.
âThankfully, the same GPUs that are already running in crypto-coordinated DePIN networks, performing tasks like video transcoding and 3D rendering, are well-suited for AI.â
âThatâs why these crypto-networks have become such a critical component of the AI boom,â he said. Industry experts have already been pointing to this yearâs emerging âpower coupleâ of DePINs and AI.
A conscious consumerÂ
While computing power is becoming a major concern for those developing the technology, Petkanics explained that, per usual, consumers are not in the loop.
âMost people donât give much thought to infrastructure: where the power comes from, how the internet works, the cost or carbon footprint of a Google search,â he said. âThey want utilities to work on demand, every time.â
He said the same goes for AI computing power. Most users donât think about the energy, usage or computing cost of the AI prompts they input into their favorite chatbot. âThey care about speed and relevancy,â he said.
âConsumers wonât notice issues with computing power until the costs get passed on in the form of an increased quantity of ads, decreased speed/quality of responses or rising subscription costs.â
National and global implications
Addressing broader implications, Petkanics expressed concerns about the monopolization of scaled AI platforms.
Just as the dawn of the internet ushered in a group of companies that are now collectively known as âBig Techâ â Alphabet (Google), Amazon, Apple, Meta (Facebook and Instagram) and Microsoft â the beginning of the AI era sees the same mega companies racing for dominance.
âThese handful of big tech companies already own large proprietary data sets farmed from customer data to train models; they have spent billions training these modelsâŚâ
Since these companies own this full stack, it would essentially allow them to âarbitrarily insertâ their own biases into how the models perform on a given set of inputs. This would also theoretically allow them to censor what inputs and outputs can be provided to these models by users.
Thus, Petkanics stressed the importance of the open-source AI movement, which he said can be seen already with the aforementioned DePIN infrastructure that can help ensure accessibility and mitigate risks that are associated with centralization.
âCountries should be supporting these movements to ensure this wave of innovation is accessible to all and benefits a worldwide population.â
The AI boom is only just beginning. As it continues to evolve, Petkanics and many others at the forefront of these developments already predict a âwhole new setâ of economic, environmental and social considerations.
Magazine: Scam AI âkidnappings,â $20K robot chef, Ackmanâs AI plagiarism war: AI Eye