🎉 Gate Square Growth Points Summer Lucky Draw Round 1️⃣ 2️⃣ Is Live!
🎁 Prize pool over $10,000! Win Huawei Mate Tri-fold Phone, F1 Red Bull Racing Car Model, exclusive Gate merch, popular tokens & more!
Try your luck now 👉 https://www.gate.com/activities/pointprize?now_period=12
How to earn Growth Points fast?
1️⃣ Go to [Square], tap the icon next to your avatar to enter [Community Center]
2️⃣ Complete daily tasks like posting, commenting, liking, and chatting to earn points
100% chance to win — prizes guaranteed! Come and draw now!
Event ends: August 9, 16:00 UTC
More details: https://www
AI and Web3 Convergence: New Opportunities from Computing Power Sharing to Data Incentives
AI+Web3: Towers and Squares
TL;DR
Web3 projects based on AI concepts have become targets for capital attraction in both primary and secondary markets.
The opportunities of Web3 in the AI industry are reflected in: using distributed incentives to coordinate potential supply in the long tail, across data, storage, and computation; at the same time, establishing an open-source model and a decentralized marketplace for AI Agents.
AI is mainly used in the Web3 industry for on-chain finance ( cryptocurrency payments, trading, data analysis ), and assisting development.
The utility of AI+Web3 is reflected in the complementarity of the two: Web3 is expected to counteract AI centralization, while AI is expected to help Web3 break out of its boundaries.
Introduction
In the past two years, the development of AI has accelerated as if a fast-forward button has been pressed. This butterfly effect, instigated by ChatGPT, has not only opened up a new world of generative artificial intelligence but has also created a tidal wave in the Web3 domain.
With the support of AI concepts, the financing boost in the slowing cryptocurrency market is evident. Media statistics show that in the first half of 2024 alone, 64 Web3+AI projects completed financing, with the AI-based operating system Zyber365 achieving a maximum financing amount of $100 million in its Series A round.
The secondary market is more prosperous. According to data from the crypto aggregation website Coingecko, the total market value of the AI sector has reached 48.5 billion USD in just over a year, with a 24-hour trading volume close to 8.6 billion USD; the positive impact of mainstream AI technology advancements is evident, as the average price of the AI sector rose by 151% following the release of OpenAI's Sora text-to-video model; the AI effect has also extended to the crypto fundraising sector, one of which is Meme: the first AI Agent concept MemeCoin GOAT has quickly gained popularity and achieved a valuation of 1.4 billion USD, successfully sparking the AI Meme craze.
The research and topics around AI + Web3 are equally hot, from AI + Depin to AI Memecoin and now to the current AI Agent and AI DAO, the FOMO sentiment can no longer keep up with the speed of the new narrative rotation.
AI + Web3, this combination of terms filled with hot money, opportunities, and future fantasies is inevitably seen as a marriage arranged by capital. It seems difficult for us to discern whether beneath this splendid robe lies the playground of speculators or the eve of an explosion at dawn?
To answer this question, a key consideration for both parties is whether it would be better with the other. Can they benefit from each other's models? In this article, we also attempt to examine this pattern by standing on the shoulders of our predecessors: how Web3 can play a role in various aspects of the AI technology stack, and what new vitality AI can bring to Web3.
Part.1 What opportunities does Web3 have under the AI stack?
Before diving into this topic, we need to understand the technology stack of AI large models:
To put the entire process in more straightforward terms: "Large models" are like the human brain. In the early stages, this brain belongs to a newborn baby who has just come into the world and needs to observe and absorb massive amounts of information from the surroundings to understand this world. This is the "data collection" stage. Since computers do not possess human senses like vision and hearing, before training, the vast amounts of unlabelled external information need to be transformed into a format that computers can understand and use through "preprocessing."
After inputting the data, the AI constructs a model with understanding and prediction capabilities through "training", which can be seen as the process of an infant gradually understanding and learning about the outside world. The parameters of the model are like the language abilities that the infant continuously adjusts during the learning process. When the learning content begins to become specialized, or when feedback is received from communication with others and corrections are made, it enters the "fine-tuning" stage of the large model.
As children gradually grow up and learn to speak, they can understand meanings and express their feelings and thoughts in new conversations. This stage is similar to the "reasoning" of large AI models, which can predict and analyze new language and text inputs. Infants express feelings, describe objects, and solve various problems through language abilities, which is also akin to how large AI models apply reasoning to various specific tasks after completing training and being put to use, such as image classification and speech recognition.
The AI Agent is more akin to the next form of large models - capable of independently executing tasks and pursuing complex goals, possessing not only the ability to think but also to remember, plan, and interact with the world using tools.
Currently, in response to the pain points of AI across various stacks, Web3 has preliminarily formed a multi-layered, interconnected ecosystem that encompasses all stages of the AI model process.
1. Basic Layer: Computing Power and Data's Airbnb
Hashrate
Currently, one of the highest costs of AI is the computational power and energy required for training and inference models.
One example is that Meta's LLAMA3 requires 16,000 H100 GPUs produced by NVIDIA(, which is a top-tier graphics processing unit designed specifically for artificial intelligence and high-performance computing workloads. Training takes 30 days to complete. The unit price of the latter's 80GB version ranges from $30,000 to $40,000, requiring an investment of $400 million to $700 million in computing hardware), including GPUs and network chips(. Additionally, the monthly training consumes 1.6 billion kilowatt-hours, with energy expenditures nearing $20 million per month.
The unburdening of AI computing power is precisely one of the earliest intersections of Web3 and AI - DePin) decentralized physical infrastructure network(. Currently, the DePin Ninja data website has displayed over 1,400 projects, among which representative projects for GPU computing power sharing include io.net, Aethir, Akash, Render Network, and more.
The main logic is that the platform allows individuals or entities with idle GPU resources to contribute their computing power in a permissionless decentralized manner, similar to an online marketplace for buyers and sellers like Uber or Airbnb, increasing the utilization of underused GPU resources, and end users thereby obtain more cost-effective and efficient computing resources; at the same time, the staking mechanism ensures that if there are violations of quality control mechanisms or network interruptions, resource providers face corresponding penalties.
Its characteristics are:
Gather idle GPU resources: The suppliers are mainly excess computing power resources from third-party independent small and medium-sized data centers, cryptocurrency mining farms, and other operators, with mining hardware based on the PoS consensus mechanism, such as FileCoin and ETH miners. Currently, there are also projects dedicated to launching devices with lower entry barriers, such as exolab, which utilizes local devices like MacBook, iPhone, and iPad to establish a computing power network for running large model inference.
Facing the long tail market of AI computing power:
a. "From a technical perspective," decentralized computing power markets are more suitable for inference steps. Training relies heavily on the data processing capabilities brought by ultra-large cluster scale GPUs, while inference requires relatively lower GPU computing performance, such as Aethir focusing on low-latency rendering work and AI inference applications.
b. "From the demand side perspective," small to medium computing power demanders will not train their own large models individually, but will instead choose to optimize and fine-tune around a few leading large models, and these scenarios are naturally suitable for distributed idle computing resources.
Data
Data is the foundation of AI. Without data, computation is as useless as a floating duckweed, and the relationship between data and models is akin to the saying "Garbage in, Garbage out"; the quantity of data and the quality of input determine the final output quality of the model. For the training of current AI models, data determines the model's language ability, comprehension ability, and even values and human-like performance. Currently, the data demand dilemma for AI is primarily focused on the following four aspects:
Data hunger: AI model training relies on a large amount of data input. Public information shows that OpenAI trained GPT-4 with a parameter count reaching trillions.
Data Quality: With the integration of AI in various industries, the timeliness of data, the diversity of data, the professionalism of vertical data, and the incorporation of emerging data sources such as social media sentiment have all raised new demands for its quality.
Privacy and compliance issues: Currently, various countries and enterprises are gradually recognizing the importance of high-quality datasets and are imposing restrictions on dataset crawling.
High costs of data processing: large data volume and complex processing. Public information shows that more than 30% of the R&D costs of AI companies are used for basic data collection and processing.
Currently, web3 solutions are reflected in the following four aspects:
Allowing users who make real contributions to also participate in the value creation brought by data, and to obtain more private and valuable data from users in a low-cost manner through distributed networks and incentive mechanisms, is the vision of Web3.
Grass is a decentralized data layer and network, where users can run Grass nodes to contribute idle bandwidth and relay traffic to capture real-time data from the entire internet and receive token rewards;
Vana introduces a unique data liquidity pool )DLP( concept, where users can upload their private data ) such as shopping records, browsing habits, social media activities, etc. ( to a specific DLP and flexibly choose whether to authorize these data for use by specific third parties;
In PublicAI, users can use )Web3 as a classification tag on social platforms and @PublicAI to achieve data collection.
Currently, Grass and OpenLayer are both considering adding data annotation as a key step.
Synesis proposed the concept of "Train2earn", emphasizing data quality, where users can earn rewards by providing labeled data, annotations, or other forms of input.
The data labeling project Sapien gamifies the labeling tasks and allows users to stake points to earn more points.
The currently common privacy technologies in Web3 include:
Trusted Execution Environment ) TEE (, such as Super Protocol;
Fully Homomorphic Encryption ) FHE (, for example BasedAI, Fhenix.io or Inco Network;
Zero-knowledge technology ) zk (, such as Reclaim Protocol using zkTLS technology, generates zero-knowledge proofs of HTTPS traffic, allowing users to securely import activity, reputation, and identity data from external websites without exposing sensitive information.
However, the field is still in its early stages, and most projects are still exploring. One current dilemma is that the computing costs are too high. Some examples are:
The zkML framework EZKL takes about 80 minutes to generate a proof for a 1M-nanoGPT model.
According to data from Modulus Labs, the overhead of zkML is more than 1000 times higher than pure computation.