The current AI landscape is dominated by centralized, closed source and oligopolistic tech giants. A small number of corporations control the most high performant models and this is largely due to the extreme centralizing forces that are conducive to model development and inference.
In the first part of this two-part series, we walk through the benefits of decentralized AI and practical challenges impeding its deployment. In the next part, we shall explore the different techniques being explored to make decentralized AI a reality and benchmark Oyster, a TEE-based coprocessor against popular zkML libraries.
How AI models work today?
Creating a machine learning model typically involves three main phases: pre-training, fine-tuning, and inference. These phases are crucial for the development of a robust and accurate model that can generalize well to new, unseen data.
The pre-training phase
During the pre-training phase, the model is trained on a large, general dataset. This dataset is not specific to the task the model will eventually perform but is instead designed to help the model learn a wide array of features and patterns. For example, in the case of language models, this might involve learning language structure, grammar, and a broad vocabulary from a massive corpus of text. The goal here is to develop a model that has a good understanding of the underlying structure of the data it will work with, whether that's text, images, or any other form of data.
Within the pre-training phase, there exist a number of centralizing forces:
1. Data collection and sorting - The pre-training phase is critically dependent on aggregating extensive datasets from varied sources, including literature, digital articles, and specialized databases. Industry behemoths, such as Google, have historically capitalized on user-generated content to craft models of unparalleled efficacy—a practice that continues today with entities like Microsoft and OpenAI, which secure top-tier data through exclusive alliances and proprietary platforms. This aggregation of capabilities within a handful of corporations contributes to a pronounced centralization within the AI sector.
Moreover, the reliance on proprietary datasets for model training introduces significant ethical considerations and the risk of perpetuating biases. AI algorithms, by design, derive their operational paradigms from the underlying data, rendering them susceptible to embedding and replicating any inherent biases. This scenario underscores the imperative for meticulous scrutiny and ethical oversight in the development process, ensuring that models reflect equitable and intentional patterns and associations.
2. Resource requirements - It has been established that the efficacy of models enhances logarithmically with the volume of training data, indicating that models benefiting from the most extensive GPU compute cycles generally perform the best. This dynamic introduces a significant centralizing influence in the pre-training phase, driven by the economies of scale and productivity advantages held by major tech and data corporations. This trend is evident in the dominance of industry giants such as OpenAI, Google, Amazon, Microsoft, and Meta in the field. These corporations own and operate the majority of data centers worldwide and have access to the newest, top of the line hardware coming out of NVIDIA.
The fine-tuning phase
Once the model has been pre-trained, it undergoes fine-tuning. In this phase, the model is further trained on a smaller, task-specific dataset. The purpose is to adjust the weights and features that the model has learned during pre-training to make it better suited for the specific task at hand. This might involve teaching a language model to understand medical terminology or training an image recognition model to differentiate between different species of birds.
The fine-tuning phase allows the model to specialize and improve its performance on tasks that are of interest to the end-user. Again within the fine-tuning phase there are a number of centralizing forces, the most important one being closed source models & verifiability.
In the fine-tuning stage, the model's parameters are refined and set, shaping its functionality and performance. The prevailing trend towards proprietary AI models, such as OpenAI's GPT series and Google's Gemini, means that the inner workings and parameters of these models are not disclosed. As a result, when users request inferences, they are unable to verify if the responses truly come from the model they believe they are interacting with.
This lack of transparency can disadvantage users, particularly in scenarios where trust and verifiability are paramount. For instance, in healthcare, where AI models might assist in diagnosing diseases or recommending treatments, the inability for practitioners to confirm the source and accuracy of the model's inference could lead to mistrust or even misdiagnosis. If a medical professional cannot ascertain that the AI's recommendation is based on the most reliable and updated model, the consequences could directly impact patient care and outcomes, underscoring the critical need for transparency and accountability in AI deployments.
The inference phase
The inference phase is where the model is actually put to use. At this point, the model has been trained and fine-tuned, and it's ready to make predictions on new data. In the case of an AI model, this could mean answering questions, translating languages, or providing recommendations. This phase is all about applying the trained model to real-world problems and is typically where the value of the model is realized.
Within the inference phase, the factors leading to centralization are:
1. Access - Centralized frontends for AI model access pose a risk by potentially cutting users off from API or inference access. When a few entities control these gateways, they can, at their discretion, deny access to essential AI services for various reasons, including policy changes or disputes. This centralization highlights the need for decentralized approaches to ensure broader, more resilient access to AI technologies, mitigating risks of censorship and access inequity.
2. Resource requirements - The cost of inference in AI, particularly for tasks requiring extensive computational resources, acts as a centralizing force within the technology sector. High inference costs mean that only large companies with substantial financial resources can afford to run advanced AI models at scale. This financial barrier limits the ability of smaller entities or individual developers to leverage the full potential of cutting-edge AI technologies.
As a result, the landscape becomes increasingly dominated by a few powerful players, stifling competition and innovation. This centralization not only impacts the diversity of AI development but also limits access to AI's benefits to a select group of well-funded organizations, creating a significant imbalance in the technological ecosystem.
Within the centralized AI landscape, several recurring themes emerge, particularly regarding the evolution of Web2 companies. Initially founded as open networks, these entities often shift their focus towards maximizing shareholder value. This transition frequently leads to the closing off of their networks, with algorithms being adjusted to discourage out-linking, often acting contrary to the best interests of their users.
This misalignment between company incentives and user needs typically occurs as the organization matures and secures external funding. In fact, we have already seen this phenomenon with OpenAI, which began as a non-profit with the ambition to democratize AI access, illustrating how such shifts in focus can manifest in the industry. It is easy to blame these individual companies, but we believe this reflects a systemic issue stemming from the centralizing forces within the tech industry, where the concentration of power and resources often leads to misalignments between company incentives and broader user needs.
Possible future - AI 🤝 Web 3
Crypto furnishes artificial intelligence with a foundation that allows for seamless, open, and secure exchanges of information and value. Blockchain technology delivers a clear and accountable system for managing transactions and recording data. At the convergence of crypto and AI, numerous opportunities arise where both domains can mutually enhance and benefit from each other's capabilities.
Incentive aligned
Decentralized computing holds significant value in the pre-training and fine-tuning phases of model development. Foundation models, known for their extensive need for GPU compute cycles, usually run these processes in centralized data centers. Decentralised Physical Infrastructure Networks (DePIN) can offer decentralized, permissionless computing access. Through crypto economic incentives, softwares can autonomously compensate for hardware usage without the need for a centralized governing entity. This enables the users of the network to control the network, aligning incentives and ensuring both data and model providers are sufficiently compensated for their contribution.
Verifiability
The current landscape of artificial intelligence infrastructure predominantly leans towards proprietary models, wherein users are required to place their trust in the inference providers to execute their queries through the designated model and generate a legitimate output. In this context, cryptographic proving systems emerge as a pivotal technology, offering a mechanism to validate model outputs on the blockchain. This process enables a user to submit a query, which the inference provider processes using the agreed-upon model, subsequently producing an output accompanied by a cryptographic proof. This proof serves as verifiable evidence that the query was indeed processed through the specified model.
The primary aim of these initiatives is to offload the burden of heavy computational tasks to off-chain environments, while ensuring that the results can be verified on-chain. This approach not only alleviates the computational load on the blockchain but also introduces a layer of transparency and trustworthiness by providing immutable proof of the accuracy and completion of the computations conducted off-chain.
Incorporating these cryptographic proofs into the AI model validation process addresses several key concerns associated with closed source AI systems. It mitigates the risk of opaque or unverified computations, enhances the integrity of the computational process, and fosters a trust-based relationship between the users and the inference providers. Furthermore, this methodology aligns with the broader movement towards decentralized and transparent systems, echoing the foundational principles of blockchain technology.
Composability
One of the primary advantages of decentralized finance and blockchain networks has been the composability they enable. Composability allows for the proliferation of "Money Legos" in DeFi, i.e. the combination of different protocols and outputs, to create new applications. Although this composability introduces new forms of risk into the system, it also simplifies application development for developers, increasing innovation and the speed of development, and enables for easier user experience and convenience.
Similar to how crypto has enabled composability for financial products, it will also create composable AI networks and applications, by acting as a permissionless and trustless layer on top of which AI modules can build on and work independently while staying interconnected with other modules to form a network capable of providing various services. Through blockchain network effects and crypto, decentralized AI projects and applications can connect with each other to complete the AI stack.
For example, pre-processed data using Akash or Filecoin can be used to train models using Marlin, Gensyn or Together. Following fine-tuning, these trained models can respond to user queries (inference) through Bittensor. Although seemingly more complicated, an end user would only interact with a single front-end, while developers and applications benefit off of building on top of various stacks and applications.
Another important aspect of composability enabled through decentralized networks is data composability. As users become increasingly interested in owning the data they generate, and demand that they are able to carry it around from AI protocol to AI protocol, they will require that their data is not constrained within walled gardens. Decentralized and open source AI applications enable data portability.
Data protection
Decentralized compute, augmented with external data and privacy solutions, offers users more autonomy over their data, making it a more compelling option than its centralized counterparts. In particular, methods such as Fully Homomorphic Encryption (FHE), allow for computations on encrypted data without needing to decrypt it first.
Through FHE, the training of machine learning models using encrypted data sets, thereby maintaining data privacy and security throughout the entire training process is enabled. This provides an end-to-end security solution with robust cryptographic assurances, allowing for privacy-preserving model training, particularly in edge networks, and allows for the development of AI systems that protect user privacy while leveraging advanced AI capabilities.
The role of FHE extends to running large language models securely on encrypted data in cloud environments. This not only safeguards user privacy and sensitive information but also enhances the ability to run models on applications with inherent privacy. As the integration of AI in various sectors, especially sensitive ones like finance, intensifies, the need for technologies like FHE that protect against potential information leaks becomes paramount.
Auto-upgradeability
AI can be used to maintain, update and auto-upgrade smart contracts based on a set of changes and conditions. For example, AI can be used on the protocol side to adjust risk parameters, based on changes in risk and other market conditions. A commonly given example here is in the case of money markets. Money markets currently rely on external organizations or DAO decisions to adjust risk parameters for lending and borrowing on assets. AI agents could streamline updating and upgrading certain parameters instead, which would be a stark improvement when compared to humans and DAOs, which can be slow and inefficient.
Challenges of decentralized AI
Decentralized AI presents a range of challenges, particularly in balancing the open-source nature of cryptography with the security concerns in AI, as well as the computational requirement for AI. In cryptography, open source is crucial for ensuring security, but in AI, making a model or its training data publicly available increases its vulnerability to adversarial machine learning attacks. This dichotomy poses a significant challenge in developing applications that leverage both technologies.
Furthermore, applications of AI in blockchain, such as AI-driven arbitrage bots, prediction markets, and decision-making mechanisms, raise issues around fairness and manipulation. AI can potentially improve efficiency and decision-making in these areas, but there's a risk of AI not fully grasping the nuances of human-driven market dynamics, leading to unforeseen consequences.
Another area of concern is using AI as an interface in crypto applications. While AI can assist users in navigating the complex world of cryptocurrency, it also poses risks like being susceptible to adversarial inputs or causing users to over-rely on AI for critical decisions. Moreover, integrating AI into the rules of blockchain applications, such as DAOs or smart contracts, is fraught with risks. Adversarial machine learning could exploit AI model weaknesses, resulting in manipulated or incorrect outcomes. Ensuring that AI models are accurate, user-friendly, and secure from manipulation is a significant challenge.
Additionally, the integration of AI with ZK proofs or MPC is not only computationally intensive but also faces hurdles such as high computational costs, memory limitations, and model complexity. The current state of tooling and infrastructure for zero-knowledge machine learning (zkML) is still underdeveloped, and there's a lack of skilled developers in this niche area. These factors contribute to the considerable amount of work required before zkML can be implemented at a scale suitable for consumer products.
Conclusion
Balancing decentralization and trust, especially in cases where AI uses trusted hardware or specific data governance models, remains a critical challenge in maintaining the decentralized ethos of blockchain while ensuring AI system reliability. In the next part of this article, we shall delve deeper into the technologies that can power decentralized AI and how Marlin’s infrastructure plays a crucial role in making it a reality.
Follow our official social media channels to get the latest updates as and when they come out!
Twitter | Telegram Announcements | Telegram Chat | Discord | Website