top of page

The evolution of AI decision making: From centralized to distributed

Artificial intelligence (AI) is rapidly transforming nearly every industry, from transportation to healthcare to agriculture. AI-powered systems are getting incredibly good at understanding human language, answering questions, making recommendations, and even generating original content.

Behind the scenes, most of this AI magic is made possible by large language models (LLMs) like GPT. These foundation models are trained on massive datasets using hundreds of billions of parameters on powerful cloud computing infrastructure. They develop a broad understanding of language that allows them to perform various natural language tasks.

The most advanced LLMs are developed by big tech companies like OpenAI, Google, and Microsoft. They operate server farms with thousands of high-end GPUs and TPUs that can handle the intense computational requirements of training large neural networks.

This centralized approach to developing and deploying AI has some important advantages. The cloud offers practically unlimited compute scalability for training state-of-the-art models. Centralization also simplifies model development, maintenance, and updates for tech companies.

However, as AI becomes more pervasive and gets embedded into all facets of our lives, maintaining all AI in the centralized cloud creates some emerging issues around privacy, security, accuracy, and efficiency. These issues are driving an important evolution in AI architectures from centralized to distributed.

  • Privacy and Security Risks: With centralized AI, user data and model inputs get sent to the cloud to generate outputs. This creates privacy and security risks, as personal data is concentrated into large centralized repositories. High-profile data breaches have underscored these dangers. As people become more privacy conscious, there is a growing preference for solutions that keep data local and make decisions on-device without sending data to the cloud. Federated learning, split learning, and on-device inference offer more secure and privacy-preserving alternatives.

  • Lack of Personalization: Centralized models create one-size-fits-all outputs. But for many applications, it is important to provide personalized and context-aware results tailored to each user. For example, speech recognition for mobile assistants needs to adapt to an individual's voice patterns and dialect. Centrally-hosted models often lack the local signals and adaptability to provide truly personalized results.

  • Accuracy Limitations: Centralized AI models lack context, which can limit their accuracy on many tasks. For example, autonomous vehicles need to perceive and understand objects in their local surroundings. For computer vision applications, localized training on specific environmental conditions can improve accuracy over centralized models.

  • The inefficiency of constant cloud connection: Having to send data back and forth to the cloud for processing creates latency, consumes bandwidth, and is costly. Inference done locally on device or on edge servers nearer to the user is more efficient, providing quicker response times and lowering data transfer costs.

Given these considerations, we will see AI systems evolve to adopt more distributed architectures. However, centralized and distributed AI both have their own advantages. The optimal architecture will be a hybrid model that combines both approaches:

  • Model training/development will remain largely centralized. The cloud offers unmatched scale for training the most advanced models with massive datasets and compute.

  • Inference will shift towards on-device and edge servers close to the user. This distributes compute across endpoints for lower latency, privacy, and personalization.

  • - Some model fine-tuning may occur at the edge to adapt models to local data and conditions.

  • Core model components will be developed centrally and then compressed or partitioned to distribute inference.

  • Centralized and distributed models will intelligently coordinate with aggregation frameworks to benefit from collective learning.

Enterprises will need to evolve their AI infrastructure to support this hybrid approach. This includes:

  • Investing in on-device capabilities and edge infrastructure.

  • Developing compression techniques to deploy production models locally.

  • Building privacy-preserving analytics and transfer learning capabilities.

  • Creating frameworks to manage models and coordination between the cloud, edge, and endpoints.

Consumer tech giants like Google and Apple are already pioneering this hybrid approach. Apple is performing more Siri processing on-device to improve latency, privacy, and reliability. Google is pushing BERT models to Android phones to enable sentence completion with Gboard even without an internet connection. Waymo trains its autonomous vehicle models in the centralized cloud but performs key inference locally for each vehicle.


As AI becomes ubiquitous, distributed and hybrid AI architectures will be critical for improving privacy, security, accuracy, and efficiency. The exact balance between centralized and distributed components may vary across different applications and use cases based on the tradeoffs. But the overarching trend is clear – AI is getting localized.

Commentaires


bottom of page