White Paper
  • 1. Introduction
  • 2. The Volunteer Computing Network
  • 3. Training Platform
  • 4. Inference Platform
  • 5. General Features
  • 6. NetMind Chain
  • 7. NetMind Token
  • 8. User types
  • 9. Future Direction
  • 10. Conclusion
Powered by GitBook
On this page
  • 4.1 Inference Platform Introduction
  • 4.2 Key features and technical details of the inference platform

4. Inference Platform

Previous3. Training PlatformNext5. General Features

Last updated 1 year ago

4.1 Inference Platform Introduction

The inference platform is designed to complement the training platform by providing a seamless way for users to deploy and run inference on their own models, models created by others, and open source off-the-shelf models.

The platform's innovative architecture and features allow for easy integration, production level performance, and reduced costs, making it an ideal solution for organizations seeking to harness the power of AI in a decentralized, secure, and scalable manner.

4.2 Key features and technical details of the inference platform

  1. Model Deployment: Users can deploy their trained AI models on the inference platform, making them accessible to other users and applications via API. The platform supports containerization, allowing for the packaging of AI models and their dependencies into lightweight, portable containers that can be easily deployed across the network.

  2. Scalability: The inference platform is designed to handle varying workloads, automatically scaling up or down based on demand. It employs distributed computing and load balancing techniques to distribute the inference workload across multiple GPUs in the network, ensuring efficient use of resources and minimal latency.

  3. Cost Optimization: By leveraging the decentralized nature of the platform and the idle resources of participants, the inference platform provides cost-effective access to computing power for running AI models. This reduces operational expenses for users while maintaining high performance. NetMind Power's resource allocation algorithm dynamically assigns inference tasks to the most suitable GPUs in the network, taking into account factors such as computational capacity, latency, and availability, to optimize costs and resource utilization.

  4. Security: The inference platform employs state-of-the-art security measures to protect both the AI models and the data being processed. This includes techniques such as encryption, secure enclaves for model execution, and secure multi-party computation to maintain data privacy and model integrity during the inference process.

When there is a large amount of work waiting to be processed, the system will scale up according to demand as represented by the “+” sign. On the inverse, when demand is low, the system will scale down accordingly, as represented by the “-” sign.
A secure enclave provides CPU hardware-level isolation and memory encryption on every server, by isolating application code and data from anyone without privileges, and encrypting its memory.