Get in Touch

Course Outline

Introduction to Huawei CloudMatrix

  • Overview of the CloudMatrix ecosystem and deployment flow
  • Supported models, formats, and deployment modes
  • Typical use cases and supported chipsets

Preparing Models for Deployment

  • Exporting models from training tools (MindSpore, TensorFlow, PyTorch)
  • Using ATC (Ascend Tensor Compiler) for format conversion
  • Distinctions between static and dynamic shape models

Deploying to CloudMatrix

  • Creating services and registering models
  • Deploying inference services via UI or CLI
  • Configuring routing, authentication, and access control

Serving Inference Requests

  • Comparing batch and real-time inference flows
  • Implementing data preprocessing and postprocessing pipelines
  • Invoking CloudMatrix services from external applications

Monitoring and Performance Tuning

  • Accessing deployment logs and tracking requests
  • Managing resource scaling and load balancing
  • Optimizing latency and throughput

Integration with Enterprise Tools

  • Connecting CloudMatrix with OBS and ModelArts
  • Leveraging workflows and model versioning
  • Implementing CI/CD for model deployment and rollback

End-to-End Inference Pipeline

  • Deploying a complete image classification pipeline
  • Benchmarking and validating accuracy
  • Simulating failover scenarios and system alerts

Summary and Next Steps

Requirements

  • Familiarity with AI model training workflows
  • Experience with Python-based ML frameworks
  • Basic understanding of cloud deployment concepts

Target Audience

  • AI operations teams
  • Machine learning engineers
  • Cloud deployment specialists working with Huawei infrastructure
 21 Hours

Number of participants


Price per participant

Testimonials (2)

Upcoming Courses

Related Categories