Назад към всички

Edge Deployment Skill

// ML model optimization and deployment on robot edge devices (Jetson, embedded)

$ git log --oneline --stat
stars:384
forks:73
updated:March 4, 2026
SKILL.mdreadonly
SKILL.md Frontmatter
nameEdge Deployment Skill
descriptionML model optimization and deployment on robot edge devices (Jetson, embedded)
slugedge-deployment
categoryDeployment
allowed-toolsBash,Read,Write,Edit,Glob,Grep

Edge Deployment Skill

Overview

Expert skill for optimizing and deploying machine learning models on robot edge devices including NVIDIA Jetson and embedded systems.

Capabilities

  • Configure TensorRT optimization for NVIDIA Jetson
  • Set up ONNX model conversion and optimization
  • Implement INT8 and FP16 quantization
  • Configure DeepStream for video analytics
  • Set up CUDA graph optimization
  • Implement model pruning and distillation
  • Configure DLA (Deep Learning Accelerator) deployment
  • Set up multi-stream inference
  • Implement ROS2 inference nodes
  • Profile and benchmark on target hardware

Target Processes

  • nn-model-optimization.js
  • object-detection-pipeline.js
  • rl-robot-control.js
  • field-testing-validation.js

Dependencies

  • TensorRT
  • ONNX Runtime
  • NVIDIA Jetson SDK
  • DeepStream

Usage Context

This skill is invoked when processes require deploying ML models on edge devices with optimized inference performance.

Output Artifacts

  • TensorRT engine files
  • ONNX optimized models
  • Quantization configurations
  • DeepStream pipeline configs
  • Inference benchmark reports
  • ROS2 inference node implementations