Openpi-flash: Real-time inference engine for openpi
Real-time inference engine for openpi. Contribute to Hebbian-Robotics/openpi-flash development by creating an account on GitHub.
Opening excerpt (first ~120 words) tap to expand
openpi-flash Real-time inference engine for openpi. Optimized for low-latency policy serving over QUIC and WebSocket. Deploy on AWS EC2 (Docker) or Modal. openpi-flash gives robots the task-specific brain they need to actually ship into production environments like fulfillment, retail, and other commercial deployments where a general-purpose policy isn't enough. Key features If you're coming from the upstream openpi server, these are the key additions openpi-flash includes: Planner module — specialize a general VLA to your task distribution with a fine-tuned pi0.5 subtask generator using language coaching. Loads alongside the action policy, and jointly conditions the high level task prompts with the subtask steering before action inference. See Subtask generation (planner).
…
Excerpt limited to ~120 words for fair-use compliance. The full article is at GitHub.