What is OpenMind (OM1)

OM1 allows AI agents to be configured and then deployed in both the digital and physical world. You can create one AI agent and run it in the cloud but also on physical robot hardware such as Quadrupeds, and, soon, TurtleBot 3 and Humanoids.

For example, an AI agent built on OM1 can ingest data from multiple sources (the web, X/Twitter, cameras, and LIDAR) and can then Tweet and explore your house, shake your hand, or talk to you. In another example, with OM1, you can talk with OpenAI’s gpt-4o and literally shake hands with it.

Capabilities of OM1

Simple, modular architecture

Easy-to-understand, independent components that work together seamlessly.

All python

Independent modules that are easy maintain, and extend.

Easy to add new data inputs

Seamlessly integrate new data without major changes to the existing structure.

Easy to support new hardware

via plugins for API endpoints and specific robot hardware

Can be connected to:

ROS2, Zenoh, and CycloneDDS

Includes a simple web-based debug display

to watch the system work (WebSim at http://localhost:8000)

Preconfigured endpoints

for Voice-to-Speech, OpenAI’s gpt-4o, DeepSeek, and multiple VLMs

Was this page helpful?