Installation
Learn how to install, set up and configure OM1 on your robots.
System Requirements
Operating System
- macOS 12.0+
- Linux (Ubuntu 20.04+)
Hardware
- Memory (RAM): 8GB
- Storage: 16GB
- Camera, speakers, microphone etc as robots sensors
Software
- Python 3.10+
- uv 0.6.2
- Openmind API key
Prerequisites
Ensure you have the following installed on your machine:
Package Manager
UV (A Rust and Python package manager)
PortAudio Library
This will let you speak to the LLM and it will generate voice outputs. On Mac and Linux, you need portaudio
.
ffmpeg
FFmpeg is the leading multimedia framework, able to decode, encode, transcode, mux, demux, stream, filter and play pretty much anything that humans and machines have created.
Installation and Setup
- Clone the repository
Run the following commands to clone the repository and set up the environment:
- Set the configuration variables
Locate the config
folder and add your Openmind API key in /config/spot.json
. If you do not already have one, you can obtain a free access key at https://portal.openmind.org/. Note: Using the placeholder key openmind-free will generate errors.
- Run the Spot Agent
Run the following command to start the Spot Agent:
- WebSim to check input and output
Go to http://localhost:8000 to see real time logs along with the input and output in the terminal. For easy debugging, add --debug
to see additional logging information.
Congratulations! - you just got started with OM1 and can now explore its capabilities.
Some necessary packages will be installed during this process, the first time you run the command. This might take a little time. Please be patient. Then you will see the system come to life:
Explanation of the log data
The response above provides insight into how the spot agent processes its environment and decides on its next actions.
- First, it detects a person using vision.
- Next, it decides on a friendly action (dancing and speaking).
- Expresses emotions via facial displays.
- Logs latency and processing times to monitor system performance.
- Communicates with an external AI API for response generation.
Overall, the system follows a predefined behavior where spotting a person triggers joyful interactions, driven by the LLM-assisted decision-making process.
Was this page helpful?