Open Interpreter
From Notes_Wiki
Home > Local system based AI tools > Open Interpreter
As a non-root / local user we can install interpreter via:
pip install open-interpreter
This install interpreter in ~/.local/bin path. Then if this is already in path we can run interpreter via:
interpreter --local
Ideally for this local Ollama should be setup with at least one model eg phi4 for it to work properly.
Refer: