v1.0.0 MVP Release

The Local-First
Autonomous Computer Operator

OpenPC is a Python-based autonomous AI agent that runs entirely on your local machine. It uses a Synchronous Lock-Step State Machine to perceive your screen, plan complex tasks, and execute actions—without ever sending data to the cloud.

Powered by local inferences:
GUI-Actor Qwen 2.5 Llama 3.2 LLaVA 7b

Our Mission: Democratize AI Autonomy

We believe autonomous AI should be accessible, private, and user-controlled. OpenPC challenges the cloud-dominated landscape by proving that sophisticated AI agents can run entirely on local hardware, giving users complete ownership of their digital workflows.

Research Foundations

  • Lock-Step State Machine Architecture — Synchronous perception, planning, and execution cycles prevent race conditions and ensure deterministic behavior
  • Coordinate-Free Visual Grounding — GUI-Actor's heatmap approach eliminates brittle DOM parsing and UI element dependencies
  • Multi-Model Collaboration — Specialized models for vision, planning, verification, and clarification create robust decision-making pipelines
  • Local-First Design Philosophy — Zero telemetry, complete data privacy, and offline operation as core architectural principles

Impact & Vision

100%
Local Processing
0
API Calls
Task Possibilities

By 2025, we aim to make local AI autonomy the standard for personal computing, enabling anyone to automate their digital life without compromising privacy or control.

Capabilities

Zero Telemetry. Unlimited Potential.

100% Local

Vision via GUI-Actor and Hugging Face. Planning via Qwen2.5 and Llama-3.2 through Ollama. Zero API tokens required.

Full Autonomy

No hardcoded shortcuts. The planner decides every step from the goal and the screen context using its own intelligence.

Dual Execution

Standard GUI mode using PyAutoGUI or "God Mode" utilizing direct shell and HTTP commands for blazing fast navigation.

Self-Correcting

Failed actions are logged and never repeated. Stuck detection forces strategy changes autonomously.

Join the core

Shape the Future of Autonomy

Deep AI Remote

AI Core Researcher

Optimize vision grounding and strategic reasoning loops. Work directly with GUI-Actor and Qwen2.5 fine-tuning.

View Details →
Systems Remote

Systems Engineer

Master of GPU offloading and local state orchestration. help us squeeze every millisecond out of consumer hardware.

View Details →
UX/UI Remote

Product Designer

Craft the next generation of human-AI interfaces. Minimalist, high-performance, and deeply intuitive design.

View Details →

Want to see all open roles?

Explore Full Careers Page
openpc-agent — cmd
The Brain Stack

Purpose-Built Local Intelligence

Vision ~8GB VRAM

GUI-Actor-7B

Chosen for its best-in-class, coordinate-free heatmap grounding. It maps screenshots and goals to precise screen pixels instantly, avoiding the overhead of heavy DOM manipulation or UI element trees.

Planner ~9GB VRAM

Qwen2.5 14B

Selected for exceptional strategic reasoning capabilities. It is the perfect engine for reading associative memory, evaluating vision outputs, and picking precise atomic actions with complete autonomy.

Verifier ~5GB VRAM

LLaVA 7B

Integrated for fast, reliable visual validation. Highly optimized for before/after screenshot comparisons, instantly catching wrong-UI interactions and providing actionable failure logs to the planner.

Clarifier ~2GB VRAM

Llama 3.2 3B

Utilized as an ultra-lightweight disambiguation gate. Extremely efficient at pre-loop analysis, asking clarifying questions only when a goal lacks clear deterministic intent.

Hardware Requirements

Built for Enthusiast Hardware

GPU Constraints

An NVIDIA GPU with 16 GB+ VRAM is strongly recommended (system tested extensively on RTX 4060 Ti 16GB). OpenPC forces all inferences onto the GPU for rapid, lock-step processing.

Local Stack Prerequisites

Requires Ollama running locally (with num_gpu bypassed) to host the Planner, Verifier, and Clarifier. Vision is handled through a Hugging Face PyTorch installation.