Portal Foundation

PortalOS

Your locally-hosted AI command-line interface for enhanced privacy and productivity.
PortalOS project image

PortalOS: Your Local AI Command-Line Interface


What Is It

PortalOS is a privacy-first, locally-hosted AI command-line interface that enables users to control their system, search files, process documents, and automate workflows through natural language. With support for powerful models like LLaMA, Mistral, and Phi via Ollama or llama.cpp, PortalOS runs entirely on your machine — combining AI capabilities with full data control.


Vision

We believe everyone should be able to use AI without compromising their privacy.

PortalOS empowers users to run high-performance AI tasks directly on their local device — whether that's file operations, system commands, document summarisation, or personal knowledge management — without relying on external servers or cloud APIs.


Why It Matters

Most AI tools are:

  • Cloud-based and privacy-invasive
  • Resource-hungry and overcomplicated
  • Locked into black-box workflows with limited control

PortalOS flips that paradigm by:

  • Running entirely locally with no cloud data leakage
  • Supporting open-source models tailored to your hardware
  • Providing powerful command-line automation and smart file/document control
  • Giving developers, researchers, and privacy advocates full transparency and flexibility

Problem & Solution

The Problem

Users face a tradeoff between using AI and protecting their data. Cloud-based AI tools expose sensitive files and interactions to external systems, while local alternatives are often fragmented, technical, or limited in functionality.

Our Solution

PortalOS is an AI operating layer that:

  • Works fully offline
  • Interfaces with your local system securely
  • Allows natural-language interaction with files, documents, and tools
  • Supports multiple backends and hardware setups
  • Enables task automation, document summarisation, and semantic search

How It Works

  1. Choose your model backend (Ollama, llama.cpp, or custom)
  2. Launch PortalOS via terminal and load your preferred model
  3. Input commands in plain language (e.g., "Summarise this PDF", "Search all files mentioning 'quarterly earnings'")
  4. PortalOS processes the request locally, executes system commands or parses documents
  5. You receive a private, context-aware AI output in your terminal

Key Features

💬 Natural Language Interface

  • Intuitive, context-aware CLI
  • Everyday language command parsing
  • Command history + smart interpretation
  • Custom command creation and automation

🔐 Local Processing & Privacy

  • 100% local execution
  • No external API calls or cloud dependencies
  • Secure file handling and encryption
  • Total data ownership and visibility

🧠 Flexible Model Options

  • Ollama backend: For modern systems with 8GB+ RAM
  • llama.cpp backend: Optimised for older or resource-constrained machines
  • Custom model support: For fine-tuned or domain-specific deployments
  • Automatic hardware-aware model selection

📄 Advanced Document Processing

  • PDF, Markdown, TXT, and more
  • OCR support via Tesseract
  • Intelligent summarisation and key point extraction
  • Document comparison, insight extraction, and analysis
  • Natural language file discovery
  • Semantic search across file content
  • Smart folder organisation suggestions
  • Batch operations using simple prompts

⚙️ System Control & Automation

  • Execute system-level commands via text
  • Monitor CPU, RAM, and disk usage
  • Automate recurring tasks with workflows
  • Background task management and scripting

Technical Infrastructure

🧠 AI Engine Architecture

  • Modular NLP and command parsing pipeline
  • Contextual memory management
  • Extensible plugin support
  • Language model serving via REST or CLI
  • Hardware-based performance optimization

📚 Knowledge Management

  • Document loader and chunker
  • Personal retrieval-augmented generation (RAG)
  • Metadata and context tagging
  • Long-term information retention

💻 System Integration

  • Cross-platform OS interaction (Linux, macOS, Windows)
  • Process control and background service management
  • System health monitoring
  • Shell command execution and sandboxing

🔌 Model Options

Ollama

  • Models: LLaMA, Mistral, Phi, etc.
  • Modern hardware optimized
  • REST serving and easy switching

llama.cpp

  • Lightweight, quantized model support
  • Fast inference on limited devices
  • Optimized for low-memory environments

Custom Backend

  • Fine-tuned model support
  • Custom architecture integration
  • Local inference server with CLI control

Use Cases

For Personal Productivity

  • Search and summarise large documents
  • Organise folders with AI
  • Schedule daily task automations
  • Maintain a local knowledge base

For Knowledge Work

  • Research and content extraction
  • Technical document comparison
  • Offline article or meeting summary generation
  • Secure data analysis workflows

For Developers & Sysadmins

  • CLI-based system control
  • Code and config file documentation
  • Log analysis and troubleshooting
  • Custom workflow scripting

For Privacy-Conscious Users

  • Work with sensitive documents without cloud risk
  • Build local knowledge databases
  • Process, search, and query private data safely
  • Retain full visibility over AI model behaviour

Roadmap

MilestoneETAStatus
Core CLI, file operations, model setupQ2 2025✅ Live
Ollama + llama.cpp integrationQ2 2025✅ Live
Advanced document parsing + OCRQ3 2025🔄 In Progress
Plugin system + custom workflowsQ3 2025🔄 In Progress
Personal data training (RAG)Q4 2025🔜 Planned
Cross-device sync + multimodal supportQ1 2026🔜 Planned