Project Aeon - Modular local LLM system

2025-ongoing
Project Aeon - Modular local LLM system

About the Project

My personal AI assistant that runs entirely offline. No data leaves your machine. Uses FastAPI, ChromaDB for semantic search, and Vue 3 for the UI. Integrated with Ollama for local LLMs. Still building it.

Key Learnings

  • Vector databases and semantic search (RAG architecture)
  • FastAPI and modern Python backend patterns
  • Vue 3 Composition API with TypeScript
  • Local LLM deployment and optimization
  • Privacy-preserving AI system architecture
  • Integration of multiple AI frameworks

Tech Stack

FastAPIChromaDBVue 3TypeScriptNaive UIFabric AIOllamasentence-transformers