Skip to content

AI Design

AI Server

AI models, especially LLMs, come in many variants, and typically you need to specialize what models you are running. The models require GPU resources and will serve a number of use cases across TAPPaaS.

Architecture

The architecture uses an LLM server setup:

  • One VM on a machine with resources for AI, running Ollama as the central point for loading and accessing LLMs
  • No users directly interact with this server - interactions are done via AI client programs

AI Clients

Several client types exist in the TAPPaaS ecosystem:

Chat Clients

TAPPaaS installs OpenWebUI as default in a dedicated VM. This provides:

  • Web-based chat interface for interacting with LLMs
  • RAG (Retrieval Augmented Generation) capabilities via SearXNG

Workflow Automation

n8n provides workflow automation with AI capabilities, allowing:

  • Automated processing pipelines
  • Integration with other TAPPaaS services
  • Custom AI-powered workflows

Application Integration

Regular clients that need OpenAI API access include:

  • Home Assistant - Home butler function for smart home automation
  • Immich - Picture classification and search capabilities