Behind the scenes, every pipeline step runs as a Modal job that talks to managed LanceDB, Kuzu, and PostgreSQL clusters.
System Overview
Cognee Cloud’s architecture centers around three main layers that work together to provide a managed knowledge processing platform:Modal (Managed Infrastructure)
Modal provides the compute foundation for all Cognee Cloud operations:- API Services: Hosts the FastAPI service that handles all REST endpoints and authentication (see Cognee Cloud SDK)
- Notebook Sandbox: Provides isolated environments for running user code with 24-hour timeout support (see Cognee Cloud Notebooks)
- Container Orchestration: Every API request runs inside a Modal container with secrets managed internally by Cognee Cloud
- Code Execution: Notebook code runs in short-lived sandboxes that forward the user’s Cognee Cloud API key to the managed API
Storage Services (Managed by Cognee Cloud)
All data persistence is handled through Cognee Cloud’s managed storage infrastructure:- S3 – Central storage for all raw uploads, LanceDB tables, and Kuzu graph files in Cognee Cloud’s managed S3 infrastructure
- LanceDB – Vector database that stores embeddings generated during the cognify process
- Kuzu – Graph database that maintains knowledge graph relationships and entities
- PostgreSQL – Relational database for users, datasets, permissions, quotas, and billing records
Key Architectural Principles
- Dataset Isolation: All processing happens at the dataset level, with separate storage namespaces (see permissions & security for details)
- Managed Infrastructure: Users don’t configure Modal, S3, or database credentials—everything is managed by Cognee Cloud
- Compatibility: Storage schemas remain compatible with self-hosted Cognee for easy migration