cognee.update()
Description
Update existing data in Cognee. Supported Input Types:- Text strings: Direct text content (str) - any string not starting with ”/” or “file://”
- File paths: Local file paths as strings in these formats:
- Absolute paths: “/path/to/document.pdf”
- File URLs: “file:///path/to/document.pdf” or “file://relative/path.txt”
- S3 paths: “s3://bucket-name/path/to/file.pdf”
- Binary file objects: File handles/streams (BinaryIO)
- Lists: Multiple files or text strings in a single call
- Text files (.txt, .md, .csv)
- PDFs (.pdf)
- Images (.png, .jpg, .jpeg) - extracted via OCR/vision models
- Audio files (.mp3, .wav) - transcribed to text
- Code files (.py, .js, .ts, etc.) - parsed for structure and content
- Office documents (.docx, .pptx) Workflow:
- Data Resolution: Resolves file paths and validates accessibility
- Content Extraction: Extracts text content from various file formats
- Dataset Storage: Stores processed content in the specified dataset
- Metadata Tracking: Records file metadata, timestamps, and user permissions
- Permission Assignment: Grants user read/write/delete/share permissions on dataset
- Single text string: “Your text content here”
- Absolute file path: “/path/to/document.pdf”
- File URL: “file:///absolute/path/to/document.pdf” or “file://relative/path.txt”
- S3 path: “s3://my-bucket/documents/file.pdf”
- List of mixed types: [“text content”, “/path/file.pdf”, “file://doc.txt”, file_handle]
- Binary file object: open(“file.txt”, “rb”) dataset_name: Name of the dataset to store data in. Defaults to “main_dataset”. Create separate datasets to organize different knowledge domains. user: User object for authentication and permissions. Uses default user if None. Default user: “default_user@example.com” (created automatically on first use). Users can only access datasets they have permissions for. node_set: Optional list of node identifiers for graph organization and access control. Used for grouping related data points in the knowledge graph. vector_db_config: Optional configuration for vector database (for custom setups). graph_db_config: Optional configuration for graph database (for custom setups). dataset_id: Optional specific dataset UUID to use instead of dataset_name.
- Pipeline run ID for tracking
- Dataset ID where data was stored
- Processing status and any errors
- Execution timestamps and metadata
Parameters
UUID of the data item to update.
New data to replace the existing data.
UUID of the dataset containing the data.
User performing the operation.
List of node set names to associate.
Override vector database configuration.
Override graph database configuration.
Custom loader configuration.
If true, skip unchanged data.
Returns
Union[Dict[str, PipelineRunInfo], List[PipelineRunInfo]]
How it works
update() replaces an existing data item: it deletes the old data and adds the
new data, then re-runs processing. The data item keeps the same UUID.