Deployment Decision Guide
Pixeltable supports two production deployment patterns. Choose based on your constraints:| Question | Answer | Recommendation |
|---|---|---|
| Existing production DB that must stay? | Yes | Orchestration Layer |
| Building new multimodal app? | Yes | Full Backend |
| Need semantic search (RAG)? | Yes | Full Backend |
| Only ETL/transformation? | Yes | Orchestration Layer |
| Expose Pixeltable as MCP server for LLM tools? | Yes | Full Backend + MCP Server |
Technical Capabilities (Both)
Regardless of deployment mode, you get:- Multimodal Types: Native handling of Video, Document, Audio, Image, JSON.
- Computed Columns: Automatic incremental updates and dependency tracking.
- Views & Iterators: Built-in logic for chunking documents, extracting frames, etc.
- Model Orchestration: Rate-limited API calls to OpenAI, Anthropic, Gemini, local models.
- Data Interoperability: Import/export Parquet, PyTorch, LanceDB, pandas.
- Configurable Media Storage: Per-column destination (local or cloud bucket).
Deployment Strategies
Approach 1: Pixeltable as Orchestration Layer
Use Pixeltable for multimodal data orchestration while retaining your existing data infrastructure.Use When
Use When
- Existing RDBMS (PostgreSQL, MySQL) and blob storage (S3, GCS, Azure Blob) must remain
- Application already queries a separate data layer
- Incremental adoption required with minimal stack changes
Architecture
Architecture
- Deploy Pixeltable in Docker container or dedicated compute instance
- Define tables, views, computed columns, and UDFs for multimodal processing
- Process videos, documents, audio, images within Pixeltable
- Export structured outputs (embeddings, metadata, classifications) to RDBMS
- Export generated media to blob storage
- Application queries existing data layer, not Pixeltable
What This Provides
What This Provides
- Native multimodal type system (Video, Document, Audio, Image, JSON)
- Declarative computed columns eliminate orchestration boilerplate
- Incremental computation automatically handles new data
- UDFs encapsulate transformation logic
- LLM call orchestration with automatic rate limiting
- Iterators for chunking documents, extracting frames, splitting audio
Approach 2: Pixeltable as Full Backend
Use Pixeltable for both orchestration and storage as your primary data backend.Use When
Use When
- Building new multimodal AI application
- Semantic search and vector similarity required
- Storage and ML pipeline need tight integration
- Stack consolidation preferred over separate storage/orchestration layers
Architecture
Architecture
- Deploy Pixeltable on persistent instance (EC2 with EBS, EKS with persistent volumes, VM)
- Build API endpoints (FastAPI, Flask, Django) that interact with Pixeltable tables
- Frontend calls endpoints to insert data and retrieve results
- Query using Pixeltable’s semantic search, filters, joins, and aggregations
- All data stored in Pixeltable: metadata, media references, computed column results
What This Provides
What This Provides
- Unified storage, computation, and retrieval in single system
- Native semantic search via embedding indexes (pgvector)
- No synchronization layer between storage and orchestration
- Automatic versioning and lineage tracking
- Incremental computation propagates through views
- LLM/agent orchestration
- Data export to PyTorch, Parquet, LanceDB