Quick Start

Deploy DataMate in 5 minutes

This guide will help you deploy DataMate platform in 5 minutes.

DataMate supports two main deployment methods:

  • Docker Compose: Suitable for quick experience and development testing
  • Kubernetes/Helm: Suitable for production deployment

Prerequisites

Docker Compose Deployment

  • Docker 20.10+
  • Docker Compose 2.0+
  • At least 4GB RAM
  • At least 10GB disk space

Kubernetes Deployment

  • Kubernetes 1.20+
  • Helm 3.0+
  • kubectl configured with cluster connection
  • At least 8GB RAM
  • At least 20GB disk space

5-Minute Quick Deployment (Docker Compose)

1. Clone the Code

git clone https://github.com/ModelEngine-Group/DataMate.git
cd DataMate

2. Start Services

Use the provided Makefile for one-click deployment:

make install

After running the command, the system will prompt you to select a deployment method:

Choose a deployment method:
1. Docker/Docker-Compose
2. Kubernetes/Helm
Enter choice:

Enter 1 to select Docker Compose deployment.

3. Verify Deployment

After services start, you can access them at:

  • Frontend: http://localhost:30000
  • API Gateway: http://localhost:8080
  • Database: localhost:5432

4. Check Service Status

docker ps

You should see the following containers running:

  • datamate-frontend (Frontend service)
  • datamate-backend (Backend service)
  • datamate-backend-python (Python backend service)
  • datamate-gateway (API gateway)
  • datamate-database (PostgreSQL database)
  • datamate-runtime (Operator runtime)

Optional Components Installation

Install Milvus Vector Database

Milvus is used for vector storage and retrieval in knowledge bases:

make install-milvus

Select Docker Compose deployment method when prompted.

Install Label Studio Annotation Tool

Label Studio is used for data annotation:

make install-label-studio

Access: http://localhost:30001

Default credentials:

Install MinerU PDF Processing Service

MinerU provides enhanced PDF document processing:

make build-mineru
make install-mineru

Install DeerFlow Service

DeerFlow is used for enhanced workflow orchestration:

make install-deer-flow

Using Local Images for Development

If you’ve modified local code, use local images for deployment:

make build
make install dev=true

Offline Environment Deployment

For offline environments, download all images first:

make download SAVE=true

Images will be saved in the dist/ directory. Load images on the target machine:

make load-images

Uninstall

Uninstall DataMate

make uninstall

The system will prompt whether to delete volumes:

  • Select 1: Delete all data (including datasets, configurations, etc.)
  • Select 2: Keep volumes

Uninstall Specific Components

# Uninstall Label Studio
make uninstall-label-studio

# Uninstall Milvus
make uninstall-milvus

# Uninstall DeerFlow
make uninstall-deer-flow

Next Steps

Common Questions

Q: What if service startup fails?

First check if ports are occupied:

# Check port usage
lsof -i :30000
lsof -i :8080

If ports are occupied, modify port mappings in deployment/docker/datamate/docker-compose.yml.

Q: How to view service logs?

# View all service logs
docker compose -f deployment/docker/datamate/docker-compose.yml logs

# View specific service logs
docker compose -f deployment/docker/datamate/docker-compose.yml logs -f datamate-backend

Q: Where is data stored?

Data is persisted through Docker volumes:

  • datamate-dataset-volume: Dataset files
  • datamate-postgresql-volume: Database data
  • datamate-log-volume: Log files

View all volumes:

docker volume ls | grep datamate

Installation Guide

Detailed installation and configuration instructions for DataMate

System Architecture

DataMate system architecture design documentation

Development Environment Setup

Local development environment configuration guide for DataMate


Last modified February 9, 2026: :memo: add english docs (3868c82)