Running with Docker Compose
Start here once you have SSH access and sudo
privileges for a server or VM.
Note
Docker server installation is only supported on Linux distributions
Container Images
A DIVE Web deployment consists of 2 main services.
- kitware/viame-web - the web server
- kitware/viame-worker - the queue worker
In addition, a database (MongoDB) and a queue service (RabbitMQ) are required.
Install dependencies
SSH into the target server and install these system dependencies.
Tip
You can skip this section if you used Ansible to configure your server, as it already installed all necessary dependencies.
- Install NVIDIA Driver (version specified in VIAME)
sudo ubuntu-drivers install
usually works.
- Install
docker
version 19.03+ guide - Install
docker-compose
version 1.28.0+ guide - Install nvidia-container-toolkit
Basic deployment
Clone this repository and configure options in .env
.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 |
|
VIAME server will be running at http://localhost:8010. You should see a page that looks like this. The default username and password is admin:letmein
.
Production deployment
If you have a server with a public-facing IP address and a domain name that points to it, you should be able to use our production deployment configuration. This is the way we deploy viame.kitware.com.
containrrr/watchtower
updates the running containers on a schedule using automated image builds from docker hub (above).linuxserver/duplicati
is included to schedule nightly backups, but must be manually configured.
You should scale the girder web server up to an appropriate number. This stack will automatically load-balance across however many instances you bring up.
1 2 3 4 5 6 7 8 |
|
Splitting services
It's possible to split your web server and task runner between multiple nodes. This may be useful if you want to run DIVE Web without a GPU or if you want to save money by keeping your GPU instance stopped when not in use. You could also increase parallel task capacity by running task runners on multiple nodes.
- Make two cloud VM instances, one with NVIDIA drivers and container toolkit, and one without. This is still a special case of scenario 1 from the Provisioning Guide
- Clone the dive repository on both, and set up
.env
on both with the same configuration. - Be sure that
WORKER_API_URL
andCELERY_BROKER_URL
in particular are uncommented and set to the IP or domain name of your web server. This is how the worker will talk to the web server, so the web server must be network accessible from the worker.
1 2 3 4 5 |
|
In order to run any jobs (video transcoding, pipelines, training, addon upgrades) the GPU server will need to be running.
Addon management
After initial deployment, DIVE Server will require an addon upgrade in order to download and scan for VIAME addons. Run this by issuing a POST /dive_configuration/upgrade_pipelines
request from the swagger UI at http://{server_url}:{server_port}/api/v1
.
- Whether you
force
or not, only those pipelines from addons from the exact urls passed will be enabled on the server. - An old addon can be disabled by simply omitting its download from the upgrade payload.
force
should be used to force re-download of all URLs in the payload even if their zipfiles have been cached.- An upgrade run is always required if the "common" pipelines in the base image change. These are updated for every run, and do not require
force
. - See the job log to verify the exact actions taken by an upgrade job.
- Optional patches are updated occasionally and you can find the latest urls here.
Configuration Reference
Server branding config
You can configure the brand and messaging that appears in various places in the DIVE Web UI using the config API.
- Open the swagger page at /api/v1
PUT /dive_configuration/brand_data
where the body is a JSON object from the template below. If you do not want to set a value and use the default, omit the key and value from the config body.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 |
|
Web Server config
This image contains both the backend and client.
Variable | Default | Description |
---|---|---|
GIRDER_MONGO_URI | mongodb://mongo:27017/girder |
a mongodb connection string |
GIRDER_ADMIN_USER | admin |
admin username |
GIRDER_ADMIN_PASS | letmein |
admin password |
CELERY_BROKER_URL | amqp://guest:guest@default/ |
rabbitmq connection string |
WORKER_API_URL | http://girder:8080/api/v1 |
Address for workers to reach web server |
There is additional configuration for the RabbitMQ Management plugin. It only matters if you intend to allow individual users to configure private job runners in standalone mode, and can otherwise be ignored.
Variable | Default | Description |
---|---|---|
RABBITMQ_MANAGEMENT_USERNAME | guest |
Management API username |
RABBITMQ_MANAGEMENT_PASSWORD | guest |
Management API password |
RABBITMQ_MANAGEMENT_VHOST | default |
Virtual host should match CELERY_BROKER_URL |
RABBITMQ_MANAGEMENT_URL | http://rabbit:15672/ |
Management API Url |
You can also pass girder configuration and celery configuration.
Worker config
This image contains a celery worker to run VIAME pipelines and transcoding jobs.
Note: Either a broker url or DIVE credentials must be supplied.
Variable | Default | Description |
---|---|---|
WORKER_WATCHING_QUEUES | null | one of celery , pipelines , training . Ignored in standalone mode. |
WORKER_CONCURRENCY | # of CPU cores |
max concurrnet jobs. Lower this if you run training |
WORKER_GPU_UUID | null | leave empty to use all GPUs. Specify UUID to use specific device |
CELERY_BROKER_URL | amqp://guest:guest@default/ |
rabbitmq connection string. Ignored in standalone mode. |
KWIVER_DEFAULT_LOG_LEVEL | warn |
kwiver log level |
DIVE_USERNAME | null | Username to start private queue processor. Providing this enables standalone mode. |
DIVE_PASSWORD | null | Password for private queue processor. Providing this enables standalone mode. |
DIVE_API_URL | https://viame.kitware.com/api/v1 |
Remote URL to authenticate against |
You can also pass regular celery configuration variables.
Running the GPU Job Runner in standalone mode
Linux Only.
Individual users can run a standalone worker to process private jobs from VIAME Web.
- Install VIAME from the github page to
/opt/noaa/viame
. - Activate the install with
source setup_viame.sh
. - Install VIAME pipeline addons by running
cd bin && download_viame_addons.sh
from the VIAME install directory. - Enable the private user queue for your jobs by visiting the jobs page
- Run a worker using the docker command below
Note: The
--volume
mount maps to the host installation. You may need to change the source from/opt/noaa/viame
depending on your install location, but you should not change the destination from/tmp/addons/extracted
.
1 2 3 4 5 6 7 8 9 |
|