US Department of Transportation (USDOT) Intelligent Transportation Systems (ITS) Joint Program Office (JPO) Utilities
The JPO ITS utilities repository serves as a central location for deploying open-source utilities used by other JPO-ITS repositories.
Table of Contents
- Minimum RAM: 16 GB
- Supported operating systems:
- Ubuntu 22.04 Linux (Recommended)
- Windows 10/11 Professional (Professional version required for Docker virtualization)
- OSX 10 Mojave
- NOTE: Not all images have ARM64 builds (they can still be ran through a compatibility layer)
- Docker-compose V2 - version 3.4 or newer
The jpo-utils repository is intended to be ran with docker-compose v2 as it uses functionality added in the v2 release.
Read the following guides to familiarize yourself with the jpo-utils Docker configuration.
Important!
You must rename sample.env
to .env
for Docker to automatically read the file. Do not push this file to source control.
A MongoDB instance that is initialized as a standalone replica-set and has configured users is configured in the docker-compose-mongo file. To use a different setup_mongo.sh
or create_indexes.js
script, pass in the relative path of the new script by overriding the KAFKA_INIT_SCRIPT_RELATIVE_PATH
or MONGO_CREATE_INDEXES_SCRIPT_RELATIVE_PATH
environmental variables. These scripts facilitate the initialization of the MongoDB Database along with the created indexes.
Where the COMPOSE_PROFILES
variable in you're .env
file are as follows:
mongo_full
- deploys all resources in the docker-compose-mongo.yml filemongo
- only deploys themongo
andmongo-setup
servicesmongo_express
- only deploys themongo-express
service
- Create a copy of
sample.env
and rename it to.env
. - Update the variable
DOCKER_HOST_IP
to the local IP address of the system running docker which can be found by running theifconfig
command- Hint: look for "inet addr:" within "eth0" or "en0" for OSX
- Set the password for
MONGO_ADMIN_DB_PASS
andMONGO_READ_WRITE_PASS
environmental variables to a secure password. - Set the
COMPOSE_PROFILES
variable to:mongo_full
- Run the following command:
docker-compose up -d
- Go to
localhost:8082
in your browser and verify thatmongo-express
can see the created database
The Bitnami Kafka is being used as a hybrid controller and broker in the docker-compose-kafka file. To use a different kafka_init.sh
script, pass in the relative path of the new script by overriding the KAFKA_INIT_SCRIPT_RELATIVE_PATH
environmental variable. This can help in initializing new topics at startup.
An optional kafka-init
, schema-registry
, and kafka-ui
instance can be deployed by configuring the COMPOSE_PROFILES
as follows:
kafka_full
- deploys all resources in the docker-compose-kafka.yml filekafka
- only deploys thekafka
serviceskafka_setup
- deploys akafka-setup
service that creates topics in thekafka
service.kafka_schema_registry
- deploys akafka-schema-registry
service that can be used to manage schemas for kafka topicskafka_ui
- deploys a web interface to interact with the kafka cluster
The Kafka topics created by the kafka-setup
service are configured in the kafka-topics-values.yaml file. The topics in that file are organized by the application, and sorted into "Stream Topics" (those with cleanup.policy
= delete
) and "Table Topics" (with cleanup.policy
= compact
).
The following enviroment variables can be used to configure Kafka Topic creation.
Environment Variable | Description |
---|---|
KAFKA_TOPIC_CREATE_ODE |
Whether to create topics for the ODE |
KAFKA_TOPIC_CREATE_GEOJSONCONVERTER |
Whether to create topics for the GeoJSON Converter |
KAFKA_TOPIC_CREATE_CONFLICTMONITOR |
Whether to create topics for the Conflict Monitor |
KAFKA_TOPIC_CREATE_DEDUPLICATOR |
Whether to create topics for the Deduplicator |
KAFKA_TOPIC_PARTITIONS |
Number of partitions |
KAFKA_TOPIC_REPLICAS |
Number of replicas |
KAFKA_TOPIC_MIN_INSYNC_REPLICAS |
Minumum number of in-sync replicas (for use with ack=all) |
KAFKA_TOPIC_RETENTION_MS |
Retention time for stream topics, milliseconds |
KAFKA_TOPIC_DELETE_RETENTION_MS |
Tombstone retention time for compacted topics, milliseconds |
- Create a copy of
sample.env
and rename it to.env
. - Update the variable
DOCKER_HOST_IP
to the local IP address of the system running docker which can be found by running theifconfig
command- Hint: look for "inet addr:" within "eth0" or "en0" for OSX
- Set the
COMPOSE_PROFILES
variable to:kafka_full
- Run the following command:
docker-compose up -d
- Go to
localhost:8001
in your browser and verify thatkafka-ui
can see the created kafka cluster and initialized topics
The mongo-connector service connects to specified Kafka topics (as defined in the mongo-connector/connect_start.sh script) and deposits these messages to separate collections in the MongoDB Database. The codebase that provides this functionality comes from Confluent using their community licensed cp-kafka-connect image. Documentation for this image can be found here.
Provided in the mongo-connector directory is a sample configuration shell script (connect_start.sh) that can be used to create kafka connectors to MongoDB. The connectors in kafka connect are defined in the format that follows:
declare -A config_name=([name]="topic_name" [collection]="mongo_collection_name"
[convert_timestamp]=true [timefield]="timestamp" [use_key]=true [key]="key" [add_timestamp]=true)
The format above describes the basic configuration for configuring a sink connector, this should be placed at the beginning of the connect_start.sh file. In general we recommend to keep the MongoDB collection name the same as the topic name to avoid confusion. Additionally, if there is a top level timefield set convert_timestamp
to true and then specify the time field name that appears in the message. This will allow MongoDB to transform that message into a date object to allow for TTL creation and reduce message size. To override MongoDB's default message _id
field, set use_key
to true and then set the key
property to "key". The "add_timestamp" field defines whether the connector will add a auto generated timestamp to each document. This allows for creation of Time To Live (TTL) indexes on the collections to help limit collection size growth.
After the sink connector is configured above, then make sure to call the createSink function with the config_name of the configuration like so:
createSink config_name
This needs to be put after the createSink function definition. To use a different connect_start.sh
script, pass in the relative path of the new script by overriding the CONNECT_SCRIPT_RELATIVE_PATH
environmental variable.
Set the COMPOSE_PROFILES
environmental variable as follows:
kafka_connect
will only spin up thekafka-connect
service in docker-compose-connect- NOTE: This implies that you will be using a separate Kafka and MongoDB cluster
kafka_connect_standalone
will run the following:kafka-connect
service from docker-compose-connectkafka
service from docker-compose-kafkamongo
andmongo-setup
services from docker-compose-mongo
- Create a copy of
sample.env
and rename it to.env
. - Update the variable
DOCKER_HOST_IP
to the local IP address of the system running docker - Set the password for
MONGO_ADMIN_DB_PASS
andMONGO_READ_WRITE_PASS
environmental variables to a secure password. - Set the
COMPOSE_PROFILES
variable to:kafka_connect_standalone,mongo_express,kafka_ui,kafka_setup
- Navigate back to the root directory and run the following command:
docker compose up -d
- Produce a sample message to one of the sink topics by using
kafka_ui
by:- Go to
localhost:8001
- Click local -> Topics
- Select
topic.OdeBsmJson
- Select
Produce Message
- Leave the defaults except set the
Value
field to{"foo":"bar"}
- Click
Produce Message
- Go to
- View the synced message in
mongo-express
by:- Go to
localhost:8082
- Click
ode
-- Or click whatever value you set theMONGO_DB_NAME
to - Click
OdeBsmJson
, and now you should see your message!
- Go to
- Feel free to test this with other topics or by producing to these topics using the ODE