Current state
flowchart LR
digiwf-gateway <--> digiwf-tasklist
digiwf-gateway <--> digiwf-engine-service
digiwf-engine-service <-- to be replaced by connector --> kafka[(Kafka)]
%% integrations
subgraph digiwf-integrations
digiwf-alw-integration
digiwf-cosys-integration
digiwf-email-integration
digiwf-s3-integration
digiwf-verification-integration
end
kafka <--> digiwf-alw-integration
kafka <--> digiwf-cosys-integration
kafka <--> digiwf-email-integration
kafka <--> digiwf-s3-integration
kafka <--> digiwf-verification-integration
%% sync connection to s3-integration
digiwf-engine-service -- synchronous connection --> digiwf-s3-integration
%% connector tbd.
digiwf-connector <-.-> digiwf-engine-service
digiwf-connector <-.-> kafka
Future state
flowchart LR
subgraph digiwf-tasklist
frontend
end
digiwf-gateway <--> digiwf-tasklist
digiwf-gateway <--> digiwf-process
digiwf-gateway <--> digiwf-task
kafka[(Kafka)]
digiwf-task <--> kafka
digiwf-process <--> kafka
%% connector
digiwf-connector <--> camunda
digiwf-connector <--> kafka
%% integrations
subgraph digiwf-integrations
digiwf-alw-integration
digiwf-cosys-integration
digiwf-email-integration
digiwf-s3-integration
digiwf-verification-integration
end
kafka <--> digiwf-alw-integration
kafka <--> digiwf-cosys-integration
kafka <--> digiwf-email-integration
kafka <--> digiwf-s3-integration
kafka <--> digiwf-verification-integration
%% sync connection to s3-integration
digiwf-task -- synchronous connection --> digiwf-s3-integration
%% future tbd.
subgraph digiwf-task
polyflow
end
For local development, you need to start the digiwf-tasklist (the Vue and Vuetify-based frontend application), the digiwf-engine-service (Spring Boot and Camunda-based backend application), and the digiwf-gateway. Additionally, you can start the digiwf-integrations if you need features from any of the integration services.
You may run the applications with the default settings (from application.yml
) and customize the configurations by
adding environment variables. You always have to set the env variable DIGIWF_ENV
(for local
development DIGIWF_ENV=local-01
). Additionally, some services require a few more environment variables.
Under (All env variables)[#all-env-variables], you will find a list of all possible env variables.
You can use the docker-compose stack to start all the necessary infrastructure components (like e.g. Kafka and Minio). For more information, see here.
If you want to develop the frontend, you can also use the dev server.
- Make sure you start the stack without the tasklist profile (
docker compose up -d
). - Go to digiwf-apps
- Run
npm run serve:tasklist
- If the command fails on the first attempt, trigger a build (
npm run build
) and try it again - If you want to debug any of the components, run additionally
npm run dev
to auto-rebuild the components
- If the command fails on the first attempt, trigger a build (
- Go to localhost:8081
Run the DigiWF-Engine with the profiles local
, streaming
, no-security
,
and no-ldap
. Additionally, set the environment variable DIGIWF_ENV=local-01
.
Run the DigiWF-Gateway with the profiles local
and no-security
.
Run the integrations you need. You may run the applications with the default settings, but you probably have to set a few environment variables ( described above).
Applications and profiles
To start any of the applications, you can use the profiles:
local
for local development settingsstreaming
to enable event streaming to Kafka
Additionally, you can use the following profiles to disable a few features to simplify development:
no-security
to disable Spring Securityno-mail
to prevent the applications from sending emails- (only digiwf-engine-service)
no-ldap
to use a fake user service instead of the actual LHM LDAP
./mvnw clean verify
To use Jacoco with a Maven multi-module project, it's recommended to create a dedicated module for the test coverage generation (see Jacoco docs). Therefore, we created the digiwf-coverage module which depends on all modules that should get tested.
If you want to track the test coverage of additional modules you have to add the maven module as dependency in digiwf-coverage.
After executing the tests, a Jacoco coverage report is generated in digiwf-coverage target directory.
For updating the version of the project, we use the Maven-versions plugin.
- Run
./mvnw versions:set -DprocessAllModules -DnextSnapshot -Pfrontend -Pdocs
. - Check if the update is correct.
- If the update is correct, run
./mvnw versions:commit
; otherwise, run./mvnw versions:revert
.
- Run
./mvnw versions:set -DprocessAllModules -Pfrontend -Pdocs -DnewVersion=<your version>
. - Check if the update is correct.
- If the update is correct, run
./mvnw versions:commit
; otherwise, run./mvnw versions:revert
.
- Execute the release pipeline manually and set
Release services (y/n)?
toy
. -SNAPSHOT
will be removed automatically by the pipeline ifSnapshot build (y/n)?
is set ton
.
- Execute
npm run versioning
in the digiwf-apps folder. - Execute the release pipeline manually and set
Release apps (y/n)?
toy
. -SNAPSHOT
suffix will be set ifSnapshot build (y/n)?
is set toy
.
Env variable | Example | Description |
---|---|---|
DIGIWF_ENV | local-01, dev, ... | The environment digiwf runs in |
DIGIWF_S3_HTTPAPI | http://localhost:8086 | Url to the http endpoints of the document-storage-service (s3-integration) |
SSO_BASE_URL | http://localhost:8085/auth | Base URL to the SSO |
SSO_REALM | DEPARTMENT-Softwaredevelopment | The realm |
SSO_ISSUER_URL | ${SSO_BASE_URL}/realms/${SSO_REALM} |
The issuer URL |
SSO_ENGINE_CLIENT_ID | digiwf | The SSO client ID of the digiwf-engine application |
SSO_ENGINE_CLIENT_SECRET | The SSO client secret of the digiwf-engine application | |
ENGINE_DATASOURCE_URL | jdbc:h2:file:~/backenddb;DB_CLOSE_ON_EXIT=FALSE;AUTO_SERVER=TRUE | |
ENGINE_DATASOURCE_USER | johndoe | |
ENGINE_DATASOURCE_PASSWORD | ||
ENGINE_DATABASE_PLATFORM | org.hibernate.dialect.H2Dialect | |
ENGINE_DATABASE_DRIVERCLASSNAME | org.h2.Driver | |
MAIL_PORT | 1025 | |
MAIL_HOST | localhost | |
MAIL_USERNAME | digiwf@core.com | |
MAIL_PASSWORD | ||
KAFKA_BOOTSTRAP_SERVER | localhost | |
KAFKA_BOOTSTRAP_SERVER_PORT | 29092 | |
KAFKA_SECURITY_PROTOCOL | SSL | |
DOCUMENT_STORAGE_HOST | http://localhost | The host of the document storage service (s3-integration) |
DOCUMENT_STORAGE_PORT | 8080 | The port of the document storage service (s3-integration) |
S3_PROXY_HOST | localhost | Host of the S3 proxy service |
S3_PROXY_PORT | 9000 | Port of the S3 proxy service |
S3_BUCKETNAME | my-s3-bucket | The name of the S3 bucket |
S3_ACCESSKEY | ||
S3_URL | http://localhost:9000 | |
S3_SECRETKEY | ||
S3INTEGRATION_DATASOURCE_URL | jdbc:h2:mem:testdb;DB_CLOSE_ON_EXIT=FALSE;AUTO_SERVER=TRUE | The S3 integrations datasource URL |
S3INTEGRATION_DATASOURCE_USER | johndoe | The S3 integrations database user |
S3INTEGRATION_DATASOURCE_PASSWORD | ||
S3INTEGRATION_DATABASE_PLATFORM | org.hibernate.dialect.H2Dialect | |
S3INTEGRATION_DATABASE_DRIVERCLASSNAME | org.h2.Driver | |
ADDRESS_INTEGRATION_SERVER_PORT | 8088 | The address integration services server port |
ADDRESS_SERVICE_URL | http://localhost:8089 | The URL of the Stadt München Address Service |
Legacy EAI Properties
Env variable | Example | Description |
---|---|---|
ALW_DMS_EAIHOST | http://localhost | |
ALW_DMS_EAIPORT | 9010 | |
ALW_DMS_UIURL | ||
ALW_PERSONENINFO_EAIHOST | http://localhost | |
ALW_PERSONENINFO_EAIPORT | 8091 | |
DIGIWF_FRONTEND_URL | http://localhost:8081/ | |
COSYS_EAIHOST | http://localhost | |
COSYS_EAIPORT | 8091 | |
COCKPIT_URL | ||
DMS_HOST | http://localhost | |
DMS_PORT | 8191 | |
DMS_UIURL | ||
DMS_USERNAME | eaiclient | |
PARKAUSWEIS_HOST | http://localhost | |
PARKAUSWEIS_PORT | 9011 |