I know, Kubernetes is what all the cool kidz are using nowadays! And Kubernetes is cool too... until it explodes and you spend hours trying to debug the issue. In fact, I decided to learn about Docker Swarm after getting burned by Kubernetes for the nth time...
Kubernetes may be painless if you are using a managed Kubernetes instance, but if you are selfhosting it... it is hard!
Let's face it: You ain't insert huge company name here, you just care about hosting some containers' replicas, you don't care about all the fancy big corp stuff that Kubernetes tries to provide to you... because you ain't a big corp!
So why not use Docker Swarm?
If you are just like me that likes to scour the internet to see if users think that insert tech here is good, here are some posts on the internet about people talking about their experience with Docker Swarm!
- https://blog.kronis.dev/articles/docker-swarm-over-kubernetes
- https://news.ycombinator.com/item?id=29448182
- https://news.ycombinator.com/item?id=32306857
- https://news.ycombinator.com/item?id=32305800
- https://news.ycombinator.com/item?id=32305952
- https://news.ycombinator.com/item?id=32307229
- https://user-images.githubusercontent.com/9496359/183654510-568ddef6-68e8-4555-8380-429ad827d270.png
- https://www.reddit.com/r/devops/comments/qxq1q9/is_docker_swarm_good_enough_for_production/hlg4xeu/
- https://www.reddit.com/r/devops/comments/wwpo91/k8s_encourages_people_to_deploy_really_complex/ilmjhl3/ (well that's my own comment but shhh just look at the replies)
- https://www.reddit.com/r/devops/comments/wwpo91/k8s_encourages_people_to_deploy_really_complex/ilmw6uu/
Docker Swarm is a orchestrator, the "best little multi node orchestrator that could". It is very useful if you are already using Docker and you need to replicate your containers in multiple nodes/VMs.
- If you are self hosting your services in a dedicated server or in a VPS. (Example: dedicated servers @ OVH, Hetzner, etc)
- If you are already using Docker Compose.
- Docker Swarm uses Docker Compose files for deployment too, so it is a natural step from Docker Compose -> Swarm! And I mean it! You just need to add a new
deploy
key to your already existing compose files to setup how do you want your container to be replicated and deployed... and that's it!
- Docker Swarm uses Docker Compose files for deployment too, so it is a natural step from Docker Compose -> Swarm! And I mean it! You just need to add a new
- If your services don't require persistent storage.
As with any tool, there are some disadvantages to Docker Swarm, here are some reasons that you may want to avoid using Docker Swarm:
- You are using AWS/GCP/Azure/etc.
- Just use their managed Kubernetes service. As I said before: Kubernetes is cool, except when it breaks. However using a Kubernetes instance that is managed by someone else lifts all the heavyweight from you, in exchange of a pretty hefty price increase.
- You need to auto scale according to demand.
- However if you are using a rented dedicated server, why would tou want to auto scale? And if you are using a host that charges per usage (like AWS, GCP, Azure, etc), then they probably has a managed Kubernetes instance.
- You need to have distributed storage because you want to have a distributed database.
- Docker Swarm does support distributed storage, however Kubernetes' solutions is more rock solid so just go with Kubernetes. But then again, do you really need a distributed storage for your smol service? You can go pretty far with a single instance database!
- You need a feature that isn't supported by Docker Swarm.
- While Docker Swarm isn't deprecated, it is also not actively developed, so if you require a feature that isn't supported by it, it may take a loooong time until it is implemented.
- You like to have shiny buzzwords in your CV.
People talk about Docker Swarm being dead because, back in the day, Docker Inc. had a orchestrator called "Docker Swarm", and that got deprecated. Docker Inc., in their infinite marketing wisdom, later created a orchestrator called "Swarm Mode" for Docker, and that's why a lot of people get confused about if Swarm is dead or not.
Contrary to the popular belief, Swarm Mode is not deprecated! Yeah, sure, sadly there isn't too much dev work on Swarm nowadays, but it isn't dead or deprecated, and after all, if it fits your needs, does it really need to be actively worked on? As long as it works, and if you get to a point where Swarm is not fitting your needs, then move to Kubernetes! Don't overcomplicate your life right now just because some day, maybe, you would need Kubernetes.
If it does get deprecated, you can migrate off of it by converting your Docker Compose deployment files with Kompose, so it is not like you are going to end up being stuck on a dead platform forever.
Docker has a Swarm Mode tutorial on their website, check it out!
Docker also has a tutorial on how to deploy Docker Compose files (also known as "Stack") to Docker Swarm, check it out!
This tutorial is mostly a "I'm writing things as I'm learning", focused more on the Docker Swarm + Docker Compose combo than the official tutorial's Docker Swarm + Docker CLI combo, so there may be things that are incorrect or misleading, however I think that other people may think that this tutorial is also useful!
I'm using a Ubuntu 22.04 VM for this tutorial.
https://docs.docker.com/engine/install/
Start your Docker Swarm with docker swarm init
!
If you have multiple network interfaces, Docker will ask you to choose what IP to use. You need to choose the IP that can access your other Docker instances!
By default, Docker will use IPs on the 10.0.0.0 range, if your machine's interface is behind NAT and it also uses the 10.0.0.0 range, it will cause issues when you try to access containers hosted in your Docker Swarm!
To fix this, initialize your Docker Swarm with the --default-addr-pool
parameter! Let's suppose we want to use 192.168.128.0/18
for our containers. [🐳 Learn more]
sudo docker swarm init --default-addr-pool 192.168.128.0/18
Warning
If you change your
default-addr-pool
, check if the IPs aren't being used by another network!ip a
shows what IP ranges are being used. If you use an IP range that is already being used by something else, your containers won't start due toPool overlaps with other one on this address space
!
Warning
If you have multiple network interfaces, Docker will say
Error response from daemon: could not choose an IP address to advertise since this system has multiple addresses on different interfaces (10.29.10.1 on ens18 and 172.29.10.1 on ens19) - specify one with --advertise-addr
. In this case, the--advertise-addr
parameter should be the IP that can communicate with other nodes! So, if172.29.10.1
is the IP that can access other nodes, then that should be your--advertise-addr
.
Warning
If your Docker instance is communicating to other Docker instances via VXLAN or any other network that has a different MTU than the default 1500, you need to delete the default ingress network and create a new one! This is needed because Docker doesn't inherit the MTU of your networking interface, there will be intermittent packet losses when communicating between nodes! Read more about how to fix it here
If you get Swarm initialized: current node (qgsfyhmhwtpkp7zpo7lts2vhp) is now a manager.
, then your Docker instance is in a swarm, and your node is a manager node, sweet!
While not related to Docker Swarm, I thought it was nice to talk about this. :3
docker login ghcr.io
Use your GitHub username as your Username, and a Personal Access Token as your Password.
When deploying your stack, add the parameter --with-registry-auth
!
Once again something not related to Docker Swarm, but I thought it was nice to talk about this. :3
By default, Docker uses json-file
as its logging driver, however Docker recomemnds changing the log driver to local
because the json-file
driver is only set by default for backwards compatibility.
To do this, edit your /etc/docker/daemon.json
and insert the following contents, if the file already exists, add the log-driver
field to the already existing JSON.
{
"log-driver": "local"
}
Did you know that Docker Swarm uses docker-compose.yml
files??? Crazy huh?
version: '3.9'
services:
helloworld:
# This will set the hostname to helloworld-ReplicaID
hostname: "helloworld-{{.Task.Slot}}"
# The image, we will use a helloworld http image
image: strm/helloworld-http
# We will expose the service at port 8080 on the host
ports:
- "8080:80"
# Docker Swarm configuration deployment configurations!
deploy:
# We want to replicate our service...
mode: replicated
# And it will have two instances of the container!
replicas: 2
Now let's deploy our service with docker stack deploy --compose-file docker-compose.yml stackdemo
!
If everything goes well, you will be able to see your stack on the docker stack ls
command!
mrpowergamerbr@docker-swarm-test:~$ sudo docker stack ls
NAME SERVICES ORCHESTRATOR
stackdemo 1 Swarm
And you can also see the service status with docker stack services stackdemo
mrpowergamerbr@docker-swarm-test:~$ sudo docker stack services stackdemo
ID NAME MODE REPLICAS IMAGE PORTS
totx5zzra290 stackdemo_helloworld replicated 2/2 strm/helloworld-http:latest *:8080->80/tcp
You can also view events about the stack with docker stack ps stackdemo
swarm@docker-swarm-manager-1:~$ sudo docker stack ps powercms
ID NAME IMAGE NODE DESIRED STATE CURRENT STATE ERROR PORTS
ztf65gnt1fjc powercms_powercms.1 ghcr.io/mrpowergamerbr/powercms@sha256:41e5cf391194792b4ddff8d0a5561122213cb793ada75edc1e4e6a5ba9b90e16 docker-swarm-manager-1 Ready Rejected 4 seconds ago "Pool overlaps with other one …"
9oiwb29bc1eu \_ powercms_powercms.1 ghcr.io/mrpowergamerbr/powercms@sha256:41e5cf391194792b4ddff8d0a5561122213cb793ada75edc1e4e6a5ba9b90e16 docker-swarm-manager-1 Shutdown Rejected 9 seconds ago "Pool overlaps with other one …"
ezif348rn177 \_ powercms_powercms.1 ghcr.io/mrpowergamerbr/powercms@sha256:41e5cf391194792b4ddff8d0a5561122213cb793ada75edc1e4e6a5ba9b90e16 docker-swarm-worker-1 Shutdown Rejected 14 seconds ago "Pool overlaps with other one …"
xszs8mbxrzdh \_ powercms_powercms.1 ghcr.io/mrpowergamerbr/powercms@sha256:41e5cf391194792b4ddff8d0a5561122213cb793ada75edc1e4e6a5ba9b90e16 docker-swarm-worker-1 Shutdown Rejected 19 seconds ago "Pool overlaps with other one …"
nnq6i1ybb4v4 \_ powercms_powercms.1 ghcr.io/mrpowergamerbr/powercms@sha256:41e5cf391194792b4ddff8d0a5561122213cb793ada75edc1e4e6a5ba9b90e16 docker-swarm-manager-1 Shutdown Rejected 24 seconds ago "Pool overlaps with other one …"
Let's also check the containers running on our Docker instance...
mrpowergamerbr@docker-swarm-test:~$ sudo docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
25088b353de7 strm/helloworld-http:latest "/main.sh" 41 minutes ago Up 41 minutes 80/tcp stackdemo_front.2.wvzs76sk9iq05xozgz4x9cdqe
18588ab9c07c strm/helloworld-http:latest "/main.sh" 41 minutes ago Up 41 minutes 80/tcp stackdemo_front.1.nmreutrz6q4l57rolejuco6xa
Warning
docker ps
only shows the containers running on the current Docker instance! Don't forget about this if you include multiple Docker instances on the swarm!
Now let's access the service with curl
!
mrpowergamerbr@docker-swarm-test:~$ curl 127.0.0.1:8080
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-2</h1></body></html
Now, what happens if we do a bunch of requests?
mrpowergamerbr@docker-swarm-test:~$ curl 127.0.0.1:8080
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-1</h1></body></html
mrpowergamerbr@docker-swarm-test:~$ curl 127.0.0.1:8080
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-2</h1></body></html
mrpowergamerbr@docker-swarm-test:~$ curl 127.0.0.1:8080
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-1</h1></body></html
mrpowergamerbr@docker-swarm-test:~$ curl 127.0.0.1:8080
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-2</h1></body></html
As you can see, the hostname is alternating between helloworld-1
and helloworld-2
, showing that Docker Swarm is load balancing between your two helloworld-http
replicas!
From within any of the containers of the same stack, you can access the other instance by its hostname!
mrpowergamerbr@docker-swarm-test:~$ sudo docker exec -it e4999d8d5cb9 /bin/bash
root@helloworld-1:/www# ping helloworld-2
PING helloworld-2 (192.168.2.4): 56 data bytes
64 bytes from 192.168.2.4: icmp_seq=0 ttl=64 time=0.093 ms
64 bytes from 192.168.2.4: icmp_seq=1 ttl=64 time=0.073 ms
64 bytes from 192.168.2.4: icmp_seq=2 ttl=64 time=0.071 ms
^C--- helloworld-2 ping statistics ---
3 packets transmitted, 3 packets received, 0% packet loss
round-trip min/avg/max/stddev = 0.071/0.079/0.093/0.000 ms
root@helloworld-1:/www# apt install curl
... installing curl ...
root@helloworld-1:/www# curl helloworld-2
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-2</h1></body></html
root@helloworld-1:/www# curl helloworld-2
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-2</h1></body></html
root@helloworld-1:/www# curl helloworld-2
<html><head><title>HTTP Hello World</title></head><body><h1>Hello from helloworld-2</h1></body></html
You can also access the service externally at YourMachineIP:8080
!
You can apply updates to your stack with docker stack deploy --compose-file docker-compose.yml stackdemo
!
Warning
If you had an service, removed it from the Compose file and used
stack deploy
, Docker will not remove the already running services! To remove an service, usesudo docker service rm servicename
, you can see all of your stack's running services withsudo docker stack services stackdemo
!
Here's our Web Server. It is very simple and it takes a bit of time to be ready, because it needs to setup database connections and other thingamajigs.
fun main() {
println("Setting up Web Server...")
// Imagine that this is initializing db connections and stuff like that
Thread.sleep(15_000)
println("Finished setting up Web Server!")
println("Starting Web Server...")
val server = embeddedServer(Netty) {
routing {
get("/") {
call.respondText("Hello World!")
}
}
}
server.start(true)
println("Finished Starting Web Server!")
}
And then have our Docker Compose file with this configuration.
version: '3.9'
services:
slowstartwebserver:
image: sha256:7ffe31bc8eb30ba35b98b25a13bd748a7b5cd284826100656c2c6ffcfe9630d1
ports:
- "33333:80"
Let's check that it works...
mrpowergamerbr@PhoenixWhistler:/mnt/c/Windows/system32$ curl 127.0.0.1:33333
Hello World!
Now, we have a new version of our web server. Because sending Hello World!
is boring, we changed it to Hello World! Loritta is so cute!! :3
!
version: '3.9'
services:
slowstartwebserver:
image: sha256:1c00b295e356ad17156d2c12e856132c4bbde008c862ab1e2d449afadbfda36b
ports:
- "33333:80"
However, after deploying the stack update, you will notice that the old version will be shut down first, then the new version will be deployed.
This can be changed by changing update_config
's order
option!
stop-first
: old task is stopped before starting new one (default)start-first
: new task is started first, and the running tasks briefly overlap
version: '3.9'
services:
slowstartwebserver:
image: sha256:1c00b295e356ad17156d2c12e856132c4bbde008c862ab1e2d449afadbfda36b
ports:
- "33333:80"
deploy:
update_config:
order: start-first
mrpowergamerbr@PhoenixWhistler:/mnt/c/Windows/system32$ docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
c7acffd6aac7 7ffe31bc8eb3 "java -cp @/app/jib-…" 2 seconds ago Up Less than a second slow-start-web-server-stack_slowstartwebserver.1.imx9ubf9sa4nbl92cc657fq1x
1f94ead29bb0 1c00b295e356 "java -cp @/app/jib-…" 3 minutes ago Up 3 minutes slow-start-web-server-stack_slowstartwebserver.1.gewj7xl809tlzo0mvnkkuiy4s
Now, when applying updates, both instances will briefly overlap. If you have used Kubernetes before, you will notice that this is how Kubernetes also works.
But wait, what about the web server's slow start? The previous instance will be shut down before our new instance is ready to serve new connections!
That's where Docker's HEALTHCHECK comes in!
version: '3.9'
services:
slowstartwebserver:
image: sha256:7ffe31bc8eb30ba35b98b25a13bd748a7b5cd284826100656c2c6ffcfe9630d1
ports:
- "33333:80"
deploy:
update_config:
order: start-first
healthcheck:
# What command will be used to check if the system is healthy or not
# Keep in mind that the command will be executed within your container! So, if you are using curl to check if your service is alive,
# you need to be sure that curl is also installed on your container!
test: ["CMD", "curl", "-f", "http://localhost"]
interval: 5s # The check interval
timeout: 10s # How much time the healthcheck process will wait before timing out
retries: 3 # How many times the healthcheck will retry before considering the service as unhealthy
start_period: 15s # Healthcheck initial delay
Now, the previous instance won't be shut down until the new instance is healthy!
PS C:\Users\Leonardo\Documents\Docker> docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
53f4b4ec536c 7ffe31bc8eb3 "java -cp @/app/jib-…" 11 seconds ago Up 9 seconds (health: starting) slow-start-web-server-stack_slowstartwebserver.1.om1wcqjibzylpsnhmnn46twf2
c7acffd6aac7 7ffe31bc8eb3 "java -cp @/app/jib-…" 2 minutes ago Up 2 minutes slow-start-web-server-stack_slowstartwebserver.1.imx9ubf9sa4nbl92cc657fq1x
But will Docker Swarm try to forward requests to the non-healthy instance while it is booting up?
mrpowergamerbr@PhoenixWhistler:/mnt/c/Windows/system32$ while sleep 1; do curl 127.0.0.1:33333; echo ''; done
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World!
Hello World! Loritta is so cute!! :3
Hello World! Loritta is so cute!! :3
Hello World! Loritta is so cute!! :3
Hello World! Loritta is so cute!! :3
Thankfully, Docker Swarm only forwards traffic after the instance is healthy and ready to rock!
mrpowergamerbr@docker-swarm-test:~$ echo "Loritta is so cute! :3" > loritta.txt
version: '3.9'
services:
helloworld:
hostname: "helloworld-{{.Task.Slot}}"
image: strm/helloworld-http
ports:
- "8080:80"
deploy:
mode: replicated
replicas: 2
# The configuration!
configs:
- source: my_first_config # The source should match the config name in the "configs" section
target: /loritta_cute.txt # Where the config should be mounted
configs:
my_first_config:
file: ./loritta.txt
docker stack deploy --compose-file docker-compose.yml stackdemo
Don't worry if two applications have a config named my_first_config
! Docker prefixes the service name before the config name.
mrpowergamerbr@docker-swarm-test:~$ sudo docker exec -it 66726bffdd6d /bin/bash
root@helloworld-2:/www# cat /loritta_cute.txt
Loritta is so cute! :3
Warning
If you update your configuration file, the
docker stack deploy --compose-file docker-compose.yml stackdemo
command will fail! This is because there is another container that is already using the configuration file. To workaround this issue, change the configuration name (my_first_config
) after changing anything in the configuration file!Idea: Suffix the file's hash at the end of the configuration file, that's what Kubernetes' Kustomize does, and then create a script that periodically tries to delete all configs from your Docker Swarm cluster, configs that are in use won't be deleted by Docker. :P
While you can merge compose files with docker compose -f docker-compose.stats-collector.yml -f patch.yml config
, it doesn't work that great, because Docker ends up mangling the file too much to the point that docker stack deploy
rejects the file with "nuh huh, that ain't a Docker Compose file my dawg!" (example: It removes the version
from the Compose file, so Docker Swarm thinks that's not a Docker Compose v3 file)
So you can use a tool like yq to merge multiple yaml files, and then deploy that file.
Just like Kubernetes, you can also set resources requests and limits to your application!
If you already used Kubernetes before...
-
k8s'
limits
= Swarm'slimits
-
k8s'
requests
= Swarm'sreservations
-
reservations
: When scheduling a container, Swarm MUST guarantee container can allocate at least the configured amount- If your container is configured to
reservations.memory: 4G
, and none of your Swarm nodes have 4GB+ of RAM available, the node won't be scheduled due toinsufficient resources on X nodes
!
- If your container is configured to
-
limits
: When scheduling a container, Swarm MUST prevent container to allocate more- If your container is configured to
limits.memory: 4G
, and your container is using more than 4GBs of RAM, Swarm will terminate your container automatically!
- If your container is configured to
You can set both, only one of them, or none! Same goes for the resources specified within the section.
version: '3.9'
services:
helloworld:
hostname: "helloworld-{{.Task.Slot}}"
image: strm/helloworld-http
ports:
- "8080:80"
deploy:
mode: replicated
replicas: 2
resources:
limits:
cpus: '0.50'
memory: 50M
reservations:
cpus: '0.25'
memory: 4G
If you are running apps on the JVM (Java Virtual Machine), check out our Java Resources Recommendations guide to help you understand how setting reservations
and limits
affect how the JVM allocates and behaves!
You can delete the stack with docker stack rm stackdemo
, this will delete the stack + service + all containers associated with it!
mrpowergamerbr@docker-swarm-test:~$ sudo docker stack rm stackdemo
Removing service stackdemo_helloworld
Removing network stackdemo_default
Having only one Docker instance in your swarm is kinda pointless, the fun really begins after we add other Docker instances to our swarm! This way, replicas can be hosted in different nodes, and this is all handled by Docker Swarm.
sudo docker swarm join-token worker
Copy the command and execute it on your VM!
docker swarm join --token SWMTKN-1-2s4fg0ctnjcgy2hei784kz1c3lmc15axeozlfgs08obm6cajzj-7ycl00ssan3glz3309ac79kl4 10.0.12.10:2377
If you get This node joined a swarm as a worker.
, then it means that now your Docker instance is now in the Swarm!
On your manager node, use docker node ls
to view all nodes on the swarm.
swarm@docker-swarm-manager-1:/etc/netplan$ sudo docker node ls
ID HOSTNAME STATUS AVAILABILITY MANAGER STATUS ENGINE VERSION
qgsfyhmhwtpkp7zpo7lts2vhp * docker-swarm-manager-1 Ready Active Leader 20.10.17
d0o8mcxcjb4fwk8jxqxym0kcc docker-swarm-worker-1 Ready Active 20.10.17
TODO
You can access your service at YourMachineIP:8080
, but what if I told you that there is another way to access your service?
You can access the service at any node IP! Not just your machine IP, and it can even be in a machine that isn't hosting your service at the moment! As long as it is connected to your swarm, you are able to access it.
If you ever used k3s' Load Balancer "Klipper", it works in the exact same way!
With this, we can load balance our service with nginx!
upstream powercms_backend {
server 172.29.10.1:8080;
server 172.29.11.1:8080;
}
server {
listen 443 ssl;
server_name mrpowergamerbr.com;
include mrpowergamerbr_ssl.conf;
location / {
proxy_pass http://powercms_backend;
proxy_set_header Host $host;
proxy_set_header X-Real-IP $remote_addr;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Forwarded-Proto $scheme;
}
}
https://betterprogramming.pub/docker-tips-access-the-docker-daemon-via-ssh-97cd6b44a53
- Generate a public/private key with
ssh-keygen
- Login to your Docker instance
- Add the public key (
cat ~/.ssh/id_rsa.pub
) to your Docker instance AS ROOTsudo nano /root/.ssh/authorized_keys
(just append it) docker -H ssh://root@127.0.0.1:2222 ps
orexport DOCKER_HOST=ssh://root@127.0.0.1:2222
- yay!
bye bye~
If you are deploying a service that uses an image that is hosted in a private repository, you need to docker login
in the machine that you are triggering the deploy!
TODO
Docker by default doesn't clean up unused files, so maybe it could be a good idea to setup a cron job to automatically run docker system prune
on your nodes to avoid running out of disk space.
If your node is not queueing new containers, or if you have containers that show up in docker ps
but when trying to view their log, Docker says that the container doesn't exist
If this is happening, probably your node's disk is full! Clean it up and it will magically fix itself :3
exec
bug: OCI runtime exec failed: exec failed: unable to start container process: open /dev/pts/0: operation not permitted: unknown
If you are getting OCI runtime exec failed: exec failed: unable to start container process: open /dev/pts/0: operation not permitted: unknown
when trying to exec
into a container, downgrade your containerd.io
to version 1.6.6 until version 1.6.8 is released. https://askubuntu.com/questions/1424317/docker-20-10-ubuntu-22-04-oci-runtime-exec-failed
TODO