Reply from customer at 2023-09-29 about WeKan v7.08:
The good thing is that that new version is working really well and being able to spread the load between 4 OpenShift pods changes everything.
Maybe related to these improvements:
Recommended specs:
free -h
is server is using any swap. If it is, add more RAM.nproc
how many CPUs you have. Look with top
or htop
is server using max 100% CPUs, if it is, add higher performance optimized CPUs (or more CPUs). But if it looks like Wekan not using some of those added CPUs, then adding more CPUs is not useful.[Document](https://example.com/files/document.doc)
. Click Wekan board
=> ☰
=> ⚙
=> Board Settings
=> Card Settings
. There uncheck [_] Attachments
to hide attachments at card.Admin Panel / Settings / Accounts / Hide System Messages of All Users
. If someone needs to show system messages, they can click slider at opened card to show them. Sometime later, if many have manually enabled showing system messages, click that same button at Admin Panel again.Admin Panel / Settings / Global Webhooks
(that sends most board actions to webhook) or at each board, per-board webhooks (that sends most one board actions to webhook, more info at wiki right menu Webhooks) at Wekan board
=> ☰
=> ⚙
=> Outgoing Webhooks
. You also see this with DBGate at port localhost:27019 / database: wekan / table: integrations. Each webhook should immediately return 200 response before processing any data, because otherwise it will slow down Wekan a lot.Minimum specs:
At https://wekan.github.io / Download / Kubernetes or OpenShift, etc
From Tampa:
Hey,
... (about other tries) ...
Last month I threw all this out, recreated all the boards and connected them centrally to a single instance of mongo running on a dedicated server with custom hardware. This was like stepping into the light almost. Since then not a single machine has sent me a mail that it reached 50% usage. It seems insignificant, but the results speak for themselves.
The cloud instances are all shared 1vcpu, 1GB, 10GB storage, they just run wekan natively piped to the background, no docker, no snap, native install. They are connected to the central DB server sitting in the same datacenter. I stuffed a Raid6 with solid disks in that and gave it a hardware controller with a nice big cache. The latency being below 5ms over the network and the DB server having plenty of IO to go around it almost never has a queue of commits going to it and from the cache and IO use I suspect I could grow this tenfold easily.
With this setup each board essentially runs on half the hardware, in terms of cost anyways, yet it works so much better. There seems to be some magic ingredient here, being really fast IO for mongo that reduces system load of wekan by such a large amount that is can practically run even large boards with 40+ concurrent users on the least hardware most cloud providers even offer. With the central server setting up backups has become so much easier, I no longer need to wait for low usage to do backups either.
For any large scale usage, you can:
a) scale with Docker Swarm, etc
b) for big reads or writes, do it on replica
c) for big reads or writes, do it at small amounts at a time, at night, or when database CPU usage seems to be low
Related to docker-compose.yml at https://github.com/wekan/wekan , using Docker Swarm: