

From your description I would gues that the affected trackers have some rate or connection limits, and your qbittorrent announces are exceeding them. try setting a higher announce interval, like 1+ hours


From your description I would gues that the affected trackers have some rate or connection limits, and your qbittorrent announces are exceeding them. try setting a higher announce interval, like 1+ hours


I’m surprised no one mentioned ansible yet. It’s meant for this (and more).
By ssh keys I assume you’re talking about authorized_keys, not private keys. I agree with other posters that private keys should not be synced, just generate new ones and add them to the relevant servers authorized_keys with ansible.


Insurance is privatized with many providers, however the ‘base insurance’ covers the same at all providers and is mandatory to have for all residents. However base insurance prices still vary between providers every year, so people are encouraged to switch the provider regularily, which is a business on its own with brokers / call centers getting bonuses for every ‘sign up’, which means a lot of wasted money:
https://www.emolument.com/salary-reports/jobs/insurance-broker/45892


no, its personal preference


I’m using organizr, it embeds your apps in tabs/iframes and allows you to configure them in the UI.


deleted by creator
It would not be for me, but they just sent me this chat message which is concerning:
We are currently seeing unexpected growth across Dropbox Advanced, and as a result are currently only able to grant 1 TB per month per team. We understand this may be frustrating and are working to resolve this for our customers.


thank you, just subsribed


750gb upload, 5tb download per day. However this seems to be another limit, maybe file based max sharing or something.


guys are not downloading enough 😅
User statistics
All-time upload: 143.678 TiB
All-time download: 112.403 TiB
All-time share ratio: 1.27

you need to improve your search skills:


It matters only if “the docker hosts external IP” your dns resolves is a public IP. In that case packets travel to the router which needs to map/send them back to the docker hosts LAN IP (NAT-Reflection). With cgnat this would need to be enabled on the carrier side, where you set up the port forwarding. If that’s not possible, split-DNS may be an alternative.
If “the docker hosts external IP” is actually your docker hosts LAN IP, all of that is irrelevant. Split-DNS would accomplish that.


Are you hosting behind NAT / at home? If so, you may need to enable NAT reflection on your router.


It depends on your usage. If you are downloading hundreds of GB per month or more, a block account does not make sense.
Personally I get almost everything off torrents, so I also have some Block accounts which last me many years for the occassional use.
I guess your OPNSense rule from Edit3 is not working because the source is not your mailu instance, because connections are initiated from the outside and mailu only answers (TCP ACK). So you have asynchornous routing.
You may get this working if you set the “reply-to” option to the wg gateway on the firewall rule that allows VPS -> wg -> mailu traffic.
However there is a much cleaner solution using the PROXY protocol, which mailu seems to support: https://mailu.io/master/reverse.html
They are using traefik, but nginx also supports the PROXY protocol.