I’ve recently been moving a few of my services from bare metal installations over to docker containers. Normally I use ansible to deploy everything in the right place (and you should be doing that too), but I have a “playground” to try out stuff before promoting it to “properly deployed on a different VM with ansible”.
The following script came in handy to simplify the process of creating systemd service files for the docker services.
It assumes that you are in a directory with a docker-compose.yml and the directory name will be the service name, e.g. you are in /opt/watchtower/ and there is a docker-compose.yml here -> the service name will be watchtower .
echo"Switch into the directory with your docker-compose.yml file"
I work in the gaming industry, but normally don’t talk about games much because a) I work on the technical side and b) there are so many other blogs, outlets, news and review sites that do this every day that I don’t feel I can add a perspective that is unique (and there is no sense in writing redundant blog posts).
That being said, I really enjoyed one of the games I recently played, so I decided to write about it anyway. The game is called The Ascent and it’s an isometric top-down shooter in a cyberpunk setting.
There is a wide variety of weapons that you can level-up throughout the game (although the resources required to level up weapons are scarce at higher levels, so you can only bring a few of the weapons to max level). A few different weapon damage types are available, that work better or worse against different enemies. And to further customize your character, there are a few abilities and skills available.
The story was engaging, kept me busy with both main missions and side missions. Missions always have a suggested level with is useful because if you head into an area you aren’t ready for you will be killed swiftly 😉 Aside from the story, I also enjoyed the graphical level of detail. No obvious repeating of elements, each area/level had a very distinct feeling to it, even backgrounds or parts of the map where players can’t navigate to were very detailed. Music was also very well picked for the different game areas.
Difficulty scales fairly moderate. The beginning can be tough at times, but as long as you keep leveling up your weapons, and making sure you have the right gear equipped, it’s pretty smooth sailing. Also, remember to match up weapon type with enemy type.
A recent update introduced a “Game+” mode with tougher enemies that allows you to re-play the game with you character from the normal playthrough (i.e. keep your stats/skills/inventory, but the enemies also start out at around level 30). Game+ mode is nice if you really want to max out all the skills or level up a few more weapons.
There were a few cons too, in coop multiplayer mode we had a few crashes and quests getting stuck (requiring us to restart the game). But all in all it was playable and we mad it through the whole game. Multiplayer had a very “added in later” feel to it, not something designed to be a integral feature of the game (e.g. “guests” don’t see any of the useful mini-map markers). Also, there is little to no replay value in the game. I did a second playthrough in Game+ mode, and it felt more like a speedrun focusing on the main mission and picking up side missions in the same area for extra EXP.
All in all I really enjoyed the game, it wasn’t too difficult or grindy, and travel times not too bad as long as you kept an eye on your missions and made liberal use of the taxi and subway system. I got all the achievements in the game, and the one that took the longest was “Die 100 times”.
Paperless is a document management system that helps you manage digital scans of your documents. I’ve been using it for a while, but as with many projects the developer(s) lost motivation/time to keep the project up-to-date, a fork was made (paperless-ng) which after time also died off. Now a few developers got together, worked through the backlog of issues and forked the next generation as “paperless-ngx”.
Short version: I have my scanner set up to be able to store scans on a network share as PDFs, and paperless monitors the same share for new documents. If it finds a new file, it performs OCR on it (saves the PDF as well as the text) and runs user defined rules against it (e.g. detecting the date of the document, or correspondents, or what kind of document it is, …). It actually makes it very easy to manage physical documents. I just pop them in the scanner, select the destination share, and that’s it, everything else happens automatically. So if I need to search for documents needed to do my taxes. I can do it digitally and not have to go through folders with physical paper.
It can also ingest images or office documents, just toss them in the directory the application monitors for ingesting documents (or use the “Upload” button in the application).
I’m not a big fan of the default dark theme, but a normal light theme is still available.
All in all, it’s pretty nifty, so have a look if this is something that you might find useful.
I spent some time the last few days tweaking my mail server settings since there has been an annoying rise in spam lately. Nothing special, mostly spring-cleaning of the blocklists and Spam Assassin settings. But as I was going over my config, I realized I didn’t have any way to measure “success”. I don’t really know which Blocklists work well for me and which don’t.
I use Graylog to collect logs from my systems and applications. But as far as my Exim logs are concerned my setup was pretty barebones (i.e. not parsing any fields, just dumping them as they were into Graylog). So I spent some time setting up proper extractors for my Exim logs to store everything useful in fields. A lot of the Exim logs use a straightforward key=value structure, making them easy to parse.
(spoiler: I bundled up everything here as a Graylog Content pack in case anyone wants to use it. Link at the bottom of the post)
Once the logs were properly parsed, I moved on to Dashboards to visualize the data. I started out with a visualization of the score Spam Assassin assigns to incoming emails (negative is good, positiv is bad, it’s been years since I’ve seen anything above a score of 5.0 that wasn’t spam). This gives me an indication of the quality of the mail making it through filters to my mailbox. Then a little overview of incoming and outgoing mail, and how much is discarded by SPF and DNSBL.
This dashboard s the most interesting one when it comes to deciding which DNSBL lists are useful and which aren’t. It shows which lists are finding spambot globally as well as over time. All my dashboards also have a widget with the relevant logs from the dashboard underneath to have easy access to the raw logs.
Since I had the data anyway, I also create a dashboard to show transport encryption information. About 60% of mail servers seem to support transport encryption, which is a lot lower than I would have expected (since it is easy to configure). I didn’t dig deeper into this, but I wouldn’t be surprised if the 40% sending email using plain unencrypted methods are mostly spammers that have very simple bots running to send their email.
This dashboard is technically also not related to spam, it’s bots trying to brute force user accounts on my mail server to abuse them to send more spam. Fairly aggressive fail2ban settings take care of that though. It’s interesting to see, that the botnets aren’t used solely for sending spam, they are also used to try and compromise mail server accounts to increase the volume of mails they can send.
Recently I’ve been noticing my Sonarr and Radarr applications behaving erratic (sometimes not responsive, sometimes not performing tasks, not searching or adding content, but at other times behaving totally fine). A quick look at the logs told me the applications were crashing and being restarted by systemctl, after a few crashes they seemed to stabilize. Today I had some time to dig deeper into the issue. I had already searched the Internet for general issues, but it didn’t seem to be a widespread problem. I assumed it might have to do with the ARM architecture, but Raspberry Pi users didn’t seem to be having these issues.
In the past I had difficulty reproducing the issue, but today I was in luck, every time I tried to kick off the “Process Monitored Downloads” task in Radarr, it would start working on the task and then crash and restart. The core issue turned out to be oddly specific to the Odroid XU4 hardware.
The XU4 has eight CPU cores, four A7 cores running at 1.4GHz, and four A15 cores running at 2GHz.
Whenever the mono process moved from an A7 to an A15 core (or vice versa) the process crashed. Since both Sonarr and Radarr are Mono applications, they were both affected. Pinning the applications to either the A15 or the A7 CPUs resolved the problem. taskset--cpu-list can be used to change the CPU affinity of a process.
First look up the systemctl service files for Radarr and Sonarr (e.g. via systemctl status radarr ). Edit the service files by prefixing the ExecStart command with /usr/bin/taskset-c0-3 (for the A7 cores) or /usr/bin/taskset-c4-7 (for the A15 cores). Then reload the systemctl files ( systemctl daemon-reload ) and restart the service.