![](https://feddit.uk/pictrs/image/5d293428-9697-45e8-9cc4-f7b2b9c16462.png)
![](https://beehaw.org/pictrs/image/fed50129-04e7-4dbc-8f54-4ba5bae58370.png)
I was going to do an origin character as a solo play-through and a custom character for a group play-through with my mates, but now I might do it the other way around… which means hours in the character creator! Ha.
I was going to do an origin character as a solo play-through and a custom character for a group play-through with my mates, but now I might do it the other way around… which means hours in the character creator! Ha.
Red Hat saying that argument in-particular shows they’ve pivoted their philosophy significantly, it’s a seemingly subtle change but is huge - presumably due to the IBM acquisition, but maybe due to the pressures in the market right now.
It’s the classic argument against FOSS, which Red Hat themselves have argued against for decades and as an organisation proved that you can build a viable business on the back of FOSS whilst also contributing to it, and that there was indirect value in having others use your work. Only time will tell, but the stage is set for Red Hat to cultivate a different relationship with FOSS and move more into proprietary code.
Don’t roll your own if you can help it, just use a distribution dedicated for use as a thin client. I was co-incidentally just looking into this last week and came across ThinStation which looks really good. There are other distro’s too, search for “linux thin client”.
How do Linux distro’s deal with this? I feel like however that’s done, I’d like node packages to work in a similar way - “package distro’s”. You could have rolling-release, long-term service w/security patches, an application and verification process for being included in a distro, etc.
It wouldn’t eliminate all problems, of course, but could help with several methods of attack, and also help focus communities and reduce duplication of effort.
I personally found Fedora to be rock solid, and along with Ubuntu provided the best hardware support out of the box on all my computers - though it’s been a couple of years since I used it. I did end up on Ubuntu non-LTS in the end as I now run Ubuntu LTS on my servers and find having the same systems to be beneficial (from a knowledge perspective).
If I’m okay with the software (not just trying it out) am I missing out by not using dockers?
No, I think in your use case you’re good. A lot of the key features of containers, such as immutability, reproduceability, scaling, portability, etc. don’t really apply to your use case.
If you reach a point where you find you want a stand-alone linux server, or an auto-reconfiguring reverse proxy to map domains to your services, or something like that - then it starts to have some additional benefit and I’d recommend it.
In fact, using native builds of this software on Windows is probably much more performant.
Containers can be based on operating systems that are different to your computer.
Containers utilise the host’s kernel - which is why there needs to be some hoops to run Linux container on Windows (VM/WSL).
That’s one of the most key differences between VMs and containers. VMs virtualise all the hardware, so you can have a totally different guest and host operating systems; whereas because a container is using the host kernel, it must use the same kind of operating system and accesses the host’s hardware through the kernel.
The big advantage of that approach, over VMs, is that containers are much more lightweight and performant because they don’t have a virtual kernel/hardware/etc. I find its best to think of them as a process wrapper, kind of like chroot for a specific application - you’re just giving the application you’re running a box to run in - but the host OS is still doing the heavy lifting.
I was using file merging, but one issue I found was that arrays don’t get merged - and since switching to use Traefik (which is great) there are a lot of arrays in the config! And I’ve since started using labels for my own tooling too.
I was recently helping someone working on a mini-project to do a bit of parsing of docker compose files, when I discovered that the docker compose spec is published as JSON Schema here.
I converted that into TypeScript types using JSON Schema to TypeScript. So I can create docker compose config in code and then just export it as yaml - I have a build/deploy script that does this at the end.
But now the great thing is that I can export/import that config, share it between projects, extend configs, mix-in, and so on. I’ve just started doing it and it’s been really nice so far, when I get a chance and it’s stabilised a bit I’m going to tidy it up and share it. But there’s not much I’ve added beyond the above at the moment (just some bits to mix-in arrays, which was what set me off on this whole thing!)
I hear they have improved performance now though
It’s still not great. Better, but still slow enough to make you question whether you’ve actually launched the app or not.
Yes, I love it! Really it’s the MacOS-like “Expose” feature that I find to be essential.
I would advise against using workspaces though, I find those actually sort of go against the core idea of it IMO. There are a few things I’d really like added to it, but for the most-part when you get into it it’s great.
My main desktop I have 4 monitors (I know, but once you start a monitor habit it’s really hard to not push it to the limit - this is only the beginning!) It roughly breaks down into:
The key, literally, is you just press the Super key and boom, you can see everything and if you want to interact with something it’s all available in just one click or a few of key presses away.
On my laptop with just one screen, I find it equally invaluable, and is actually where I started to use it the most - once again, just one press of Super and I can see all the applications I have open and quickly select one or launch something.
It’s replaced Alt + Tab for me - and I know they’ve made that better, and added Super + Tab, but none of them are as good as just pressing Super.
The things I’d really love added to it are:
I just have a static page that I randomly change - you can see mine here. In this case I was testing the idea of having text within an SVG for better scaling from mobile to desktop, and also I’m loving orange and purple at the moment for some reason! Oh, and I was testing automated deployments from CI/CD, so I always use my own base domain with those first tests!
Glad you sorted it though! It’s a nightmare when you get such an opaque error and there’s so many moving parts that could be responsible!
Assume nothing! Test every little assumption and you’ll find the problem. Some things to get you started:
While not a direct solution to your problem, I no longer manually configure my reverse proxies at all now and use auto-configuring ones instead. The nginx-proxy image is great, along with it’s ACME companion image for automatic SSL cert generation with certbot - you’ll be up and running in under 30 mins. I used that for a long time and it was great.
I’ve since moved to using Traefik as it’s more flexible and offers more features, but it’s a bit more involved to configure (simple, but the additional flexibility means everything requires more config).
That way you just bring up your container and the reverse proxy pulls meta-data from it (e.g. host to map/certbot email) and off it goes.
It does feel like there’s been a shift, especially in organisations that use the work of others for their own benefit (e.g. open source, community produced content, etc). It seems like there’s been a real move to have their cake and eat it.
Oracle has just made an aggressive move with regards to Java licensing too, they’re now charging as much as $15/month/employee to use their Java runtime on the desktop/server. Their FAQ even points you to OpenJDK if you don’t want to pay, which is strange - it makes me think the relationship between Oracle and the OpenJDK will be ending sometime in the not-so-distant future. There are several Java projects I’ve done where that would just become non-viable as it was a project for a single department in a large company.
Software developers are one of the most altruistic groups of people - it’s amazing just how much time and effort they put into things that they get no financial return on, only the love of actually doing it. And people that dedicate their time and effort to online communities, education, and so on are equally amazing.
But I think it’s time to stop being so naive and realise that many large corporate entities are abusing this relationship for their own gain.
Definitely give Ruthless a go, I love it… reminds me of early game ARPG’s on higher difficulties. Positioning really matters, you have to adapt based on what you get. It seems to have been the proving ground for PoE2’s new tempo.