You still had a 4GB memory limit for processes, as well as a total memory limit of 64GB. Especially the first one was a problem for Java apps before AMD introduced 64bit extensions and a reason to use Sun servers for that.
I was referring to work setups with the overengineering - if I had a cent for every time I had to argue with somebody at work to not make things more complex than we actually need I’d have retired a long time ago.
Unless you are gunning for a job in infrastructure you don’t need to go into kubernetes or terraform or anything like that,
Even then knowing when not to use k8s or similar things is often more valuable than having deep knowledge of those - a lot of stuff where I see k8s or similar stuff used doesn’t have the uptime requirements to warrant the complexity. If I have something that just should be up during working hours, and have reliable monitoring plus the ability to re-deploy it via ansible within 10 minutes if it goes poof maybe putting a few additional layers that can blow up in between isn’t the best idea.
Everything is deployed via ansible - including nameservices. So I already have the description of my infra in ansible, and rest is just a matter of writing scripts to pull it in a more readable form, and maybe add a few comment labels that also get extracted for easily forgettable admin URLs.
As a non-Windows-user I see that as a good thing. LLMs are not going away - but that kind of nonsense at least will make sure all PCs will eventually have cheap and reasonably fast AI acceleration. Which is required for killing off centrally hosted LLMs (plus nvidias cash grabbing)
Intel is well known for requiring a new board for each new CPU generation, even if it is the same socket. AMD on the other hand is known to push stuff to its physical limits before they break compatibility.
Had to look that lawyer bit up as it just sounded too much like Gravenreuth - and indeed it was.
I nowadays manage my private stuff with the ansible scripts I develop for work - so mostly my own stuff is a development environment for work, and therefore doesn’t need to be done on private time.
There is nothing like this availlable currently. Framework probably comes closest, but they only sell in a few countries, and there is lots of stuff to dislike about their solutions - but building your own around a framework board might be feasible.
I have two mnt reforms - as you said, slow and expensive. They have their use for work prototyping for me, but generally wouldn’t recommend. They also have the worst keyboard I’ve encountered in a notebook in the last decade.
Generally yes, but you still need hardware support (mostly kernel and mesa). They upstream - but generally you currently want packages built from their git for that.
Also the installer is very mac hardware specific.
A lot of the Zen based APUs don’t support ECC. The next thing is if it supports registered or unregistered modules - everything up to threadripper is unregistered (though I think some of the pro parts are registered), Epycs are registered.
That makes a huge difference in how much RAM you can add, and how much you pay for it.
Not just that - intel did dual core CPUs as a response to AMD doing just that, by gluing two cores together. Which is pretty funny when you look at intels 2017 campaign of discrediting ryzen by calling it a glued together CPU.
AMDs Opteron was wiping the floor with intel stuff for years - but not every vendor offered systems as they got paid off by intel. I remember helping a friend with building a kernel for one of the first available Opteron setups - that thing was impressive.
And then there’s the whole 64bit thing which intel eventually had to license from AMD.
Most of the big CPU innovations (at least in x86 space) of the last decade were by AMD - and the chiplet design of ryzen is just another one.
Is it a ‘death by quantity’ thing?
Pretty much that - those companies rely on open projects to sort it for them, so they’re pretty much scraping open databases, and selling good data they pull from there. That’s why they were complaining about the kernel stuff - the info required was there already, just you needed to put effort in, so they were asking for CVEs. Now they got their CVEs - but to profit from it they’d still need to put the same effort in as they’d had to without CVEs in place.
Short version: A bunch of shitty companies have as business model to sell open databases to companies to track security vulnerabilities - at pretty much zero effort to themselves. So they’ve been bugging the kernel folks to start issuing CVEs and do impact analysis so they have more to sell - and the kernel folks just went “it is the kernel, everything is critical”
tl;dr: this is pretty much an elaborate “go fuck yourself” towards shady ‘security’ companies.
Making an exception for one organisation, pressured by politicians, would be harmful. BBC has the following policy about neutral reporting:
We don’t use loaded words like “evil” or “cowardly”. We don’t talk about “terrorists”. And we’re not the only ones to follow this line. Some of the world’s most respected news organisations have exactly the same policy
One fascinating example is one owner that replaced the DC barrel jack with a USB-C port, so they could utilize USB-PD for external power.
Oddly enough that’s also an example for bad design in that notebook: The barrel jack is soldered in. With a module that is plugged into the board that’d be significantly easier to replace - and also provide strain relief for power jack abuse. All my old thinkpads were trivial to move to USB-C PD because they use a separate power jack with attached cable.
The transparent bottom also isn’t very functional - it is pretty annoying to remove and put back, due to the large amount of screws required. For a notebook designed for tinkering I’d have wanted some kind of quick release for that. Also annoying is the lack of USB ports on the board - there’s enough space to integrate a USB hub, but just doing that on the board and providing extra ports would’ve been way more sensible.
The CPU module also is a bit of a mixed bag - it pretty much is designed for the first module they developed, and later modules don’t have full support for the existing ports. I was expecting that, though - many projects trying to offer that kind of modular upgrade path run into that sooner or later, and for that kind of small project with all its teething problems ‘sooner’ was to be expected. It still is very interesting for some prototyping needs - but that’s mostly companies or very dedicated hackers, not the average linux user.
No, most companies also have mostly incompetent engineers.
Roku always was a company with great engineers and shitty money grabbing management. The new user creation always requested data not necessary for basic operation.
There are extensions for that. Which are worse than they used to be because they didn’t provide APIs enabling to do that properly, about 10 fucking years after they dropped the old APIs. There are a lot of other feature requests from back then open, often even filed years before they went through with dropping the old APIs. The best way of doing custom keyboard shortcuts in Firefox is still injecting Javascript into each page, with all the shortcomings this has. Usability of Firefox is way worse nowadays than what it was 10 years ago - and I do understand (and agree) with the decision to dump the legacy APIs, but you can’t just break functionality lots of people use, and not provide APIs in over a decade to fix that.
I’m trying other browsers now and then, but every single one is a dumpster fire. At least the Firefox dumpster fire is a bit less out of control - but that’s the most positive thing I can say about it nowadays.