

Is there any meaningful difference between installing CachyOS and installing Arch but setong up the CachyOS repositories and using CachyOS’ kernel? I did the latter, and things seem to work fine
Is there any meaningful difference between installing CachyOS and installing Arch but setong up the CachyOS repositories and using CachyOS’ kernel? I did the latter, and things seem to work fine
Or both! Debian on my server, arch on my desktop, btw
I realize I sounded overly critical in my comment. Open source sustainability, including in monetary form, is a real issue. It’s very important that we as a community try to find working models, and I commend posts like yours that investigate the possibilities and traverse the solution space. Unfortunately it’s a very difficult problem to solve. I think a model that can work is yet to be found
I don’t think either of these models can work. Fund-to-release is basically the same as crowdfunding, except painted as more focused. Now users, rather than “donate 5$ every month to fooProject” need to deal with a constant stream of “donate X$ to fooProject for Y feature / bug”, so it sharply increases subscription fatigue. I guess it wouldn’t be subscription fatigue then. Shopping fatigue?
And what if bugfix X reaches 90% of the funding, and feature Y reaches 90% of the funding, but neither reaches 100%? With a simpler subscription the project would have a set amount of money to distribute across its internal needs.
And this isn’t even touching on the subject of cost overruns. What happens when your feature estimated at 2 months of dev time is 60% done after 7 weeks? Do you ask for a second donation round?
Rather, for this kind of focused work a project should keep one single treasury to distribute as needed, and have polls for contributors (monetary or otherwise) to vote on which parts to focus on first.
The fund-to-release model shifts all the risk on the authors, who don’t see any monetary reward while the work is ongoing and are not guaranteed any even when the work is finished. Dev work needs a constant stream of funding (to eat, pay rent etc) unless the author starts with a sizeable initial treasury, in which case they can deal with big lump sumps to distribute as they need. But this requires at the very least the guarantee of payment once the work is done which, again, this model does not guarantee.
Sorry I don’t have solutions to propose, but I think the flaws of these alternatives far outweigh their pros.
If the time is off by that much after being powered off, this tells me two things:
An incorrect clock can absolutely cause network issues, so I would bet that’s what is causing you trouble
Why are you using networkd instead of networkmanager on a desktop?
What a weird question. Networkd works anywhere systemd works, why whould desktops be any different.
It’s the same as asking someone “why are you using systemd-boot instead of grub?” Because I like systemd boot better and it’s easier to configure. Same with networkd, configuration is stupid simple, I have installed it on my work machine even.
As for op: since you can manually ping ip addresses and the issue seems to be time-based, could it be that your machine is somehow not renegotiating a dhcp lease?
That’s not enough, a better idea is to, somehow, poison the location data. Otherwise by disabling location tracking you still leak the information that you are going to a clinic.
Yeah that’s what we did last time. I implemented a basic framework on top of a very widespread system in our codebase, which would allow a number of requested minor features to be implemented similarly, with the minimal amount of required boilerplate, and leaving the bulk of the work to implementing the actual meat of the requests.
These requests were completely independent and so could be parallelized easily. The “framework” I implemented was also incredibly thin (basically just a helper function and an human instruction in the shape of “do this for this usecase”) over a system that is preexisting knowledge. My expectation was to have to bring someone up to speed on certain things and then let them loose on this collection of tasks, maybe having to answer some question a couple times a day.
Instead, since the assigned colleague is basically just a copilot frontend, I had to spend 80% or more of my days explaining exactly what needed to be done (I would always start with the whys od things since the whats are derived from them, but this particular colleague seems uninterested in that).
So I was basically spending my time programming a set of features by proxy, while I was ostensibly working on a different set of features.
So yeah, splitting work only works if you also have people capable of doing it in the first place. Of course I couldn’t not help this colleague either, that’s a bad mark on performance review you know. Even when the colleagues have no intention of learning or being productive in any way (I live in a country with strong employee regulations so almost nobody can be fired for anything concerning actual work performance, and this particular colleague doesn’t hide that they don’t care about actually doing a good job, except to managers so they still get pay raises for “improving”).
Yeah, you can tell I’m unhappy
who is actually stopping them from dealing with it?
Management. Someone in management sets idiotic deadlines, then someone tells you “do X”, you estimate and come up with “it will take T amount of time” and production simply tells you “that’s too long, do it faster”
they don’t care about the details or maintenance
They don’t, they care about time. If there are 6 weeks to implement a feature that requires reworking half the product, they don’t care to know half the product needs to be reworked. They only care to hear you say that you’ll get it done in 6 weeks. And if you say that’s impossible, they tell you to do it anyway
you have to include the cost of managing technical debt
I do, and when I get asked why my time estimations are so long compared to those of other colleagues I say I include known costs that are required to develop the feature, as well as a buffer for known unknowns and unknown unknowns which, historically, has been necessary 100% of the time and never included causing us development difficulties and us running over cost and over time causing delays and quality issues that caused internal unhappiness, sometimes mandatory overtime, and usually a crappy product that the customers are unhappy with. That’s me doing a good job right? Except I got told to ignore all of that and only include the minimum time to get all of the dozens of tiny pieces working. We went over time, over cost, and each tiny piece “works” when taken in isolation but doesn’t really mix with everything else because there was no integration time and so each feature kinda just exists there on its own.
Then we do retrospectives in which we highlight all the process mistakes that we ran into only to do them all again next time. And I get blamed come performance review time because I was stressed and I wasn’t at the top of my game in the last year due to being chronically overburdened, overworked, and underpaid.
deleted by creator
deleted by creator
Until you write a compiler error in some deeply templated C++ code, in which case just reading every word takes all day
/s but not too much
As a dev, I had to fix an O( n! ) algorithm once because the outsourced developer that wrote it had no clue about anything. This algorithm was making database queries. To an on-device database, granted, so no network requests, but jesus christ man. I questioned the sanity of the world that time, and haven’t stopped since.
Thank you for the explanation, now I understand the context on the original message. It’s definitely an entirely different environment, especially the kind of software that runs on a bunch of servers.
I have built business programs before being a game dev, still the kinds that runs on device rather than on a server. Even then, I always strived to write the most correct and performant code. Of course, I still wrote bugs like that time that a release broke the app for a subset of users because one of the database migrations didn’t apply to some real-world use case. Unfortunately, that one was due to us not having access to real world databases pr good enough surrogates due to customer policy (we were writing an unification software of sorts, up until this project every customer could give different meanings to each database column as they were just freeform text fields. Some customers even changed the schema). The migrations ran perfectly on each one of the test databases that we did have access to, but even then I did the obvious: roll the release back, add another test database that replicated the failing real world use case, fixed the failing migrations, and re released.
So yeah, from your post it sounds that either the company is bad at hiring, bad at teaching new hires, or simply has the culture of “lol who cares someone else will fix it”. You should probably talk to management. It probably won’t do anything in the majority of cases, but it’s the only way change can actually happen.
Try to schedule one on one session with your manager every 2 to 3 weeks to assess which systematic errors in the company are causing issues. 30 minutes sessions, just to make them aware of which parts of the company need fixing.
Sorry, this comment is causing me mental whiplash so I am either ignorant, am subject to non-standard circumstances, or both.
My personal experience is that developers (the decent ones at least) know hardware better than IT people. But maybe we mean different things by “hardware”?
You see, I work as a game dev so a good chunk of the technical part of my job is thinking about things like memory layout, cache locality, memory access patterns, branch predictor behavior, cache lines, false sharing, and so on and so forth. I know very little about hardware, and yet all of the above are things I need to keep in mind and consider and know to at least some usable extent to do my job.
While IT are mostly concerned on how to keep the idiots from shooting the company in the foot, by having to roll out software that allows them to diagnose, reset, install or uninstall things on, etc, to entire fleets of computers at once. It also just so happens that this software is often buggy and uses 99% of your cpu taking it for spin loops (they had to roll that back of course) or the antivirus rules don’t apply on your system for whatever reason causing the antivirus to scan all the object files generated by the compiler even if they are generated in a whitelisted directory, causing a rebuild to take an hour rather than 10 minutes.
They are also the ones that force me to change my (already unique and internal) password every few months for “security”.
So yeah, when you say that developers often have no idea how the hardware works, the chief questions that come to mind are
It’s one thing to pay, and another to be squeezed dry.
When ads were mostly static banners on websites almost nobody was blocking them, because they were mostly unobtrusive.
However, they would often link to shady websites that would install random crap, so the usecase for blocking them was already there.
Then they became animated, and they multiplied. It was one at the bottom of content at first. Then a couple. Then two vertical banners on the sides too. Then more rectangular banners here and there for good measure.
Then they became unkillable javascript popups, then proper new browser windows. Then autoplaying videos with audio were added. And this is just the visible stuff. Add tracking pixels, tracking cookies, browser fingerprinting, and tons of other spying technology deployed under the guise of “but the content is free”.
After every step the use of ad and tracking blockers became more legitimate as serving ads moved further and further away from paying for free content and squarely in the space of selling user data collected without consent for huge profit margins.
If ads and subscriptions were enough to just make a normal amount of profit, very few would be blocking ads or pirating content, because the amount of ads or the price of subscriptions would be reasonable and affordable.
But since everyone wants to make a 1000% markup on the content they generate, they will drive their very own paying customers away.
Youtube could have served me a couple ads per video and I would have kept using it forever. Instead they served me a minimum of 20 ads per video, so now they will serve me zero, forever.
Netflix could have gotten 12 euros every month out of me for their dwindling and dwindling content selection. Instead they wanted 14 after a while. And 17 after a while. And 19 after a little while more. All the while refusing to serve me the 4k content I paid for.
So instead they now get zero too.
I am very happy to pay for content, and a lot of people like me. But the comment you originally replied to was in reference to youtube increasing the price of their subscription by ludicrous amounts. You replied there content isn’t free, and I replied that youtube has no problem making money. The increases are not to keep youtube afloat, is to make youtube make 10 billions in profit rather than 8 next year.
It’s not about paying a fair amount of money for content, it’s about making you pay all that you can give and suck you dry.
So to your question “how do you pay for content/services in general?” I answer “with money”, but that is not what is happening here.
+1. Arch is super easy to install, just open the install guide on the wiki and do what it says.
It’s also really stable nowadays, I can’t actually remember the last time something broke.
As a counterpoint, on ubuntu I constantly had weird issues where the system would change something apparently on its own. Like the key repeat resetting every so often (I mean multiple times an hour), weirdness with graphic drivers, and so on.
That said, I also appreciate debian for server usage. Getting security updates only can be desirable for something that should be little more than an appliance. Doing a dist upgrade scares the shit out of me though, while on arch that’s not even close to a concern.
I did seem to gain around 10s in a compilation workload that takes just under 2 minutes (after the switch), but I didn’t perform any actual benchmark