Cool. I’ve been using croc lately to move files around, but this differs in that they are supplying temporary off-site storage so that the host doesn’t have to remain online.
Cool. I’ve been using croc lately to move files around, but this differs in that they are supplying temporary off-site storage so that the host doesn’t have to remain online.
I’ve been playing around with the disabled GIL build and though I use threads fairly extensively in my projects, it’s been smooth sailing so far. I feel like my GUI scripts might be a bit more responsive now? (I tend to farm out user events to dedicated threads, so this is entirely possible.)
But overall, everything is stable and awesome! I’m so excited! This has been a long time coming for Python.
But the mining, milling, and production of nuclear fuel, as well as the construction and decommissioning of nuclear plants, emit greenhouse gases at levels ranging from 10 to 130 grams of carbon dioxide per kilowatt hour of power — lower than fossil fuels but higher than wind and hydroelectricity (and roughly on par with solar).
That’s interesting. The article they link gives a bit more detail:
These energy intensities translate into greenhouse gas intensities for LWR and HWR of between 10 and 130 g CO2-e/kWhel, with an average of 65 g CO2-e/kWhel.
While these greenhouse gases are expectedly lower than those of fossil technologies (typically 600–1200 g CO2-e/kWhel), they are higher than reported figures for wind turbines and hydroelectricity (around 15–25 g CO2-e/kWhel) and in the order of, or slightly lower than, solar photovoltaic or solar thermal power (around 90 g CO2-e/kWhel).
The wide range for nuclear apparently comes from difficulties in estimating the carbon footprint of mining/processing the uranium, but that nuclear is sort of in the middle of the pack in carbon footprint relative to renewables in spite of the fueling costs is good to know.
I suppose these sort of numbers may change dramatically in years to come. Take solar. A lot of focus seems to be on the efficiency of panels, which would almost certainly lower the carbon cost per unit of energy as it improves, but a breakthrough in panel longevity would also do that in an amortized emissions sort of way.
My most common use case is probably looking up stuff that may or may not be in a dict
.
if (val := dct.get(key)) is not None:
# do stuff with val
I guess that’s pretty similar to what you were doing?
Sometimes I also use it in some crazy list comprehension thing when I get backed into a corner, though it’s hard to think of an example off the top of my head? It usually happens when I’m in a rush and desperate to get something working, but it has an uncanny way of being just the thing you need at that point.
Something to be said for the wfh movement too.
I thought I read somewhere that when they were making one of the Toy Story movies, there was some catastrophic data loss that nearly tanked the whole production. But then one of the animators came back from maternity and said wait, I think I have most of it synced to my home server? And the next thing you know, John Lasseter himself is barrelling down the highway to her place and it turned out yeah, she did have it.
Cool! I can see how optical media could, in theory, be very long-lasting as long as you don’t use materials that oxidize or otherwise degrade over time.
Well I guess I’m picturing DNA encoding like a RAID billion in terms of redundancy, so with some checksumming, you ought be able to sort out any mutations? But I’m no geneticist.
That’s why I back up my data on stone tablets in Cunieform.
Seriously though, if you wanted data to last for centuries, what would be your best bet? Would it be some sort of 3D-printed mechanical storage? At least plastics are generally not biodegradable, though they are photodegradable, so I guess you’d want to stick your archive in a dry cave somewhere?
Or what about this idea of encoding the data in the DNA of some microbe and cutting it loose? What could possibly go wrong?
You know, I’m not actually quite sure what I’m doing, but I can tell you I am not looking at the keyboard. I suppose it’s similar to how I play violin? I don’t look at where my hand is but it shifts to different positions depending on what makes the most sense for the pattern I’m trying to play, and yes, a different position does imply a different fingering to reach the same notes.
When learning to program, I initially tried to follow the touch typing guidelines, but they say that you should use the right pinky to reach every key towards the upper right end of the keyboard, which gets old fast given how frequently you need to access them. And just as with music, there are patterns. In programming, you may frequently need to type {}
, :=
, or even something like \{\}
, and flailing around with the pinky is a good way to give yourself carpal tunnel. So your right hand learns to shift to hit those keys using a combination of fingers.
As a Gen X, I think my typing speed peaked around late high school/early university? I tried to teach myself touch typing and got moderately proficient. Then I got into programming where you need to reach all of those punctuation marks. So my right hand has drifted further to the right over the years, which is better for code but suboptimal for regular text.
One thing that’s really tanked for me though is writing in cursive. I used to be able to take notes in class as fast as the prof could speak. Now I can scarcely sign my own name.
I suppose it depends on how you look at it. Take solar, for example. On the one hand, you could argue that if your primary goal is to generate heat, you might as well use a solar thermal plant with lots of focusing mirrors over photovoltaics. The conversion to electricity first would inevitably be far less efficient.
On the other hand, if you’ve got your PV plants for electricity already but they are overproducing at times, there is the question of what to do with the excess power, and using it to run heat pumps may actually be a pretty efficient application at the point?
I paid a visit to Green Bank WV once out of an interest in astronomy. The giant radio telescopes are truly a sight to behold!
Less impressive were the people camped out nearby who saw the place as the promised land where they could cast off their tinfoil hats in the cellular-banned zone surrounding the complex.
I suppose the same could be said on the lemmy side. There’s no reason someone couldn’t write a lemmy app that lets you do what an RSS client does in terms of only showing content from a selected subgroup of communities.
You raise a good point that it would be nice to have more control over which group of communities you are drawing from at a given time. (Is there a way to group subscriptions and switch between them?) It’s a bit disconcerting to see 5 tech headlines and then suddenly something about the war in Ukraine or whatever. It jars my train of thought. With an RSS client, you can group feeds however you want.
That said, my experience with RSS readers is not quite so idyllic. In the end, rather than having nicely partitioned feed groups by topic, I wind up having to separate the ones that produce content frequently but with a poor signal-to-noise from those that post once in awhile but are generally worth your time. With something like lemmy, people are helping you do the work of finding the more interesting content from that site that posts every 10 minutes.
deleted by creator
Anyways, did I miss anything?
I think the big problem in link aggregation is how to sort/prioritize content for the end user. RSS does not provide a way to do this, nor should it as far as I’m concerned. It should simply be about public content being tagged in a standardized way for any app to come along and organize it using whatever algorithm.
A simple RSS reader has the problem that the more prolific sites will tend to flood your feed and make it tedious to scroll through miles of links. Commercial news portals try to learn your tastes through some sort of machine learning algorithm and direct content accordingly. This sounds like a good idea in theory, but tends to build echo chambers around people that reinforce their biases, and that hasn’t done a lot of good for the world to put it mildly.
The lemmy approach is to use one of a number of sorting algorithms built atop a crowd-sourced voting model. It may not be perfect, but I prefer it to being psychoanalyzed by an AI.
Btw there was a post from about a month ago where someone was offering to make any RSS feed into a community. I’ve subscribed to a few of them and it’s actually pretty awesome.
Didn’t kbin have a separate mechanism for supporting a post in a more public way? I can’t remember how that worked now, but it was in addition to the regular voting I think?
The thing about the MPW Shell is it was sort of the only game in town if you actually wanted a command line with the classic Mac OS. (There’s an awesome little emulator called SheepShaver if you ever want to explore it btw.) Well, I suppose there was A/UX. I thought it was a miracle when that came out. You have to realize in those early days a good chunk of the operating system itself was actually baked in to ROM. (You had to do desperate things to squeeze a GUI out of such limited resources as existed back then!) So to this day I have no idea how they managed to spin off a 'nix despite that.
Anyways. I wonder, if you made some sort of template format today, to what extent you could write some sort of conversion tool that would scrape a man page or whatever to rough it in and then you could tweak it to get what you want? man pages aren’t super standardized in their format I guess, so it’s probably more trouble than it’s worth. I like to use Python’s argparse
when rolling out scripts myself, and its --help
format is pretty rigid given that it’s algorithmically generated. Might be more plausible with something like that? I had a quick look just now to see if you can drill down into the argparse.ArgumentParser
class itself to pull out the info more directly, but it seems a rather opaque thing that doesn’t expose public APIs for that. Oh well…
Oh I see. syncthing is more for keeping 2 directories synched between machines. I guess in a sense, it’s more of a Dropbox competitor while croc and sharr are for one-off file transfers. For awhile, I ran an owncloud server at work for internal use. It was pretty good for file synching, but required some port forwarding through the router. These solutions mentioned here seem to all have a public host somewhere to eliminate that need.