I agree, so much legislation is broken, the legislators aren’t doing shit, so we citizens need to fix it!
But we could start with the right to repair.
I agree, so much legislation is broken, the legislators aren’t doing shit, so we citizens need to fix it!
But we could start with the right to repair.
I’m going to suggest food items that we still take from nature and eat with minimal preparation:
We have evidence of shellfish and fish being eaten for a very long time - at least the middle stone age at 140kya - in middens which are 10s of thousands of years old.
Honey is likely to have been a food source - a treat even - even before humans left Africa (so before 100kya) but sadly this would be invisible in the archeological record
There is non-zero risk in every surgery, and this is a major surgery. There is non-zero risk of very very severe consequences: brain infection, stroke being just some. While these risks are low, they are non-zero. The volunteers have the possibility of losing everything.
Java programmers are also functionally illiterate
Disagree with your disagreement. I also have an M1 and was a quite early adopter (within 3 months of launch). It was really snappy compared to my Intel Air it replaced. From the get-go. Even for apps that were still x86 code.
Things definitely improved over the next 9 months, but I was and am a really happy camper.
This is exactly the answer.
I’d just expand on one thing: many systems have multiple apps that need to run at the same time. Each app has its own dependencies, sometimes requiring a specific version of a library.
In this situation, it’s very easy for one app to need v1 of MyCleverLibrary (and fails with v2) and another needs v2 (and fails with v1). And then at the next OS update, the distro updates to v2.5 and breaks everything.
In this situation, before containers, you will be stuck, or have some difficult workrounds including different LD_LIBRARY_PATH settings that then break at the next update.
Using containers, each app has its own libraries at the correct and tested versions. These subtle interdependencies are eliminated and packages ‘just work’.
Exactly. And all the core internet encryption and signing algorithms are fully open source. Eg RSA, AES, DIffie Helman. And these are the algorithms the US (and most other western) governments require when sending data to or from or within there servers.
It’s all software, even the stuff on the graphics cards. Those are the rasterisers, shaders and so on. In fact the graphics cards are extremely good at running these simple (relatively) programs in an absolutely staggering number of threads at the same time, and this has been taken advantage of by both bitcoin mining and also neural net algorithms like GPT and Llama.
Here’s a video with a lot of guidance and examples of use: https://youtu.be/EAQI2ZSmxPU?si=lF5xr8fKyCRHEk7F
Typically you need about 1GB graphics RAM for each billion parameters (i.e. one byte per parameter). This is a 405B parameter model. Ouch.
Edit: you can try quantizing it. This reduces the amount of memory required per parameter to 4 bits, 2 bits or even 1 bit. As you reduce the size, the performance of the model can suffer. So in the extreme case you might be able to run this in under 64GB of graphics RAM.