People haven’t really changed. As always, power corrupts. When the rewards are great enough, it seems people are often enough willing to compromise their integrity.
People haven’t really changed. As always, power corrupts. When the rewards are great enough, it seems people are often enough willing to compromise their integrity.
I have LTS and zen kernels installed in addition to the default Arch one, that should prevent this yes?
What do you mean by “this stuff?” Machine learning models are a fundamental part of spam prevention, have been for years. The concept is just flipping it around for use by the individual, not the platform.
If by reliably you mean 99% certainty of one particular review, yeah I wouldn’t believe it either. 95% confidence interval of what proportion of a given page’s reviews are bots, now that’s plausible. If a human can tell if a review was botted you can certainly train a model to do so as well.
Cool it with the universal AI hate. There are many kinds of AI, detecting fake reviews is a totally reasonable and useful case.
Have you tried recent models? They’re not perfect no, but they can usually get you most of the way there if not all the way. If you know how to structure the problem and prompt, granted.
Them using Google indexes anonymously isn’t intending to solve the problem you think it is. It’s more about incentive structures. Google’s “free” search optimizes for ad revenue now. The API access doesn’t as much, and Kagi certainly doesn’t have an ad incentive. So privacy is a nice bonus, but the real benefit is a customer serving incentive structure.
A major caveat I’ve noticed some people misunderstand: it’s corporate CLAs that are problematic. The Apache Foundation also requires contributors sign a CLA, but it’s to provide a legal fail safe and a way to update to say Apache 3.0 if need be one day. Apache’s non profit, open source mission aligns with respecting the rights of contributors and the community. Corporations, on the other hand, not so much.
Codeberg is run off of donations, they have no service contract revenue. Nobody, much less a volunteer, wants to commit to a 5 or 10 year service plan like that, it’s not sustainable for a small project from a non profit.
CLAs can be abusive, but not necessarily. Apache Foundation contributors need to sign CLAs, which essentially codify in contract form the terms of the Apache 2.0 license. It’s a precaution, in case some jurisdiction doesn’t uphold the passive licensing scheme used otherwise. There’s also a relicensing clause, but that’s restricted to keeping in spirit, they can’t close the source.
The idea that “it’s ok cause we’d do the same” is ridiculous. There is no comparison: China is an authoritarian government and the parent company is practically an arm of the state. There are legitimate criticisms of American tech companies obviously, but they’re ultimately subject to the market and democratic governments. We shouldn’t be doing any business with authoritarians in the first place, much less inviting them to control a significant social media app in the guise of a legitimate business.
Arch for stuff I have physical access to. Nothing’s ever gone wrong, so it’s worth it for the immediate updates and consistency with my other systems. For VPS I use Debian though, occasionally the unstable/Sid branch if I really need the latest updates. There are almost always Debian images available on a VPS.
If I’m understanding this right, and this basically an API that lets you pick which app store administers an app, that could be quite helpful, not harmful. I currently have fdroid, play store, and Samsung store, and I assume they try to update apps by the fully qualified name, as multiple stores show and try to update a single app instance, sometimes with weird results.
This is a classic problem for machine learning systems, sometimes called over fitting or memorization. By analogy, it’s the difference between knowing how to do multiplication vs just memorizing the times tables. With enough training data and large enough storage AI can feign higher “intelligence”, and that is demonstrably what’s going on here. It’s a spectrum as well. In theory, nearly identical recall is undesirable, and there are known ways of shifting away from that end of the spectrum. Literal AI 101 content.
Edit: I don’t mean to say that machine learning as a technique has problems, I mean that implementations of machine learning can run into these problems. And no, I wouldn’t describe these as being intelligent any more than a chess algorithm is intelligent. They just have a much more broad problem space and the natural language processing leads us to anthropomorphize it.
I think about it like a tree structure for both. With a gui you have to move your mouse around to various places, with a cli each character branches off into another tree. Mathematically you can handle more options faster with a CLI.
Curiosity, back around 2010 before I was a teenager. No clue how I heard about it, but the concept of replacing the entire operating system was fascinating. I figured it must be really good if it was such a well kept secret.
A few years later, when I started to learn programming, Linux was the obvious winner. The online course taught C in a Linux environment, and I was amazed that the default Ubuntu build at the time had everything built in, whereas a Windows equivalent required visual studio and licensing adventures.
It really stuck as a daily driver after Windows 7, where a clear trend emerged: Windows got in my way, Linux got out of my way. Simple as.
I use Arch for my daily, and I would highly recommend against it for new users. 99% of the time it’s just fine. 1% of the time some edge case sneaks by and you update before a fix is pushed. In those cases, I’ve had installations be deeply broken, far beyond my expectations of normal users.
For actual recommendations, something Debian based for sure. Vanilla Debian, Mint, or Mint Debian edition. If you wanna live on the edge, Sid is rolling but in my experience was more stable than Arch.
I use netdata, it’s very good at digesting thousands of metrics to sharing actionable. The cloud portion is proprietary, but you can toggle off the data collection. I did turn on the cloud portion though, I get email notifications when something breaks. Might sound counter to the self hosted mantra, but a self hosted monitoring system isn’t very helpful when your own systems go down.
I recently removed in editor AI cause I noticed I was acquiring muscle memory for my brain, not thinking through the rest past the start of a snippet that would get an LLM to auto complete. I’m still using LLMs, particularly for languages and libraries I’m not familiar with, but using the artifacts editors in ChatGPT and Claude.