![](https://lemmy.blahaj.zone/pictrs/image/vjZxbFTXph.jpg)
![](https://fry.gs/pictrs/image/c6832070-8625-4688-b9e5-5d519541e092.png)
I mean yeah, but I they were talking about net neutrality, preventing ISPs from unilaterally making those decisions, not that there would be Literally No restrictions.
20, they/she, math+CS student
I mean yeah, but I they were talking about net neutrality, preventing ISPs from unilaterally making those decisions, not that there would be Literally No restrictions.
I mean, LLaMA is open-source and it’s made by Facebook for profit, there’s grey areas. Imo tho, any service that claims to be anything more than a fancy wrapper for OpenAI, Anthropic, etc. API calls is possibly a scam. Especially if they’re trying to sell you hardware, or the service costs more than like $10/month, LLM API calls are obscenely cheap. I use a local frontend as an AI assistant that works by making API calls through a service called openrouter (basically a unified service that makes API calls to all the major cloud LLM providers for you). I put like $5 in it 3 or 4 months ago and it still hasn’t run out.
This is because dedicated consumer AI hardware is a dumb idea. If it’s powerful enough to run a model locally, you should be able to use it for other things (like, say, as a phone or PC) and if it’s sending all its API requests to the cloud, then it has no business being anything but a smartphone app or website.
If the social biases of the model put a hard limit on your ability to write a good woman character, I question how much it’s really you that’s “writing” the story. I’m not against using LLMs in writing, but it’s a tool, not a creative partner. They can be useful for brainstorming and as a sounding-board for ideas (potentially even editing), but imo you need to write the actual prose yourself to claim you’re writing something.
I mean, the correct answer is just to pirate it, but you could probably fool it into thinking you’re on windows by changing the user agent string of your browser to Chrome Windows x86_64
Imo a non-market-based socialist economy doesn’t require AI, it just requires extensively documenting inventory/production and a good mechanism for gauging consumer demand, in other words a good economic planning mechanism. Because if you break it down to its simplest function, the capitalist market is just an economic planning mechanism, it uses consumer purchases to judge demand and adjust production accordingly, but it’s more difficult to control since it can’t account for negative externalities (effects of production that don’t have a direct impact on sales), and when you introduce wage labor into a market, the incentive structure encourages those with more resources to spend those resources on labor and then exploit that labor to the maximum possible extent.
To implement a non-market socialist planning system, you could have a broad plan that specifies large macroeconomic goals to be voted on democratically (e.g. increase investment in clean energy, increase investment in a certain popular sector of consumer goods like electronics, etc.) And then use data from the past to estimate future demand for consumer goods. Then you could calculate the demand for intermediate commodities (things used to produce consumer goods and accomplish larger state infrastructure projects) based on that. You could then put all of those into a really big matrix and then row-reduce that matrix to solve for any areas where the necessary resource allocation is uncertain and then use that to refine the initial estimates for production you started with, and iterate this process many times. At the end you end up with a table that shows exactly which resources need to be allocated where to meet production targets and consumer demand.
Obviously consumer demand might differ from your prior estimates, so you also have a system for monitoring how much of each good is purchased at each storefront, and then make minor adjustments to the global production targets to meet the actual demand, which would help mitigate the recurring shortages that occurred in the USSR due to inefficient resource allocation from their oversimplified planning system.
Basically, it’s all about having good-quality, real-time data on economic activity so that the planned economy can respond as dynamically as a capitalist market, but without the negative externalities and worker exploitation that come along with capitalism.
I read a really good book that outlines in more detail how this would work and even gives the algorithm for efficiently manipulating the economy-wide resource allocation matrix, it’s called Towards a New Socialism. Apparently the guys who wrote it are weird transphobes now, so I don’t endorse them personally, but it’s the most well thought out, concrete plan for a workable socialist economy I’ve seen so it’s worth a read. Also look into project CyberSyn in Chile under Allende, it’s the closest attempt irl to do something like this.
We’d probably be fucked, but it’d be interesting to see. The professed ideology of the Chinese state is very different from their actual actions – for example, they mandate study of texts by Marx and Lenin that advocate for worker-controlled militant labor organizing while at the same time banning any non-state-controlled unions. An actually Marxist AI might turn on them lmao
That depends on your use case, I just did a simple zpool with no redundancy because I wanted maximum speed/capacity and all my data is backed up on an external HDD. If you need redundancy, I would look online for how to configure that and what the optimal setup is.
Actually, I assumed you just had the SSD, if you have more than 256gb of free space between those HDDs, you can go ahead and remove the SSD from your zpool right now (unless your bootloader is there, then you’ll have to make an EFI system partition on one of the HDDs and install a bootloader first)
Fair, haven’t use Ubuntu or any of it’s derivatives in years
You need to add the new drive to your existing pool because ZFS stores data across all drives by default, similar to a RAID0. Then you remove the old drive and ZFS will automatically copy the data off the failing drive onto the healthy one and allow you to remove the failing drive with no data loss.
Manjaro is basically just arch Linux on a 1-2 week update lag, so you’d have just as much if not more success with EndeavorOS or raw Arch.
Honestly my main issue with Manjaro is still that they hold updates for a week or two for “testing” which tends to break certain AUR packages. I’d be less mad if the testing actually amounted to anything, but half the time they basically do nothing, and if there were any bugs Arch has released updates that resolve them already, which you won’t get for another week because of their update schedule. Anytime anyone talks about being interested in Manjaro, I just recommend they get EndeavorOS instead, it’s basically stock arch with a fancy installer and sane defaults which is great for anyone who mostly knows what they’re doing with Linux (or is at least capable of opening a terminal window and pasting error messages into google or, failing that, ChatGPT and following basic instructions)
Doesn’t Mint hold back kernel updates to major version upgrades like Ubuntu though? That could be problematic if they have newer hardware that’s better supported (or only supported at all) in newer kernel releases.
Anything with a recent kernel is fine. If you’re not very experienced, I’d recommend something like Fedora or OpenSUSE (both semi-rolling releases so you’ll get new kernels, graphics drivers, etc. but less likely to break for no reason than arch/gentoo derivatives).
Manjaro is fine if you don’t use the AUR, but arch/manjaro repositories on their own will be inadequate, and it will be so easy to get what’s missing from the AUR, which will eventually break something. This is because Manjaro holds back arch Linux updates for a week or two for “testing” purposes, but the AUR expects precisely the latest arch packages. If you’re thinking about Manjaro, do EndeavorOS instead. It’s the same thing (arch Linux with a more user friendly installer and relatively sane default apps/configs) with infinitely less hassle. Plus there’s really no point to using an arch-based distro without the AUR imo.
Garuda is also cool, I haven’t used it myself, but it’s supposed to be another preconfigured version of Arch more targeted towards gamers. YMMV, I’d probably just stick with EndeavorOS.
If you want an Ubuntu or Debian derivative, I’d go with Pop!OS. It’s basically Ubuntu without all the Ubuntu bullshit (snaps ludicrously out of date packages, etc), and they keep the kernel and video drivers pretty recent, unlike stock Ubuntu. Plus they have a cool desktop environment. Currently it’s a fork of GNOME, but they’re working on rewriting it from scratch and are making great progress, which will be interesting once it’s more developed.
I mean yeah. The only reason anyone is surprised when apple pulls shit like this is because they’re aggressively anti-consumer in their pricing and hardware design (parts pairing, poor reparability, etc.). People assume because they’re so flagrantly anti-consumer with their hardware, they can afford to not be so anti-consumer with their software. This is wrong, of course. They’re a publicly traded company, they’ll milk their users for every cent they can.
Use an Ubuntu live USB, all recent versions of Ubuntu have ZFS drivers baked into the live environment. Then you should add your new SSD to the ZFS pool, and remove the old one from the ZFS pool. Your m.2 WiFi slot should be able to host the 2nd drive while you do this, but if not you can use an external USB housing for it, you’ll just have to make sure that the ZFS pool knows its UUID so that it knows it’s the same drive.
Eh, open-sourcing is just good business, the only reason every big tech company doesn’t is that loads of executives are stuck in the past. Of course having random people on the internet do labor for you for free is something Google would want. They get the advantage of tens of thousands of extra eyes on their code pointing out potential security vulnerabilities and they can just put all the really shady shit in proprietary blobs like Google Play Services, they’re getting the best of both worlds as far as they’re concerned.
Large publicly-traded companies do not do anything for the good of anyone but themselves, they are literally Legally Obligated to make the most profitable decisions for themselves at all times. If they’re open-sourcing things it’s to make money, not because they were “good guys”.
I think it’ll end up like Facebook (the social media platform, not the company). Eventually you’ll hit model collapse for new models trained off uncurated internet data once a critical portion of all online posts are made by AI, and it’ll become Much more expensive to create quality, up-to-date datasets for new models. Older/less tech literate people will stay on the big, AI-dominated platforms getting their brains melted by increasingly compelling, individually-tailored AI propaganda and everyone else will move to newer, less enshittified platforms until the cycle repeats.
Maybe we’ll see an increase in discord/matrix style chatroom type social media, since it’s easier to curate those and be relatively confident everyone in a particular server is human. I also think most current fediverse platforms are also marginally more resistant to AI bots because individual servers can have an application process that verifies your humanity, and then defederate from instances that don’t do that.
Basically anything that can segment the Unceasing Firehose of traffic on the big social media platforms into smaller chunks that can be more effectively moderated, ideally by volunteers because a large tech company would probably just automate moderation and then you’re back at square 1.
That’s fair, dedicated ASICs for AI acceleration are totally a valid consumer product, but I meant more along the lines of independent devices (like Rabbit R1 and the AI Pin), not components you can add to an existing device. I should have been more clear.