Oh hey, I love your work on Plasma’s HDR and color management. Glad to see you on Lemmy.
Oh hey, I love your work on Plasma’s HDR and color management. Glad to see you on Lemmy.
Phoronix is the ONLY website I disable uBlock Origin for.
I thought it was weird such an old piece of software had so much Rust in it. I noticed all the Rust-related things while Firefox Librewolf compiles but never looked into it further.
There is a pretty big difference in terms of usability between Arch and everything else because of the rolling release model and the AUR. Lots of things you would have to manually install from a git repo or track down a PPA for can be installed like a normal package.
Plasma actually has a UI for smart TVs if you weren’t aware, although I have never used it myself so I’m not sure how good it is. https://plasma-bigscreen.org
Really? For me rspamd blocks at least 15 spam emails a day, usually from China or Russia. An additional 2-3 go to the junk folder, and some still slip through the cracks especially if it’s coming from a gmail address.
But it could be as simple as it being because my email is publicly available (github, my website, etc.) so scrapers are picking it up.
Sorry, I misinterpreted what you meant. You said “any AI models” so I thought you were talking about the model itself should somehow know where the data came from. Obviously the companies training the models can catalog their data sources.
But besides that, if you work on AI you should know better than anyone that removing training data is counter to the goal of fixing overfitting. You need more data to make the model more generalized. All you’d be doing is making it more likely to reproduce existing material because it has less to work off of. That’s worse for everyone.
What you’re asking for is literally impossible.
A neural network is basically nothing more than a set of weights. If one word makes a weight go up by 0.0001 and then another word makes it go down by 0.0001, and you do that billions of times for billions of weights, how do you determine what in the data created those weights? Every single thing that’s in the training data had some kind of effect on everything else.
It’s like combining billions of buckets of water together in a pool and then taking out 1 cup from that and trying to figure out which buckets contributed to that cup. It doesn’t make any sense.
For me on Arch, Flatpaks are kinda useless. I can maybe see the appeal for other distros but Arch already has up-to-date versions of everything and anything that’s missing from the main repos is in the AUR.
I also don’t like how it’s a separate package manager, they take up more space, and to run things from the CLI it’s flatpak run com.website.Something
instead of just something
. It’s super cumbersome compared to using normal packages.
but after fresh install
See, there’s your problem. If you never re-install this is longer a factor. Sure I had to do those things, but I had to do them exactly once like 8 years ago…
deleted by creator