![](https://lemdro.id/pictrs/image/bde232c1-cd9b-4367-8f80-96ae69a76aee.jpeg)
![](https://fry.gs/pictrs/image/c6832070-8625-4688-b9e5-5d519541e092.png)
Yeah as someone who has worked in web development for over 20 years everything in here is completely standard. Almost every major website in existence collects this kind of analytical data.
Yeah as someone who has worked in web development for over 20 years everything in here is completely standard. Almost every major website in existence collects this kind of analytical data.
This happens to me constantly. Just the other day I asked some friends for something and then they sent the literal exact opposite of that thing. Pretend I asked for blue with red stripes they gave me green with yellow polka dots. And it wasn’t just one person it was three separate people who all decided that made sense for some reason.
I was extremely specific too, even more than usual because I know people constantly misinterpret me. I made extra sure to not use any language with vague meanings and it still happened anyway. It’s like we live in alternate realities where words have completely different meanings.
It makes me not want to talk to people at all.
Again, even an exact copy is not stealing. It’s copyright infringement. Theft is a different crime.
But paraphrasing is not copyright infringement either. It’s no different than Wikipedia having a synopsis for every single episode of a TV series. Telling someone about what a work contains for informational purposes is perfectly fine.
Sorry, I misinterpreted what you meant. You said “any AI models” so I thought you were talking about the model itself should somehow know where the data came from. Obviously the companies training the models can catalog their data sources.
But besides that, if you work on AI you should know better than anyone that removing training data is counter to the goal of fixing overfitting. You need more data to make the model more generalized. All you’d be doing is making it more likely to reproduce existing material because it has less to work off of. That’s worse for everyone.
What you’re asking for is literally impossible.
A neural network is basically nothing more than a set of weights. If one word makes a weight go up by 0.0001 and then another word makes it go down by 0.0001, and you do that billions of times for billions of weights, how do you determine what in the data created those weights? Every single thing that’s in the training data had some kind of effect on everything else.
It’s like combining billions of buckets of water together in a pool and then taking out 1 cup from that and trying to figure out which buckets contributed to that cup. It doesn’t make any sense.
If the model isn’t overfitted it’s also not even copying. By their nature LLMs are transformative which is the whole point of fair use.
No it works perfectly fine with a mod for uncapped FPS
I pretty much never reboot the Pi. It currently has over 18 months of uptime on it. My NAS on the other hand I probably restart for one reason or another maybe once every 6 months. So yeah I’d say I reboot it minimum 3x more often.
Plus a reboot takes much longer on my NAS than on the Pi. The server board is slow to start, the SAS cards are slow to start, and unRAID is slow to start. Then I need to manually enter the password for disk encryption. Then wait for the array to start up. Then wait a bit more for the docker containers to start. Add all of that up and even the absolute fastest reboot is like 10 minutes while the Pi probably takes 30 seconds.
And what if I want to swap hard drives? Now it’s down for an hour. I guess I could wait until 3am to do all my upgrades so everyone is asleep, but I’d rather not. I suppose if it were just for myself it would matter a lot less. But again, it’s only $15 to not have to think about it at all.
I used to do that, but it comes with the problem of your DNS going down any time you want to restart or do a hardware swap on your NAS. Or since it was running in docker something as simple as reloading docker would knock out the internet for a few minutes. It’s worth the $15 to have them operate separately.
$80? I run mine on a Pi Zero that I got for $9 with a $6 wired network adapter for a grand total of $15. No problems for a household of five with one of us (me) being an extremely heavy user.
I work from home and never call anyone so on Tello I pay $6/mo for 100 minutes + 1GB of data that pretty much functions as a 2FA delivery system.
I have actually been seeing some timestamps that are completely wrong lately, maybe this is why.
I’ve never had a job where we celebrate Shitty Ass Day. What’s that like?
You are looking for S0und
I also use the Jellyfin+Symfonium combo. There is also Finamp as the free and open source option. And Feishin on desktop.
Same story for me on a OnePlus 5T which is the even older Snapdragon 835. Firefox is genuinely unusable. I tried Mull and Iceraven too. For several months I tried to put up with it, but they were all a slow and buggy mess. Switched to Brave and it works fine.
I use Librewolf on my desktop for the record.
There is also FreshTomato if your router has Broadcom wifi chipset like mine does.
That’s my experience as well. Now I use a SearXNG instance with Google as the only source and it works well I’d say roughly 97% of the time.
Like the comment I replied to already explained, this information is necessary to make informed development decisions. If you don’t know who is using what feature you might be wasting resources on something barely anyone uses while neglecting something everyone needs.
You also need some of that data for security purposes. You can’t implement rate limiting or prevent abuse if you can’t log and track how your services are being interacted with.
And this is aggregate data. I can promise you not a single person cares about what any individual user is doing (assuming it’s not illegal)