• 0 Posts
  • 122 Comments
Joined 1 year ago
cake
Cake day: July 28th, 2023

help-circle








  • Like the comment I replied to already explained, this information is necessary to make informed development decisions. If you don’t know who is using what feature you might be wasting resources on something barely anyone uses while neglecting something everyone needs.

    You also need some of that data for security purposes. You can’t implement rate limiting or prevent abuse if you can’t log and track how your services are being interacted with.

    And this is aggregate data. I can promise you not a single person cares about what any individual user is doing (assuming it’s not illegal)




  • This happens to me constantly. Just the other day I asked some friends for something and then they sent the literal exact opposite of that thing. Pretend I asked for blue with red stripes they gave me green with yellow polka dots. And it wasn’t just one person it was three separate people who all decided that made sense for some reason.

    I was extremely specific too, even more than usual because I know people constantly misinterpret me. I made extra sure to not use any language with vague meanings and it still happened anyway. It’s like we live in alternate realities where words have completely different meanings.

    It makes me not want to talk to people at all.



  • Sorry, I misinterpreted what you meant. You said “any AI models” so I thought you were talking about the model itself should somehow know where the data came from. Obviously the companies training the models can catalog their data sources.

    But besides that, if you work on AI you should know better than anyone that removing training data is counter to the goal of fixing overfitting. You need more data to make the model more generalized. All you’d be doing is making it more likely to reproduce existing material because it has less to work off of. That’s worse for everyone.


  • What you’re asking for is literally impossible.

    A neural network is basically nothing more than a set of weights. If one word makes a weight go up by 0.0001 and then another word makes it go down by 0.0001, and you do that billions of times for billions of weights, how do you determine what in the data created those weights? Every single thing that’s in the training data had some kind of effect on everything else.

    It’s like combining billions of buckets of water together in a pool and then taking out 1 cup from that and trying to figure out which buckets contributed to that cup. It doesn’t make any sense.



  • ayaya@lemdro.idtoLinux@lemmy.mlFlathub has passed 2 billion downloads
    link
    fedilink
    English
    arrow-up
    35
    arrow-down
    5
    ·
    edit-2
    3 months ago

    For me on Arch, Flatpaks are kinda useless. I can maybe see the appeal for other distros but Arch already has up-to-date versions of everything and anything that’s missing from the main repos is in the AUR.

    I also don’t like how it’s a separate package manager, they take up more space, and to run things from the CLI it’s flatpak run com.website.Something instead of just something. It’s super cumbersome compared to using normal packages.


  • ayaya@lemdro.idtolinuxmemes@lemmy.worldlow effort maymay
    link
    fedilink
    English
    arrow-up
    13
    ·
    3 months ago

    Same here. Switched to Arch in 2015 so I am also coming up on the 9 year mark. I have had very few issues, and the ones I have had were usually my fault for doing something stupid. I used Windows, OS X, and Ubuntu previously and compared to those Arch is a dream. Hence why I’ve stuck with it for so long now.