• 0 Posts
  • 87 Comments
Joined 1 year ago
cake
Cake day: June 15th, 2023

help-circle

  • There is so much wrong with this…

    AI is a range of technologies. So yes, you can make surveillance with it, just like you can with a computer program like a virus. But obviously not all computer programs are viruses nor exist for surveillance. What a weird generalization. AI is used extensively in medical research, so your life might literally be saved by it one day.

    You’re most likely talking about “Chat Control”, which is a controversial EU proposal to scan either on people’s devices or from provider’s ends for dangerous and illegal content like CSAM. This is obviously a dystopian way to achieve that as it sacrifices literally everyone’s privacy to do it, and there is plenty to be said about that without randomly dragging AI into that. You can do this scanning without AI as well, and it doesn’t change anything about how dystopian it would be.

    You should be using end to end regardless, and a VPN is a good investment for making your traffic harder to discern, but if Chat Control is passed to operate on the device level you are kind of boned without circumventing this software, which would potentially be outlawed or made very difficult. It’s clear on it’s own that Chat Control is a bad thing, you don’t need some kind of conspiracy theory about ‘the true purpose of AI’ to see that.



  • I feel you in avoiding public transit. That’s where my hate comes from as well. And yes, many people that do these things have have excuses. Because they need to, to justify doing their business in a place where their habit unavoidably harms and frustrates other people. I hate the fact we still allow that so readily as society. Or at least we should restrict it further to the point a normal person doesn’t have to be bothered by people like that in public. It undermines public services to an extent.

    But after I no longer needed to use public transit, I did start to see things in a slightly different light. And that’s the only thing I wanted to say. People that are conscientious about enjoying any kind of mind altering substances will choose to do so safely and harmlessly outside of public, or in designated places like clubs specifically for that substance. Harm reduction must be central to substance use. And I know now that many people have that mentality. But that mentality is somewhat threatened exactly because they make sure nobody is bothered by them. It causes the experience to be defined by those people in public places, the loud minority.


  • I hate public smokers with a passion. But you must realize that you have effectively zero exposure to people that contain their smoke by doing it at home or using a method without smoke production. And there could be a lot more of those.

    The last line is especially golden for me since I live in the Netherlands so we have plenty of weed being smoked but the vast vast majority of public smoke hinderance is from tobacco smokers. If they decide to smoke in public they have absolutely no shame and will literally do it at places like bus stops and just outside restaurants. Weed smokers rarely do that here. So if I were to believe you it seems to just be correlated to people with shitty attitudes rather than the substance.

    But there’s no denying that if everyone would drop alcohol for weed, it would be better. Not because weed is harmless but because alcohol is pretty terrible health wise.


  • I never anthropomorphized the technology, unfortunately due to how language works it’s easy to misinterpret it as such. I was indeed trying to explain overfitting. You are forgetting the fact that current AI technology (artificial neural networks) are based on biological neural networks. There is a range of quirks that it exhibits that biological neural networks do as well. But it is not human, nor anything close. But that does not mean that there are no similarities that can be rightfully pointed out.

    Overfitting isn’t just what you describe though. It also occurs if the prompt guides the AI towards a very specific part of it’s training data. To the point where the calculations it will perform are extremely certain about what words come next. Overfitting here isn’t caused by an abundance of data, but rather a lack of it. The training data isn’t being produced from within the model, but as a statistical inevitability of the mathematical version of your prompt. Which is why it’s tricking the AI, because an AI doesn’t understand copyright - it just performs the calculations. But you do. And so using that as an example is like saying “Ha, stupid gun. I pulled the trigger and you shot this man in front of me, don’t you know murder is illegal buddy?”

    Nobody should be expecting a machine to use itself ethically. Ethics is a human thing.

    People that use AI have an ethical obligation to avoid overfitting. People that produce AI also have an ethical obligation to reduce overfitting. But a prompt quite literally has infinite combinations (within the token limits) to consider, so overfitting will happen in fringe situations. That’s not because that data is actually present in the model, but because the combination of the prompt with the model pushes the calculation towards a very specific prediction which can heavily resemble or be verbatim the original text. (Note: I do really dislike companies that try to hide the existence of overfitting to users though, and you can rightfully criticize them for claiming it doesn’t exist)

    This isn’t akin to anything human, people can’t repeat pages of text verbatim like this and no toddler can be tricked into repeating a random page from a random book as you say.

    This is incorrect. A toddler can and will verbatim repeat nursery rhymes that it hears. It’s literally one of their defining features, to the dismay of parents and grandparents around the world. I can also whistle pretty much my entire music collection exactly as it was produced because I’ve listened to each song hundreds if not thousands of times. And I’m quite certain you too have a situation like that. An AI’s mind does not decay or degrade (Nor does it change for the better like humans) and the data encoded in it is far greater, so it will present more of these situations in it’s fringes.

    but it isn’t crafting its own sentences, it’s using everyone else’s.

    How do you think toddlers learn to make their first own sentences? It’s why parents spend so much time saying “Papa” or “Mama” to their toddler. Exactly because they want them to copy them verbatim. Eventually the corpus of their knowledge grows big enough to the point where they start to experiment and eventually develop their own style of talking. But it’s still heavily based on the information they take it. It’s why we have dialects and languages. Take a look at what happens when children don’t learn from others: https://en.wikipedia.org/wiki/Feral_child So yes, the AI is using it’s training data, nobody’s arguing it doesn’t. But it’s trivial to see how it’s crafting it’s own sentences from that data for the vast majority of situations. It’s also why you can ask it to talk like a pirate, and then it will suddenly know how to mix in the essence of talking like a pirate into it’s responses. Or how it can remember names and mix those into sentences.

    Therefore it is factually wrong to state that it doesn’t keep the training data in a usable format

    If your arguments is that it can produce something that happens to align with it’s training data with the right prompt, well yeah that’s not incorrect. But it is so heavily misguided and borders bad faith to suggest that this tiny minority of cases where overfitting occurs is indicative of the rest of it. LLMs are a prediction machines, so if you know how to guide it towards what you want it to predict, and that is in the training data, it’s going to predict that most likely. Under normal circumstances where the prompt you give it is neutral and unique, you will basically never encounter overfitting. You really have to try for most AI models.

    But then again, you might be arguing this based on a specific AI model that is very prone to overfitting, while I am arguing this out of the technology as a whole.

    This isn’t originality, creativity or anything that it is marketed as. It is storing, encoding and copying information to reproduce in a slightly different format.

    It is originality, as these AI can easily produce material never seen before in the vast, vast majority of situations. Which is also what we often refer to as creativity, because it has to be able to mix information and still retain legibility. Humans also constantly reuse phrases, ideas, visions, ideals of other people. It is intellectually dishonest to not look at these similarities in human psychology and then treat AI as having to be perfect all the time, never once saying the same thing as someone else. To convey certain information, there are only finite ways to do so within the English language.





  • Your first point is misguided and incorrect. If you’ve ever learned something by ‘cramming’, a.k.a. just repeating ingesting material until you remember it completely. You don’t need the book in front of you anymore to write the material down verbatim in a test. You still discarded your training material despite you knowing the exact contents. If this was all the AI could do it would indeed be an infringement machine. But you said it yourself, you need to trick the AI to do this. It’s not made to do this, but certain sentences are indeed almost certain to show up with the right conditioning. Which is indeed something anyone using an AI should be aware of, and avoid that kind of conditioning. (Which in practice often just means, don’t ask the AI to make something infringing)


  • This would be a good point, if this is what the explicit purpose of the AI was. Which it isn’t. It can quote certain information verbatim despite not containing that data verbatim, through the process of learning, for the same reason we can.

    I can ask you to quote famous lines from books all day as well. That doesn’t mean that you knowing those lines means you infringed on copyright. Now, if you were to put those to paper and sell them, you might get a cease and desist or a lawsuit. Therein lies the difference. Your goal would be explicitly to infringe on the specific expression of those words. Any human that would explicitly try to get an AI to produce infringing material… would be infringing. And unknowing infringement… well there are countless court cases where both sides think they did nothing wrong.

    You don’t even need AI for that, if you followed the Infinite Monkey Theorem and just happened to stumble upon a work falling under copyright, you still could not sell it even if it was produced by a purely random process.

    Another great example is the Mona Lisa. Most people know what it looks like and if they had sufficient talent could mimic it 1:1. However, there are numerous adaptations of the Mona Lisa that are not infringing (by today’s standards), because they transform the work to the point where it’s no longer the original expression, but a re-expression of the same idea. Anything less than that is pretty much completely safe infringement wise.

    You’re right though that OpenAI tries to cover their ass by implementing safeguards. Which is to be expected because it’s a legal argument in court that once they became aware of situations they have to take steps to limit harm. They can indeed not prevent it completely, but it’s the effort that counts. Practically none of that kind of moderation is 100% effective. Otherwise we’d live in a pretty good world.



  • Every piece of legislature ever needs to deal with the emotions of it’s subjects. An unemphatic, but cold hard rational law, will be nothing less of tyrannical most of the time. Laws are for humans to follow, and humans have emotions that need to be understood for a law to be successful and supported to last into the future. A law that isn’t supported by it’s subjects eventually leads to revolution (big and small).

    How logic and rational a person can be is highly dependent on their emotional intelligence. You might be able to suppress your emotions when there is no stress at all, but if you cave during a stressful situation and start lashing out, that does impact your overall intelligence. Intelligence is just the collection of behaviors and training that make you effective at doing what you want to do, and being rational and logical is definitely good, but not the end of it all.





  • Exactly, thinking that is what I was getting pulled me over the edge, I sometimes remember a music video I want to listen to on my phone during my commute and I don’t want to spend 30 minutes either getting on my PC to download it with a tool, or using a third party downloader which can at times be shady. So upgrading that to a single click in the app seemed like a great deal. Crushingly disappointed when I found out how it actually was. Turns out the real answer was NewPipe, which I don’t even have to pay for.


  • Yeah, I thought it was a nice compromise. It seemed sensible that if Premium is the ‘compliant’ response to not wanting ads, the ‘compliant’ response to using third party tools to download videos was to just be able to do things more easily and with more options through Premium as well. But apparently they wanted to advertise something anyone who’s wanted to download a youtube video would not describe as ‘downloading’, which is easily out competed by free (but at times shady) tools.



  • ClamDrinker@lemmy.worldtoMemes@lemmy.mlMe but ublock origin
    link
    fedilink
    arrow-up
    9
    ·
    edit-2
    3 months ago

    I like those perks too, but if I pay more to be able to download videos (which again, I could’ve used a free tool for) I want to be able to do whatever I want with it. Download means getting a file I can watch using my own video player and store for later even if Youtube dies tomorrow, If I go on holiday without internet, or if my internet goes down for a week. Anything.

    If Google is going to be “Uhm aksually, you are technically downloading it, thats why we can advertise it like that”, then I’m already downloading literally every video I watch. And thats not the kind of bullshit you give to a paying customer. That is spitting in my face for paying you. Why does a non-Premium user get better service with free third party youtube downloaders?

    It’s a matter of principle.