• 1 Post
  • 57 Comments
Joined 1 year ago
cake
Cake day: June 26th, 2023

help-circle



  • But personally I’ll continue to advocate for technology which empowers people and culture, and not the other way around.

    You won’t achieve this goal by aiding the gatekeepers. Stop helping them by trying to misapply copyright.

    Any experienced programmer knows that GPL code is still subject to copyright […]

    GPL is a clever hack of a bad system. It would be better if copyright didn’t exist, and I say that as someone that writes AGPL code.

    I think you misunderstood what I meant. We should drop copyright, and pass a new law where if you use a model, or contribute to one, or a model is used against you, that model must be made available to you. Similar in spirit to the GPL, but not a reliant on an outdated system.

    This would catch so many more use cases than trying to cram copyright where it doesn’t apply. No more:

    • Handful of already-rich companies building an AI moat that keeps newcomers out
    • Credit agencies assigning you a black box score that affects your entire life
    • Minorities being denied bail because of a black box model
    • Being put on a no-fly list with no way to know that you’re on it or why
    • Facebook experimenting on you to see if they can make you sad without your knowledge

  • Why should they? Copyright is an artificial restriction in the first place, that exists “To promote the Progress of Science and useful Arts” (in the US, but that’s where most companies are based). Why should we allow further legal restrictions that might strangle the progress of science and the useful arts?

    What many people here want is for AI to help as many people as possible instead of just making some rich fucks richer. If we try to jam copyright into this, the rich fucks will just use it to build a moat and keep out the competition. What you should be advocating for instead is something like a mandatory GPL-style license, where anybody who uses the model or contributed training data to it has the right to a copy of it that they can run themselves. That would ensure that generative AI is democratizing. It also works for many different issues, such as biased models keeping minorities in jail longer.

    tl;dr: Advocate for open models, not copyright










  • BitSound@lemmy.worldtoTechnology@lemmy.mlGPT-4 Understands
    link
    fedilink
    arrow-up
    2
    arrow-down
    2
    ·
    9 months ago

    Your concept of a chair is an abstract thought representation of a chair. An LLM has vectors that combine or decompose in some way to turn into the word “chair,” but are not a concept of a chair or an abstract representation of a chair. It is simply vectors and weights, unrelated to anything that actually exists.

    Just so incredibly wrong. Fortunately, I’ll have save myself time arguing with such a misunderstanding. GPT-4 is here to help:

    This reads like a misunderstanding of how LLMs (like GPT) work. Saying an LLM’s understanding is “simply vectors and weights” is like saying our brain’s understanding is just “neurons and synapses”. Both systems are trying to capture patterns in data. The LLM does have a representation of a chair, but it’s in its own encoded form, much like our neurons have encoded representations of concepts. Oversimplifying and saying it’s unrelated to anything that actually exists misses the point of how pattern recognition and information encoding works in both machines and humans.


  • BitSound@lemmy.worldtoTechnology@lemmy.mlGPT-4 Understands
    link
    fedilink
    arrow-up
    2
    arrow-down
    2
    ·
    edit-2
    9 months ago

    You really, truly don’t understand what you’re talking about.

    The vectors do not represent concepts. The vectors are math

    If this community values good discussion, it should probably just ban statements that manage to be this wrong. It’s like when creationists say things like “if we came from monkeys why are they still around???”. The person has just demonstrated such a fundamental lack of understanding that it’s better to not engage.


  • BitSound@lemmy.worldtoTechnology@lemmy.mlGPT-4 Understands
    link
    fedilink
    arrow-up
    3
    arrow-down
    1
    ·
    9 months ago

    processed into a script for ELIZA

    That wouldn’t accomplish anything. I don’t know why the OP brought it up, and that subject should just get dropped. Also yes, you can use your intelligence to string together multiple tools to accomplish a particular task. Or you can use the intelligence of GPT-4 to accomplish the same task, without any other tools

    LLMs lack the capability of understanding and comprehension

    Also not true

    states that it is not using an accepted definition of intelligence.

    Nowhere does it state that. It says “There is no generally agreed upon definition of intelligence”. I’m not sure why you’re bringing up a physical good such as leather here. Two things: a) grab a microscope and inspect GPT-4. The comparison doesn’t make sense. b) “Is” should be banned, it encourages lazy thought and pointless discussion (Yes I’m guilty of it in this comment, but it helps when you really start asking what “is” means in context). You’re wandering into p-zombie territory, and my answer is that “is” means nothing. GPT-4 displays behaviors that are useful because of their intelligence, and nothing else matters from a practical standpoint.

    it is clear that LLMs may be useful components in building actual general intelligence.

    You’re staring the actual general intelligence in the face already, there’s no need to speculate about perhaps being components. There’s no reason right now to think that we need anything more than better compute. The actual general intelligence is yet a baby, and has experienced the world through the tiny funnel of human text, but that will change with hardware advances. Let’s see what happens with a few orders of magnitude more computing power.


  • BitSound@lemmy.worldtoTechnology@lemmy.mlGPT-4 Understands
    link
    fedilink
    arrow-up
    1
    arrow-down
    3
    ·
    9 months ago

    For your edit: Yes, that’s what’s known as the context window limit. ChatGPT has an 8k token “memory” (for most people), and older entries are dropped. That’s not an inherent limitation of the approach, it’s just a way of keeping OpenAI’s bills lower.

    Without an example I don’t think there’s anything to discuss. Here’s one trivial example though where I altered ChatGPT’s understanding of the world:

    If I continued that conversation, ChatGPT would eventually forget that due to the aforementioned context window limit. For a more substantive way of altering an LLM’s understanding of the world, look at how OpenAI did RLHF to get ChatGPT to not say naughty things. That permanently altered the way GPT-4 responds, in a similar manner to having an angry nun rap your knuckles whenever you say something naughty.


  • BitSound@lemmy.worldtoTechnology@lemmy.mlGPT-4 Understands
    link
    fedilink
    arrow-up
    3
    arrow-down
    2
    ·
    9 months ago

    LLMs can certainly do that, why are you asserting otherwise?

    ChatGPT can do it for a single session, but not across multiple sessions. That’s not some inherent limitations to LLMs, that’s just because it’s convenient for OpenAI to do it that way. If we spun up a copy of a human from the same original state every time you wanted to ask it a question and then killed it after it was done responding, it similarly wouldn’t be able to change its behavior across questions.

    Like, imagine we could do something like this. You could spin up a copy of that brain image, alter its understanding of the world, then spin up a fresh copy that doesn’t have that altered understanding. That’s essentially what we’re doing with LLMs today. But if you don’t spin up a fresh copy, it would retain its altered understanding.


  • BitSound@lemmy.worldtoTechnology@lemmy.mlGPT-4 Understands
    link
    fedilink
    arrow-up
    5
    arrow-down
    1
    ·
    9 months ago

    Give Eliza equivalent compute time and functionality to interpret the data type and it probably could get something approaching a result.

    Sorry, but this is simply incorrect. Do you know what Eliza is and how it works? It is categorically different from LLMs.

    That’s not something that is seriously debated academically

    This is also incorrect. I think the issue that many people have is that they hear “AI” and think “superintelligence”. What we have right now is indeed AI. It’s a primitive AI and certainly no superintelligence, but it’s AI nonetheless.

    There is no known reason to think that the approach we’re taking now won’t eventually lead to superintelligence with better hardware. Maybe we will hit some limit that makes the hype die down, but there’s no reason to think that limit exists right now. Keep in mind that although this is apples vs oranges, GPT-4 is a fraction of the size of a human brain. Let’s see what happens when hardware advances give us a few more orders of magnitude. There’s already a huge, noticeable difference between GPT 3.5 and GPT 4.


  • BitSound@lemmy.worldtoTechnology@lemmy.mlGPT-4 Understands
    link
    fedilink
    arrow-up
    5
    arrow-down
    1
    ·
    9 months ago

    What exactly do you think would happen if you could make an exact duplicate of a human and run it from the same state multiple times? They would generate exactly the same output every time. How could you possibly think differently without turning to human exceptionalism and believing in magic meat?