• 1 Post
  • 24 Comments
Joined 1 year ago
cake
Cake day: June 29th, 2023

help-circle

  • Excrubulent@slrpnk.nettoProgrammer Humor@programming.devWhy spend money on ChatGPT?
    link
    fedilink
    English
    arrow-up
    88
    arrow-down
    3
    ·
    edit-2
    4 months ago

    I asked this question ages ago and it was pointed out that “sub” isn’t a reddit specific term. It’s been short for “subforum” since the first BBSes, so it’s basically a ubiquitous internet term.

    “Sub” works because everybody already knows what you mean and it’s the word you intuitively reach for.

    You can call them “communities” if you want, but it’s longer and can’t easily be shortened.

    I just call them subs now.



















  • It’s definitely a qualitative shift. I suspect most of the fundamental maths of neural network matrices won’t need to change, because they are enough to emulate the lower level functions of our brains. We have dedicated parts of our brain for image recognition, face recognition, language interpretation, and so on, very analogous to the way individual NNs do those same functions. We got this far with biomimicry, and it’s fascinating to me that biomimicry on the micro level is naturally turning into biomimicry on a larger scale. It seems reasonable to believe that process will continue.

    Perhaps some subtle tuning of those matrices is needed to really replicate a mind, but I suspect the actual leap will require first of all a massive increase in raw computation, as well as some new insight into how to arrange all of those subsystems within a larger structure.

    What I find interesting is the question of whether AI can actually fully replace a person in a job without crossing that threshold and becoming AGI, and I genuinely don’t think it can. Sure it’ll be able to automate some very limited tasks, but without the capacity to understand meaning it can’t ever do real problem solving. I think past that point it has to be considered a person with all of the ethical implications that has, and I think tech bros intentionally avoid acknowledging that, because that would scare investors.


  • You’re describing an arms race, which makes me wonder if that’s part of the path to AGI. Ultimately the only way to truly detect a fake is to compare it to reality, and the only way to train a model to understand whether it is looking at reality or a generated image is to teach it to understand context and meaning, and that’s basically the ballgame at that point. That’s a qualitative shift, and in that scenario we get there with opposing groups each pursuing their own ends, not with a single group intentionally making AGI.