• Kit Sorens@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    5
    ·
    11 months ago

    Yet language and abstraction are the core of intelligence. You cannot have intelligence without 2 way communication, and if anything, your brain contains exactly that dictionary you describe. Ask any verbal autistic person, and 90% of their conversations are scripted to a fault. However, there’s another component to intelligence that the Turing Test just scrapes against. I’m not philosophical enough to identify it, but it seems like the turing test is looking for lightning by listening for rumbling that might mean thunder.

      • 0ops@lemm.ee
        link
        fedilink
        arrow-up
        2
        ·
        11 months ago

        This is what it comes down to. Until we agree on a testable definition of “intelligence” (or sentience, sapience, consciousness or just about any descriptor of human thought), it’s not really science. Even in nature, what we might consider intelligence manifests in different organisms in different ways.

        We could assume that when people say intelligence they mean human-like intelligence. That might be narrow enough to test, but you’d probably still end up failing some humans and passing some trained models

          • 0ops@lemm.ee
            link
            fedilink
            arrow-up
            1
            ·
            edit-2
            11 months ago

            You’re right, it’s very much context dependent, and I appreciate your incite on how this clash between psychology and computer science muddies the terms. As a CS guy myself who’s just dipping my toes into NN’s, I lean toward the psychology definition, where intelligence is measured by behavior.

            In an artificial neural network, the algorithms that wrangle data and build a model aren’t really what makes the decisions, they just build out the “body” (model, generator functions) and “environment” (data format), so to speak. If anything that code is more comparable to DNA than any state of mind. Training on data is where the knowledge comes from, and by making connections the model can “reason” a good answer with the correlations it found. Those processes are vague enough that I don’t feel comfortable calling them algorithms, though. It’s pretty divorced from cold, hard code.