• Ferk@kbin.social
    link
    fedilink
    arrow-up
    11
    arrow-down
    1
    ·
    edit-2
    1 year ago

    A test that didn’t require a human could theoretically be tested automatically by the machine preemptively and solved easily.

    I can’t imagine how would you test this in a way that wouldn’t require a human.

      • bedrooms@kbin.social
        link
        fedilink
        arrow-up
        6
        ·
        edit-2
        1 year ago

        Bro, humans literally don’t have that capability (that’s the presumption here). Or are you saying that many of us don’t have better consciousness than AIs? I might agree with that!

      • Ferk@kbin.social
        link
        fedilink
        arrow-up
        2
        ·
        edit-2
        1 year ago

        The AI can only judge by having a neural network trained on what’s a human and what’s an AI (and btw, for that training you need humans)… which means you can break that test by making an AI that also accesses that same neural network and uses it to self-test the responses before outputting them, providing only exactly the kind of output the other AI would give a “human” verdict on.

        So I don’t think that would work very well, it’ll just be a cat & mouse race between the AIs.