cross-posted from: https://lemmy.world/post/2312869

AI researchers say they’ve found ‘virtually unlimited’ ways to bypass Bard and ChatGPT’s safety rules::The researchers found they could use jailbreaks they’d developed for open-source systems to target mainstream and closed AI systems.

  • LoafyLemon@kbin.social
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    1 year ago

    Not really. This isn’t AGI but a text transformer. They trained it so the most probable answer to unwanted questions is ‘I’m sorry but as an AI…’.

    However, if you phrase your question in a way researchers haven’t thought about, you will bypass the filter.

    There’s not an ounce of intelligence in LLMs, it’s all statistics.