• squiblet@kbin.social
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    1 year ago

    It’s all about the models and training, though. People thinking ChatGPT 3.5/4 can write their legal papers get tripped up because it confabulates (‘hallucinates’) when it isn’t thoroughly trained on a subject. If you fed every legal case for the past 150 years into a model, it would be very effective.

      • squiblet@kbin.social
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        That’s a matter of working on the prompt interpreter.

        For what I was saying, there’s no assumption: models trained on more data and more specific data can definitely do the usual information summary tasks more accurately. This is already being used to create specialized models for legal, programming and accounting.

          • squiblet@kbin.social
            link
            fedilink
            arrow-up
            3
            ·
            1 year ago

            I agree that while it’s powerful and the capabilities are novel, it’s more limited than many think. Some people believe current “ai” systems/models can do just anything, like legal briefs or entire working programs in any language.The truth and accuracy flaws necessitate some serious rethinking. There are, like your above example, major flaws when you try to do something like simple arithmetic, since the system is not really thinking about it.