• magic_lobster_party@kbin.run
    link
    fedilink
    arrow-up
    3
    ·
    edit-2
    4 months ago

    The explanation is not that simple. Some model configurations work well. Others don’t. Not all continuous and differentiable models cut it.

    It’s not given a model can generalize the problem so well. It can just memorize the training data, but completely fail on any new data it hasn’t seen.

    What makes a model be able to see a picture of a cat it has never seen before, and respond with “ah yes, that’s a cat”? What kind of “cat-like” features has it managed to generalize? Why does these features work well?

    When I ask ChatGPT to translate a script from Java to Python, how is it able to interpret the instruction and execute it? What features has it managed to generalize to be able to perform this task?

    Just saying “why wouldn’t it work” isn’t a valid explanation.