• Assian_Candor [comrade/them]@hexbear.net
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 month ago

    LLMs don’t train to “solve” anything. They’re just sequence predictors. They predict the next item in sequences of words, that’s it. Some predict better than others in specific scenarios.

    Through techniques like RAG and multi-model frameworks you can tailor the output to fit specific tasks or use cases. This is very powerful for automating routine workflows.