• sga@piefed.social
    link
    fedilink
    English
    arrow-up
    1
    ·
    3 days ago

    pretty much this. I use smolllm (a 3B param model, trained only on openly available datasets)

      • sga@piefed.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        22 hours ago

        i rarly use it, mostly to do sentiment/grammar analysis for some formal stuff/legalese. I kinda rarely use llms (1 or 2 times a month)(i just do not have a usecase). As for how good, tiny models are not good in general, but that is because they do not have enough knowledge to store info, so my use case often is purely language processing. though i have previously used it to do some work demo to generate structured data from unstructured data. basically if you provide info, they can perform well (so you can potentially build something to fetch web search results, feed into context, and use it(many such projects are available, basically something like perplexity but open)).