• @BaroqueInMind@lemmy.one
        link
        fedilink
        3
        edit-2
        2 months ago

        What are you talking about? It follows the Llama 3 Meta license which is pretty fucking open, and essentially every LLM that isn’t a dogshit copyright-stealing Alibaba Quen model uses it.

        Edit: Mistral has an almost similar license that Meta released Llama 3 with.

        Both Llama 3 and Mistral AI’s non-production licenses restrict commercial use and emphasize ethical responsibility, Llama 3’s license has more explicit prohibitions and control over specific applications. Mistral’s non-production license focuses more on research and testing, with fewer detailed restrictions on ethical matters. Both licenses, however, require separate agreements for commercial usage.

        Tl:Dr Mistral doesn’t give two fucks about ethics and needs money more than Meta

        • Possibly linux
          link
          fedilink
          English
          82 months ago

          Mistral is licensed under the Apache license version 2.0. This license is recognized under the GNU project and under the Open source initiative. This is because it protects your freedom.

          Meanwhile the Meta license places restrictions on use and arbitrary requirements. It is those requirements that lead me to choose not to use it. The issue with LLM licensing is still open but I certainly do not want a EULA style license with rules and restrictions.

          • @BaroqueInMind@lemmy.one
            link
            fedilink
            32 months ago

            You are correct. I checked HuggingFace just now and see they are all released under Apache license. Thank you for the correction.

    • bruhduh
      link
      fedilink
      5
      edit-2
      2 months ago

      Is hermes 8b is better than mixtral 8x7b?

      • @BaroqueInMind@lemmy.one
        link
        fedilink
        -8
        edit-2
        2 months ago

        Hermes3 is based on the latest Llama3.1, Mixtral 8x7B is based on Llama 2 released a while ago. Take a guess which one is better. Read the technical paper, it’s only 12 fucking pages.