This is again a big win on the red team at least for me. They developed a “fully open” 3B parameters model family trained from scratch on AMD Instinct™ MI300X GPUs.

AMD is excited to announce Instella, a family of fully open state-of-the-art 3-billion-parameter language models (LMs) […]. Instella models outperform existing fully open models of similar sizes and achieve competitive performance compared to state-of-the-art open-weight models such as Llama-3.2-3B, Gemma-2-2B, and Qwen-2.5-3B […].

As shown in this image (https://rocm.blogs.amd.com/_images/scaling_perf_instruct.png) this model outperforms current other “fully open” models, coming next to open weight only models.

A step further, thank you AMD.

PS : not doing AMD propaganda but thanks them to help and contribute to the Open Source World.

    • Ulrich@feddit.org
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      9
      ·
      3 hours ago

      What’s malicious about AI and LLMs? Have you been living under a rock?

      At best it is useless, and at worst it is detrimental to society.

      • Domi@lemmy.secnd.me
        link
        fedilink
        English
        arrow-up
        10
        ·
        3 hours ago

        I disagree, LLMs have been very helpful for me and I do not see how an open source AI model trained with open source datasets is detrimental to society.

        • Ulrich@feddit.org
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          3
          ·
          2 hours ago

          I don’t know what to say other than pull your head outta the sand.

          • sugar_in_your_tea@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            6
            ·
            1 hour ago

            No you.

            Explain your exact reasons for thinking it’s malicious. There’s a lot of FUD surrounding “AI,” a lot of which come from unrealistic marketing BS and poor choices by C-suite types that have nothing to do with the technology itself. If you can describe your concerns, maybe I or others can help clarify things.

            • frezik@midwest.social
              link
              fedilink
              English
              arrow-up
              3
              ·
              1 hour ago

              These models are trained on human creations with the express intent to drive out those same human creators. There is no social safety net available so those creators can maintain a reasonable living standard without selling their art. It won’t even work–the models aren’t good enough to replace these jobs, but they’re good enough to fool the C-suite into thinking they can–but they’ll do lots of damage in the attempt.

              The issues are primarily social, not technical. In a society that judges itself on how well it takes care of the needs of everyone, I would have far less of an issue with it.

      • ZeroOne@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 hours ago

        So in a nutshell, it’s malicious because you said so

        Ok gotcha Mr/Ms/Mrs TechnoBigot