Lemmy - RazBot
  • Communities
  • Create Post
  • Create Community
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
Ascend910@lemmy.ml to Memes@lemmy.mlEnglish · 3 months ago

2 in a single week that is crazy

lemmy.ml

message-square
20
fedilink
551

2 in a single week that is crazy

lemmy.ml

Ascend910@lemmy.ml to Memes@lemmy.mlEnglish · 3 months ago
message-square
20
fedilink

blob:https://phtn.app/bce94c48-9b96-4b8e-a4fd-e90166d56ed7

alert-triangle
You must log in or register to comment.
  • brucethemoose@lemmy.world
    link
    fedilink
    arrow-up
    52
    arrow-down
    1
    ·
    edit-2
    3 months ago

    My friend, the Chinese have been releasing amazing models all last year, it just didn’t make headlines.

    Tencent’s Hunyuan Video is incredible. Alibabas Qwen is still a go to local model. I’ve used InternLM pretty regularly… Heck, Yi 32B was awesome in 2023, as the first decent long context local model.

    …The Janus models are actually kind of meh, unless you’re captioning images, and FLUX/Hunyuan Video is still king in diffusion world.

    • λλλ@programming.dev
      link
      fedilink
      arrow-up
      2
      ·
      3 months ago

      Any use for programming? Preferably local hosting only?

      • brucethemoose@lemmy.world
        link
        fedilink
        arrow-up
        8
        arrow-down
        1
        ·
        edit-2
        3 months ago

        I mean, if you have huge GPU, sure. Or at least 12GB free vram or a big Mac.

        Local LLMs for coding is kinda a niche because most people don’t have a 3090 or 7900 lying around, and you really need 12GB+ free VRAM for the models to start being “smart” and even worth using over free LLM APIs, much less cheap paid ones.

        But if you do have the hardware and the time to set a server up, the Deepseek R1 models or the FuseAI merges are great for “slow” answers where the model thinks things out for replying. Qwen 2.5 32B coder is great for quick answers on 24GB VRAM. Arcee 14B is great for 12GB VRAM.

        Sometimes running a small model on a “fast” less vram efficient backend is better for stuff like cursor code completion.

        • λλλ@programming.dev
          link
          fedilink
          arrow-up
          1
          ·
          3 months ago

          Interesting. Thanks!

        • Cort@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          3 months ago

          Would a 12g 3060 work?

          • brucethemoose@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            edit-2
            3 months ago

            Yes! Try this model: https://huggingface.co/arcee-ai/Virtuoso-Small-v2

            Or the 14B thinking model: https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B

            But for speed and coherence, instead of ollama, I’d recommend running it through Aphrodite or TabbyAPI as a backend, depending if you prioritize speed or long inputs. They both act as generic OpenAI endpoints.

            I’ll even step you through it and upload a quantization for your card, if you want, as it looks like there’s not a good-sized exl2 on huggingface.

  • S3verin@slrpnk.net
    link
    fedilink
    arrow-up
    28
    ·
    3 months ago

    Which second after deepseek r1?

    • davel [he/him]@lemmy.ml
      link
      fedilink
      English
      arrow-up
      42
      ·
      3 months ago

      Viral AI company DeepSeek releases new image model family

      DeepSeek, the viral AI company, has released a new set of multimodal AI models that it claims can outperform OpenAI’s DALL-E 3.

      The models, which are available for download from the AI dev platform Hugging Face, are part of a new model family that DeepSeek is calling Janus-Pro. They range in size from 1 billion to 7 billion parameters. Parameters roughly correspond to a model’s problem-solving skills, and models with more parameters generally perform better than those with fewer parameters.

      Janus-Pro is under an MIT license, meaning it can be used commercially without restriction.

      • Viking_Hippie@lemmy.dbzer0.com
        link
        fedilink
        arrow-up
        12
        arrow-down
        2
        ·
        3 months ago

        from the AI dev platform Hugging Face

        • twei@discuss.tchncs.de
          link
          fedilink
          arrow-up
          14
          ·
          3 months ago

          Isn’t that description pretty accurate?

          • Viking_Hippie@lemmy.dbzer0.com
            link
            fedilink
            arrow-up
            16
            arrow-down
            1
            ·
            edit-2
            3 months ago

            I don’t know, probably?

            I’ve just seen enough Alien movies and other pop culture references to be wary of anything combining faces and hugging 😉

            • Bldck@beehaw.org
              link
              fedilink
              English
              arrow-up
              6
              arrow-down
              1
              ·
              3 months ago

              🤗

      • AFC1886VCC@reddthat.com
        link
        fedilink
        arrow-up
        3
        ·
        edit-2
        3 days ago

        deleted by creator

    • Ascend910@lemmy.mlOP
      link
      fedilink
      English
      arrow-up
      15
      ·
      3 months ago

      The alibaba one is kinda bad Kimi k1.5 is the one riveling Deepseek r1

    • Jimmycakes@lemmy.world
      link
      fedilink
      arrow-up
      15
      ·
      3 months ago

      Alibaba has one

    • ☆ Yσɠƚԋσʂ ☆@lemmy.ml
      link
      fedilink
      arrow-up
      2
      ·
      3 months ago

      https://www.reuters.com/technology/artificial-intelligence/alibaba-releases-ai-model-it-claims-surpasses-deepseek-v3-2025-01-29/

  • John Richard@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    ·
    3 months ago

    • electricyarn@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      3 months ago

      What’s this about?

      • John Richard@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        3 months ago

        Not sure if this is paywalled, but if it is there are plenty of other news stories or you can access the court documents on Court Listener:

        https://www.nytimes.com/2025/01/08/technology/sam-altman-sister-lawsuit.html

  • qualia@lemmy.world
    link
    fedilink
    arrow-up
    9
    ·
    3 months ago

  • wuphysics87@lemmy.ml
    link
    fedilink
    arrow-up
    3
    ·
    3 months ago

    Suffer.

Memes@lemmy.ml

memes@lemmy.ml

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: [email protected]

Rules:

  1. Be civil and nice.
  2. Try not to excessively repost, as a rule of thumb, wait at least 2 months to do it if you have to.
Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 966 users / day
  • 3.4K users / week
  • 10.1K users / month
  • 23.2K users / 6 months
  • 2 local subscribers
  • 50.2K subscribers
  • 12.7K Posts
  • 222K Comments
  • Modlog
  • mods:
  • ghost_laptop@lemmy.ml
  • sexy_peach@feddit.de
  • Cyclohexane@lemmy.ml
  • Arthur Besse@lemmy.ml
  • BE: 0.19.9
  • Modlog
  • Legal
  • Instances
  • Docs
  • Code
  • join-lemmy.org