AzamServer
  • Communities
  • Create Post
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
@Ascend910@lemmy.ml to Memes@lemmy.mlEnglish • 7 months ago

2 in a single week that is crazy

lemmy.ml

message-square
21
fedilink
557

2 in a single week that is crazy

lemmy.ml

@Ascend910@lemmy.ml to Memes@lemmy.mlEnglish • 7 months ago
message-square
21
fedilink

blob:https://phtn.app/bce94c48-9b96-4b8e-a4fd-e90166d56ed7

  • @Cort@lemmy.world
    link
    fedilink
    1•7 months ago

    Would a 12g 3060 work?

    • @brucethemoose@lemmy.world
      link
      fedilink
      1•
      edit-2
      7 months ago

      Yes! Try this model: https://huggingface.co/arcee-ai/Virtuoso-Small-v2

      Or the 14B thinking model: https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B

      But for speed and coherence, instead of ollama, I’d recommend running it through Aphrodite or TabbyAPI as a backend, depending if you prioritize speed or long inputs. They both act as generic OpenAI endpoints.

      I’ll even step you through it and upload a quantization for your card, if you want, as it looks like there’s not a good-sized exl2 on huggingface.

Memes@lemmy.ml

!memes@lemmy.ml

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !memes@lemmy.ml

Rules:

  1. Be civil and nice.
  2. Try not to excessively repost, as a rule of thumb, wait at least 2 months to do it if you have to.
  • 891 users / day
  • 3.36K users / week
  • 7.51K users / month
  • 22K users / 6 months
  • 52.2K subscribers
  • 12.3K Posts
  • 155K Comments
  • Modlog
  • mods:
  • ghost_laptop
  • @sexy_peach@feddit.de
  • Cyclohexane
  • Arthur Besse
  • BE: 0.19.3
  • Modlog
  • Instances
  • Docs
  • Code
  • join-lemmy.org