[-] pepperfree@sh.itjust.works 1 points 5 days ago

I wonder if we can extend the context length. It already fine-tuned with YaRN so we can't get free extend with that method.

[-] pepperfree@sh.itjust.works 1 points 5 days ago

Everybody been rumoring about R2. So releasing this thing kinda unexpected

40

Not what we expected...

[-] pepperfree@sh.itjust.works 1 points 2 weeks ago

There is new project which they share fine-tuned modernbert on some task. Here is the org https://huggingface.co/adaptive-classifier

187
When DeepSeek V4 and R2? (sh.itjust.works)
[-] pepperfree@sh.itjust.works 4 points 2 weeks ago* (last edited 2 weeks ago)

No, you can run sd, flux based model inside the koboldcpp. You can try it out using the original koboldcpp in google colab. It loads gguf model. Related discussion on Reddit: https://www.reddit.com/r/StableDiffusion/comments/1gsdygl/koboldcpp_now_supports_generating_images_locally/

Edit: Sorry, I kinda missed the point, maybe I'm sleepy when writing that comment. Yeah, I agree that LLM need big memory to run which is one of it's downside. I remember someone doing comparison that API with token based pricing is cheaper that to run it locally. But, running image generation locally is cheaper than API with step+megapixel pricing.

[-] pepperfree@sh.itjust.works 1 points 2 weeks ago

Skywork downfall

[-] pepperfree@sh.itjust.works 3 points 2 weeks ago

There is koboldcp-rocm fork. Koboldcpp itself has basic image generation. https://github.com/YellowRoseCx/koboldcpp-rocm

19

Built on Qwen, these models incorporate our latest advances in post-training techniques. MindLink demonstrates strong performance across various common benchmarks and is widely applicable in diverse AI scenarios.

72B 32B

pepperfree

joined 2 years ago