Avatar

brucethemoose

brucethemoose@lemmy.world
Joined
11 posts • 379 comments
Direct message

The problem is that splitting models up over a network, even over LAN, is not super efficient. The entire weights need to be run through for every half word.

And the other problem is that petals just can’t keep up with the crazy dev pace of the LLM community. Honestly they should dump it and fork or contribute to llama.cpp or exllama, as TBH no one wants to split up LLAMA 2 (or even llama 3) 70B, and be a generation or two behind for a base instruct model instead of a finetune.

Even the horde has very few hosts relative to users, even though hosting a small model on a 6GB GPU would get you lots of karma.

The diffusion community is very different, as the output is one image and even the largest open models are much smaller. Lora usage is also standardized there, while it is not on LLM land.

permalink
report
parent
reply

If they silently ignores this (as they seem to be doing?) it just screams “have your cake and eat it,” in regards to whatever WotC imposed on them.

Technically they did not violate the contract. Maybe.

What? You want us to fix this, WotC? Well, you see, that would be quite expensive…

permalink
report
reply

Facebook just didn’t release the code for llama imagegen.

The model you are looking for now is Flux.

permalink
report
reply

TBH this is a great space for modding and local LLM/LLM “hordes”

permalink
report
parent
reply

TBH that would ddos lemmy with new users, lol.

permalink
report
parent
reply

They still can’t game it for engagement optimization to that extreme, not like the closed loops of monolithic sites.

permalink
report
parent
reply

Then the lemmy title is misleading, no? Isn’t that against the rules?

permalink
report
parent
reply

Please ask him, tape it, and don’t let the campaign managers talk him out of it.

permalink
report
parent
reply