• 3 Posts
  • 322 Comments
Joined 9 months ago
cake
Cake day: March 22nd, 2024

help-circle
  • NFT artists kind of hide behind the oldschool art world to justify it…

    I think it shows how stupid that world is.

    The point of art is to move people, to make them think and feel something by conveying the artist’s thoughts, not be a store of value. And while I’ve been to museums with gorgeous pieces, if you’re paying millions for a painting, at some point most of that value is the gratification of hoarding it. That same money could buy you an incredible experience in today’s art landscape, but it’s not about the experience, is it? And NFTs are like the perfect deconstruction of that.

    That being said, OP I am downvoting your post because eyeballs are exactly what crypto bros want, no offense :P








  • How useful would the training data be

    Open datasets are getting much better (Tulu for an instruct database/recipe is a great example), but its clear the giants still have “secret sauce” that gives them at least a small edge over open datasets.

    There actually seems to be some vindication of using massively multilingual datasets as well, as the hybrid chinese/english models are turning out very good.



  • It turns out these clusters are being used very inefficiently, seeing how Qwen 2.5 was trained with a fraction of the GPUs and is clobbering models from much larger clusters.

    One could say Facebook, OpenAI, X and such are “hoarding” H100s but are not pressured to utilize them efficiently since they are so GPU unconstrained.

    Google is an interesting case, as Gemini is getting better quickly, but they presumably use much more efficient/cheap TPUs to train.







  • For me it’s not efficiency, if anything it takes much longer running the LLM notepad because I will randomly ask it about passages/chapters and revise stuff.

    It’s kinda fun having an assistant to just bounce ideas off of on a whim. You can’t get that with beta readers, as they don’t just sit there while you write (and the LLM is much faster), and I don’t feel like I’m being charged for every response with an API model, especially if it ingests the entire story every time.

    It’s also “smart” beyond me. For instance, sometimes I wanna come up with a name for a character, city or whatever, and I can ask it “what’s an interesting name for this engineer character’s brother, from X city in the story, maybe something mythological and fire themed,” and it will crank out tons of examples and start a little conversation about it. It takes me places I never would have googled, much less known off the top of my head, because everything is on the top of an LLM’s head.



  • I suspect Lemmy is going to dislike this, but local LLMs are great writing helpers.

    When you’re stuck on a sentence or a blank paragraph, get them to continue it, and rewrite it once it jogs your mind. If you’re drafting ideas for characters or chapters, you can sanity check them or sometimes get ideas. They can reword and clean up your writing and improve it beyond what self experimentation can do… just keeping in mind that its like an idiot intern that tends to repeat itself and hallucinate.

    And this is very different from an API model like ChatGPT because:

    • It won’t refuse you.

    • It’s formatted as a notebook you can continue at any arbitrary point, rather than a user/chatbot type format.

    • The writing isn’t so dry, and it isn’t filled with AI slop, especially with cutting edge sampling

    • All its knowledge is “internal,” with no reaching out to the web or hidden prompting under your nose

    Along with all the usual reasons, namely being free, self-hosted, more ethically trained, fast and efficient with long context thanks to caching, and has nothing to do with Sam Altman’s nightmarish visions.

    I’d recommend: https://huggingface.co/nbeerbower/Qwen2.5-Gutenberg-Doppel-32B

    And once the story gets long: https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2

    I’d recommend LanguageTool (with a local server and the browser extension) for locally hosted spelling/grammar/style as well.

    I have ADD, which may be why I find this setup to be so therapeutic.