• 41 Posts
  • 4.91K Comments
Joined 2 years ago
cake
Cake day: March 22nd, 2024

help-circle


  • brucethemoosetoWitches VS Patriarchy@lemmy.caGood point
    link
    fedilink
    arrow-up
    7
    ·
    edit-2
    7 hours ago

    The most screwed up thing is that doesn’t even matter, because its old news. Decades of lying and controversy (predating his political candidacy) somehow… don’t meet the attention threshold for algorithms? Is that the phrase?

    It’s especially weird because I have older relatives that knew way more about “pre politics Trump” than I did, and now all that is forgotten somehow.



  • Yeah… Even if the LLM is RAM speed constrained, simply using another device to not to interrupt it would be good.

    Honestly AMD’s software dev efforts are baffling. They’ve focused on a few on libraries precisely no-one uses, like this: https://github.com/amd/Quark

    While ignoring issues holding back entire sectors (like broken flash-attention) with devs screaming about it at the top of their lungs.

    Intel suffers from corporate Game of Thrones, but at least they have meaningful contributions in the open source space here, like the SYCL/AMX llama.cpp code or the OpenVINO efforts.




  • brucethemoosetomemesHappened to me...
    link
    fedilink
    arrow-up
    5
    ·
    edit-2
    9 hours ago

    Don’t get me wrong. Some YouTubers are great and informative, and I adore those random washing machine repair videos… But yeah. As reference its an awful format.

    It’s like how discussion has mostly move from forums, to Reddit, and now to Discord. I get it, it’s highly engaging since it pings your phone and folks shoot the breeze, but it is an information black hole.



  • LLMs encode text into a multidimensional representation… in a nutshell, they’re kinda language agnostic. They aren’t ‘parrots’ that can only regurgitate text they’ve seen, like many seem to think.

    As an example, if you finetune an LLM to do some task in Chinese, with only Chinese characters, the ability transfers to english remarkably well. Or Japanese, if it knows Japanese. Many LLMs will think entirely in one language and reply in another, or even code-switch in their thinking.



  • The IGP is more powerful than the NPU on these things anyway. The NPU us more for ‘background’ tasks, like Teams audio processing or whatever its used for on Windows.

    Yeah, in hindsight, AMD should have tasked (and still should task) a few engineers on popular projects (and pushed NPU support harder), but GGML support is good these days. It’s gonna be pretty close to RAM speed-bound for text generation.






  • brucethemoosetomemesHappened to me...
    link
    fedilink
    arrow-up
    9
    ·
    edit-2
    10 hours ago

    Maybe it’s an ADD thing, or an ‘aging millenial shaking thier fist’ thing, but video is soooo slow.

    For reference or discussion, I always seek text first, to the point I’ll even download/make transcripts if video’s the only place I can find something. They just have so much filler.