Hackworth@lemmy.worldtoTechnology@lemmy.world•Lebanon’s health minister says 8 killed, 2,750 wounded by exploding pagersEnglish
112·
1 day agoCalling what attention transformers do memorization is wildly inaccurate.
*Unless we’re talking about semantic memory.
It’s called learning, and I wish people did more of it.
Equating LLMs with compression doesn’t make sense. Model sizes are larger than their training sets. if it requires “hacking” to extract text of sufficient length to break copyright, and the platform is doing everything they can to prevent it, that just makes them like every platform. I can download © material from YouTube (or wherever) all day long.