• 7 Posts
  • 140 Comments
Joined 1 year ago
cake
Cake day: June 15th, 2023

help-circle

  • It’s a systemic issue going back decades. To me, it seems the Dutch government always wants to fix it with a hammer. Repeatedly. Discrimination increases, no REAL effort for integration is made (forcing people to take totally-not-racist “civic integration exams” is not an effort), and over the years the divide increases. Tell people they are monsters long enough, and that’s what they’ll become. But no one wants to hear that fixing it would take years or even decades of sustained effort and change. They just want it fixed. And fixed now.

    There is no one magic bullet solution, unfortunately. And then it all comes to a head with the events in Amsterdam. The instigators need to be arrested and tried, but society needs to take a close look at what caused this to happen to begin with. And I doubt that will happen. Just more hammers.






  • Well VTR is a roleplaying game. It’s similar to Vampire the Masquerade, but different setting and somewhat different mechanics. I guess it’s best explained as “nutrients.” Animal blood and blood from e.g. blood bags gives less Vitae (magic blood points resource) than blood harvested from living humans. And as the character becomes more powerful, eventually that “lesser” blood can’t actually give them Vitae.

    The vampiric curse in VTR is explicitly stated to be supernatural, though, so there’s not a necessary scientific explanation for it. The curse imparts the Beast, which is the predator in all vampires.






  • LLMs are statistical word association machines. Or tokens more accurately. So if you tell it to not make mistakes, it’ll likely weight the output towards having validation, checks, etc. It might still produce silly output saying no mistakes were made despite having bugs or logic errors. But LLMs are just a tool! So use them for what they’re good at and can actually do, not what they themselves claim they can do lol.



  • Context was set to anywhere between 8k and 16k. It was responding in English properly, and then about halfway to 3/4s of the way through a response, it would start outputting tokens in either a foreign language (Russian/Chinese in the case of Qwen 2.5) or things that don’t make sense (random code snippets, improperly formatted text). Sometimes the text was repeating as well. But I thought that might have been a template problem, because it seemed to be answering the question twice.

    Otherwise, all settings are the defaults.