• 0 Posts
  • 10 Comments
Joined 2 years ago
cake
Cake day: June 20th, 2023

help-circle


  • Anecdotally, I use it a lot and I feel like my responses are better when I’m polite. I have a couple of theories as to why.

    1. More tokens in the context window of your question, and a clear separator between ideas in a conversation make it easier for the inference tokenizer to recognize disparate ideas.

    2. Higher quality datasets contain american boomer/millennial notions of “politeness” and when responses are structured in kind, they’re more likely to contain tokens from those higher quality datasets.

    I haven’t mathematically proven any of this within the llama.cpp tokenizer, but I strongly suspect that I could at least prove a correlation between polite token input and dataset representation output tokens







  • Your dad probably got lucky, and your router’s firewall probably did a lot of the heavy lifting. If you were to connect a win 2000/XP computer to the internet today without a firewall between, it would be compromised in minutes (there are loads of videos of people demoing this).

    While I don’t have proof that 7 would be the same, I strongly suspect it would be the same. 10 will get there soon too. Firewalls will stop most of the low hanging fruit, but an application that bridges connections through the firewall are that much more vulnerable to exploitations that won’t be integrated by your running kernel.