Notes
I am fascinated by prompt hacking that uses emotional appeals to affect the behavior of LLMs. A long-lived LLM might be used to prototype manipulation techniques, perhaps by other models in an adversarial training arrangement.
In any event, many humans may learn social interactions by interacting with them as much as or instead of with peers. I’m not sure if that’s troubling or promising!
It is both obvious and deeply unsettling that AI will serve as a channel for our biases. What really concerns me is internet-scale models may homogenize and amplify our cultural flaws to a degree that will make Facebook seem like a tiny harbinger.
If the EU starts regulating attention-manipulating design we are going to end up with a fascinating arms race between dark pattern innovation and regulatory agility.
Simon Willison’s state-of-AI talk is getting attention for many good points, but my favorite is his observation that we really need to consider new UIs for LLM interaction. Because:
chat is kind of a terrible interface for these tools.
Maggie Appleton has already explained why and I’m looking forward to new methods emerging.
Why is it so easily accepted that pricing is limited by “what the market will bear” but scale is not?
Is anyone using AI to make graphic games out of the old Infocom text adventures?