Back to archive

Thread

7 tweets

1
Prompt engineering on a completely new level. I cannot keep me from seeing GPT-3 as an entity that would say anything to bullshit its way through a conversation. In a way, that‘s exactly what it has been trained to do :) twitter.com/goodside/statu…
2
In order to understand complex deep learning models, understanding what the task and data it was trained on is even more important than understanding the architecture.
4
Large language models are trained on tasks like predicting missing words from context, or deciding whether a sentence follows another one. The current models got quite good at this, but ultimately they are producing plausible sequences of sentences.
5
Put differently, they say what you want to hear (as encoded by heaps of human generated examples).
7
Btw, I cannot believe that this was how they asked. This whole thing is entirely unscientific and probably just a publicity stunt. That said, I should probably stop giving this whole thing more attention.
Media