Je suis malade = I am sick
And “He’s calm” = “il est calme” | “Elle court” = “She’s running”
I’m pretty sure there will be some additions to this section about GPT-4 and I’m looking forward to reading that! While that happens, may I ask the admins to check whether the Figure with the description “Zero-shot, one-shot, few-shot in-context learning with language models…” is correctly depicting what’s happening in the Decoder when it says “(no parameter update).”
If I’m not misunderstanding the concepts, I think when we’re giving the pre-trained Decoder - a task, an example, and a prompt - and ask it to learn, thus fine-tune the model, we are indeed updating the Decoder parameters so that it “learns” how to translate sequences. Is that not true? What am I missing here?
I think the decoder in Fig. 11.9.3 should start with a beginning of sentence tag (e.g. <bos>). At inference time, we won’t know what the first decoded token will be (currently it shows <X>), hence why we would start decoding with <bos>.
I think the key is that, this doesn’t involving back propagating process.
My solutions to the exs: 11.9