Home | Power BI | Excel | Python | SQL | Generative AI | Visualising Data | Analysing Data
This is in three stages
Download a lot of text (a corpus), preferably the entire internet.
Go through the next steps repeatedly for over the whole corpus in a process called self-supervised learning. This will cost $100m, takes several months and generate several hundred tons of CO2.
At this point the model is pre-trained and will just continue / complete text. Now we fine-tune the model to make it useful. Humans provide questions and finetune the model so that it responds with something close to a model answer. Another technique to fine tune is for the LLM to generate two different answers. People then indicate which response they prefer and this is fed back into the fine training.
As the model grows in size, it shows surprising “emergent behaviours”: