Do you want BuboFlash to help you learning these things? Or do you want to add or correct something? Click here to log in or create user.



Large language models are built by using supervised learning to train a model to repeatedly predict the next word. For example, if an AI system has read on the Internet a sentence like my favorite drink is lychee bubble tea, then the single sentence would be turned into a lot of A to B data points for the model to learn to predict the next word.
If you want to change selection, open document below and click on "Move attachment"

Unknown title
ate new texts by repeatedly predicting what is the next word they should output. Given the widespread attention on LLMs, let's look briefly on the next slide in greater detail at how they work. <span>Large language models are built by using supervised learning to train a model to repeatedly predict the next word. For example, if an AI system has read on the Internet a sentence like my favorite drink is lychee bubble tea, then the single sentence would be turned into a lot of A to B data points for the model to learn to predict the next word. Specifically, given this sentence, we now have one data point that says, given the phrase my favorite drink, what do you predict is the next word? In this case, the right answer is give


Summary

statusnot read reprioritisations
last reprioritisation on suggested re-reading day
started reading on finished reading on

Details



Discussion

Do you want to join discussion? Click here to log in or create user.