Prompt Engineering
11 Nov 2025
So far, we’ve talked a lot about skills and applications, but not so much about fundamental concepts. Now let’s turn our attention to Xiao and Zhu (2025) to learn about fundamental concepts. This is because if we only look at skills and applications, we may overlook why certain things happen. If we examine the fundamentals and see why things happen, we can better extrapolate the examples to our future experiences.
This chapter mentions several concepts that come from different domains
Xiao and Zhu (2025) covers a lot of ground but begins with pretraining before explaining the foundations of LLMs. They focus on NLP (natural language processing) tasks because these tasks are where many breakthroughs were made.
Note that they’re mostly talking about transformer models, which were introduced in the famous paper Attention is all you need in 2017, although pre-training is much older.
We assume that it’s harder to obtain labeled data than unlabeled data due to the effort involved in labeling. For example, if we want to classify tweets as positive, negative, or neutral, we can ask humans to label some (which constitutes supervision) and use that data as input to a related task, such as rating product reviews as positive, negative, or neutral. This exemplifies supervised pre-training.
Unsupervised pre-training occurs without a human in the first phase, but humains in the training phase.
Self-supervised pre-training (without a human) allows us to go to a second phase involving a human doing prompting or training.
This slideshow was produced using quarto
Fonts are Roboto, Roboto Light, and Victor Mono Nerd Font