Inside GPT – Large Language Models Demystified

E8 May 17, 2024, 3:30 PM - 4:20 PM

Alan Smith

Azure Developer, Active Solution

Natural language processing using generative pre-trained transformers (GPT) algorithms is a rapidly evolving field that offers many opportunities and challenges for application developers. But what is a generative pre-trained transformer, and how does it work? How can you leverage the latest advances in GPT algorithms to create engaging and useful applications? Can my business benefit from creating a GPT powered chat bot?

In this demo intensive session Alan will take a deep dive into the architecture of GPT algorithms and the inner workings of ChatGPT. The journey will begin by looking at the fundamental concepts of natural language processing, such as word embedding, vectorization and tokenization. He will then demonstrate how you can apply these techniques to train a GPT2 model that can generate song lyrics, showing the internals of how word sequences are predicted.

Alan will then shift the focus to larger language models, such as ChatGPT and GPT4, demonstrating their power, capabilities, and limitations. The use of hyperparameters such as temperature and frequency penalty will be explained and their effect on the generated output demonstrated. He will then cover the concepts of prompt engineering and demonstrate how Retrieval Augmented Generation (RAG) patterns can be leveraged to create a ChatGPT experience based on your own textual data.

Join me for this session if you want to learn how to harness the power of GPT algorithms in your own solutions.

AI & Machine Learning

Get inspiration & news from us

I agree that Cornerstone will send me news via e-mail