Updated Fall 2024
ChatGPT and other large language models (LLMs) are relatively new and rapidly evolving tools for scientists and engineers. They can be powerful for summarizing documentation in large technical projects or computer languages. For instance, you can use it to explain functions and methods in Python. We encourage you to use ChatGPT as you code. Perhaps start by prompting it to summarize the libraries we will use: numpy, scipy, and matplotlib, or the basics of Markup in Jupyter notebooks.
Some things to keep in mind as you experiment with ChatGPT or other LLMs:
LLMs are basically statistical models that derive from very large sets of text. Well, don't take my word for it, see what ChatG has to say:
Prompt
ChatG, please summarize large language models for my thermodynamics students in two hundred words or less. I want to convey how LLMs are basically statistical models trained on large sets of text documents.
ChatGPT
Large language models (LLMs) are advanced statistical models trained on vast amounts of text data. They function by predicting the probability of a sequence of words based on patterns learned from the training data. These models use deep learning techniques, specifically neural networks with many layers, to capture complex relationships between words and phrases.
When an LLM processes a text input, it breaks it down into tokens (words or subwords) and predicts the most likely next tokens based on its training. This prediction is repeated iteratively to generate coherent and contextually relevant text. The "language" in these models refers to their ability to understand and generate human language, but fundamentally, they operate on statistical principles, relying on the frequencies and associations of words observed in the training data.
LLMs are powerful because they can generalize from the data they were trained on to produce meaningful responses to new, unseen text inputs. However, their accuracy and reliability are limited by the data they were trained on, which means they can sometimes generate incorrect or biased information.
Engineers don't take technical solutions for granted! We generally like to "look under the hood" and see how things work. So, if you are interested in learning more about the technical underpinnings of LLMs and their uses, here are a few resources: