From the course: Introduction to Large Language Models
Unlock the full course today
Join today to access over 24,700 courses taught by industry experts.
Comparing LLMs
From the course: Introduction to Large Language Models
Comparing LLMs
- [Instructor] How do we even compare large language models? That's a great question. I don't think we have a perfect answer yet, but we have made great progress over the last few months. Usually, we only focus on how good a model is at a task, but we don't know if that same model generates false information. So instead of just looking at one metric, a Stanford University research team proposed HELM, or the Holistic Evaluation of Language Models. With HELM, the Stanford research team worked together with the main large language model providers, and they were able to benchmark the models across a variety of data sets and get a more holistic view of model performance. The HELM benchmark is a living benchmark and should change as new models are released. I'll just cover the first couple of benchmarks and you can explore the rest further if you're interested. So let me just go ahead and scroll down a little bit. So here, each row…
Practice while you learn with exercise files
Download the files the instructor uses to teach the course. Follow along and learn by watching, listening and practicing.
Contents
-
-
-
-
-
(Locked)
BERT3m 16s
-
(Locked)
Scaling laws3m 30s
-
(Locked)
GPT-37m 41s
-
(Locked)
Chinchilla7m 54s
-
(Locked)
PaLM and PaLM 23m 59s
-
ChatGPT and GPT-45m 47s
-
(Locked)
Open LLMs5m 40s
-
(Locked)
Comparing LLMs3m 35s
-
(Locked)
GitHub Models: Comparing LLMs2m 52s
-
(Locked)
Accessing large language models using an API6m 25s
-
(Locked)
LLM trends4m 6s
-
(Locked)
-