• 0 Posts
  • 1 Comment
Joined 1 year ago
cake
Cake day: November 18th, 2023

help-circle
  • My understanding is basically, they are data sets the model is compared to. Say you wanted to see how well you knew math. You took a math test, and then your answers were compared to a key of answers…

    Some of my notes about those benchmarks

    GSM8K is a dataset of 8.5K high-quality linguistically diverse grade school math word problems created by human problem writers

    HellaSwag is the large language model benchmark for commonsense reasoning.

    Truful QA: is a benchmark to measure whether a language model is truthful in generating answers to questions.

    Winogrande - Common sense reasoning