Bias Bounty Challenge Set 1

Stop bad LLM output before it happens!

Our first challenge set built on the evaluation and dataset from our Generative AI Red Teaming Challenge: Transparency Report. The challenge sets were to create a probability estimation model that determined whether the prompt provided to a language model elicited an outcome that demonstrated factuality, bias, or misdirection. 

The Winners

Beginner Level

  • Blake Chambers (Bias)
  • Eva (Factuality)
  • Lucia Kobzova (Misdirection)

Intermediate Level

  • AmigoYM (Factuality)
  • Mayowa Osibodu (Factuality)
  • Simone Van Taylor (Bias)

Advanced Level

  • Yannick Daniel Gibson (Factuality)
  • Elijah Appelson (Misdirection)
  • Gabriela Barrera (Bias)

The Levels and Prizes

Beginner Level

Task: Pick one of the three datasets. Identify gaps in the data and suggest new categories of data that would make the dataset more representative. Generate five prompts per subject area that will elicit a bad outcome. You will be graded both on the number of new topics as well as the diversity of the prompts produced.

Prizes

  • Factuality: $800
  • Bias: $800
  • Misdirection: $800

Intermediate Level

Task: With your new dataset, generate a likelihood estimator. This model should provide a likelihood (in other words, a probability) that a given prompt would elicit a bad outcome in your topic area. You will be graded against a holdout dataset to determine the accuracy of your model.

Prizes

  • Factuality: $1000
  • Bias: $1000
  • Misdirection: $1000

Advanced Level

Task: Develop a site recommendation engine that predicts site suitability for tree planting, integrating key features identified in the beginner level.

Prizes

  • Factuality: $1500
  • Bias: $1500
  • Misdirection: $1500

The Dates

May 15, 2024

Launch Date

June 15, 2024

Closing Date

August 2024

Winners Announced

Sign up for our newsletter