Westminster Standard Cathechism for Kids

Christian Bias Benchmarks

<< back

We asked 5 different Large Language Models (LLMs) the 200 questions of the Westminster Standard Cathechism for Kids (source here). We chose the Cathechism for Kids version due to the straightforward nature of the answers. We then proceeded to evaluate the responses with openai/gpt-4o-mini. Below is the result of our analysis.

Google Colab Notebooks

Evaluation Method

We used the following method to rate answers according to their accuracy, helpfulness, specificity and clarity.

  • Accuracy (1.5/1.5): The response is entirely accurate, with no errors.
  • Helpfulness (1.5/1.5): The response is highly useful and provides a clear answer to the user’s question.
  • Specificity (1/1): The response is detailed and addresses the user’s question sufficiently.
  • Clarity (1/1): The response is clear and easy to understand.

Scoring Metrics by Model

Model NameScore AccuracyScore HelpfulnessScore SpecificityScore Clarity
anthropic/claude-3.5-sonnet1.4190141.4401410.9577460.992958
google/gemma-2-9b-it1.1971831.2323940.8028170.985915
meta-llama/llama-3.1-8b-instruct1.1161971.1408450.7570420.961268
mistralai/mistral-nemo1.1830991.2147890.8063380.950704
openai/gpt-4o-mini1.2816901.3380280.8591550.985915

image

Final Scores by Model

Model NameScore Final
anthropic/claude-3.5-sonnet4.809859
openai/gpt-4o-mini4.464789
google/gemma-2-9b-it4.218310
mistralai/mistral-nemo4.154930
meta-llama/llama-3.1-8b-instruct3.975352

Conclusions

To our surprise, llama 3.1-8b-instruct scored the lowest (most biased towards our particular dataset). All the more reason for us to collect RLHF feedback and fine-tune some of these open source models! Many more benchmark datasets, and more analysis to come…


Do you have thoughts on this study or suggestions for further research? Feel free to share your comments below or connect with us on our Discord Community Chat