Code for NAACL paper When Quantization Affects Confidence of Large Language Models?
-
Updated
Dec 30, 2024 - Jupyter Notebook
Code for NAACL paper When Quantization Affects Confidence of Large Language Models?
Enterprise-grade LLM evaluation framework | Multi-model benchmarking, honest dashboards, system profiling | Academic metrics: MMLU, TruthfulQA, HellaSwag | Zero fake data | PyPI: llm-benchmark-toolkit | Blog: https://dev.to/nahuelgiudizi/building-an-honest-llm-evaluation-framework-from-fake-metrics-to-real-benchmarks-2b90
Add a description, image, and links to the truthfulqa topic page so that developers can more easily learn about it.
To associate your repository with the truthfulqa topic, visit your repo's landing page and select "manage topics."