GitHub topics: llms-evalution
ronniross/llm-heatmap-visualizer
A set of scripts to generate full attention-head heatmaps for transformer-based LLMs
Language: Jupyter Notebook - Size: 0 Bytes - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 1 - Forks: 0

Fbxfax/llm-confidence-scorer
A set of auxiliary systems designed to provide a measure of estimated confidence for the outputs generated by Large Language Models.
Language: Python - Size: 96.7 KB - Last synced at: 1 day ago - Pushed at: 1 day ago - Stars: 0 - Forks: 0

ronniross/llm-confidence-scorer
A set of auxiliary systems designed to provide a measure of estimated confidence for the outputs generated by Large Language Models.
Language: Python - Size: 0 Bytes - Last synced at: 4 days ago - Pushed at: 4 days ago - Stars: 0 - Forks: 0

ALEX-nlp/MUI-Eval
Repository for the paper: Revisiting LLM Evaluation through Mechanism Interpretability: a New Metric and Model Utility Law
Language: Python - Size: 7.36 MB - Last synced at: 6 days ago - Pushed at: 21 days ago - Stars: 7 - Forks: 0
