LLM Evaluation Dashboard
7.3
A centralized dashboard for evaluating Large Language Models (LLMs). Automates the tracking and comparison of various metrics and benchmarks, helping developers improve model performance and select the best models for their applications.
150h
mvp estimate
7.3
viability grade
14
views
technology stack
Python
PostgreSQL
NodeJS
Medium