Select two models to compare their capabilities head-to-head.
SUPA's LLM Comparison Tool provides the insights you need to identify the best model for your specific use case. All comparison data will be open sourced to support LLM research and benchmarking.
We're excited to introduce a new open-source tool that empowers users to evaluate and compare LLM performance firsthand. While LLMs have shown immense potential across various applications, the ability to meaningfully assess their capabilities—especially in specific domains and use cases—has remained a challenge.
Our tool aims to bridge this gap by providing a transparent, hands-on comparison environment where users can pit different models against each other and draw their own conclusions.
What sets our tool apart is its inclusive approach to model evaluation. We've noticed a gap in representation, particularly for Southeast Asian LLMs, in existing comparison platforms. Our tool will include models from the SEA region alongside global options, ensuring a more comprehensive evaluation landscape. We believe this diversity is crucial for users working across different linguistic and cultural contexts.
In line with our commitment to advancing the field, all data generated through our platform will be made publicly available for research purposes and model fine-tuning.
Made with love by the SUPA Team