GLUE
The General Language Understanding Evaluation (GLUE) benchmark is a collection of resources for training, evaluating, and analyzing natural language understanding systems.
5d ago
Best for NLP Evaluation
PricingFree
Free
Evaluating natural language understanding models
Training NLP models on diverse datasets
Comparing model performance across different tasks