GLUE
The General Language Understanding Evaluation (GLUE) benchmark is a collection of resources for training, evaluating, and analyzing natural language understanding systems.
Updated 10d ago
PricingFree
Free
Evaluating natural language understanding models
Training NLP models on diverse datasets
Comparing model performance across different tasks