This is a brief summary of paper for me to study and organize it, GLUE- A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding (Wang et al., arXiv 2019) I read and studied.

This paper show a platform for Natural Language understaing tasks as follows:

GLUE Benchmark

If you want to submission the platform to test your model across a variety of tasks for NLU.

Visit GLUE benchmark site.

They provide each 9 bechmark sets about NLU tasks.

To recap, the set is composed of the following:

Wang et al. 2019

Reference