2020
DOI: 10.1609/aaai.v34i05.6519
|View full text |Cite
|
Sign up to set email alerts
|

JEC-QA: A Legal-Domain Question Answering Dataset

Abstract: We present JEC-QA, the largest question answering dataset in the legal domain, collected from the National Judicial Examination of China. The examination is a comprehensive evaluation of professional skills for legal practitioners. College students are required to pass the examination to be certified as a lawyer or a judge. The dataset is challenging for existing question answering methods, because both retrieving relevant materials and answering questions require the ability of logic reasoning. Due to the hig… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
38
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 77 publications
(38 citation statements)
references
References 28 publications
0
38
0
Order By: Relevance
“…Besides, COLIEE (Kano et al, 2018) contains about 500 yes/no questions. Moreover, the bar exam is a professional qualification examination for lawyers, so bar exam datasets (Fawei et al, 2016;Zhong et al, 2019a) may be quite hard as they require professional legal knowledge and skills.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Besides, COLIEE (Kano et al, 2018) contains about 500 yes/no questions. Moreover, the bar exam is a professional qualification examination for lawyers, so bar exam datasets (Fawei et al, 2016;Zhong et al, 2019a) may be quite hard as they require professional legal knowledge and skills.…”
Section: Related Workmentioning
confidence: 99%
“…We select JEC-QA (Zhong et al, 2019a) as the dataset of the experiments, as it is the largest dataset collected from the bar exam, which guarantees its difficulty. JEC-QA contains 28, 641 multiple-choice and multiple-answer questions, together with 79, 433 relevant articles to help to answer the questions.…”
Section: Experiments and Analysismentioning
confidence: 99%
See 2 more Smart Citations
“…Though several pre-trained language models have been introduced for domain-specific MRC, BERT based models are not as consistently dominant as they are in open field MRC tasks (Zhong et al, 2020;Yue et al, 2020). Another challenge is that medical questions are often more difficult; no labeled paragraph contains the answer to a given question.…”
Section: Introductionmentioning
confidence: 99%