SlideShare a Scribd company logo
Generation of Assessment Questions from
Textbooks Enriched with Knowledge Models
Lucas Dresscher, Isaac Alpizar Chacon, Sergey Sosnovsky
1a. Introduction
● Assessment question in digital textbooks:
○ opportunity to practice
○ opportunity to receive feedback
○ opportunity to engage in interactive learning
1a. Introduction
● Assessment question in digital textbooks:
○ opportunity to practice
○ opportunity to receive feedback
○ opportunity to engage in interactive learning
● Assessment questions in adaptive textbooks:
○ opportunity to assess student knowledge
➢ opportunity to adapt
1a. Introduction
● Assessment question in digital textbooks:
○ opportunity to practice
○ opportunity to receive feedback
○ opportunity to engage in interactive learning
● Assessment questions in adaptive textbooks:
○ opportunity to assess student knowledge
➢ opportunity to adapt
● Three primary ways to add assessment questions:
○ create them manually
○ integrate external assessment material
○ generate them automatically
1b. Introduction
● AQG: current state-of-the-art
○ High-quality factoid questions
○ Limitations:
■ Simplicity
■ Few generic systems
■ Little adaptivity
● Contributions
○ Unique source type
○ Open domain
○ Variety of questions
○ Supports adaptivity
2. Intextbooks Platform
● Knowledge models
○ Extraction: create semantic model
○ Enrichment: link additional information
○ Serialization: export model
2. Intextbooks Platform
● Knowledge models
○ Extraction: create semantic model
○ Enrichment: link additional information
○ Serialization: export model
3. Automatic Question Generation System
● Generic semantic rule-based AQG system
○ Domain independent
○ Uses textual and semantic information
○ Can generate questions for a part of a book or a (set of) concept(s)
3A. Source Extraction
● Input
○ Knowledge model of a textbook
● Extraction
○ Relevant sentences
○ Index terms’ enrichments
■ Domain specificity
● Output
○ Initial set of sentences
3B. Preprocessing
● Annotation pipeline Stanford CoreNLP
○ Parts of Speech (POS)
○ Dependency parsing
○
● Filtering
○ Incorrect form (questions, imperative, grammatically incomplete,..)
○ Context reference (needs preceding sentences to make sense)
○ Visual references (figures, tables,...)
○ Numerical examples,
○ etc.
● Output
○ set of grammatically-fit annotated sentences
3C. Sentence Selection
● Select most appropriate sentences
● Scoring process
○ Weighted average of individual features
■ Different aspect
■ Feature score f [0, 1]
■ Weight w
○ Sentence score s [0, 1]
○ Threshold comparison
● Output
○ Potential source phrases
● Determine generatable question types
● Question types
○ True-false (unmodified)
○ True-false (negated)
○ True-false (substituted)
○ Cloze
○ Multiple-choice
● Output
○ Definitive source sentences
The mean is the average.
The mean is not the average.
The median is the average.
The ______ is the average.
The ______ is the average.
3D. Question Type Selection
A. Median
B. Mean
C. Mode
D. Variance
3E. Question Construction
● Create questions in surface form
● TFU
○ Directly use sentence
○ Answer: true
● TFN
○ Negate question stem
○ Answer: false
● TFS
○ Substitute target concept
○ Answer: false
The mean is the average.
The mean is not the average.
The mean is the average.
The mean is the average.
The median is the average.
3E. Question Construction
● CQ
○ Replace target concept by gap
○ Gap selection
● MCQ
○ CQ with options
■ Key
■ Distractors
○ Distractor generation
■ Related elements
■ Scoring procedure
The mean is the average.
The _____ is the average.
The mean is the average.
The _____ is the average.
A. Median
B. Mean
C. Mode
D. Variance
4. Evaluation
● Selection procedure
○ Three university-level statistics textbooks
○ Ten randomly selected co-occurring concepts
■ Five for automatic generation
■ Five for manual creation
○ 50 questions in total
4. Evaluation
● Selection procedure
○ Three university-level statistics textbooks
○ Ten randomly selected co-occurring concepts
■ Five for automatic generation
■ Five for manual creation
○ 50 questions in total
● Evaluation approach
○ Expert evaluation
○ Metrics
■ General: wording, assessment value, difficulty
■ Specific: gap quality, distractor quality
4. Evaluation
● Research questions:
1. Is the approach conceptually sound?
2. Is the approach practically sound?
4. Evaluation
● Inter-rater agreement (Fleiss’ Kappa)
○ 0.24 wording
○ 0.27 assessment value
○ -0.02 difficulty
4. Evaluation
● Inter-rater agreement (Fleiss’ Kappa)
○ 0.24 wording
○ 0.27 assessment value
○ -0.02 difficulty
● Comparison between handcrafted and generated questions
(Mann-Whitney U test)
○ Statistically significant difference for the overall assessment value (Handcrafted > Generated)
■ 0.32 (U = 413.5, P = 0.048)
■ No statistically significant difference per question type
○ No significant differences for the overall wording
4. Evaluation
● Assessment value
needs improvements
● Largest difference
TFUs and MCQs
● TFSs particularly poor
4. Evaluation
● Good overall wording
● Small differences
● TFSs, CQs and MCQs
particularly good
4. Evaluation
● Easy to medium
● Small differences
5. Conclusion
● Limitations
○ Nature of education textbooks
○ Weights of feature sets
● Future work
○ Other domains
○ Additional features
○ Closer Intextooks integration
Time to Generate Some Questions
4. Evaluation
● Question type specific metrics
● Gap quality good
○ No difference
○ Comment: ambiguous question
● Distractor quality mediocre
○ 2 out of 3 good handcrafted
○ 1-2 out of 3 good generated
○ Comment: unrelated to key

More Related Content

Similar to Generation of Assessment Questions from Textbooks Enriched with Knowledge Models (20)

PDF
Development of system for generating questions, answers, distractors using tr...
IJECEIAES
 
PDF
Exploring Semantic Question Generation Methodology and a Case Study for Algor...
IJCI JOURNAL
 
PDF
AN AUTOMATED MULTIPLE-CHOICE QUESTION GENERATION USING NATURAL LANGUAGE PROCE...
kevig
 
PDF
Comparison Intelligent Electronic Assessment with Traditional Assessment for ...
CSEIJJournal
 
PDF
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...
cseij
 
PDF
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...
cseij
 
PDF
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...
cseij
 
PDF
20433-39028-3-PB.pdf
IjictTeam
 
PDF
Automated Question Paper Generator And Answer Checker Using Information Retri...
Sheila Sinclair
 
PDF
Automated Essay Scoring Using Generalized Latent Semantic Analysis
Gina Rizzo
 
PDF
Exploiting Distributional Semantic Models in Question Answering
Pierpaolo Basile
 
PDF
Enhancing Video Understanding: NLP-Based Automatic Question Generation
IRJET Journal
 
PDF
IRJET- Implementation of Automatic Question Paper Generator System
IRJET Journal
 
PDF
Apply deep learning to improve the question analysis model in the Vietnamese ...
IJECEIAES
 
PDF
Autonomous Educational Testing System Using Unsupervised Feature Learning.
IJITE
 
PDF
Autonomous Educational Testing System Using Unsupervised Feature Learning
IJITE
 
PDF
Autonomous Educational Testing System Using Unsupervised Feature Learning
IJITE
 
PPTX
Advancing Intelligent Textbooks with Automatically Generated Practice: A Larg...
Sergey Sosnovsky
 
PDF
IRJET- An Automated Approach to Conduct Pune University’s In-Sem Examination
IRJET Journal
 
PPTX
Using Semantics of Textbook Highlights to Predict Student Comprehension and K...
Sergey Sosnovsky
 
Development of system for generating questions, answers, distractors using tr...
IJECEIAES
 
Exploring Semantic Question Generation Methodology and a Case Study for Algor...
IJCI JOURNAL
 
AN AUTOMATED MULTIPLE-CHOICE QUESTION GENERATION USING NATURAL LANGUAGE PROCE...
kevig
 
Comparison Intelligent Electronic Assessment with Traditional Assessment for ...
CSEIJJournal
 
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...
cseij
 
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...
cseij
 
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...
cseij
 
20433-39028-3-PB.pdf
IjictTeam
 
Automated Question Paper Generator And Answer Checker Using Information Retri...
Sheila Sinclair
 
Automated Essay Scoring Using Generalized Latent Semantic Analysis
Gina Rizzo
 
Exploiting Distributional Semantic Models in Question Answering
Pierpaolo Basile
 
Enhancing Video Understanding: NLP-Based Automatic Question Generation
IRJET Journal
 
IRJET- Implementation of Automatic Question Paper Generator System
IRJET Journal
 
Apply deep learning to improve the question analysis model in the Vietnamese ...
IJECEIAES
 
Autonomous Educational Testing System Using Unsupervised Feature Learning.
IJITE
 
Autonomous Educational Testing System Using Unsupervised Feature Learning
IJITE
 
Autonomous Educational Testing System Using Unsupervised Feature Learning
IJITE
 
Advancing Intelligent Textbooks with Automatically Generated Practice: A Larg...
Sergey Sosnovsky
 
IRJET- An Automated Approach to Conduct Pune University’s In-Sem Examination
IRJET Journal
 
Using Semantics of Textbook Highlights to Predict Student Comprehension and K...
Sergey Sosnovsky
 

More from Sergey Sosnovsky (20)

PPTX
Harnessing Textbooks for High-Quality Labeled Data: An Approach to Automatic ...
Sergey Sosnovsky
 
PDF
Toward Eliminating Hallucinations: GPT-based Explanatory AI for Intelligent T...
Sergey Sosnovsky
 
PDF
Layout- and Activity-based Textbook Modeling for Automatic PDF Textbook Extra...
Sergey Sosnovsky
 
PPTX
Exploring the Content Ecosystem of the First Open-source Adaptive Tutor and i...
Sergey Sosnovsky
 
PPTX
Creating Session Data from eTextbook Event Streams
Sergey Sosnovsky
 
PDF
Augmenting Digital Textbooks with Reusable Smart Learning Content: Solutions ...
Sergey Sosnovsky
 
PDF
Interactions of reading and assessment activities
Sergey Sosnovsky
 
PDF
Parallel Construction: A Parallel Corpus Approach for Automatic Question Gene...
Sergey Sosnovsky
 
PDF
YAI4Edu: an Explanatory AI to Generate Interactive e-Books for Education
Sergey Sosnovsky
 
PDF
Automatic Question Generation for Evidence-based Online Courseware Engineering
Sergey Sosnovsky
 
PDF
Reading Comprehension Quiz Generation using Generative Pre-trained Transformers
Sergey Sosnovsky
 
PPTX
Mathematical Language Processing via Tree Embeddings
Sergey Sosnovsky
 
PPTX
Contextual Definition Generation
Sergey Sosnovsky
 
PPTX
Transforming Textbooks into Learning by Doing Environments: An Evaluation of ...
Sergey Sosnovsky
 
PPTX
Dental TutorBot: Exploitation of Dental Textbooks for Automated Learning
Sergey Sosnovsky
 
PDF
What's in a textbook
Sergey Sosnovsky
 
PPTX
Using Programmed Instruction to Help Students Engage with eTextbook Content
Sergey Sosnovsky
 
PPTX
Adding Intelligence to a Textbook for Human Anatomy with a Causal Concept Map...
Sergey Sosnovsky
 
PPTX
Interlingua: Linking Textbooks Across Different Languages
Sergey Sosnovsky
 
PPTX
Student Modeling with Automatic Knowledge Component Extraction for Adaptive T...
Sergey Sosnovsky
 
Harnessing Textbooks for High-Quality Labeled Data: An Approach to Automatic ...
Sergey Sosnovsky
 
Toward Eliminating Hallucinations: GPT-based Explanatory AI for Intelligent T...
Sergey Sosnovsky
 
Layout- and Activity-based Textbook Modeling for Automatic PDF Textbook Extra...
Sergey Sosnovsky
 
Exploring the Content Ecosystem of the First Open-source Adaptive Tutor and i...
Sergey Sosnovsky
 
Creating Session Data from eTextbook Event Streams
Sergey Sosnovsky
 
Augmenting Digital Textbooks with Reusable Smart Learning Content: Solutions ...
Sergey Sosnovsky
 
Interactions of reading and assessment activities
Sergey Sosnovsky
 
Parallel Construction: A Parallel Corpus Approach for Automatic Question Gene...
Sergey Sosnovsky
 
YAI4Edu: an Explanatory AI to Generate Interactive e-Books for Education
Sergey Sosnovsky
 
Automatic Question Generation for Evidence-based Online Courseware Engineering
Sergey Sosnovsky
 
Reading Comprehension Quiz Generation using Generative Pre-trained Transformers
Sergey Sosnovsky
 
Mathematical Language Processing via Tree Embeddings
Sergey Sosnovsky
 
Contextual Definition Generation
Sergey Sosnovsky
 
Transforming Textbooks into Learning by Doing Environments: An Evaluation of ...
Sergey Sosnovsky
 
Dental TutorBot: Exploitation of Dental Textbooks for Automated Learning
Sergey Sosnovsky
 
What's in a textbook
Sergey Sosnovsky
 
Using Programmed Instruction to Help Students Engage with eTextbook Content
Sergey Sosnovsky
 
Adding Intelligence to a Textbook for Human Anatomy with a Causal Concept Map...
Sergey Sosnovsky
 
Interlingua: Linking Textbooks Across Different Languages
Sergey Sosnovsky
 
Student Modeling with Automatic Knowledge Component Extraction for Adaptive T...
Sergey Sosnovsky
 
Ad

Recently uploaded (20)

PDF
Supply Chain Security A Comprehensive Approach 1st Edition Arthur G. Arway
rxgnika452
 
PDF
Genomics Proteomics and Vaccines 1st Edition Guido Grandi (Editor)
kboqcyuw976
 
PPTX
How Physics Enhances Our Quality of Life.pptx
AngeliqueTolentinoDe
 
PDF
Quiz Night Live May 2025 - Intra Pragya Online General Quiz
Pragya - UEM Kolkata Quiz Club
 
PDF
Cooperative wireless communications 1st Edition Yan Zhang
jsphyftmkb123
 
DOCX
Lesson 1 - Nature and Inquiry of Research
marvinnbustamante1
 
PDF
Andreas Schleicher_Teaching Compass_Education 2040.pdf
EduSkills OECD
 
PDF
Indian National movement PPT by Simanchala Sarab, Covering The INC(Formation,...
Simanchala Sarab, BABed(ITEP Secondary stage) in History student at GNDU Amritsar
 
PPTX
How to Create & Manage Stages in Odoo 18 Helpdesk
Celine George
 
PPTX
Elo the Hero is an story about a young boy who became hero.
TeacherEmily1
 
PPTX
PLANNING FOR EMERGENCY AND DISASTER MANAGEMENT ppt.pptx
PRADEEP ABOTHU
 
PPTX
How to Add a Custom Button in Odoo 18 POS Screen
Celine George
 
PPTX
Lesson 1 Cell (Structures, Functions, and Theory).pptx
marvinnbustamante1
 
PPTX
How to Setup Automatic Reordering Rule in Odoo 18 Inventory
Celine George
 
PPTX
Urban Hierarchy and Service Provisions.pptx
Islamic University of Bangladesh
 
DOCX
MUSIC AND ARTS 5 DLL MATATAG LESSON EXEMPLAR QUARTER 1_Q1_W1.docx
DianaValiente5
 
PDF
COM and NET Component Services 1st Edition Juval Löwy
kboqcyuw976
 
PDF
Lesson 1 : Science and the Art of Geography Ecosystem
marvinnbustamante1
 
PPTX
How to Configure Refusal of Applicants in Odoo 18 Recruitment
Celine George
 
PPTX
Practice Gardens and Polytechnic Education: Utilizing Nature in 1950s’ Hu...
Lajos Somogyvári
 
Supply Chain Security A Comprehensive Approach 1st Edition Arthur G. Arway
rxgnika452
 
Genomics Proteomics and Vaccines 1st Edition Guido Grandi (Editor)
kboqcyuw976
 
How Physics Enhances Our Quality of Life.pptx
AngeliqueTolentinoDe
 
Quiz Night Live May 2025 - Intra Pragya Online General Quiz
Pragya - UEM Kolkata Quiz Club
 
Cooperative wireless communications 1st Edition Yan Zhang
jsphyftmkb123
 
Lesson 1 - Nature and Inquiry of Research
marvinnbustamante1
 
Andreas Schleicher_Teaching Compass_Education 2040.pdf
EduSkills OECD
 
Indian National movement PPT by Simanchala Sarab, Covering The INC(Formation,...
Simanchala Sarab, BABed(ITEP Secondary stage) in History student at GNDU Amritsar
 
How to Create & Manage Stages in Odoo 18 Helpdesk
Celine George
 
Elo the Hero is an story about a young boy who became hero.
TeacherEmily1
 
PLANNING FOR EMERGENCY AND DISASTER MANAGEMENT ppt.pptx
PRADEEP ABOTHU
 
How to Add a Custom Button in Odoo 18 POS Screen
Celine George
 
Lesson 1 Cell (Structures, Functions, and Theory).pptx
marvinnbustamante1
 
How to Setup Automatic Reordering Rule in Odoo 18 Inventory
Celine George
 
Urban Hierarchy and Service Provisions.pptx
Islamic University of Bangladesh
 
MUSIC AND ARTS 5 DLL MATATAG LESSON EXEMPLAR QUARTER 1_Q1_W1.docx
DianaValiente5
 
COM and NET Component Services 1st Edition Juval Löwy
kboqcyuw976
 
Lesson 1 : Science and the Art of Geography Ecosystem
marvinnbustamante1
 
How to Configure Refusal of Applicants in Odoo 18 Recruitment
Celine George
 
Practice Gardens and Polytechnic Education: Utilizing Nature in 1950s’ Hu...
Lajos Somogyvári
 
Ad

Generation of Assessment Questions from Textbooks Enriched with Knowledge Models

  • 1. Generation of Assessment Questions from Textbooks Enriched with Knowledge Models Lucas Dresscher, Isaac Alpizar Chacon, Sergey Sosnovsky
  • 2. 1a. Introduction ● Assessment question in digital textbooks: ○ opportunity to practice ○ opportunity to receive feedback ○ opportunity to engage in interactive learning
  • 3. 1a. Introduction ● Assessment question in digital textbooks: ○ opportunity to practice ○ opportunity to receive feedback ○ opportunity to engage in interactive learning ● Assessment questions in adaptive textbooks: ○ opportunity to assess student knowledge ➢ opportunity to adapt
  • 4. 1a. Introduction ● Assessment question in digital textbooks: ○ opportunity to practice ○ opportunity to receive feedback ○ opportunity to engage in interactive learning ● Assessment questions in adaptive textbooks: ○ opportunity to assess student knowledge ➢ opportunity to adapt ● Three primary ways to add assessment questions: ○ create them manually ○ integrate external assessment material ○ generate them automatically
  • 5. 1b. Introduction ● AQG: current state-of-the-art ○ High-quality factoid questions ○ Limitations: ■ Simplicity ■ Few generic systems ■ Little adaptivity ● Contributions ○ Unique source type ○ Open domain ○ Variety of questions ○ Supports adaptivity
  • 6. 2. Intextbooks Platform ● Knowledge models ○ Extraction: create semantic model ○ Enrichment: link additional information ○ Serialization: export model
  • 7. 2. Intextbooks Platform ● Knowledge models ○ Extraction: create semantic model ○ Enrichment: link additional information ○ Serialization: export model
  • 8. 3. Automatic Question Generation System ● Generic semantic rule-based AQG system ○ Domain independent ○ Uses textual and semantic information ○ Can generate questions for a part of a book or a (set of) concept(s)
  • 9. 3A. Source Extraction ● Input ○ Knowledge model of a textbook ● Extraction ○ Relevant sentences ○ Index terms’ enrichments ■ Domain specificity ● Output ○ Initial set of sentences
  • 10. 3B. Preprocessing ● Annotation pipeline Stanford CoreNLP ○ Parts of Speech (POS) ○ Dependency parsing ○ ● Filtering ○ Incorrect form (questions, imperative, grammatically incomplete,..) ○ Context reference (needs preceding sentences to make sense) ○ Visual references (figures, tables,...) ○ Numerical examples, ○ etc. ● Output ○ set of grammatically-fit annotated sentences
  • 11. 3C. Sentence Selection ● Select most appropriate sentences ● Scoring process ○ Weighted average of individual features ■ Different aspect ■ Feature score f [0, 1] ■ Weight w ○ Sentence score s [0, 1] ○ Threshold comparison ● Output ○ Potential source phrases
  • 12. ● Determine generatable question types ● Question types ○ True-false (unmodified) ○ True-false (negated) ○ True-false (substituted) ○ Cloze ○ Multiple-choice ● Output ○ Definitive source sentences The mean is the average. The mean is not the average. The median is the average. The ______ is the average. The ______ is the average. 3D. Question Type Selection A. Median B. Mean C. Mode D. Variance
  • 13. 3E. Question Construction ● Create questions in surface form ● TFU ○ Directly use sentence ○ Answer: true ● TFN ○ Negate question stem ○ Answer: false ● TFS ○ Substitute target concept ○ Answer: false The mean is the average. The mean is not the average. The mean is the average. The mean is the average. The median is the average.
  • 14. 3E. Question Construction ● CQ ○ Replace target concept by gap ○ Gap selection ● MCQ ○ CQ with options ■ Key ■ Distractors ○ Distractor generation ■ Related elements ■ Scoring procedure The mean is the average. The _____ is the average. The mean is the average. The _____ is the average. A. Median B. Mean C. Mode D. Variance
  • 15. 4. Evaluation ● Selection procedure ○ Three university-level statistics textbooks ○ Ten randomly selected co-occurring concepts ■ Five for automatic generation ■ Five for manual creation ○ 50 questions in total
  • 16. 4. Evaluation ● Selection procedure ○ Three university-level statistics textbooks ○ Ten randomly selected co-occurring concepts ■ Five for automatic generation ■ Five for manual creation ○ 50 questions in total ● Evaluation approach ○ Expert evaluation ○ Metrics ■ General: wording, assessment value, difficulty ■ Specific: gap quality, distractor quality
  • 17. 4. Evaluation ● Research questions: 1. Is the approach conceptually sound? 2. Is the approach practically sound?
  • 18. 4. Evaluation ● Inter-rater agreement (Fleiss’ Kappa) ○ 0.24 wording ○ 0.27 assessment value ○ -0.02 difficulty
  • 19. 4. Evaluation ● Inter-rater agreement (Fleiss’ Kappa) ○ 0.24 wording ○ 0.27 assessment value ○ -0.02 difficulty ● Comparison between handcrafted and generated questions (Mann-Whitney U test) ○ Statistically significant difference for the overall assessment value (Handcrafted > Generated) ■ 0.32 (U = 413.5, P = 0.048) ■ No statistically significant difference per question type ○ No significant differences for the overall wording
  • 20. 4. Evaluation ● Assessment value needs improvements ● Largest difference TFUs and MCQs ● TFSs particularly poor
  • 21. 4. Evaluation ● Good overall wording ● Small differences ● TFSs, CQs and MCQs particularly good
  • 22. 4. Evaluation ● Easy to medium ● Small differences
  • 23. 5. Conclusion ● Limitations ○ Nature of education textbooks ○ Weights of feature sets ● Future work ○ Other domains ○ Additional features ○ Closer Intextooks integration
  • 24. Time to Generate Some Questions
  • 25. 4. Evaluation ● Question type specific metrics ● Gap quality good ○ No difference ○ Comment: ambiguous question ● Distractor quality mediocre ○ 2 out of 3 good handcrafted ○ 1-2 out of 3 good generated ○ Comment: unrelated to key

Editor's Notes

  • #2: Welcome, glad you’re all here, advantage of a digital version we don’t have to worry about traffic jams all around Utrecht Excited to talk about research of past 7-8 months about automated generation of assessment questions from textbook models First introduce research area Why? Problem aims to resolve Current state of the art Then a quick look at earlier work which my research is built-upon Before moving on to the actual AQG system I researched Finalize with results from a pilot evaluation and some concluding words
  • #3: Motivation for this research area begins with the general value of assessment questions Improve learning process: interactivity, reinforce learning process by repeating concepts, provides evidence of a student’s knowledge MOOCS platforms, like Coursera or Duolingo, rapidly grown in popularity, further enhanced by the covid-19 epidemic. With millions of users it is no longer feasible to manually develop assessment questions at such large scale → So, needs a solution. New techniques and improvements from other research areas, e.g. neural networks
  • #4: Motivation for this research area begins with the general value of assessment questions Improve learning process: interactivity, reinforce learning process by repeating concepts, provides evidence of a student’s knowledge MOOCS platforms, like Coursera or Duolingo, rapidly grown in popularity, further enhanced by the covid-19 epidemic. With millions of users it is no longer feasible to manually develop assessment questions at such large scale → So, needs a solution. New techniques and improvements from other research areas, e.g. neural networks
  • #5: Motivation for this research area begins with the general value of assessment questions Improve learning process: interactivity, reinforce learning process by repeating concepts, provides evidence of a student’s knowledge MOOCS platforms, like Coursera or Duolingo, rapidly grown in popularity, further enhanced by the covid-19 epidemic. With millions of users it is no longer feasible to manually develop assessment questions at such large scale → So, needs a solution. New techniques and improvements from other research areas, e.g. neural networks
  • #6: Current state-of-the-art Capable of generating high-quality factoid questions using different types of systems and generation sources Limitations General shortcoming is the simplicity of the questions, targeting only lower cognitive levels Research commonly focuses on just a single question type with systems specifically designed for this type, sometimes single domain Little adaptivity in question scope and difficulty
  • #7: Most importantly for this research is the extracted domain knowledge, obtained by processing the index at the end of the textbook Each Individual index terms are identified. Using the page numbers, the term is recognized in the sentences that are about this index term, which are then extracted. Terms are connected to DBPedia resources, adding semantic information: Abstracts, categories, other concepts to which the term is related, domain specificity: the relationship the index term has with the domain of the textbook.
  • #8: Most importantly for this research is the extracted domain knowledge, obtained by processing the index at the end of the textbook Each Individual index terms are identified. Using the page numbers, the term is recognized in the sentences that are about this index term, which are then extracted. Terms are connected to DBPedia resources, adding semantic information: Abstracts, categories, other concepts to which the term is related, domain specificity: the relationship the index term has with the domain of the textbook.
  • #9: The system operates in five major steps
  • #10: Output: initial set of relevant sentences
  • #11: Incorrect form: Questions Imperative sentences Grammatically incorrect
  • #13: Which QT can be created is mainly determined by looking at the structure of the sentence, its target concept and relations with other related terms
  • #15: Challenging task, because distractor needs to be semantically similar but not plausible answer themselves Follows same scoring procedure as sentence selection (with a feature set that combined computes the quality of the distractor) Substituted True-False question is selected in the same way, but requires just a single related element
  • #16: 1 question of each of the 5 types was created for every concept, resulting in 50 questions in total (25 handcrafted, 25 generated)
  • #17: For CQS and MCQs, they also looked at: how well is the gap chosen? how well are the selected distractors ?
  • #19: Low but expected as it’s a difficult metric to estimate objectively. → usually calibrated based on data produced by real test takers.
  • #20: Compare evaluations of handcrafted and generated questions
  • #23: Compare evaluations of handcrafted and generated questions