Implementation of Bloom's Taxonomy in Electronic Assessment: Evaluation of Subjective Type Questions using AI Techniques

Implementation of Bloom's Taxonomy in E-Assessment: Evaluation of Subjective Type Questions
Cognitive Domain of Bloom's Taxonomy 


Electronic assessment or e-assessment systems mostly use Objective Type Questions (OTQs) because these have clear answer and choices are limited.  However, Subjective Type Questions (STQs) are difficult to grade as their answers may have a difficult sequence of logic followed by the student.  These are difficult to automate due to divergent responses based on the opinion of the learner.  On the other hand, Bloom’s Taxonomy (BT)  is a systematic categorization of educational objectives (Bloom et. al., 1956) commonly used as a classification scheme to determine different levels of learner’s intellectual competencies.  Basically, it is application of three domains for learner’s learning assessment i.e. i) Cognitive ii) Affective and iii) Psychomotor .  These areas are hierarchical models used to classify educational learning objectives from basic to complex levels of learner’s thinking.  These levels are used to measure the higher order thinking level and provide guideline for design and construction appropriate depth of evaluations.

In modern age of technology, assessment has been done using automated mechanisms.  Consequently, technology and assessment create new domain known as e-assessment.  It is an important component of e-learning and has become one of the most common forms of technology enhanced assessment since the 1990s (Bruyn et. al., 2011).  E-assessment is an assessment carried out using various technology based tools, methods and devices/services. Assessment is about evaluating learning outcomes (Stodberg, 2012).  It is a tool for monitoring student’s activity using certain specified criteria.  There are several purposes of assessment as defined by Mehrens and Lehmann (cited in Cannon, 1995).  These are; judging mastery, measuring improvement over time, diagnosing student hitches, assessing the teaching methods, calculating the effectiveness of the course and motivating students to study etc.  Use of BT may help to design assessment for appropriate focus of e-assessments. 

Many common issues of e-assessment related to technology and devices were discussed by (Sangi, 2008).  Some of such issues specially associated to technology and devices have been researched over time.  However, work on e-assessment of STQs still needs more development.  Because, it is a most used and considered more valid process to assess depth of the student’s performance at higher level education.  Use of e-assessment is quick and with development of e-assessment of STQs it may be more valuable for institutions and faculty.   Additionally, use of BT will serve the base purpose of fair and valid e-assessment perhaps.   It may also provide flexibility, advance level questions capacity which can reduce the load of admin, provide more storage efficiency in systems (Nikolova, 2012).

Furthermore, E-assessment has four basic forms i) Diagnostic ii) Formative iii) Summative and iv) Integrative (Crisp, 2011).  While, formative assessment is used during the learning process for the assessment of learner. Furthermore, it also used to monitor learner’s ongoing progress and offer them feedback.   In each type of e-assessment design different formats of questions may be used. Currently many e-assessment techniques include the multiple formats of questions.  These questions formats are broadly categorized into following:

Divergent and Convergent Questions

Divergent and Convergent Questions


Figure 1: Formats of Questions 



However, objective type questions (OTQs) are difficult to construct but easy to assess automatically while subjective type questions (STQs) are easy to construct but difficult to assess automatically.  Therefore, STQs are often evaluated by tutors, teaching assistants or instructors hence these are time consuming to grade (Cannon, 1995). Additionally, issues are associated with maintaining fairness, objectivity in grading and often maintaining language sensitivity, which is influenced by teacher and learner’s ability to communicate.  The issues of variety in answer by different students make STQs difficult to understand by the machine and thus not practical in automated evaluations.  It can be assumed that introducing systematic process of BT in formatting the questions may help to overcome of some difficulties mentioned above.


Additionally, the study of (Sangi, 2008 and Bruyn et. al., 2011) mentiond for incorporation of traditional methods of assessment i.e. written tests, oral examinations, assignments, portfolios, projects, presentations etc in e-assessment. While, (Crisp, 2011) explained that it can either be part of Learning Management System (LMS) or a standalone application and also identify various other modern techniques of assessment i.e. animation, blog, cloze, discussion, free text, hot spot, matrix, MCQ, ordering, pull down, role-play, self or peer review, simulation, virtual world and wiki.  Furthermore, the study of (Nikolova, 2012 & Alruwais et. al., 2018) reveals that e-assessment offers richer assessment experience, which increased the flexibility, provides instant feedback, reduces the management burden, offer greater storage and enhanced questions styles/formats.  Hence, it evaluates the learner’s performance electronically i.e. by the use of Information and Communication Technologies (ICT).



Assessment is a procedure for measuring the knowledge, attitudes and skills of the learner’s and useful for teaching and learning environment, which promote learning, measure or certify results of the learner’s (Clements et. al., 2013).  Whereas, Scottish Qualifications Authority (SQA) publication define assessment i.e. “Assessment is the process of evaluating an individual’s attainment of Cognitive, Affective and Psychomotor.  It has many uses but these can be divided into two major categories: uses for individual candidates, and uses for external organizations” (SQA, 2017).   Hence, it is the most important and useful component and tool to measure or evaluate the quality of teaching and learning environments.  Several assessment techniques include the type of questions to be assessed by the examiner or the assessment tool in case of automated assessment.  Moreover, as per planning framework table of BT the quiz and test are the most common methods used for the evaluation of the learner’s (“Kurwongbss,” n.d).  These questions are mostly categorized into two major formats i.e. OTQs and STQs discussed in Figure 1. 



The OTQs are also known as selected-response items require their answers to be picked form a list of choices while STQs are constructed-responses items require their answers in written form (Chalkboard, 2018).  On the other hand, OTQs are difficult to construct but easy to assess automatically while STQs are easy to construct but difficult to assess automatically (Alrehily et. al., 2018).  Moreover, OTQs are Forced-choice questions not effective to assess the skills competencies and depth of knowledge (Sangi, 2008).  While, the SQA in its publications discusses the question and states as following under the heading “Non-objective questions”:




“The downside of computer marking is that sometimes the assessment engine can be less accommodating than human markers. One question type available with some assessment engines that is not usually marked automatically is the free-text extended response or essay questions. The responses entered by candidates to these questions need to be scrutinized by a human.  Candidates type their answer in a text box and submit them in the normal ways that are then assessed by teacher” (SQA, 2005).



Various research studies revealed that most of the work in e-assessment is done on assessing OTQs (Sangi, 2008; Praveen et. al., 2014; Sankar, 2015; Agrawal et. al., 2017; Alrehily et. al., 2018).  However, cognitive domain of Bloom's Taxonomy was implemented with MCQs format (Bruyn et. al., 2011 & Lajis et. al., 2018)  to assess the medical knowledge and programming skills of the learner's.  Hence, to measure the higher order thinking level i.e. cognitive, affective and psychomotor skills of the learner which cannot readily be assessed using OTQs formats and more authentic e-assessments are being proposed (Kuh et. al., 2014). However, it is a challenge to test higher levels and promote deep learning in e-assessment as identify by (Bruyn et. al.,  2011). Consequently, the study of (Agrawal et. al., 2017) recommended automatic grading of STQs because it will be very helpful for educational institutions if the process of evaluation of descriptive answers be evaluate using e-assessment.



As a result of all this work, different commercial/academic tools and software have been developed some of this software such as Moodle (Moodle, 2018), Blackboard (FinanceOnline, 2018), e-Rater (ETS, 2018), ConductExame (ConductExam, 2018), Zoho (Praveen et. al., 2014) etc have electronic assessment modules incorporated in the package. Although there are numerous tools / software available as defined by (Sankar, 2015) but e-Rater and AutoMark are only well known tools which can automatic grading STQs for English literature to check writing style, grammatical and syntactical structure of sentences through Natural Language Processing (NLP) parser.   The online tests available at different websites also include e-assessment of students using OTQs.  Even if there is some space for answering lengthy i.e. STQs, their marking is done manually (Praveen et. al., 2014) and no mechanism is provided for automatic grading of the long answers. 



Furthermore, there is a little work done in assessing the STQs electronically using BT domains due to limitations involved in automated grading of subjective contents intended to measure or verified abilities of the learner.  One possible mechanism is to electronically assess the in-depth knowledge, attitudes and skills of the learner’s abilities etc organized in the form of subjective contents may be an assessment model developed based on BT domains and the semantic analysis of sentences.  Hence, so far many Artificial intelligence (AI) techniques and algorithms have been developed/devised to mark STQs.  These involve Computational Linguistics, Natural Language Processing (NLP), Statistics, Information Extraction (IE), Pattern Matching (Dube et. al., 2014).   While (Alrehily et. al., 2018) propose another technique i.e Syemantic similarity and document simalarity to find matching the responses of the learner.  Some of these techniques and algorithms are combined in existing marking engines to achieve better accuracy. 


Whereas, there is a little work done in assessing the STQs electronically using BT domains due to limitations involved in automated grading of subjective contents intended to measure or verified abilities of the learner.  One possible mechanism is to electronically assess the in-depth knowledge, attitudes and skills of the learner’s abilities etc organized in the form of subjective contents may be an assessment model developed based on BT domains and the semantic analysis of sentences.  Hence, so far many Artificial intelligence (AI) techniques and algorithms have been developed/devised to mark STQs.  These involve Computational Linguistics, Natural Language Processing (NLP), Statistics, Information Extraction (IE), Pattern Matching (Dube et. al., 2014).   While (Alrehily et. al., 2018) propose another technique i.e Syemantic similarity and document simalarity to find matching the responses of the learner.  Some of these techniques and algorithms are combined in existing marking engines to achieve better accuracy. 


Hence, there is a need of time for implementation of Bloom's Taxonomy in electronic assessment especially for systematic (and more objective) evaluation of subjective/divergent type questions.  As these questions are regarded as the most valid and reliable tool to evaluate the learner’s deep knowledge, understanding and skills.  Hence, BT can be applied in e-assessment for STQs as well.  This would provide the theoretical base for development of a new model of e-assessment using systematic and well appreciated approach of BT.  Which may further pave way for development of Artificial Intelligence (AI) / Machine Learning (ML) based techniques for evaluation of STQs in future. Additionally, it will not only help presentation of different question types but also ensure that higher cognitive skills can also be covered, thus making such systems pedagogically valuable.



Finally, it will be a fast, accurate, cost-effective, reliable and valid assessment method. Additionally, it will replace manual grading system which is costly, time consuming, lacks of consistency and slow in feedback.   Furthermore, the ultimate beneficiary will be teachers, learners and Institutions leading to system development.  It will provide instant feedback, faster evaluation, diagnosing learner’s hitches, evaluating the teaching methods, save precious time of learners, evaluators and ensure authentic assessment.

References:

Alruwais, N., Wills, G., & Wald, M. (2018). Advantages and Challenges of Using E-assessment. International Journal of Information and Education Technology, 8(1), 34-37.
Alrehily, A. D., Siddiqui, M. A., & Buhari, S. M. (2018) INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS. Computer Science & Information Technology, 47.
Anderson, L. W., Krathwohl, D. R., Airasian, P. W., Cruikshank, K. A., Mayer, R. E., Pintrich, P. R., Raths, J & Wittrock, M. C. (2001). A taxonomy for learning, teaching, and assessing: A revision of Bloom’s taxonomy of educational objectives, abridged edition. White Plains, NY: Longman.
Agrawal, L., Mhalunge, V., Pitale, B., & Bahir, K. (2017). Subjective Answer Evaluation System. IJSRSET  Volume 3,  Issue 1, 514-517.
Bloom, B. S., Engelhart, M. D., Furst, E. J., Hill, W. H., & Krathwohl, D. R. (1984). Taxonomy of educational objectives: Handbook 1: Cognitive domain. Longman Publishing Group.
Cannon, Newble, D. & Robert. (1995). A handbook for teachers in universities & colleges. London: Kogan Page.
Chalkboard, D. (2018). Selected-Response vs. Constructed-Response Assessment Methods,  Accessed on:  October 26, Retrieved from: https://www.mydigitalchalkboard.org/portal/default/Content/Viewer/Content?action=2&scId=505706&sciId=15387.
Clements, M. D., & Cord, B. A. 2013. "Assessment guiding learning: developing graduate qualities in an experiential learning programme. ." Assessment & Evaluation in Higher Education, 38(1), 114-124.
Crisp, G. (2011). Teacher's Handbook on e-Assessment. Sydney: Australian Learning and Teaching Council Fellowship.
ConductExam. (2018). Subjective Type Questions. Accessed on: October 08, Retrieved from: https://blog.conductexam.com/subjective-type-questions.
Dube, T., & Ma, M. (2014). Marking short free text responses in e-assessment. In Workshop Chair (p. 16).
De Bruyn, E., Mostert, E., & Van Schoor, A. (2011, September). Computer-based testing-the ideal tool to assess on the different levels of Bloom's taxonomy. In Interactive Collaborative Learning (ICL), 2011 14th International Conference on (pp. 444-449). IEEE.
ETS. (2018). How the e-rater® Engine Works,   Accessed on: October 05, Retrieved from: https://www.ets.org/erater/how/.
FinanceOnline. (2018). Compare Moodle vs Blackboard, Accessed on: October 05, Retrieved from:  https://comparisons.financesonline.com/moodle-vs-blackboard-learn.
Isaacs, G. (1996). Bloom's taxonomy of educational objectives. Brisbane: The University of Queensland.
Lajis, A., Nasir, H. M., & Aziz, N. A. (2018, February). Proposed Assessment Framework Based on Bloom Taxonomy Cognitive Competency: Introduction to Programming. In Proceedings of the 2018 7th International Conference on Software and Computer Applications (pp. 97-101). ACM.
Kuh, G. D., Jankowski, N., Ikenberry, S. O., &Kinzie, J. L. (2014). Knowing what students know and can do: The current state of student learning outcomes assessment in US colleges and universities. Champaign, IL: National Institute for Learning Outcomes Assessment.
Kaur, N., & Jyoti, K. (2012). Automated assessment of short one-line free-text responses in computer science. International Journal of Computer Science and Informatics, 2(1), 105-109.
Melovitz Vasan, C. A., DeFouw, D. O., Holland, B. K., & Vasan, N. S. (2018). Analysis of testing with multiple choice versus open‐ended questions: Outcome‐based observations in an anatomy course. Anatomical sciences education, 11(3), 254-261.
Moodle. (2018). Moodle, http://moodle.org/, Accessed on: October 05, Retrieved from: http://moodle.org/.
Nikolova, M. (2012). "Characteristics and Forms of the Electronic Assessment of the Knowledge ." Scientific works of the University of Rousse 94.
Praveen, S., & Khan, A. H. J. (2014). An Approach to Evaluate Subjective Questions for Online Examination System. Published in International Journal of Innovative Research in Computer and Communication Engineering2(11).
Sangi, N. A. (2008). Electronic assessment issues and practices in Pakistan: a case study. Learning, Media and Technology33(3), 191-206.
Sankar, A. (2015). Towards an automated system for short-answer assessment using ontology mapping. Int. Arab J. e_technology4(Guha), 1989.
Kurwongbss. (n.d). Planning Sheet, Accessed on: October 19, Retrieved from: https://www.niu.edu/facdev/_doc/blooms_planningsheet.doc.
SQA. (2005). Guidelines on e-assessment for Schools . Glasgow: Scottish Qualifications Authority.
Stodberg, U. (2012). "A research review of e-assessment ." Assessment & Evaluation in Higher Education 591-604.



No comments:

Post a Comment