Back close

Inter-Rater Reliability of a Dyslexia Screening Test

Publication Type : Conference Paper

Thematic Areas : Learning-Technologies, Wireless Network and Application

Publisher : 2019 IEEE Tenth International Conference on Technology for Education (T4E) .

Source : 2019 IEEE Tenth International Conference on Technology for Education (T4E) (2019)

Url : https://ieeexplore.ieee.org/abstract/document/8983727

Keywords : Answer sheets, bilingual assessment, Computer aided instruction, consistent results, correlation coefficient, Dogs, dyslexia, error types, handicapped aids, Indian schools, inter-rater agreement, inter-rater reliability, internal consistency, internal test consistency, intraclass agreement, Malayalam-English dyslexia screening, paediatrics, paper-based tests, rater agreement, reading disability, Reliability, reproducible results, screening test, second-grade students, standardized dyslexia screening test, Task analysis, undiagnosed dyslexics, Visualization, Vocabulary, Writing.

Campus : Amritapuri

School : Department of Computer Science and Engineering, School of Engineering

Center : AI and Disability Studies, Technologies & Education (AmritaCREATE), Amrita Center For Research in Analytics, Amrita Center for Wireless Networks and Applications (AmritaWNA)

Department : Computer Science

Year : 2019

Abstract : A standardized dyslexia screening test can help in identifying the vast number of undiagnosed dyslexics in Indian schools; needless to say, such a test should produce consistent and reproducible results. This study investigated reliability and consistency among raters of a Malayalam-English dyslexia screening in India. Paper-based tests were administered to groups of students, and four raters evaluated the answer sheets of 208 second-grade students (ages 6-7). Inter-rater agreement, intraclass agreement, and internal consistency were calculated. Our findings include good agreement among raters' appraisals for most error types and tasks. Internal consistency for a few tasks was low, possibly because these tasks evaluated more than one skill. A few error types need to be redefined and a few tasks need to be more skill-specific to enable unambiguous and fruitful interpretation by different raters in the future.

Cite this Research Publication : M. Haridas, Dr. Nirmala Vasudevan, Sasikumar, L., Gutjahr, G., Raghu Raman, and Prof. Prema Nedungadi, “Inter-Rater Reliability of a Dyslexia Screening Test”, in 2019 IEEE Tenth International Conference on Technology for Education (T4E), 2019

Admissions Apply Now