ProgramsView all programs
From the news
- Chancellor Amma Addresses the Parliament of World’s Religions
- Amrita Students Qualify for the European Mars Rover Challenge
Publication Type : Conference Paper
Thematic Areas : Learning-Technologies, Wireless Network and Application
Publisher : 2019 IEEE Tenth International Conference on Technology for Education (T4E) .
Source : 2019 IEEE Tenth International Conference on Technology for Education (T4E) (2019)
Keywords : Answer sheets, bilingual assessment, Computer aided instruction, consistent results, correlation coefficient, Dogs, dyslexia, error types, handicapped aids, Indian schools, inter-rater agreement, inter-rater reliability, internal consistency, internal test consistency, intraclass agreement, Malayalam-English dyslexia screening, paediatrics, paper-based tests, rater agreement, reading disability, Reliability, reproducible results, screening test, second-grade students, standardized dyslexia screening test, Task analysis, undiagnosed dyslexics, Visualization, Vocabulary, Writing.
Campus : Amritapuri
School : Department of Computer Science and Engineering, School of Engineering
Center : Technologies & Education (AmritaCREATE), Amrita Center For Research in Analytics, Amrita Center for Wireless Networks and Applications (AmritaWNA)
Department : Computer Science
Year : 2019
Abstract : A standardized dyslexia screening test can help in identifying the vast number of undiagnosed dyslexics in Indian schools; needless to say, such a test should produce consistent and reproducible results. This study investigated reliability and consistency among raters of a Malayalam-English dyslexia screening in India. Paper-based tests were administered to groups of students, and four raters evaluated the answer sheets of 208 second-grade students (ages 6-7). Inter-rater agreement, intraclass agreement, and internal consistency were calculated. Our findings include good agreement among raters' appraisals for most error types and tasks. Internal consistency for a few tasks was low, possibly because these tasks evaluated more than one skill. A few error types need to be redefined and a few tasks need to be more skill-specific to enable unambiguous and fruitful interpretation by different raters in the future.
Cite this Research Publication : M. Haridas, Dr. Nirmala Vasudevan, Sasikumar, L., Gutjahr, G., Raghu Raman, and Prof. Prema Nedungadi, “Inter-Rater Reliability of a Dyslexia Screening Test”, in 2019 IEEE Tenth International Conference on Technology for Education (T4E), 2019