Trinh, Thanh
Unknown Affiliation

Published : 2 Documents Claim Missing Document
Claim Missing Document
Check
Articles

Found 2 Documents
Search

Variation in Linguistic Complexity and Human Ratings of Second Language (L2) Writing Trinh, Thanh
International Journal of English Language and Pedagogy Vol. 3 No. 1 (2025): International Journal of English Language and Pedagogy (IJELP)
Publisher : Universitas Terbuka

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.33830/ijelp.v3i1.12042

Abstract

Drawn upon the usage-based approach which considers inter-individual variation as an inherent feature of language use and production, this study attempted to explore what linguistic complexity measures significantly differed among individual learners of similar learning conditions and how they were compatible with rubric-based scorings. Statistical analyses confirm that variation in complexity ranged from lexical, syntactic to discoursal levels. Comparisons with human ratings reveal that the variation in linguistic complexity measures is complex in nature since complexity correlates with perceived proficiency, particularly when it comes to semantic cohesion and lexical control but in others, complexity varies independently of performance, suggesting that the developmental stage of complexity may moderate its impact on ratings.
Reliability of ChatGPT-5.0 as an Automated Essay Scoring Tool: What Matters? Trinh, Thanh
International Journal of English Language and Pedagogy Vol. 3 No. 2 (2025): International Journal of English Language and Pedagogy (IJELP)
Publisher : Universitas Terbuka

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.33830/ijelp.v3i2.13551

Abstract

The aims of the present study were twofold: Exploring the variabilities of ChatGPT-5.0’s capabilities of rubric-based essay scoring across three prompting designs and two essay feeding methods; and testing the reliability of ChatGPT-generated scores against human ratings. Drawing upon three reliability measurements, including: Spearman’s correlations, Intraclass Correlations (Koraishi, 2024; Bui & Barrot, 2025) and quadratic weighted kappa (QWK) ( Mizumoto & Eguchi, 2023; Poole & Coss, 2024), the findings revealed that although the reliability coefficients ranged from moderate to substantial, the essay scoring abilities of ChatGPT-5.0 depended greatly upon users’ expertise to engineer prompts and their choices of essay feeding. This study highlights the importance of continued effort in the validation of this technology as an automated essay scoring tool and emphasizes the irreplaceability of human professional judgment in this field.