Machine Learning Models: A Study of English Essay Text Content Feature Extraction and Automatic Scoring

Authors

  • Wei Shang School of Humanities, Shandong Agriculture and Engineering University, Jinan, Shandong 250000, China
  • Huihua Men School of Humanities, Shandong Agriculture and Engineering University, Jinan, Shandong 250000, China
  • Xiujie Du School of Humanities, Shandong Agriculture and Engineering University, Jinan, Shandong 250000, China

DOI:

https://doi.org/10.13052/jicts2245-800X.1143

Keywords:

Machine learning, English essay, feature extraction, Automatic scoring

Abstract

Accurate automatic scoring of English essay is beneficial for both teachers and students in English teaching. This paper briefly introduced an XGBoost-based automated scoring algorithm for English essay. To improve the accuracy of the algorithm, a long short-term memory (LSTM) semantic model was introduced to extract semantic scoring features from essays. Finally, the improved XGBoost algorithm was compared with the traditional XGBoost and LSTM algorithms in a simulation experiment using five types of essay prompts. The results indicate that the improved XGBoost algorithm has the best performance for automatic scoring of English essay and also requires the shortest scoring time.

Downloads

Download data is not yet available.

Author Biographies

Wei Shang, School of Humanities, Shandong Agriculture and Engineering University, Jinan, Shandong 250000, China

Wei Shang, female, graduated from Shandong Normal University, majoring in Curriculum and Teaching of English, and received Master’s Degree. Now she is working as a lecturer in School of Humanities in Shandong Agriculture and Engineering University, specializing in research of English teaching and applied linguistics. She has participated in three provincial scientific research projects and published more than ten papers.

Huihua Men, School of Humanities, Shandong Agriculture and Engineering University, Jinan, Shandong 250000, China

Huihua Men is a associate professor in the International Exchange and Cooperation Department in Shandong Agriculture and Engineering University, China. Her research interest include core competency, learning evaluation and English teaching. She has published more than 15 papers.

Xiujie Du, School of Humanities, Shandong Agriculture and Engineering University, Jinan, Shandong 250000, China

Xiujie Du graduated from Shandong Normal University, majoring in curriculum and teaching of English, and received Master’s Degree. Now she is working as a lecturer in School of Humanities in Shandong Agriculture and Engineering University, specializing in research of English teaching and translation.

References

B. Uysal, N. Doan, ‘Automated Essay Scoring Effect on Test Equating Errors in Mixed-format Test’, Int. J. Assess. Tools Educ., 8(2):222–238, 2021.

M. Beseiso, S. Alzahrani, ‘An Empirical Analysis of BERT Embedding for Automated Essay Scoring’, Int. J. Adv. Comput. Sc., 11(10): 204–210, 2020.

L. Xia, A. Mc, C. Jyn, ‘SEDNN: Shared and enhanced deep neural network model for cross-prompt automated essay scoring’, Knowl.-Based Syst., 210:106491.1–106491.14, 2020.

D. S. McNamara, S. A. Crossley, R. D. Roscoe, L. K. Allen, J. Dai, ‘A hierarchical classification approach to automated essay scoring’, Assess. Writ., 23:35–59, 2015.

H. Li, T. Dai, ‘Explore Deep Learning for Chinese Essay Automated Scoring’, J. Phys. Conf. Ser., 1631:012036, 2020.

S. Hao, Y. Xu, D. Ke, K. Su, H. Peng, ‘SCESS: a WFSA-based automated simplified chinese essay scoring system with incremental latent semantic analysis’, Nat. Lang. Eng., 22(2):291–319, 2016.

T. C. Stephen, M. C. Gierl, S. King, ‘Automated essay scoring (AES) of constructed responses in nursing examinations: An evaluation’, Nurse Educ. Pract., 54(1):103085, 2021.

L. Qian, Y. Zhao, Y. Cheng, ‘Evaluating China’s Automated Essay Scoring System iWrite’, J. Educ. Comput. Res., 2019(2):073563311988147, 2019.

R. Ahmad, ‘E-learning Automated Essay Scoring System Menggunakan Metode Searching Text Similarity Matching Text’, Jurnal Penelitian Enjiniring, 22(1):38–43, 2019.

J. Contreras, S. Hilles, Z. B. Abubakar, ‘Automated essay scoring using ontology generator and natural language processing with question generator based on blooms taxonomy’s cognitive level’, Int. J. Adv. Technol. Eng. Explor., 2019:2249–8958, 2019.

K. Rysová, M. Rysová, M. Novák, J. Mírovský, E. Hajičová, ‘EVALD – a Pioneer Application for Automated Essay Scoring in Czech’, Prague Bull. Math. Ling., 113(1):9–30, 2019.

H. M. Wang, ‘Automated Chinese Essay Scoring Based on Deep Learning’, Comput. Mater. Con., 2020(10):817–833, 2020.

M. Yamamoto, N. Umemura, H. Kawano, ‘Proposal of Japanese Vocabulary Difficulty Level Dictionaries for Automated Essay Scoring Support System Using Rubric’, J. Oper. Res. Soc. China, 8(4):601–617, 2020.

B. Uysal, N. Doan, ‘Automated Essay Scoring Effect on Test Equating Errors in Mixed-format Test’, Int. J. Assess. Tools E., 8(2):222–238, 2021.

M. Beseiso, S. Alzahrani, ‘An Empirical Analysis of BERT Embedding for Automated Essay Scoring’, Int. J. Adv. Comput. Sc., 11(10): 204–210, 2020.

Downloads

Published

2023-11-18

How to Cite

Shang, W. ., Men, H. ., & Du, X. . (2023). Machine Learning Models: A Study of English Essay Text Content Feature Extraction and Automatic Scoring. Journal of ICT Standardization, 11(04), 379–390. https://doi.org/10.13052/jicts2245-800X.1143

Issue

Section

Articles