Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
5 views20 pages

Chapter 1 Final

This study investigates the use of supervised learning techniques to predict student academic performance at Virginia Run Elementary School, aiming to identify at-risk students and inform personalized interventions. It highlights the importance of predictive analytics in education, addressing challenges such as data quality, algorithmic bias, and the gap between prediction and actionable support. The research seeks to contribute to educational equity by developing models that leverage historical performance data to enhance student outcomes.

Uploaded by

td24031
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
5 views20 pages

Chapter 1 Final

This study investigates the use of supervised learning techniques to predict student academic performance at Virginia Run Elementary School, aiming to identify at-risk students and inform personalized interventions. It highlights the importance of predictive analytics in education, addressing challenges such as data quality, algorithmic bias, and the gap between prediction and actionable support. The research seeks to contribute to educational equity by developing models that leverage historical performance data to enhance student outcomes.

Uploaded by

td24031
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 20

Predictive Analysis of Student’s Academic Performance at Virginia Run Elementary School

Using Supervised Learning

Chapter 1: Introduction

Sufiyan Ahmed

2106313

Bay Atlantic University

February 21, 2024

1
Chapter 1: Introduction

1.0 Overview

The predictive analysis of student academic performance, particularly utilizing supervised

learning techniques, stands as a critical endeavor within contemporary education. With the

increasing emphasis on personalized learning and educational equity, understanding and

addressing the needs of individual students have become paramount (Dumont & Ready, 2023;

Zhang et al., 2020). The study at hand aims to delve into this realm by focusing on the academic

performance of students at Virginia Run Elementary School. This research is necessitated by the

gap in current educational practices, where proactive measures for identifying students who may

require additional support are often lacking (Syed, 2023). While educators strive to cater to the

diverse needs of their students, the ability to predict and preemptively address academic

challenges remains somewhat elusive. By leveraging the power of predictive analytics and

supervised learning, this study seeks to bridge this gap by developing a robust model that can

forecast students' academic trajectories based on past performance data, such as grades, test

scores, and attendance records.

The implications of this study are manifold. Firstly, it holds the potential to revolutionize the way

educational interventions are designed and implemented, shifting from reactive to proactive

strategies. By identifying students who are at risk of academic underperformance early on,

educators can tailor their support mechanisms accordingly, thereby fostering a more conducive

learning environment for all students (Yağcı, 2022). Moreover, the findings of this research could

contribute to the broader discourse on data-driven decision-making in education, highlighting the

efficacy of leveraging technology to optimize student outcomes.

2
1.1 Background of the Study

The background of the study provides a comprehensive overview of the existing research

literature concerning predictive analysis of student academic performance and supervised

learning techniques in education. It traces the evolution of this field, highlighting its significance

in addressing the ongoing challenges faced by educators in effectively supporting students'

academic growth.

1.1.1 Evolution of Predictive Analytics in Education

The evolution of predictive analytics in education represents a transformative journey marked by

significant advancements in data science, machine learning, and educational research. Initially,

educational research primarily focused on descriptive and inferential statistics to understand

patterns and trends in student performance (Picciano, 2012). However, the advent of predictive

analytics introduced a paradigm shift towards proactive decision-making and personalized

interventions (Sghir et al., 2022). Early efforts in predictive analytics within education can be

traced back to the late 20th century, where researchers began exploring the use of statistical

modeling techniques to predict student outcomes (Picciano, 2012). Studies by Baker et al. (2008)

and Romero et al. (2008) laid the groundwork for predictive modeling by identifying correlations

between various student attributes and academic success. These pioneering works demonstrated

the potential of predictive analytics in forecasting student performance based on demographic

factors, prior academic achievement, and socio-economic status (Baker et al., 2008; Romero et

al., 2008).

As computational power and data availability continued to expand, researchers turned towards

more sophisticated machine learning algorithms, particularly supervised learning, to enhance

3
predictive accuracy. Supervised learning algorithms, such as decision trees, logistic regression,

and neural networks, gained prominence due to their ability to learn from labeled training data

and make predictions on unseen instances (El Aouifi et al., 2021; Sarker, 2021). Kotsiantis et al.

(2007) and Sanchez-Pozo et al. (2021) are among the researchers who explored the efficacy of

supervised learning techniques in predicting student academic performance across diverse

educational contexts. The proliferation of learning management systems (LMS), student

information systems (SIS), and educational data warehouses further accelerated the adoption of

predictive analytics in education (Picciano, 2012; Sghir et al., 2022). These systems generate vast

amounts of data, including grades, attendance records, and assessment scores, providing rich

sources of information for predictive modeling (Maraza-Quispe et al., 2021). Educational

institutions began leveraging predictive analytics to identify students at risk of academic

underperformance, intervene proactively, and allocate resources more efficiently (Maraza-Quispe

et al., 2021).

Moreover, the emergence of open-source machine learning libraries, such as scikit-learn and

TensorFlow, democratized access to predictive modeling tools, enabling educators and

researchers to develop custom predictive models tailored to their specific contexts (Wang et al.,

2019). Collaborative initiatives, such as the Predictive Analytics Reporting (PAR) Framework

and the Society for Learning Analytics Research (SoLAR), facilitated knowledge sharing and

best practices in the application of predictive analytics in education (Wang et al., 2019).

1.1.2 Current Challenges and Limitations

One of the primary challenges in predictive analytics is ensuring the quality and availability of

data (Nunn et al., 2016). Educational data often exhibit variability, incompleteness, and

inconsistencies, stemming from diverse sources such as student information systems, learning

4
management systems, and standardized assessments (Nunn et al., 2016). Poor data quality can

compromise the accuracy and reliability of predictive models, leading to erroneous predictions

and misguided interventions (Nunn et al., 2016). Moreover, data silos and interoperability issues

across different systems pose obstacles to aggregating and integrating data for comprehensive

analysis. The use of student data for predictive analytics raises significant privacy and ethical

concerns (Al-Tameemi et al., 2020). Educational institutions must navigate complex legal and

regulatory frameworks, such as the Family Educational Rights and Privacy Act (FERPA) in the

United States, to ensure compliance with data protection laws (Al-Tameemi et al., 2020).

Moreover, there is a risk of unintended consequences, such as stigmatization and discrimination,

resulting from the use of predictive models to profile and label students based on their predicted

outcomes (Al-Tameemi et al., 2020; Nunn et al., 2016).

Another critical challenge in predictive analytics is the presence of algorithmic bias, where

machine learning models perpetuate or amplify existing societal biases present in the training

data (Akter et al., 2022). Biased predictions can disproportionately impact marginalized and

underrepresented student populations, exacerbating inequities in educational opportunities and

outcomes (Akter et al., 2022). Researchers and practitioners must employ fairness-aware

techniques, such as fairness-aware learning algorithms and bias mitigation strategies, to mitigate

algorithmic bias and promote equitable decision-making in predictive analytics (Yang et al.,

2023). The black-box nature of many machine learning algorithms poses challenges in

interpreting and explaining the underlying factors driving predictive outcomes (Yang et al.,

2023). Educators and stakeholders may be hesitant to trust and act upon predictions generated by

opaque models without understanding the rationale behind them.

5
The existing literature highlights the gap between identifying students at risk of academic

underperformance and implementing timely interventions to support their learning needs. While

predictive models can effectively identify high-risk students based on historical data, the

translation of these insights into actionable interventions remains a critical challenge. Educators

often lack the resources, expertise, and infrastructure required to implement targeted

interventions at scale. Consequently, there is a pressing need for research that not only identifies

students in need of support but also provides actionable recommendations for personalized

interventions tailored to individual students' needs (Reber et al., 2018).

In conclusion, the current state of research underscores the urgent need for proactive intervention

strategies informed by predictive analytics and supervised learning techniques. By addressing the

gap between prediction and intervention, this study aims to contribute to the ongoing efforts to

promote educational equity and student success. Through a nuanced understanding of the

challenges and opportunities associated with predictive analysis in education, this research

endeavors to empower educators with the tools and insights needed to foster a culture of data-

driven decision-making and personalized learning.

1.2 Problem Statement

Recent advancements in predictive analytics have promised to revolutionize education by

providing insights into student academic performance and enabling proactive interventions.

However, despite these technological advancements, a staggering statistic reveals the persistent

challenge facing educators: Data released by the Department of Education indicate that over 40%

of students in urban public schools are at risk of academic underperformance in Virgina,

highlighting the urgent need for targeted intervention strategies (Cline, 2023). Despite extensive

research in the last five years, a meaningful gap remains in understanding how to effectively

6
translate predictive insights into actionable interventions tailored to individual student needs. The

general problem is the disconnect between predictive analytics' potential to identify at-risk

students and the implementation of proactive intervention strategies to support their academic

success. Specifically, this study aims to address the gap in the current research literature by

investigating how supervised learning techniques can be leveraged to predict student academic

performance and inform personalized intervention strategies at Virginia Run Elementary School.

1.3 Statement of Purpose

This quantitative study aims to explore the predictive capabilities of supervised learning

techniques in forecasting student academic performance at Virginia Run Elementary School.

Specifically, the study intends to develop predictive models that describe the relationship

between independent variables such as grades, test scores, and attendance records, and the

dependent variable of student academic performance. Additionally, the study will explore the

potential moderating effect of socio-economic status as a covariate variable on the relationship

between predictor variables and academic performance. By employing a quantitative approach,

this research seeks to provide actionable insights for educators to identify at-risk students early

and implement targeted interventions to support their academic success.

1.4 Research Question

The primary research question guiding this quantitative study is:

RQ: What is the predictive capability of supervised learning techniques in forecasting student

academic performance at Virginia Run Elementary School, considering grades, test scores, and

attendance records as predictor variables?

7
Null Hypothesis (H0): There is no significant relationship between grades, test scores, attendance

records, and student academic performance at Virginia Run Elementary School.

Alternative Hypothesis (H1): There is a significant relationship between grades, test scores,

attendance records, and student academic performance at Virginia Run Elementary School.

1.4.1 Variables

Independent Variables: Grades, test scores, and attendance records.

Dependent Variable: Student academic performance.

Covariate Variable: Socio-economic status.

Grades will be measured on a numerical scale (e.g., GPA), test scores will be measured as

numerical scores (e.g., standardized test scores), attendance records will be measured as the

percentage of days attended, and student academic performance will be measured using

standardized assessments or GPA.

1.4.2 Theoretical Foundation

The theoretical foundation of this study is grounded in Social Cognitive Theory (SCT), proposed

by Albert Bandura (1986). SCT posits that individuals learn through observation, imitation, and

modeling, emphasizing the reciprocal interaction between personal factors, environmental

influences, and behavior (Bandura, 1986). In the context of academic performance, SCT suggests

that students' beliefs in their capabilities (self-efficacy), environmental factors such as family

support and school resources, and behaviors such as study habits and attendance contribute to

their academic success (Bembenutty et al., 2016).

8
This study aligns with SCT as it investigates how external factors (grades, test scores,

attendance) influence students' academic performance, reflecting the interplay between personal,

environmental, and behavioral determinants of learning outcomes. By employing supervised

learning techniques to analyze historical academic data, the study seeks to elucidate the

predictive relationships between these variables, thereby contributing to our understanding of

how environmental factors shape academic achievement. Furthermore, by considering socio-

economic status as a covariate variable, the study acknowledges the socio-cultural context in

which learning occurs, echoing SCT's emphasis on the influence of environmental factors on

behavior and cognition.

1.5 Significance

By applying supervised learning techniques to predict student academic performance, this study

contributes to the advancement of educational theory, specifically in the areas of predictive

analytics and personalized learning. The findings of this research have the potential to enrich our

understanding of the complex interactions between student characteristics, environmental

factors, and academic outcomes, thus enhancing existing theoretical frameworks such as Social

Cognitive Theory and the Student Success Framework. From a practical standpoint, the insights

gained from this study can inform evidence-based decision-making and intervention strategies in

educational settings. Educators and administrators can utilize predictive models to identify

students at risk of academic underperformance early on and implement targeted interventions to

support their learning needs. Additionally, the development of a predictive analytics framework

tailored to the context of Virginia Run Elementary School can serve as a model for other

educational institutions seeking to leverage data-driven approaches to enhance student success.

9
This study addresses a significant gap in the current research literature by focusing on the

practical application of predictive analytics in predicting student academic performance at the

elementary school level. While existing studies have explored predictive modeling in higher

education contexts, there is limited research examining its effectiveness in elementary schools.

By filling this gap, the study provides valuable insights into the feasibility and efficacy of

employing supervised learning techniques to support early intervention efforts and promote

academic achievement among elementary school students.

1.5.1 Assumptions

Several assumptions underlie this study, which are critical to its meaningfulness and validity.

Firstly, it is assumed that the historical academic data collected from Virginia Run Elementary

School is accurate, reliable, and representative of students' academic performance. While efforts

will be made to ensure data quality and integrity through rigorous data cleaning and validation

processes, the accuracy of the data cannot be guaranteed. Secondly, the study assumes that the

predictive models developed using supervised learning techniques will generalize well to new

data and maintain their predictive accuracy over time. Although cross-validation techniques and

model evaluation metrics will be employed to assess the robustness of the models, there is

inherent uncertainty in their performance on unseen data. Lastly, it is assumed that the

implementation of proactive intervention strategies based on predictive insights will lead to

improvements in student academic performance. While theoretical frameworks and empirical

evidence support the efficacy of targeted interventions, the actual effectiveness of these

interventions may vary depending on various contextual factors and implementation fidelity.

1.5.2 Limitations

10
Despite the rigorous methodology employed in this study, several limitations warrant

consideration. Internal validity refers to the degree to which the observed effects can be

attributed to the independent variables rather than confounding factors or measurement errors

(Andrade, 2018). One potential limitation is the inability to control for all possible confounding

variables that may influence student academic performance, such as teacher quality, family

dynamics, and individual student characteristics. To mitigate this limitation, statistical techniques

such as multivariate analysis and propensity score matching will be employed to account for

known confounders and minimize their impact on the study outcomes (Andrade, 2018). External

validity pertains to the generalizability of study findings to other populations, settings, and

contexts (Andrade, 2018). The generalizability of the study findings may be limited by the

specific characteristics of Virginia Run Elementary School and its student population. To

enhance external validity, efforts will be made to ensure the representativeness of the sample and

the transparency of the research methodology, allowing other researchers to replicate the study in

different educational settings (Andrade, 2018).

Construct validity refers to the degree to which the operationalizations of the variables accurately

measure the theoretical constructs they are intended to represent (Westen & Rosenthal, 2003).

While measures of academic performance, such as grades and test scores, are commonly used

indicators, they may not fully capture the multidimensional nature of student achievement. To

address this limitation, multiple measures of academic performance will be utilized, including

standardized assessments, teacher evaluations, and longitudinal academic records, to provide a

more comprehensive assessment of student achievement. Biases inherent in the data collection

and analysis process may influence study outcomes. For instance, selection bias may arise if

certain groups of students are disproportionately included or excluded from the analysis

11
(Pannucci & Wilkins, 2011). To mitigate selection bias, random sampling techniques will be

employed to ensure the representativeness of the study sample. Additionally, researcher bias may

occur if subjective judgments or preconceived notions influence data interpretation (Baldwin et

al., 2022). To minimize researcher bias, inter-rater reliability checks and independent validation

of data coding will be conducted to enhance the reliability and objectivity of the study findings

(Baldwin et al., 2022).

To address these limitations, transparency and rigor will be prioritized throughout the research

process. Detailed documentation of the study methodology, data collection procedures, and

analytical techniques will be provided to enhance the transparency and reproducibility of the

study findings. Additionally, sensitivity analyses and robustness checks will be conducted to

assess the robustness of the results to different methodological assumptions and data

specifications. By acknowledging and addressing these limitations, this study aims to enhance

the validity and reliability of its findings, thereby contributing to the advancement of knowledge

in the field of educational research.

1.5.3 Definition of Terms

Confounding Variables: Confounding variables are extraneous variables that may influence the

relationship between the independent and dependent variables, leading to spurious or misleading

associations. In quantitative research, confounding variables are sources of bias that can distort

the true effect of the independent variable on the dependent variable. Controlling for

confounding variables is essential to establish causal relationships and ensure the internal

validity of the study (Pearl, 2010).

12
Construct Validity: Construct validity refers to the extent to which a measurement accurately

assesses the theoretical construct or concept it intends to measure. It involves demonstrating that

the operationalization of the construct aligns with its theoretical definition and that the

measurement instrument effectively captures the intended construct. Establishing construct

validity is essential for ensuring the meaningfulness and relevance of research findings. (Westen

& Rosenthal, 2003).

Predictive Analytics: Predictive analytics refers to the process of using historical data, statistical

algorithms, and machine learning techniques to forecast future events or outcomes. In the context

of this study, predictive analytics involves analyzing past academic performance data, such as

grades, test scores, and attendance records, to predict students' future academic performance.

(Al-Tameemi et al., 2020).

Supervised Learning: Supervised learning is a machine learning paradigm where the algorithm

learns from labeled training data, with each example being a pair consisting of an input object

(typically a vector) and a desired output value (also called the supervisory signal). The algorithm

then generalizes from the training data to make predictions or decisions about unseen data. In

this study, supervised learning techniques, such as decision trees, support vector machines, and

neural networks, are utilized to develop predictive models based on past academic performance

data (Kotsiantis et al., 2007).

Validity: Validity refers to the extent to which a research study measures what it claims to

measure. In quantitative research, validity encompasses various aspects, including internal

validity, external validity, and construct validity. Internal validity relates to the accuracy of the

causal inferences drawn within the study, external validity pertains to the generalizability of the

study findings to other populations or contexts, and construct validity concerns the extent to

13
which the study accurately measures the theoretical constructs under investigation. Ensuring

validity is crucial for establishing the credibility and trustworthiness of research findings (Patino

& Ferreira, 2018).

1.6 Organization of the Remaining Chapters

In Chapter 2, the literature review will delve into the existing research and scholarly works

relevant to predictive analytics in education, supervised learning techniques, and their

application in forecasting student academic performance. The chapter will synthesize key

findings from recent studies, theories, and frameworks related to the predictive modeling of

student outcomes. It will explore theoretical foundations, methodological approaches, and

practical implications of predictive analytics in educational settings. Additionally, the literature

review will identify gaps, inconsistencies, and areas for further investigation, setting the stage for

the current study's research questions and objectives. Chapter 3 will provide a detailed overview

of the methodology employed in conducting the study. It will outline the research design,

sampling procedures, data collection methods, and operationalization of variables. The chapter

will describe the specific machine learning algorithms and techniques utilized for predictive

modeling, including data preprocessing, model selection, and evaluation metrics. Additionally, it

will discuss ethical considerations, potential biases, and limitations inherent in the methodology,

along with strategies to mitigate these challenges. The methodology chapter aims to provide a

transparent and systematic framework for replicating the study and ensuring the validity and

reliability of the findings.

In Chapter 4, the focus will be on the data analysis process, where the predictive models

developed in Chapter 3 will be applied to the collected data from Virginia Run Elementary

School. The chapter will present descriptive statistics, model performance metrics, and

14
interpretation of results, highlighting significant findings and patterns identified through the

analysis. It will discuss the implications of the findings for theory, practice, and future research,

addressing the research questions posed in Chapter 1. Additionally, the chapter may include

visualizations, tables, and graphs to enhance the presentation of results and facilitate

comprehension. Chapter 5 will serve as the culmination of the dissertation, providing a

comprehensive summary of the study's objectives, methodology, findings, and implications. The

chapter will revisit the research questions and hypotheses, summarizing the key findings and

contributions of the study to the field of education and predictive analytics. It will discuss the

theoretical and practical implications of the findings, as well as their limitations and areas for

future research. Finally, the chapter will conclude with reflections on the significance of the

study and recommendations for educators, policymakers, and researchers interested in leveraging

predictive analytics to support student success.

15
References

Akter, S., Dwivedi, Y. K., Sajib, S., Biswas, K., Bandara, R. J., & Michael, K. (2022).

Algorithmic bias in machine learning-based marketing models. Journal of Business

Research, 144(144), 201–216. https://doi.org/10.1016/j.jbusres.2022.01.083

Al-Tameemi, G., Xue, J., Ajit, S., Kanakis, T., & Hadi, I. (2020, June 1). Predictive Learning

Analytics in Higher Education: Factors, Methods and Challenges. IEEE Xplore.

https://doi.org/10.1109/ICACCE49060.2020.9154946

Andrade, C. (2018). Internal, external, and ecological validity in research design, conduct, and

evaluation. Indian Journal of Psychological Medicine, 40(5), 498–499.

https://doi.org/10.4103/IJPSYM.IJPSYM_334_18

Baker, R. S. J. d., Corbett, A. T., Roll, I., & Koedinger, K. R. (2008). Developing a generalizable

detector of when students game the system. User Modeling and User-Adapted

Interaction, 18(3), 287–314. https://doi.org/10.1007/s11257-007-9045-6

Baldwin, J. R., Pingault, J.-B., Schoeler, T., Sallis, H. M., & Munafò, M. R. (2022). Protecting

against researcher bias in secondary data analysis: Challenges and potential solutions.

European Journal of Epidemiology, 37(1), 1–10. https://doi.org/10.1007/s10654-021-

00839-0

Bandura, A. (1986). Social foundations of thought and action: A social cognitive theory.

Prentice-Hall.

Bembenutty, H., White, M. C., & DiBenedetto, M. K. (2016). Applying Social Cognitive Theory

in the Development of Self-Regulated Competencies Throughout K-12 Grades. The

16
Springer Series on Human Exceptionality, 215–239. https://doi.org/10.1007/978-3-319-

28606-8_9

Cline, N. (2023, September 8). Despite post-pandemic efforts, Virginia student test scores remain

mostly flat. Virginia Mercury. https://virginiamercury.com/2023/09/07/despite-post-

pandemic-efforts-virginia-student-test-scores-remain-mostly-flat/

Dumont, H., & Ready, D. D. (2023). On the promise of personalized learning for educational

equity. Npj Science of Learning, 8(1). https://doi.org/10.1038/s41539-023-00174-x

El Aouifi, H., El Hajji, M., Es-Saady, Y., & Douzi, H. (2021). Predicting learner’s performance

through video sequences viewing behavior analysis using educational data-mining.

Education and Information Technologies, 26(5), 5799–5814.

https://doi.org/10.1007/s10639-021-10512-4

Kotsiantis, S. B., Zaharakis, I. D., & Pintelas, P. E. (2007). Machine learning: a review of

classification and combining techniques. Artificial Intelligence Review, 26(3), 159–190.

https://doi.org/10.1007/s10462-007-9052-3

Maraza-Quispe, B., Valderrama-Chauca, E. D., Cari-Mogrovejo, L. H., & Apaza-Huanca, J. M.

(2021). Predictive Model of Student Academic Performance from LMS data based on

Learning Analytics. ICETC ’21: Proceedings of the 13th International Conference on

Education Technology and Computers, 13–19. https://doi.org/10.1145/3498765.3498768

Nunn, S., Avella, J. T., Kanai, T., & Kebritchi, M. (2016). Learning Analytics Methods, Benefits,

and Challenges in Higher Education: A Systematic Literature Review. Online Learning,

20(2). https://doi.org/10.24059/olj.v20i2.790

17
Pannucci, C. J., & Wilkins, E. G. (2011). Identifying and Avoiding Bias in Research. Plastic and

Reconstructive Surgery, 126(2), 619–625.

https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2917255/

Patino, C. M., & Ferreira, J. C. (2018). Internal and external validity: Can you apply research

study results to your patients? Jornal Brasileiro de Pneumologia, 44(3), 183–183.

https://doi.org/10.1590/S1806-37562018000000164

Pearl, J. (2010). Confounding, and Collapsibility In Causality: Models, Reasoning and Inference.

Cambridge University Press.

Picciano, A. G. (2012). The Evolution of Big Data and Learning Analytics in American Higher

Education. Online Learning, 16(3). https://doi.org/10.24059/olj.v16i3.267

Reber, R., Canning, E. A., & Harackiewicz, J. M. (2018). Personalized Education to Increase

Interest. Current Directions in Psychological Science, 27(6), 449–454.

https://doi.org/10.1177/0963721418793140

Romero, C., Ventura, S., & García, E. (2008). Data mining in course management systems:

Moodle case study and tutorial. Computers & Education, 51(1), 368–384.

https://doi.org/10.1016/j.compedu.2007.05.016

Sanchez-Pozo, N. N., Mejia-Ordonez, J. S., Chamorro, D. C., Mayorca-Torres, D., & Peluffo-

Ordonez, D. H. (2021). Predicting High School Students’ Academic Performance: A

Comparative Study of Supervised Machine Learning Techniques. 2021 Machine

Learning-Driven Digital Technologies for Educational Innovation Workshop.

https://doi.org/10.1109/ieeeconf53024.2021.9733756

18
Sarker, I. H. (2021). Machine Learning: Algorithms, Real-World Applications and Research

Directions. SN Computer Science, 2(3), 1–21. Springer. https://doi.org/10.1007/s42979-

021-00592-x

Sghir, N., Adadi, A., & Lahmer, M. (2022). Recent advances in Predictive Learning Analytics: A

decade systematic review (2012–2022). Education and Information Technologies, 28.

https://doi.org/10.1007/s10639-022-11536-0

Syed, D. (2023). Predictive Analysis of Students’ Learning Performance Using Data Mining

Techniques: A Comparative Study of Feature Selection Methods. Applied System

Innovation, 6(5), 86–86. https://doi.org/10.3390/asi6050086

Wang, Z., Liu, K., Li, J., Zhu, Y., & Zhang, Y. (2019). Various Frameworks and Libraries of

Machine Learning and Deep Learning: A Survey. Archives of Computational Methods in

Engineering, 31(9). https://doi.org/10.1007/s11831-018-09312-w

Westen, D., & Rosenthal, R. (2003). Quantifying construct validity: Two simple measures.

Journal of Personality and Social Psychology, 84(3), 608–618.

https://doi.org/10.1037/0022-3514.84.3.608

Yağcı, M. (2022). Educational data mining: prediction of students’ academic performance using

machine learning algorithms. Smart Learning Environments, 9(1).

https://doi.org/10.1186/s40561-022-00192-z

Yang, J., Soltan, A. A. S., Eyre, D. W., & Clifton, D. A. (2023). Algorithmic fairness and bias

mitigation for clinical machine learning with deep reinforcement learning. Nature

Machine Intelligence, 5(8), 884–894. https://doi.org/10.1038/s42256-023-00697-3

19
Zhang, L., Basham, J. D., & Yang, S. (2020). Understanding the implementation of personalized

learning: A research synthesis. Educational Research Review, 31, 100339.

https://doi.org/10.1016/j.edurev.2020.100339

20

You might also like