3,468 results on '"ALGORITHMS"'
Search Results
2. The Effects of Unified School Enrollment Systems on New Orleans Schools: Enrollment, Demographics, and Outcomes after the Transition to OneApp. Technical Report
- Author
-
National Center for Research on Education Access and Choice (REACH), Jane Arnold Lincove, and Jon Valant
- Abstract
Unified enrollment (UE) systems were designed to improve efficiency, equity, and transparency in school choice processes, but research has focused on efficiency gains. This study examines whether moving from decentralized enrollment processes to UE mitigates or exacerbates racial segregation that often occurs in choice systems. Specifically, we examine a subset of charter schools in New Orleans that had enrolled disproportionately high numbers of white students prior to entering UE. We find that UE entry was associated with increased enrollment of nonwhite students in these schools without offsetting declines in white enrollment, facilitated by schools also increasing total enrollment after entering UE. We find no meaningful impacts of UE on school accountability measures, student or teacher mobility, or student discipline.
- Published
- 2023
3. Setting Priorities in School Choice Enrollment Systems: Who Benefits from Placement Algorithm Preferences? Technical Report
- Author
-
National Center for Research on Education Access and Choice (REACH), Education Research Alliance for New Orleans (ERA), Jon Valant, and Brigham Walker
- Abstract
Many cities with school choice programs employ algorithms to make school placements. These algorithms use student priorities to determine which applicants get seats in oversubscribed schools. This study explores whether the New Orleans placement algorithm tends to favor students of certain races or socioeconomic classes. Specifically, we examine cases where families of Black and White or poor and non-poor children request the same elementary school as their top choice. We find that when Black and White applicants submit the same first-choice request for kindergarten, Black applicants are 9 percentage points less likely to receive it. Meanwhile, students in poverty are 6 percentage points less likely to receive a first-choice placement than other applicants for the same kindergarten program. However, these biases are not inevitable. In non-entry grades, where placement policies favor students whose schools are closing, Black and low-income applicants are more likely to obtain first-choice placements than their peers. We examine these priorities and simulate placements under alternate specifications of a deferred-acceptance algorithm to assess the potential of algorithm reform as a policymaking tool.
- Published
- 2023
4. Setting Priorities in School Choice Enrollment Systems: Who Benefits from Placement Algorithm Preferences? Policy Brief
- Author
-
National Center for Research on Education Access and Choice (REACH), Education Research Alliance for New Orleans (ERA), Jon Valant, and Brigham Walker
- Abstract
Many U.S. cities with school choice programs have adopted unified enrollment systems to manage their application and placement processes centrally. Typically, these systems use placement algorithms to assign students to schools. These algorithms make placements based on families' rank-ordered requests, seat availability in schools, and various priorities and lottery numbers that determine students' standing at each school. This study examines the placement algorithm--and broader school request, placement, and enrollment patterns--in New Orleans, which has a citywide system of charter schools. The authors explore whether the priority categories in the New Orleans placement algorithm tend to favor students of certain races or socioeconomic classes. Specifically, the authors examine cases where families of Black and white children, or lower-income and higher-income families, submit the same first-choice requests for kindergarten (a key entry grade for elementary school). In addition to examining whether certain groups of students are more likely than others to get school placements when they vie for the same seats, the authors run simulations to assess how placement patterns might differ with different policies.
- Published
- 2023
5. Content Proximity Spring 2022 Pilot Study. Research Brief
- Author
-
NWEA, Meyer, J. Patrick, Hu, Ann, and Li, Sylvia
- Abstract
The Content Proximity Project was designed to improve the content validity of the MAP® Growth™ assessments while retaining the ability for the test to adapt off-grade and meet students wherever they are in their learning. Two main features of the project were the development of an enhanced item selection algorithm, and a spring pilot study conducted in volunteer school districts. The purpose of the pilot study was to evaluate the new algorithm during live testing, study the comparability of scores with traditional MAP Growth assessments, and produce evidence of test content validity and score reliability. The pilot study began in spring 2022 with a group of NWEA Partners who volunteered to participate. The Content Proximity Project was initiated with several benefits in mind. The primary benefits are enhanced content validity, improved perceptions of test quality, and greater test taking engagement. The test will continue to adapt off grade when needed to deliver items of suitable difficulty for a student. However, this adaptation will be done in such a way that test events will be more closely aligned with grade-level content, especially for students exhibiting typical performance for a grade. The stronger preference for grade-level content means that the test more closely matches the subject matter students have an opportunity to learn in school. Subsequently, MAP Growth scores should allow for better connections to curriculum materials and resources, and produce scores that are more highly correlated with end-of-year summative tests.
- Published
- 2023
6. Automated Pipeline for Multi-Lingual Automated Essay Scoring with ReaderBench
- Author
-
Stefan Ruseti, Ionut Paraschiv, Mihai Dascalu, and Danielle S. McNamara
- Abstract
Automated Essay Scoring (AES) is a well-studied problem in Natural Language Processing applied in education. Solutions vary from handcrafted linguistic features to large Transformer-based models, implying a significant effort in feature extraction and model implementation. We introduce a novel Automated Machine Learning (AutoML) pipeline integrated into the ReaderBench platform designed to simplify the process of training AES models by automating both feature extraction and architecture tuning for any multilingual dataset uploaded by the user. The dataset must contain a list of texts, each with potentially multiple annotations, either scores or labels. The platform includes traditional ML models relying on linguistic features and a hybrid approach combining Transformer-based architectures with the previous features. Our method was evaluated on three publicly available datasets in three different languages (English, Portuguese, and French) and compared with the best currently published results on these datasets. Our automated approach achieved comparable results to state-of-the-art models on two datasets, while it obtained the best performance on the third corpus in Portuguese. [This is the online first version of an article published in "International Journal of Artificial Intelligence in Education."]
- Published
- 2024
- Full Text
- View/download PDF
7. Variational Estimation for Multidimensional Generalized Partial Credit Model
- Author
-
Chengyu Cui, Chun Wang, and Gongjun Xu
- Abstract
Multidimensional item response theory (MIRT) models have generated increasing interest in the psychometrics literature. Efficient approaches for estimating MIRT models with dichotomous responses have been developed, but constructing an equally efficient and robust algorithm for polytomous models has received limited attention. To address this gap, this paper presents a novel Gaussian variational estimation algorithm for the multidimensional generalized partial credit model (MGPCM). The proposed algorithm demonstrates both fast and accurate performance, as illustrated through a series of simulation studies and two real data analyses. [This is the online version of an article published in "Psychometrika."]
- Published
- 2024
- Full Text
- View/download PDF
8. Rise of the Machines: Navigating the Opportunities and Challenges of AI-Assisted Research and Learning
- Author
-
Justin K. Dimmel and Izge Bayyurt
- Abstract
This commentary was written by ChatGPT, an artificial intelligence language model developed by OpenAI. It was conceived by the first author as a test for how the advent of predictive language modeling will create opportunities and challenges for researchers and teachers in mathematics education. The paper consists of a commentary that was written by ChatGPT, followed by a reflection written by the authors that explains how the model was prompted to generate the text and how we worked with ChatGPT to validate and edit the text that was produced. We consider the implications of models like ChatGPT on the future of academic work. [For the complete proceedings, see ED658295.]
- Published
- 2023
9. Towards Scalable Adaptive Learning with Graph Neural Networks and Reinforcement Learning
- Author
-
Vassoyan, Jean and Vie, Jill-Jênn
- Abstract
Adaptive learning is an area of educational technology that consists in delivering personalized learning experiences to address the unique needs of each learner. An important subfield of adaptive learning is learning path personalization: it aims at designing systems that recommend sequences of educational activities to maximize students' learning outcomes. Many machine learning approaches have already demonstrated significant results in a variety of contexts related to learning path personalization. However, most of them were designed for very specific settings and are not very reusable. This is accentuated by the fact that they often rely on non-scalable models, which are unable to integrate new elements after being trained on a specific set of educational resources. In this paper, we introduce a flexible and scalable approach towards the problem of learning path personalization, which we formalize as a reinforcement learning problem. Our model is a sequential recommender system based on a graph neural network, which we evaluate on a population of simulated learners. Our results demonstrate that it can learn to make good recommendations in the small-data regime. [For the complete proceedings, see ED630829.]
- Published
- 2023
10. To Speak or Not to Speak, and What to Speak, When Doing Task Actions Collaboratively
- Author
-
Nasir, Jauwairia, Kothiyal, Aditi, Sheng, Haoyu, and Dillenbourg, Pierre
- Abstract
Transactive discussion during collaborative learning is crucial for building on each other's reasoning and developing problem solving strategies. In a tabletop collaborative learning activity, student actions on the interface can drive their thinking and be used to ground discussions, thus affecting their problem-solving performance and learning. However, it is not clear how the interplay of actions and discussions, for instance, how students performing actions or pausing actions while discussing, is related to their learning. In this paper, we seek to understand how the transactivity of actions and discussions is associated with learning. Specifically, we ask what is the relationship between discussion and actions, and how it is different between those who learn (gainers) and those who do not (non-gainers). We present a combined differential sequence mining and content analysis approach to examine this relationship, which we applied on the data from 32 teams collaborating on a problem designed to help them learn concepts of minimum spanning trees. We found that discussion and action occur concurrently more frequently among gainers than non-gainers. Further we find that gainers tend to do more reflective actions along with discussion, such as looking at their previous solutions, than non-gainers. Finally, gainers discussion consists more of goal clarification, reflection on past solutions and agreement on future actions than non-gainers, who do not share their ideas and cannot agree on next steps. Thus this approach helps us identify how the interplay of actions and discussion could lead to learning, and the findings offer guidelines to teachers and instructional designers regarding indicators of productive collaborative learning, and when and how, they should intervene to improve learning. Concretely, the results suggest that teachers should support elaborative, reflective and planning discussions along with reflective actions. [For the complete proceedings, see ED630829.]
- Published
- 2023
11. Optimizing Parameters for Accurate Position Data Mining in Diverse Classrooms Layouts
- Author
-
Shou, Tianze, Borchers, Conrad, Karumbaiah, Shamya, and Aleven, Vincent
- Abstract
Spatial analytics receive increased attention in educational data mining. A critical issue in stop detection (i.e., the automatic extraction of timestamped and located stops in the movement of individuals) is a lack of validation of stop accuracy to represent phenomena of interest. Next to a radius that an actor does not exceed for a certain duration to establish a stop, this study presents a reproducible procedure to optimize a range parameter for K-12 classrooms where students sitting within a certain vicinity of an inferred stop are tagged as being visited. This extension is motivated by adapting parameters to infer teacher visits (i.e., on-task and off-task conversations between the teacher and one or more students) in an intelligent tutoring system classroom with a dense layout. We evaluate the accuracy of our algorithm and highlight a tradeoff between precision and recall in teacher visit detection, which favors recall. We recommend that future research adjust their parameter search based on stop detection precision thresholds. This adjustment led to better cross-validation accuracy than maximizing parameters for an average of precision and recall (F1 = 0.18 compared to 0.09). As stop sample size shrinks with higher precision cutoffs, thresholds can be informed by ensuring sufficient statistical power in offline analyses. We share avenues for future research to refine our procedure further. Detecting teacher visits may benefit from additional spatial features (e.g., teacher movement trajectory) and can facilitate studying the interplay of teacher behavior and student learning. [For the complete proceedings, see ED630829.]
- Published
- 2023
12. Variational Temporal IRT: Fast, Accurate, and Explainable Inference of Dynamic Learner Proficiency
- Author
-
Kim, Yunsung, Sreechan, Piech, Chris, and Thille, Candace
- Abstract
Dynamic Item Response Models extend the standard Item Response Theory (IRT) to capture temporal dynamics in learner ability. While these models have the potential to allow instructional systems to actively monitor the evolution of learner proficiency in real time, existing dynamic item response models rely on expensive inference algorithms that scale poorly to massive datasets. In this work, we propose Variational Temporal IRT (VTIRT) for fast and accurate inference of dynamic learner proficiency. VTIRT offers orders of magnitude speedup in inference runtime while still providing accurate inference. Moreover, the proposed algorithm is intrinsically interpretable by virtue of its modular design. When applied to 9 real student datasets, VTIRT consistently yields improvements in predicting future learner performance over other learner proficiency models. [For the complete proceedings, see ED630829.]
- Published
- 2023
13. Clustering to Define Interview Participants for Analyzing Student Feedback: A Case of Legends of Learning
- Author
-
Karimov, Ayaz, Saarela, Mirka, and Kärkkäinen, Tommi
- Abstract
Within the last decade, different educational data mining techniques, particularly quantitative methods such as clustering, and regression analysis are widely used to analyze the data from educational games. In this research, we implemented a quantitative data mining technique (clustering) to further investigate students' feedback. Students played educational games within a week on the educational games platform, Legends of Learning and after a week, we asked them to fulfill the feedback survey about their feelings on the use of this platform. To analyze the collected data from students, firstly, we prepared clusters and selected one prototype student closest to the centroid of each cluster to interview. Interviews were held to explain the clusters more and due to time and resource limitations, we were unable to interview all (N=60) students, thus only the most representative students were interviewed. In addition to the students, we conducted an interview with the teacher as well to get her detailed feedback and observations on the usage of educational games. We also asked students to take an exam before and after the research to see the impact of games on their grades. Our results depict that though educational games can increase students' motivation, they may negatively impact some students' grades. And even though playing games made students feel interested and fun, they would not like to play them on a daily basis. Hence, using educational games for a certain duration such as subject revision weeks may positively influence students' grades and motivation. [For the complete proceedings, see ED630829.]
- Published
- 2023
14. Is Your Model 'MADD'? A Novel Metric to Evaluate Algorithmic Fairness for Predictive Student Models
- Author
-
Verger, Mélina, Lallé, Sébastien, Bouchet, François, and Luengo, Vanda
- Abstract
Predictive student models are increasingly used in learning environments due to their ability to enhance educational outcomes and support stakeholders in making informed decisions. However, predictive models can be biased and produce unfair outcomes, leading to potential discrimination against some students and possible harmful long-term implications. This has prompted research on fairness metrics meant to capture and quantify such biases. Nonetheless, so far, existing fairness metrics used in education are predictive performance-oriented, focusing on assessing biased outcomes across groups of students, without considering the behaviors of the models nor the severity of the biases in the outcomes. Therefore, we propose a novel metric, the Model Absolute Density Distance (MADD), to analyze models' discriminatory behaviors independently from their predictive performance. We also provide a complementary visualization-based analysis to enable fine-grained human assessment of how the models discriminate between groups of students. We evaluate our approach on the common task of predicting student success in online courses, using several common predictive classification models on an open educational dataset. We also compare our metric to the only predictive performance-oriented fairness metric developed in education, ABROCA. Results on this dataset show that: (1) fair predictive performance does not guarantee fair models' behaviors and thus fair outcomes; (2) there is no direct relationship between data bias and predictive performance bias nor discriminatory behaviors bias; and (3) trained on the same data, models exhibit different discriminatory behaviors, according to different sensitive features too. We thus recommend using the MADD on models that show satisfying predictive performance, to gain a finer-grained understanding on how they behave and regarding who and to refine models selection and their usage. Altogether, this work contributes to advancing the research on fair student models in education. Source code and data are in open access at https://github.com/melinaverger/MADD. [For the complete proceedings, see ED630829.]
- Published
- 2023
15. Proceedings of the International Conference on Educational Data Mining (EDM) (16th, Bengaluru, India, July 11-14, 2023)
- Author
-
International Educational Data Mining Society, Feng, Mingyu, Käser, Tanja, and Talukdar, Partha
- Abstract
The Indian Institute of Science is proud to host the fully in-person sixteenth iteration of the International Conference on Educational Data Mining (EDM) during July 11-14, 2023. EDM is the annual flagship conference of the International Educational Data Mining Society. The theme of this year's conference is "Educational data mining for amplifying human potential." Not all students or seekers of knowledge receive the education necessary to help them realize their full potential, be it due to a lack of resources or lack of access to high quality teaching. The dearth in high-quality educational content, teaching aids, and methodologies, and non-availability of objective feedback on how they could become better teachers, deprive our teachers from achieving their full potential. The administrators and policy makers lack tools for making optimal decisions such as optimal class sizes, class composition, and course sequencing. All these handicap the nations, particularly the economically emergent ones, who recognize the centrality of education for their growth. EDM-2023 has striven to focus on concepts, principles, and techniques mined from educational data for amplifying the potential of all the stakeholders in the education system. The spotlights of EDM-2023 include: (1) Five keynote talks by outstanding researchers of eminence; (2) A plenary Test of Time award talk and a Banquet talk; (3) Five tutorials (foundational as well as advanced); (4) Four thought provoking panels on contemporary themes; (5) Peer reviewed technical paper and poster presentations; (6) Doctoral students consortium; and (7) An enchanting cultural programme. [Individual papers are indexed in ERIC.]
- Published
- 2023
16. Human versus Machine: Do College Advisors Outperform a Machine-Learning Algorithm in Predicting Student Enrollment? EdWorkingPaper No. 23-699
- Author
-
Annenberg Institute for School Reform at Brown University, Akmanchi, Suchitra, Bird, Kelli A., and Castleman, Benjamin L.
- Abstract
Prediction algorithms are used across public policy domains to aid in the identification of at-risk individuals and guide service provision or resource allocation. While growing research has investigated concerns of algorithmic bias, much less research has compared algorithmically-driven targeting to the counterfactual: human prediction. We compare algorithmic and human predictions in the context of a national college advising program, focusing in particular on predicting high-achieving, lower-income students' college enrollment quality. College advisors slightly outperform a prediction algorithm; however, greater advisor accuracy is concentrated among students with whom advisors had more interactions. The algorithm achieved similar accuracy among students lower in the distribution of interactions, despite advisors having substantially more information. We find no evidence that the advisors or algorithm exhibit bias against vulnerable populations. Our results suggest that, especially at scale, algorithms have the potential to provide efficient, accurate, and unbiased predictions to target scarce social services and resources.
- Published
- 2023
17. Evaluating Machine Learning for Projecting Completion Rates for VET Programs. Technical Paper
- Author
-
National Centre for Vocational Education Research (NCVER) (Australia), Hall, Michelle, Lees, Melinda, Serich, Cameron, and Hunt, Richard
- Abstract
This paper summarises exploratory analysis undertaken to evaluate the effectiveness of using machine learning approaches to calculate projected completion rates for vocational education and training (VET) programs, and compares this with the current approach used at the National Centre for Vocational Education Research (NCVER) -- Markov chains methodology. While the Markov chains methodology currently used by NCVER has demonstrated that it is reliable, with predictions aligning well with the actual rates of completion for historical estimates, it has not been reviewed for some time and it does have some limitations. The evaluation of machine learning techniques for predicting VET program completion rates was undertaken to overcome some of these limitations and with a view to improving our current predictions. This report includes: (1) an overview of the methodologies: Markov chains and two machine learning algorithms that were applied to predict completion rates for VET programs (XGBoost and CatBoost); (2) a comparison of the accuracy of the predictions generated by both methodologies; and (3) an evaluation of the relative strengths and limitations of both methodologies.
- Published
- 2023
18. Mining, Analyzing, and Modeling the Cognitive Strategies Students Use to Construct Higher Quality Causal Maps
- Author
-
Allan Jeong and Hyoung Seok-Shin
- Abstract
The Jeong (2020) study found that greater use of backward and depth-first processing was associated with higher scores on students' argument maps and that analysis of only the first five nodes students placed in their maps predicted map scores. This study utilized the jMAP tool and algorithms developed in the Jeong (2020) study to determine if the same processes produce higher-quality causal maps. This study analyzed the first five nodes that students (n = 37) placed in their causal maps to reveal that: 1) use of backward, forward, breadth-first, and depth-first processing produced maps of similar quality; and 2) backward processing had three times more impact on maps scores than depth-first processing to suggest that linking events into chains using backward chaining is one approach to constructing higher quality causal maps. These findings are compared with prior research findings and discussed in terms of noted differences in the task demands of constructing argument versus causal maps to gain insights into why, how, and when specific processes/strategies can be applied to create higher-quality causal maps and argument maps. These insights provide guidance on ways to develop diagramming and analytic tools that automate, analyze, and provide real-time support to improve the quality of students' maps, learning, understanding, and problem-solving skills. [For the full proceedings, see ED636095.]
- Published
- 2023
19. On the Predictors of Computational Thinking Self-Efficacy
- Author
-
Guggemos, Josef
- Abstract
Computational thinking (CT) is an important 21st-century skill. This paper aims at investigating predictors of CT self-efficacy among high-school students. The hypothesized predictors are grouped into three areas: (1) student characteristics, (2) home environment, and (3) learning opportunities. CT self-efficacy is measured with the Computational Thinking Scales (CTS) that comprises five dimensions: creativity, algorithmic thinking, cooperativity, critical thinking, and problem solving. N = 202 high-school students act as the sample, linear regression as the analysis method. The best prediction is possible for algorithmic thinking (R[superscript 2] = 0.511). For cooperativity, the explanatory power of our model it is weak (R[superscript 2] = 0.146). Across all five CTS dimensions, CT self-concept is the best predictor for CT self-efficacy. [For the full proceedings, see ED636095.]
- Published
- 2023
20. The Effects of Age and Learning with Educational Robotic Devices on Children's Algorithmic Thinking
- Author
-
Angeli, Charoula, Diakou, Panayiota, and Anastasiou, Vaso
- Abstract
Educational Robotics is increasingly used in elementary-school classrooms to develop students' algorithmic thinking and programming skills. However, most research appears descriptive and lacks experimental evidence on the effects of teaching interventions using robotics to develop algorithmic thinking. Using the robots Dash and Dot, this study examined algorithmic thinking development in groups of children aged 6, 9, and 12. The results showed a statistically significant main effect between the age of students and algorithmic thinking skills and a statistically significant main effect between intervention and algorithmic thinking. In conclusion, the findings underscore the necessity of providing learners with structured, scaffolded activities tailored to their age to effectively nurture algorithmic thinking skills when engaging in Dash and Dot activities. [For the full proceedings, see ED636095.]
- Published
- 2023
21. Maching Learning Based Financial Management Mobile Application to Enhance College Students' Financial Literacy
- Author
-
Mohsina Kamarudeen and K. Vijayalakshmi
- Abstract
This paper presents a mobile application aimed at enhancing the financial literacy of college students by monitoring their spending patterns and promoting better decision-making. The application is developed using the agile methodology with Android Studio and Flutter as development tools and Firebase as a database. The app is divided into sub-applications, with the home page serving as the program's integration point, displaying a summary of the user's financial progress. The app generates valuable insights into the user's current and future financial success, utilizing data analytics and machine learning to provide detailed and summary insights into the user's financial progress. The machine-learning algorithm used in this app is linear regression, which predicts the user's income and expenses for the upcoming month based on their historical spending data. In addition, the app highlights deals and student discounts in the user's vicinity and links to financial articles that promote better financial planning and decision-making. By promoting responsible spending habits and providing valuable financial insights, this mobile application aims to help students become financially literate and make informed financial decisions for future. [For the full proceedings, see ED654100.]
- Published
- 2023
22. Changing the Success Probability in Computerized Adaptive Testing: A Monte-Carlo Simultion on the Open Matrices Item Bank
- Author
-
Hanif Akhtar
- Abstract
For efficiency, Computerized Adaptive Test (CAT) algorithm selects items with the maximum information, typically with a 50% probability of being answered correctly. However, examinees may not be satisfied if they only correctly answer 50% of the items. Researchers discovered that changing the item selection algorithms to choose easier items (i.e., success probability > 50%), albeit not optimum from a measurement efficiency standpoint, would provide a better experience. The current study aims to investigate the impact of changing the success probability on measurement efficiency. A Monte-Carlo simulation was performed on the Open Matrices Item Bank and simulated item bank. A total of 1500 examinees were generated. We modified the item selection algorithm with the expected success probability of 60%, 70%, and 80%. Each examinee was assigned to five item selection methods: maximum-information, random, p=0.6, p=0.7, and p=0.8. The results indicated that traditional CAT was 60-70% shorter than random item selection. Altering the success probability did not affect the estimation of the examinee's ability. Increasing the probability of success in CAT increased the number of items required to achieve specified levels of precision. Practical considerations on how to maximize the trade-off between examinees' experiences and measurement efficiency are mentioned in the discussion. [For the full proceedings, see ED654100.]
- Published
- 2023
23. GPTZero vs. Text Tampering: The Battle That GPTZero Wins
- Author
-
David W. Brown and Dean Jensen
- Abstract
The growth of Artificial Intelligence (AI) chatbots has created a great deal of discussion in the education community. While many have gravitated towards the ability of these bots to make learning more interactive, others have grave concerns that student created essays, long used as a means of assessing the subject comprehension of students, may be at risk. The bot's ability to quickly create high quality papers, sometimes complete with reference material, has led to concern that these programs will make students too reliant on their ability and not develop the critical thinking skills necessary to succeed. The rise in these applications has led to the need for the development of detection programs that are able to read the students submitted work and return an accurate estimation of if the paper is human or computer created. These detection programs use natural language processing's (NLP) ideas of perplexity, or randomness of the text, and burstiness, or the tendency for certain words and phrases to appear together, plus sophisticated algorithms to compare the essays to preexisting literature to generate an accurate estimation on the likely author of the paper. The use of these systems has been found to be highly effective in reducing plagiarism among students, however concerns have been raised about the limitations of these systems. False positives, false negatives, and cross language identification are three areas of concern amongst faculty and have led to reduced usage of the detection engines. Despite the limitations however, these systems are a valuable tool for educational institutions to maintain academic integrity and ensure that students are submitting original work. [For the full proceedings, see ED656038.]
- Published
- 2023
24. Learning Latent and Hierarchical Structures in Cognitive Diagnosis Models
- Author
-
Chenchen Ma, Jing Ouyang, and Gongjun Xu
- Abstract
Cognitive Diagnosis Models (CDMs) are a special family of discrete latent variable models that are widely used in educational and psychological measurement. A key component of CDMs is the Q-matrix characterizing the dependence structure between the items and the latent attributes. Additionally, researchers also assume in many applications certain hierarchical structures among the latent attributes to characterize their dependence. In most CDM applications, the attribute--attribute hierarchical structures, the item-attribute Q-matrix, the item-level diagnostic models, as well as the number of latent attributes, need to be fully or partially pre-specified, which however may be subjective and misspecified as noted by many recent studies. This paper considers the problem of jointly learning these latent and hierarchical structures in CDMs from observed data with minimal model assumptions. Specifically, a penalized likelihood approach is proposed to select the number of attributes and estimate the latent and hierarchical structures simultaneously. An expectation-maximization (EM) algorithm is developed for efficient computation, and statistical consistency theory is also established under mild conditions. The good performance of the proposed method is illustrated by simulation studies and real data applications in educational assessment.
- Published
- 2023
- Full Text
- View/download PDF
25. The Relationship between Knowledge Production and Google in Framing and Reframing AI Imaginary. A Comparative Algorithmic Audit between the US and Italy
- Author
-
Natalia Stanusch
- Abstract
This study offers an analysis and comparison of search results from Google concerning the topic of Artificial Intelligence (AI) in two geographically and politically different contexts: the United States and Italy. As new AI systems, tools, and solutions are developed and implemented in each sector of human life on a global scale, certain imaginaries of AI are emerging. These imaginaries constitute the ground for the public understanding, support, and disapproval of certain AI technologies and regulations. As citizens turn into users, Google remains the dominant gatekeeper of information, thus becoming an influential actor in sharping AI imaginaries. The following analysis is a response to the criticism of Google's search results, considering Google as an essential producer and certifier of AI imaginaries for general public. The comparison of search queries conducted in this analysis shows that the sources which Google presents in its search results add to different types of AI imaginaries, consequently influencing public opinion in different, often asymmetrical, ways. [For the full proceedings, see ED654100.]
- Published
- 2023
26. Structuring Topics of Philippine Universities' Introductory Programming Courses Using Semi-Supervised Pairwise-Constrained Clustering to Synthesize Alternative Course Topic Outlines
- Author
-
Kleb Dale G. Bayaras
- Abstract
In course design, topic outline organization encompasses the structuring and sequencing of topics to be delivered in a learning environment. Recent studies in topic outline optimization revolve around massive open online courses (MOOCs) due to their abundance but not much has been studied on the traditional courses. This study investigates the organization of topic outlines in traditional introductory programming courses across Philippine higher education institutions (HEIs) and evaluates the viability of the synthesized alternative topic outlines. Course syllabi were collected from 16 HEIs. A topic precedence graph (TPG) model that provides a structured overview of the introductory programming was created via a semi-supervised pairwise constrained k-means (PCK-Means) clustering to structure the topics which produced 20 topic clusters with strong topic cohesion within the clusters. The TPG showed that HEIs tend to start the outline similarly, followed by core programming topics with varied sequences, and divergent ways of ending the outline. Two anomaly clusters were identified as having topic titles grouped that do not seem to have a unifying topic. Limitations of the clustering algorithm are identified where it cannot identify semantic meaning between words which may affect its applicability in situations where topic titles are named inconsistently. From the TPG, alternative optimal and comprehensive topic outlines were synthesized via greedy and DFS graph traversal algorithms. However, these alternative outlines performed very poorly when compared with the evaluators' (n=19) arrangement of topic outlines due to some prerequisite topics being discussed in the latter part already. Overall, this study introduces a method to incorporate computer science technologies in structuring topics across HEIs and aiding educators in topic outline design but more research is needed before it can be implemented in a real classroom setting.
- Published
- 2023
27. Blueprint for an AI Bill of Rights: Making Automated Systems Work for the American People
- Author
-
White House, Office of Science and Technology Policy (OSTP)
- Abstract
Among the great challenges posed to democracy today is the use of technology, data, and automated systems in ways that threaten the rights of the American public. Too often, these tools are used to limit our opportunities and prevent our access to critical resources or services. These problems are well documented. In America and around the world, systems supposed to help with patient care have proven unsafe, ineffective, or biased. Algorithms used in hiring and credit decisions have been found to reflect and reproduce existing unwanted inequities or embed new harmful bias and discrimination. Unchecked social media data collection has been used to threaten people's opportunities, undermine their privacy, or pervasively track their activity--often without their knowledge or consent. To advance President Biden's vision, the White House Office of Science and Technology Policy has identified five principles that should guide the design, use, and deployment of automated systems to protect the American public in the age of artificial intelligence. "The Blueprint for an AI Bill of Rights" is a guide for a society that protects all people from these threats--and uses technologies in ways that reinforce our highest values. Responding to the experiences of the American public, and informed by insights from researchers, technologists, advocates, journalists, and policymakers, this framework is accompanied by a technical companion--a handbook for anyone seeking to incorporate these protections into policy and practice, including detailed steps toward actualizing these principles in the technological design process. These principles help provide guidance whenever automated systems can meaningfully impact the public's rights, opportunities, or access to critical needs.
- Published
- 2022
28. Domain Specific Languages for Geometry Processing
- Author
-
Yong Li
- Abstract
Geometry processing holds a foundational position within the realm of computer graphics, with advancements in this field regularly published at SIGGRAPH annually. The journey from writing the paper to implementing the algorithms is a meticulous and error-prone process, demanding significant dedication and attention to detail. Authors frequently encounter challenges, including inadvertent typos within formulas that can introduce discrepancies between the paper and the actual code. This discrepancy can pose a significant hurdle for readers, especially new researchers and graduate students, aiming to reproduce the results. Even when authors release their code, readers may desire versions in their preferred programming languages. My dissertation focus is on mitigating challenges faced by researchers throughout scientific computing according to a suite of domain-specific languages (DSLs). The goal is to enable authors to easily try new research ideas and compose papers with these DSLs, automating the generation of algorithmic code across diverse backend languages like C++, Python, and MATLAB. I have developed three instrumental tools with my collaborators to handle those sections in papers: I[heart]LA for compiling the implementation, I[heart]MESH for compiling the discretization, and H[heart]rtDown for compiling linear algebra papers into interactive documents. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
29. Continual Learning with Language Models
- Author
-
Zixuan Ke
- Abstract
The essence of human intelligence lies in its ability to learn continuously, accumulating past knowledge to aid in future learning and problem-solving endeavors. In contrast, the current machine learning paradigm often operates in isolation, lacking the capacity for continual learning and adaptation. This deficiency becomes apparent in the face of rapidly evolving artificial intelligence (AI) technologies, particularly large language models (LLMs), where incremental training remains a challenge. Continual learning (CL), also known as lifelong learning, is indispensable for truly intelligent systems, especially in dynamic environments where constant adaptation is necessary. This dissertation explores recent advancements in continual learning algorithms within the framework of language models. We first introduce the settings, challenges, and general approaches of CL. We then delve into our efforts to achieve both catastrophic forgetting (CF) mitigation and knowledge transfer (KT), and how we apply CL to different stages of language model development, including pre-training and end-task adaptation. With the aid of continual learning, the performance of language models is greatly improved. Finally, we will discuss the opportunities for AI autonomy and open-world continual learning. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
30. Regularized Multivariate Functional Principal Component Analysis
- Author
-
Yue Zhao
- Abstract
Multivariate Functional Principal Component Analysis (MFPCA) is a valuable tool for exploring relationships and identifying shared patterns of variation in multivariate functional data. However, interpreting these functional principal components (PCs) can sometimes be challenging due to issues such as roughness and sparsity. In this dissertation, we establish the theoretical foundations of the penalized MFPCA problem within Hilbert space and propose three novel regularized MFPCA approaches. These approaches utilize eigen decomposition and singular value decomposition (SVD) techniques to enhance the performance of MFPCA by incorporating multiple penalty terms, such as roughness and sparsity penalties. In the first method, a roughness penalty is directly imposed on functional PCs, extending the eigen decomposition problem to a Hilbert space that specifically accounts for the roughness of the functions. A parameter vector is employed as a tuning parameter to regulate the smoothness of each functional variable. Additionally, this method allows for each functional variable to be smoothed on different domains, providing greater flexibility in handling diverse functional data. In the other two methods, we establish a mathematical foundation for penalized functional SVD to address the regularized MFPCA problem. Within the functional SVD framework, we propose iterative power algorithms that offer both the flexibility to assign unique tuning parameters for each functional PC and computational efficiency. Moreover, the functional SVD approach allows for the straightforward and simultaneous incorporation of various penalties, such as smoothing and sparsity, each serving a distinct purpose. Additionally, our functional SVD approach introduces an innovative form of sparsity within PC scores, which proves beneficial for obtaining more informative PCs. Similar to the first approach, these two methods also allow each functional variable to be defined on different domains, providing greater flexibility in data analysis. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
31. Gaussian Variational Estimation of MIRT and Its Applications in Large-Scale Assessments
- Author
-
Jiaying Xiao
- Abstract
Multidimensional Item Response Theory (MIRT) has been widely used in educational and psychological assessments. It estimates multiple constructs simultaneously and models the correlations among latent constructs. While it provides more accurate results, the unidimensional IRT model is still dominant in real applications. One major reason is that the parameter estimation is still challenging because of intractable multidimensional integrals of the likelihood, especially in high dimensions. Several algorithms have been proposed to address the issue, such as adaptive Gaussian quadrature methods, Laplace approximations, and stochastic methods. However, the state-of-the-art algorithms are still time-consuming, especially when the number of latent traits exceeds 5. Recently, the Gaussian variational Expectation Maximization (GVEM) algorithm (Cho et al., 2021) was proposed as an alternative for further improving computational efficiency and estimation accuracy. The general framework allows the closed-form solutions for the expectation-maximization process by introducing a variational lower bound of the likelihood function. Although prior studies have demonstrated the superiority of the GVEM algorithm over the widely used Metropolis-Hastings Robbins-Monro algorithm (MH-RM) under various conditions, its performance across diverse practical contexts remains relatively unexplored. For instance, there is an immense need for further investigation into the robustness of the GVEM framework across various missing data scenarios. Additionally, efforts should be directed towards devising methods for estimating standard errors within the GVEM framework. Moreover, the development of an R package to facilitate the application of the GVEM algorithm would significantly augment its accessibility and utility. The purpose of this dissertation is to extend the applicability of the GVEM algorithm and investigate its performance in diverse scenarios. In the second chapter, a modified GVEM algorithm was proposed by adding the bootstrap bias correction step and denoted it as GVEM-BS. A series of simulation studies and real data analysis were conducted to compare GVEM-BS to MH-RM in terms of estimation precision under different missing data scenarios and assessment designs. The results demonstrated the robustness and precision of GVEMBS in the context of high missing proportions, especially for missing at completely random conditions. When applying the two methods to different assessment designs, both GVEM-BS and MH-RM yielded comparable results. In the third chapter, an updated supplemented expectation maximization (USEM) method and a bootstrap method were proposed for GVEM-based SE estimation. These two methods were compared in terms of SE recovery accuracy. The simulation results demonstrated that the GVEM algorithm with bootstrap and item priors (GVEM-BSP) outperformed the other methods, exhibiting less bias and relative bias for SE estimates under most conditions. Although the GVEM with USEM (GVEM-USEM) was the computationally most efficient method, it yielded an upward bias for SE estimates. In the fourth chapter, an R package, VEMIRT, was introduced by offering users efficient computational tools tailored for high-dimensional data under the GVEM framework. This package facilitates both exploratory and confirmatory analyses through the utilization of GVEM models. Additionally, it enables users to compute standard errors of item parameters and implement corrections such as bootstrap sampling and importance sampling, thereby enhancing the accuracy of estimations. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
32. Effectiveness of Machine Learning Algorithms on Predicting Course Level Outcomes from Learning Management System Data
- Author
-
Michael Wade Ashby
- Abstract
Whether machine learning algorithms effectively predict college students' course outcomes using learning management system data is unknown. Identifying students who will have a poor outcome can help institutions plan future budgets and allocate resources to create interventions for underachieving students. Therefore, knowing the effectiveness of applying the algorithms to build models will be helpful in higher education institutions. This study utilizes the probably approximately correct learning theory, which posits machines can learn any concept as long as there is a data set of examples that labels the outcome of the concept, enough examples, and the computation can be completed in a polynomial number of steps. This quantitative comparative study compared four different machine learning algorithms' (naive Bayes, decision tree, neural network, and support vector machine) ability to predict the outcome of students in college courses by training models from learning management system data across two universities. It then measured the predictions of each model at a course level to determine their effectiveness. The results showed that the probably approximately correct learning theory works even in predicting course outcomes, as the decision tree successfully predicted students with poor outcomes with an F1 value above 0.5 and significantly better than the other three algorithms. Future studies can expand on the number of institutions involved in the contributing data and different learning management system data points as they may provide better predictions. Having learned the confidence of the decision tree's ability to predict students that will have poor outcomes, higher education institutions can better plan and allocate resources. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
33. The Future of AI Can Be Kind: Strategies for Embedded Ethics in AI Education
- Author
-
Yim Register
- Abstract
The field of Data Science has seen rapid growth over the past two decades, with a high demand for people with skills in data analytics, programming, statistics, and ability to visualize, predict from, and otherwise make sense of data. Alongside the rise of various artificial intelligence (AI) and machine learning (ML) applications, we have also witnessed egregious algorithmic biases and harms - from discriminatory outputs of models to reinforcing normative ideals about beauty, gender, race, class, etc. These harms range from high profile cases such as the racial bias embedded in the COMPAS recidivism algorithm, to more insidious cases of algorithmic harm that compound over time with re-traumatizing effects (such as the mental health impacts of recommender systems, social media content organization and the struggle for visibility, and discriminatory content moderation of marginalized individuals). There are various strategies to combat and repair algorithmic harms, ranging from algorithmic audits and fairness metrics to AI Ethics Standards put forth by major institutions and tech companies. However, there is evidence to suggest that current Data Science curricula do not adequately prepare future practitioners to effectively respond to issues of algorithmic harm, "especially" the day-to-day issues that practitioners are likely to face. Through a review of AI Ethics standards and the literature, I devise a set of 9 characterizations of effective AI ethics education: "specific", "prescriptivist", "action-centered", "relatable", "empathetic", "contextual", "expansive", "preventative", and "integrated." The empirical work of this dissertation reveals the value of embedding ethical critique into technical machine learning instruction - demonstrating how teaching AI concepts using cases of algorithmic harm can boost both technical comprehension and ethical considerations [397, 398]. I demonstrate the value of relying on real-world cases and experiences that students already have (such as with hiring/admissions decisions, social media algorithms, or generative AI tools) to boost their learning of both technical and social impact topics. I explore this relationship between personal relatability and experiential learning, demonstrating how to harness students' lived experiences to relate to cases of algorithmic harm and opportunities for repair. My preliminary work also reveals significant "in-group favoritism", suggesting students find AI errors more urgent when they personally relate to them. While this may prove beneficial for engaging underrepresented students in the classroom, it must be paired with empathy-building techniques for students who relate less to cases of algorithmic harm, as well as trauma-informed pedagogical practice. My results also revealed an over-reliance on "life-or-death reasoning" when it came to ethical decision-making, along with organizational and financial pressures that might impede AI professionals from delaying harmful software. This dissertation contributes several strategies to effectively prepare Data Scientists to consider both technical and social aspects of their work, along with empirical results suggesting the benefits of embedded ethics throughout all areas of AI education. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
34. The Influence of Computational Thinking on New York State Geometry Regents Proficiency Rates
- Author
-
Joseph Crifo
- Abstract
The present study was conducted to determine how implementing computational thinking (via a proxy in AP Computer Science Principles) into a school's curriculum impacted student proficiency rates on the New York State Geometry Regents. Recent research has suggested that computational thinking is a skill that transcends specific content areas and can influence student learning outcomes across multiple disciplines. By equipping students with these skills, each individual's zone of proximal development may increase, leading to increased learning efficiency. Given the rise of technology and the need for computational literacy, schools are looking to implement courses to help students develop these skills. The target school students were compared to their fellow general education peers in their home and neighboring counties. The target school was unique because students were mandated to take AP Computer Science Principles during their freshman year, while the other students were not. Through multinomial logistic regression, the influence of computational thinking on student proficiency rates was quantified and found to be insignificant. However, the COVID-19 pandemic greatly impacted the students' performance. While the findings were insignificant, the students in the target school were likelier than the other students in their county and the neighboring county to be proficient in Geometry, according to the New York State Education Department's definition of proficiency. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
35. A Human-Centered Approach to Improving Adolescent Real-Time Online Risk Detection Algorithms
- Author
-
Ashwaq Alsoubai
- Abstract
Computational risk detection holds promise for shielding particularly vulnerable groups from online harm. A thorough literature review on real-time computational risk detection methods revealed that most research defined 'real-time' as approaches that analyze content retrospectively as early as possible or as preventive approaches to prevent risks from reaching online environments. This review provided a research agenda to advance the field, highlighting key areas: employing ecologically valid datasets, basing models and features on human understanding, developing responsive models, and evaluating model performance through detection timing and human assessment. This dissertation embraces human-centric methods for both gaining empirical insights into young people's risk experiences online and developing a real-time risk detection system using a dataset of youth social media. By analyzing adolescent posts on an online peer support mental health forum through a mixed-methods approach, it was discovered that online risks faced by youth could be laden by other factors, like mental health issues, suggesting a multidimensional nature of these risks. Leveraging these insights, a statistical model was used to create profiles of youth based on their reported online and offline risks, which were then mapped with their actual online discussions. This empirical study uncovered that approximately 20% of youth fall into the highest risk category, necessitating immediate intervention. Building on this critical finding, the third study of this dissertation introduced a novel algorithmic framework aimed at the 'timely' identification of high-risk situations in youth online interactions. This framework prioritizes the riskiest interactions for high-risk evaluation, rather than uniformly assessing all youth discussions. A notable aspect of this study is the application of reinforcement learning for prioritizing conversations that need urgent attention. This innovative method uses decision-making processes to flag conversations as high or low priority. After training several deep learning models, the study identified Bi-Long Short-Term Memory (Bi-LSTM) networks as the most effective for categorizing conversation priority. The Bi-LSTM model's capability to retain information over long durations is crucial for ongoing online risk monitoring. This dissertation sheds light on crucial factors that enhance the capability to detect risks in real time within private conversations among youth. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
36. Combinatorial Tasks as Model Systems of Deep Learning
- Author
-
Benjamin L. Edelman
- Abstract
This dissertation is about a particular style of research. The philosophy of this style is that in order to scientifically understand deep learning, it is fruitful to investigate what happens when neural networks are trained on simple, mathematically well-defined tasks. Even though the training data is simple, the training algorithm can end up producing rich, unexpected results; and understanding these results can shed light on fundamental mysteries of high relevance to contemporary deep learning. First, we situate this methodological approach in a broader scientific context, discussing and systematizing the role of "model systems" in science and in the science of deep learning in particular. We then present five intensive case studies, each of which uses a particular combinatorial task as a lens through which to demystify puzzles of deep learning. The combinatorial tasks employed are sparse Boolean functions, sparse parities, learning finite group operations, performing modular addition, and learning Markov chains in-context. Topics of explanatory interest include the inductive biases of the transformer architecture, the phenomenon of emergent capabilities during training, the nuances of deep learning in the presence of statistical-computational gaps, the tradeoffs between different resources of training, the effect of network width on optimization, the relationship between symmetries in training data and harmonic structure in trained networks, the origins of the mechanisms of in-context learning in transformers, and the influence of spurious solutions on optimization. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
37. Algorithms, Abolition, and African American Youth Development: Theorizing and Examining the Impacts of Artificial Intelligence Systems on Black Adolescents in the Age of #BlackLivesMatter
- Author
-
Avriel Epps
- Abstract
The unprecedented racial justice movements of 2020, intensified by George Floyd's tragic murder, catalyzed a global mobilization. This dissertation investigates the intersection of these movements with sociotechnical tools that shaped the movement, focusing on the youth deeply engaged with socioalgorithmic systems and their developmental experiences. Article 1 introduces a theoretical framework to understand modern developmental ecologies through a socioalgorithmic lens. It argues for a pivot in the field of developmental science toward studying the technical facets of digital environments, recognizing algorithms shape not just social interactions and political awareness, but also broader psychological and biological processes. Amid 2020's racial justice uprisings, the article underscores the urgency to investigate how algorithmic mediation influences developmental processes like socio-political identity development, moving beyond the discourse on screen time to scrutinize the roles of technological structures in shaping development. Article 2 presents an empirical study examining the impact of algorithmic bias on Black youth during 2020's tumultuous events. Investigating whether ethnic-racial identity exploration activities buffer against the detrimental mental health effects of algorithmic bias in AI selfie filters, the research integrates racial identity development theories with the psychology of systemic racism. Findings illuminate the complex interactions between socioalgorithmic systems and mental health outcomes in racially marginalized youth, highlighting the non-neutrality of algorithms and their potential to reinforce societal biases. The third article conducts an algorithmic audit of YouTube's search engine regarding Black Lives Matter movement content. This audit critically assesses the digital spaces that youth interact with, analyzing the biases embedded in such search algorithms. The findings suggest although many aspects of the Black Lives Matter movement were represented to be in alignment of the organizer's values, the more radical black liberatory ideas associated with "abolition" are represented more negatively in the content curated by YouTube search. By examining the confluence of algorithmic systems and racial justice activism in 2020, my dissertation advocates for a developmental science paradigm that integrates socioalgorithmic context. It aims to enrich understanding of the digital age's impact on the developmental trajectories of racially marginalized youth and calls for incorporating algorithmic systems analysis into the broader understanding of human development. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
38. Essays on the Economics of Education
- Author
-
Julio Rodriguez
- Abstract
In this dissertation, I present an examination of the economics of education through three chapters. In the first paper, I study the overrepresentation of elite university graduates in senior positions in public administration. Using rich administrative data from Chile, I employ a stacked fuzzy regression discontinuity design to estimate the causal effect of attending elite universities versus non-elite institutions on the likelihood of working in the public sector and attaining top positions within it. The findings suggest that while the observed disparity in top positions within public administration is largely a result of selection rather than inherent advantages of elite education, attending elite universities may enhance social mobility for students from lower socioeconomic backgrounds, particularly within specific majors. In the second paper, my coauthors and I propose an alternative approach using algorithms to predict college readiness and guide course placement. Drawing on experimental data from seven community colleges, the study shows that algorithmic placement increases placement rates into college-level courses without sacrificing pass rates. Moreover, algorithmic placement shows promise in narrowing demographic disparities in placement rates and remedial course enrollment, outperforming traditional placement tests in terms of predictive accuracy while mitigating discrimination. In the final chapter, I explore the relationship between school counselor availability and disciplinary outcomes in middle and high schools across the United States. Leveraging exogenous variations in student-to-counselor ratios driven by state recommendations and mandates, I employ administrative data from 26 states to estimate the causal impact of counselor availability on disciplinary actions such as suspensions, expulsions, and transfers. The results indicate that increased counselor availability reduces school disciplinary actions, with larger effects observed in high schools compared to middle schools. Moreover, speculative analyses suggest that the effectiveness of counselors in mitigating disciplinary issues may be complemented by the overall staffing levels in high schools. This dissertation contributes to our understanding of how educational policies and practices shape individual outcomes and societal inequalities, shedding light on avenues for promoting social mobility, improving educational access and equity, and fostering conducive learning environments. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
39. Deep Lifelong Learning with Factorized Knowledge Transfer
- Author
-
Seungwon Lee
- Abstract
Human intelligence has the ability to capture abstract concepts from experience and utilize that learned knowledge for adaptation to new situations. Lifelong machine learning aims to achieve those same properties of human intelligence by designing algorithms to learn from a sequence of tasks, extract useful knowledge of previous tasks, and re-use the extracted knowledge to learn new future tasks. Research into lifelong learning has explored various methodologies, including techniques for sharing knowledge across tasks, techniques for maintaining previously acquired skills, and techniques for actively selecting the next task to learn. This dissertation will focus on one theme of lifelong learning: the way knowledge is transferred across tasks via factorization, which breaks down the architecture of neural networks to naturally encode conceptual knowledge. The tensor factorization is capable of discovering abstract but generalizable knowledge from experiences. This dissertation investigates methods to factorize knowledge encoded in neural networks and share the knowledge across multiple tasks, as well as methods to enhance the training of these factorized knowledge transfer mechanisms. This dissertation starts by developing a lifelong learning architecture that utilizes deconvolutional operation to preserve multi-axis features of data. This deconvolution-based factorization architecture empirically shows reduced harmful interference between tasks thanks to sharing abstract knowledge via factorization. The dissertation then studies the importance of transferring the proper level of knowledge in the network for the success of lifelong learning. As a result, an expectation-maximization style algorithm is developed to discover the useful granularity of knowledge to share for each task depending on the given data. This algorithm determines which layers to share while learning tasks in parallel and reduces human intervention in selecting the knowledge transfer architecture for lifelong learning, which is critical for realistic scenarios with complex task relationships. Moreover, it applies to diverse lifelong learning architectures, augmenting existing lifelong learning works. Lastly, the dissertation investigates the use of data programming to extend existing lifelong learning algorithms into semi-supervised settings, tackling the lifelong learning challenge of data annotation. Due to the modularized framework and theoretical guarantees on the quality of generated labels, this framework can be applied to the existing supervised lifelong learning algorithms. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
40. Challenging Curriculum/Assessment and Grade Reporting Practices
- Author
-
Karen E. B. Elliott
- Abstract
When the purpose and meaning of grades is not well-defined by an organization, there is a disconnect between curricular/assessment practices and student engagement. This study sought to investigate origins and find solutions in regards to discerning a student's knowledge and mastery of a subject and how this is reflected in the grade, and how grades can be accurately reported. Participants and data collected in Cycle One consisted of professional faculty who shared personal experience with teacher/student/parent perception of grades and their value, and what their curricular and assessments were revealing about student engagement and performance. As consistent themes emerged through interviews and focus groups, Action steps were designed, implemented, and evaluated in Cycle Two to provide a framework for the high school division that would 1) provide clarity and consistency among the faculty, their subject areas, and their departments, and 2) to enhance student engagement through curricular/assessment and grade reporting practices that better reflected the student's knowledge and mastery. A ten-week book study was formed in Cycle Two that challenged current conventions and practices with discussions which evolved from Thomas Guskey's On Your Mark: Challenging the Practices of Grading and Reporting. Ten sessions were developed and co-facilitated by members of the high school faculty/administration at the research site. Evaluating the results of the Action Research study included a thorough assessment of the value of grades and how they are reported at this organization. At the conclusion of the book study, the focus group/participants delivered a formal presentation to the faculty in the high school division which focused on the purpose of grades; the subjectivity of grade reporting; the inaccuracy of percentages and algorithms; the misperception of class ranking; and elements, like behaviors, that obscure the meaning of grades. Participants provided critical feedback throughout the process, and the administration intends to continue honing how faculty arrive at a grade, its intended meaning, its purpose for the student/parent/teacher, and how the organization can better streamline curricular/assessment practices so mastery of a subject matter is clearer. Implications for the organization included an official purpose statement to define the meaning of grades and explain why they are reported; a document that clarifies vocabulary used among all high school subject areas/departments; and, a broad rubric for how to evaluate written assessments. Organizational leaders intend to make other changes that make curricular/assessment and grading practices much more meaningful for the student and their educational experience. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
41. The Impact of Algorithmic Literacy and Purpose on Educator Use of and Confidence in Algorithmic Advice
- Author
-
Jessa Henderson
- Abstract
Algorithms may be better at prediction than humans in a variety of contexts, but they are not perfect. A deeper understanding of the ways in which educators use and question algorithmic advice within their professional domain is needed. Educators are a particularly unique professional group, in comparison with the other groups studied in the literature, and will likely be the on-the-ground employees expected to integrate algorithmic advice into their pedagogical planning and decision making. This quantitative study used a 2x2 pretest-posttest experimental design to investigate the way that educators used and trusted algorithmic advice for high school science course recommendations, including the way contextual framing and individual-level differences impacted these relationships. Participants were randomly assigned to one of two purpose conditions in a simulation environment designed using the Judge-Advisor System (JAS) paradigm to assess the influence of algorithmic advice. Group comparison and multiple regression techniques were utilized to test hypotheses. Results indicated that approximately 30% of participants did not change their decision at any time and that the remaining 70% tended to not change their decision when given the chance. Additionally, a significant interaction was detected for average confidence change with participants who had more years of teaching experience having higher rates of average confidence change in the positive purpose condition. More research is suggested that investigates the potential for manipulation of educator trust based on a recommendation system's framing and/or marketing. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
42. Fair Multivariate Adaptive Regression Splines for Ensuring Equity and Transparency
- Author
-
Parian Haghighat, Denisa Gandara, Lulu Kang, and Hadis Anahideh
- Abstract
Predictive analytics is widely used in various domains, including education, to inform decision-making and improve outcomes. However, many predictive models are proprietary and inaccessible for evaluation or modification by researchers and practitioners, limiting their accountability and ethical design. Moreover, predictive models are often opaque and incomprehensible to the officials who use them, reducing their trust and utility. Furthermore, predictive models may introduce or exacerbate bias and inequity, as they have done in many sectors of society. Therefore, there is a need for transparent, interpretable, and fair predictive models that can be easily adopted and adapted by different stakeholders. In this paper, we propose a fair predictive model based on multivariate adaptive regression splines (MARS) that incorporates fairness measures in the learning process. MARS is a non-parametric regression model that performs feature selection, handles non-linear relationships, generates interpretable decision rules, and derives optimal splitting criteria on the variables. Specifically, we integrate fairness into the knot optimization algorithm and provide theoretical and empirical evidence of how it results in a fair knot placement. We apply our "fair"MARS model to real-world data and demonstrate its effectiveness in terms of accuracy and equity. Our paper contributes to the advancement of responsible and ethical predictive analytics for social good. [This paper was presented at an Association for the Advancement of Artificial Intelligence conference.]
- Published
- 2024
43. Adaptive Training with Virtual Reality for the Maintenance of Mission-Critical Skills in Long Duration Spaceflight
- Author
-
Esther Putman
- Abstract
Human space exploration is entering a new era through the design of missions that venture further from earth for longer durations, such as missions to the Moon and Mars. Mission success during long duration exploration missions (LDEM) requires astronauts to flawlessly execute exceptionally complex skills across a multitude of tasks, presenting new challenges for training. This thesis investigates virtual reality (VR) as a potential platform to support critical skill training and maintenance during LDEM. First, a VR Entry, Descent, and Landing (EDL) training scenario was created to support the investigational aims of this research. EDL represents a highly complex task that requires extensive preflight training and carries extreme consequences of failure if skill is not properly maintained over time. A collection of design objectives drove the development of the training scenario, including complex subtask integration, algorithm facilitation of adaptive difficulty, a sense of immersion and presence, allowing for repeated training without predictability or boredom, and providing feedback to support learning. The EDL task was created in a modular format that allowed for presentation in immersive VR, non-immersive screen, and physical mockup modalities, as well as with static or adaptive difficulty. In the second aim, the EDL task was then investigated in a collection of between-subjects experiments. In the first experiment, results showed promising trends that adaptive VR training supports higher skill gain and less plateau in learning as well as higher average performance when transitioning from virtual to physical task performance. In the second experiment, facilitation with an adaptive algorithm was found to significantly improve task mastery as opposed to self-led difficulty selection or static difficulty training. Next, simplified corollary task training was found to support retention of skill only if the cognitive demands of that simplified task were appropriately mapped to those of the operational task. And finally, just-in-time training was found to be an effect method to mitigate skill loss while preventing the risks of negative training. The work presented in this thesis found that VR is a valuable tool for the maintenance of mission-critical skill during LDEM. However, skill maintenance is best supported through design of VR training that utilizes adaptive algorithms to facilitate learning, presents content with cognitive demands that align to the tasks they seek to train, and when refresher training is administered in a JIT cadence. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
44. A Multidimensional Perspective of Mathematical Ability for Eliciting Innovations in Education
- Author
-
Christopher Garrido Lechuga
- Abstract
Adaptive tutoring systems often model student knowledge in ways that break away from a "one size fits all" approach to learning. Nonetheless, the strengths of these systems can often be limited, as knowledge representations are not easily interpreted by teachers, which make these systems difficult to integrate into pedagogical practices. Fortunately, as researchers, we can still take advantage of the AI in these systems to extend our innovations and address problems of practice. As such, the goal of the present work is to leverage multidimensional representations of knowledge that these systems provide to explore innovations in educational measurement, pedagogical techniques, and practice. I explore these in three studies outlined below. In Study 1 I explore "Innovations in educational measurement" by adopting an alternative perspective to measuring student mathematical ability. In this study I invite the reader to re-conceptualize mathematical ability as a multidimensional construct, which runs counter to long-established tradition in academic measurement and pedagogical practice. In doing so, I compare the variation observed in students' mathematical ability when ability is measured using different metrics that vary in dimensionality. Findings suggest that under a multidimensional view, students (including those who may be traditionally seen as "low-ability") often possess relative strengths when compared to their peers, thus suggesting that categorizations such as low- and high-ability, which are typically used in practice, may be an over simplification. In Study 2 I explore "Innovations in pedagogical techniques" as I present three novel algorithms for grouping students that leverage existing technological AI innovations that model student knowledge across hundreds of mathematical skills. These methods attempt to better align the personalization of a tutoring system with teachers' instructional practices. I evaluate each of the three methods against two alternative baseline methods--one that groups students randomly and one that groups students based on a unidimensional course score. Findings demonstrate that these novel methods, which adopt a multidimensional view of ability, were more capable than the baseline methods at grouping students with similar strengths and weaknesses on a fine-grained skill level. In Study 3 I explore "Innovations in practice" in a research-practice partnership (RPP) with curriculum specialists at a local urban school district in Southern California. Implementation strategies, including curriculum and grouping recommendations were provided to teachers for a summer intervention program. In the spirit of a design-based implementation partnership, strategies and recommendations were formed to meet the needs of partners, while simultaneously attempting to address known problems of practice when implementing a blended learning design in the classroom. In this first cycle of iteration, these attempts were examined through teacher surveys as well as student login and learning data in a tutoring system. Ultimately, these data were used to examine the fidelity of implementation so that findings may inform future iterations of this ongoing partnership. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
45. Computational Learning Theory through a New Lens: Scalability, Uncertainty, Practicality, and beyond
- Author
-
Chen Wang
- Abstract
Computational learning theory studies the design and analysis of learning algorithms, and it is integral to the foundation of machine learning. In the modern era, classical computational learning theory is growingly unable to catch up with new practical demands. In particular, problems arise in the following aspects: i). "scalability": with the massive scales of modern datasets, classical learning algorithms that use polynomial time and store the entire input are no longer efficient; ii). "uncertainty": information nowadays is usually coupled with noise and uncertainty, which renders classical algorithms that assume accurate information unreliable; and iii). "practicality": under the modern context, traditional "negligible" terms in learning theory (e.g., log n factors) can no longer be ignored, and many theoretically-efficient techniques become inapplicable in practice. There are several promising approaches to tackle the above challenges. For scalability, one of the most popular approaches is to study learning algorithms under "sublinear" models, including streaming, sublinear time, and Massively Parallel Computation (MPC) models. Learning algorithms under these models usually use resources substaintially smaller than the input size. For uncertainty, we can look into learning algorithms that naturally take noisy inputs, e.g., algorithms that deal with multi-armed bandits (MABs), or algorithms that operate with randomly corrupted inputs. Finally, for practicality, we can focus on designing algorithms that are easy to implement and aim for algorithms with both theoretical guarantees and experimental performances. In light of the above discussion, this dissertation presents three major areas of study as follows: (1) In Part I, we present recent results in streaming multi-armed bandits, where the arms arrive one by one in a stream, and the algorithm stores a much smaller number of arms than the input. We consider two fundamental learning problems in this model, namely pure exploration and regret minimization. For pure exploration, we give an ultra-efficient algorithm that finds the best arm and stores a single extra arm at any point. The algorithm uses a single pass and the information-theoretically optimal number of arm pulls. Subsequently, under various settings, we characterize the optimal sample-memory-pass trade-offs for pure exploration streaming MABs. For regret minimization, we give the optimal regret bounds for single-pass MABs. Together, these results complete a majority of the picture for classical MABs problems under the memory-constraint setting; (2) In part II, we study graph clustering in sublinear settings. We consider two important problems in practice: correlation clustering and hierarchical clustering. For correlation clustering, we design O(1)-approximation algorithms with the semi-streaming O[approximately](n) space in the graph streaming model and with O[approximately](n) time in the query model, where n is the number of "vertices" in the graph. Furthermore, for the truly-streaming polylog "n" space regime, we design algorithms that approximate the "optimal cost" of correlation clustering. We test the polylog "n"-space algorithms on various input data and find that the practical performances are better than the "worst-case" guarantees. For hierarchical clustering, we give a single-pass semi-streaming algorithm that achieves O(1)-approximation for Dasgupta's objective, and we prove the cost-space trade-off in the single-pass setting; and (3) In part III, we move to more practically-driven problems of differential privacy (DP) and weak-strong oracle learning. For the former problem, we consider the differentially private release of "range queries" on graphs--a problem with wide applications in networks--and we give optimal bounds for pure and approximate DP. For the latter problem, weak-strong oracle learning is motivated by an industry setting where information is obtained from a cheap but noisy source and an accurate but expensive source. We study (metric) k-clustering and MST in this setting and obtain nearly-optimal algorithms with strong experimental performances. [The dissertation citations contained here are published with the permission of ProQuest LLC. Further reproduction is prohibited without permission. Copies of dissertations may be obtained by Telephone (800) 1-800-521-0600. Web page: http://www.proquest.com/en-US/products/dissertations/individuals.shtml.]
- Published
- 2024
46. MSAEM Estimation for Confirmatory Multidimensional Four-Parameter Normal Ogive Models
- Author
-
Jia Liu, Xiangbin Meng, Gongjun Xu, Wei Gao, and Ningzhong Shi
- Abstract
In this paper, we develop a mixed stochastic approximation expectation-maximization (MSAEM) algorithm coupled with a Gibbs sampler to compute the marginalized maximum a posteriori estimate (MMAPE) of a confirmatory multidimensional four-parameter normal ogive (M4PNO) model. The proposed MSAEM algorithm not only has the computational advantages of the stochastic approximation expectation-maximization (SAEM) algorithm for multidimensional data, but it also alleviates the potential instability caused by label-switching, and then improved the estimation accuracy. Simulation studies are conducted to illustrate the good performance of the proposed MSAEM method, where MSAEM consistently performs better than SAEM and some other existing methods in multidimensional item response theory. Moreover, the proposed method is applied to a real data set from the 2018 Programme for International Student Assessment (PISA) to demonstrate the usefulness of the 4PNO model as well as MSAEM in practice. [This paper was published in "Journal of Educational Measurement" v61 n1 p99-124 2024.]
- Published
- 2024
- Full Text
- View/download PDF
47. A Note on Improving Variational Estimation for Multidimensional Item Response Theory
- Author
-
Chenchen Ma, Jing Ouyang, Chun Wang, and Gongjun Xu
- Abstract
Survey instruments and assessments are frequently used in many domains of social science. When the constructs that these assessments try to measure become multifaceted, multidimensional item response theory (MIRT) provides a unified framework and convenient statistical tool for item analysis, calibration, and scoring. However, the computational challenge of estimating MIRT models prohibits its wide use because many of the extant methods can hardly provide results in a realistic time frame when the number of dimensions, sample size, and test length are large. Instead, variational estimation methods, such as Gaussian Variational Expectation Maximization (GVEM) algorithm, have been recently proposed to solve the estimation challenge by providing a fast and accurate solution. However, results have shown that variational estimation methods may produce some bias on discrimination parameters during confirmatory model estimation, and this note proposes an importance weighted version of GVEM (i.e., IW-GVEM) to correct for such bias under MIRT models. We also use the adaptive moment estimation mtest bias ethod to update the learning rate for gradient descent automatically. Our simulations show that IW-GVEM can effectively correct bias with modest increase of computation time, compared with GVEM. The proposed method may also shed light on improving the variational estimation for other psychometrics models. [This paper was published in "Psychometrika" v89 p172-204 2024.]
- Published
- 2024
- Full Text
- View/download PDF
48. A Comparison of Machine Learning Algorithms for Predicting Student Performance in an Online Mathematics Game
- Author
-
Lee, Ji-Eun, Jindal, Amisha, Patki, Sanika Nitin, Gurung, Ashish, Norum, Reilly, and Ottmar, Erin
- Abstract
This paper demonstrates how to apply Machine Learning (ML) techniques to analyze student interaction data collected in an online mathematics game. We examined: (1) how different ML algorithms influenced the precision of middle-school students' (N = 359) performance prediction; and (2) what types of in-game features were associated with student math knowledge scores. The results indicated that the Random Forest algorithm showed the best performance in predicting posttest math knowledge scores among the seven algorithms employed. Out of 37 features included in the model, the validity of the students' first mathematical transformation was the most predictive of their math knowledge scores. Implications for game learning analytics and supporting students' algebraic learning are discussed based on the findings.
- Published
- 2022
- Full Text
- View/download PDF
49. Evaluating Gaming Detector Model Robustness over Time
- Author
-
Levin, Nathan, Baker, Ryan S., Nasiar, Nidhi, Fancsali, Stephen, and Hutt, Stephen
- Abstract
Research into "gaming the system" behavior in intelligent tutoring systems (ITS) has been around for almost two decades, and detection has been developed for many ITSs. Machine learning models can detect this behavior in both real-time and in historical data. However, intelligent tutoring system designs often change over time, in terms of the design of the student interface, assessment models, and data collection log schemas. Can gaming detectors still be trusted, a decade or more after they are developed? In this research, we evaluate the robustness/degradation of gaming detectors when trained on older data logs and evaluated on current data logs. We demonstrate that some machine learning models developed using past data are still able to predict gaming behavior from student data collected 16 years later, but that there is considerable variance in how well different algorithms perform over time. We demonstrate that a classic decision tree algorithm maintained its performance while more contemporary algorithms struggled to transfer to new data, even though they exhibited better performance on unseen students in both New and Old data sets by themselves. Examining the feature importance values provides some explanation for the differences in performance between models, and offers some insight into how we might safeguard against detector rot over time. [For the full proceedings, see ED623995.]
- Published
- 2022
50. Adversarial Bandits for Drawing Generalizable Conclusions in Non-Adversarial Experiments: An Empirical Study
- Author
-
Zhi-Han, Yang, Zhang, Shiyue, and Rafferty, Anna N.
- Abstract
Online educational technologies facilitate pedagogical experimentation, but typical experimental designs assign a fixed proportion of students to each condition, even if early results suggest some are ineffective. Experimental designs using multi-armed bandit (MAB) algorithms vary the probability of condition assignment for a new student based on prior results, placing more students in more effective conditions. While stochastic MAB algorithms have been used for educational experiments, they collect data that decreases power and increases false positive rates [22]. Instead, we propose using adversarial MAB algorithms, which are less exploitative and thus may exhibit more robustness. Through simulations involving data from 20+ educational experiments [29], we show data collected using adversarial MAB algorithms does not have the statistical downsides of that from stochastic MAB algorithms. Further, we explore how differences in condition variability (e.g., performance gaps between students being narrowed by an intervention) impact MAB versus uniform experimental design. Data from stochastic MAB algorithms systematically reduce power when the better arm is less variable, while increasing it when the better arm is more variable; data from the adversarial MAB algorithms results in the same statistical power as uniform assignment. Overall, these results demonstrate that adversarial MAB algorithms are a viable "off-the-shelf" solution for researchers who want to preserve the statistical power of standard experimental designs while also benefiting student participants. [For the full proceedings, see ED623995.]
- Published
- 2022
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.