Metadata Analysis of Open Educational Resources
- URL: http://arxiv.org/abs/2101.07735v1
- Date: Tue, 19 Jan 2021 17:16:44 GMT
- Title: Metadata Analysis of Open Educational Resources
- Authors: Mohammadreza Tavakoli, Mirette Elias, G\'abor Kismih\'ok, S\"oren Auer
- Abstract summary: Open Educational Resources (OERs) are openly licensed educational materials that are widely used for learning.
This work uses the metadata of 8,887 OERs to perform an exploratory data analysis on OER metadata.
Based on the results, our analysis demonstrated that OER metadata and OER content qualities are closely related, as we could detect high-quality OERs with an accuracy of 94.6%.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Open Educational Resources (OERs) are openly licensed educational materials
that are widely used for learning. Nowadays, many online learning repositories
provide millions of OERs. Therefore, it is exceedingly difficult for learners
to find the most appropriate OER among these resources. Subsequently, the
precise OER metadata is critical for providing high-quality services such as
search and recommendation. Moreover, metadata facilitates the process of
automatic OER quality control as the continuously increasing number of OERs
makes manual quality control extremely difficult. This work uses the metadata
of 8,887 OERs to perform an exploratory data analysis on OER metadata.
Accordingly, this work proposes metadata-based scoring and prediction models to
anticipate the quality of OERs. Based on the results, our analysis demonstrated
that OER metadata and OER content qualities are closely related, as we could
detect high-quality OERs with an accuracy of 94.6%. Our model was also
evaluated on 884 educational videos from Youtube to show its applicability on
other educational repositories.
Related papers
- DiscoveryBench: Towards Data-Driven Discovery with Large Language Models [50.36636396660163]
We present DiscoveryBench, the first comprehensive benchmark that formalizes the multi-step process of data-driven discovery.
Our benchmark contains 264 tasks collected across 6 diverse domains, such as sociology and engineering.
Our benchmark, thus, illustrates the challenges in autonomous data-driven discovery and serves as a valuable resource for the community to make progress.
arXiv Detail & Related papers (2024-07-01T18:58:22Z) - AI Data Readiness Inspector (AIDRIN) for Quantitative Assessment of Data Readiness for AI [0.8553254686016967]
"Garbage in Garbage Out" is a universally agreed quote by computer scientists from various domains, including Artificial Intelligence (AI)
There are no standard methods or frameworks for assessing the "readiness" of data for AI.
AIDRIN is a framework covering a broad range of readiness dimensions available in the literature.
arXiv Detail & Related papers (2024-06-27T15:26:39Z) - DACO: Towards Application-Driven and Comprehensive Data Analysis via Code Generation [83.30006900263744]
Data analysis is a crucial analytical process to generate in-depth studies and conclusive insights.
We propose to automatically generate high-quality answer annotations leveraging the code-generation capabilities of LLMs.
Our DACO-RL algorithm is evaluated by human annotators to produce more helpful answers than SFT model in 57.72% cases.
arXiv Detail & Related papers (2024-03-04T22:47:58Z) - STAR: Boosting Low-Resource Information Extraction by Structure-to-Text
Data Generation with Large Language Models [56.27786433792638]
STAR is a data generation method that leverages Large Language Models (LLMs) to synthesize data instances.
We design fine-grained step-by-step instructions to obtain the initial data instances.
Our experiments show that the data generated by STAR significantly improve the performance of low-resource event extraction and relation extraction tasks.
arXiv Detail & Related papers (2023-05-24T12:15:19Z) - Core-set Selection Using Metrics-based Explanations (CSUME) for
multiclass ECG [2.0520503083305073]
We show how a selection of good quality data improves deep learning model performance.
Our experimental results show a 9.67% and 8.69% precision and recall improvement with a significant training data volume reduction of 50%.
arXiv Detail & Related papers (2022-05-28T19:36:28Z) - Deep Transfer Learning for Multi-source Entity Linkage via Domain
Adaptation [63.24594955429465]
Multi-source entity linkage is critical in high-impact applications such as data cleaning and user stitching.
AdaMEL is a deep transfer learning framework that learns generic high-level knowledge to perform multi-source entity linkage.
Our framework achieves state-of-the-art results with 8.21% improvement on average over methods based on supervised learning.
arXiv Detail & Related papers (2021-10-27T15:20:41Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - A Survey of Deep Active Learning [54.376820959917005]
Active learning (AL) attempts to maximize the performance gain of the model by marking the fewest samples.
Deep learning (DL) is greedy for data and requires a large amount of data supply to optimize massive parameters.
Deep active learning (DAL) has emerged.
arXiv Detail & Related papers (2020-08-30T04:28:31Z) - Extracting Topics from Open Educational Resources [0.0]
We propose an OER topic extraction approach, applying text mining techniques, to generate high-quality OER metadata about topic distribution.
This is done by: 1) collecting 123 lectures from Coursera and Khan Academy in the area of data science related skills, 2) applying Latent Dirichlet Allocation (LDA) on the collected resources in order to extract existing topics related to these skills, and 3) defining topic distributions covered by a particular OER.
arXiv Detail & Related papers (2020-06-19T12:50:55Z) - OER Recommendations to Support Career Development [0.0]
Open Educational Resources (OERs) have potential to contribute to the mitigation of problems, as they are available in a wide range of learning and occupational contexts globally.
We suggest a novel, personalised OER recommendation method to match skill development targets with open learning content.
This is done by: 1) using an OER quality prediction model based on metadata, OER properties, and content; 2) supporting learners to set individual skill targets based on actual labour market information; and 3) building a personalized OER recommender to help learners to master their skill targets.
arXiv Detail & Related papers (2020-05-30T21:01:54Z) - Quality Prediction of Open Educational Resources A Metadata-based
Approach [0.0]
Metadata play a key role in offering high quality services such as recommendation and search.
We propose an OER metadata scoring model, and build a metadata-based prediction model to anticipate the quality of OERs.
Based on our data and model, we were able to detect high-quality OERs with the F1 score of 94.6%.
arXiv Detail & Related papers (2020-05-21T09:53:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.