Adaptive Learning Recommendation Strategy Based on Deep Q-learning.

Markov decision process adaptive learning recommendation system reinforcement learning

Journal

Applied psychological measurement
ISSN: 1552-3497
Titre abrégé: Appl Psychol Meas
Pays: United States
ID NLM: 7905715

Informations de publication

Date de publication:
Jun 2020
Historique:
entrez: 16 6 2020
pubmed: 17 6 2020
medline: 17 6 2020
Statut: ppublish

Résumé

Personalized recommendation system has been widely adopted in E-learning field that is adaptive to each learner's own learning pace. With full utilization of learning behavior data, psychometric assessment models keep track of the learner's proficiency on knowledge points, and then, the well-designed recommendation strategy selects a sequence of actions to meet the objective of maximizing learner's learning efficiency. This article proposes a novel adaptive recommendation strategy under the framework of reinforcement learning. The proposed strategy is realized by the deep Q-learning algorithms, which are the techniques that contributed to the success of AlphaGo Zero to achieve the super-human level in playing the game of go. The proposed algorithm incorporates an early stopping to account for the possibility that learners may choose to stop learning. It can properly deal with missing data and can handle more individual-specific features for better recommendations. The recommendation strategy guides individual learners with efficient learning paths that vary from person to person. The authors showcase concrete examples with numeric analysis of substantive learning scenarios to further demonstrate the power of the proposed method.

Identifiants

pubmed: 32536728
doi: 10.1177/0146621619858674
pii: 10.1177_0146621619858674
pmc: PMC7262997
doi:

Types de publication

Journal Article

Langues

eng

Pagination

251-266

Informations de copyright

© The Author(s) 2019.

Déclaration de conflit d'intérêts

Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Références

Appl Psychol Meas. 2018 Jan;42(1):24-41
pubmed: 29335659
Nature. 2016 Jan 28;529(7587):484-9
pubmed: 26819042
Br J Math Stat Psychol. 2019 Feb;72(1):108-135
pubmed: 30277574
Science. 2004 Dec 10;306(5703):1940-3
pubmed: 15528409
Nature. 2015 Feb 26;518(7540):529-33
pubmed: 25719670
J Finance. 2009 Dec;64(6):2515-2534
pubmed: 20352013

Auteurs

Chunxi Tan (C)

The Hong Kong University of Science and Technology, Kowloon, Hong Kong.

Ruijian Han (R)

The Hong Kong University of Science and Technology, Kowloon, Hong Kong.

Rougang Ye (R)

The Hong Kong University of Science and Technology, Kowloon, Hong Kong.

Kani Chen (K)

The Hong Kong University of Science and Technology, Kowloon, Hong Kong.

Classifications MeSH