Mobile QR Code QR CODE

2024

Acceptance Ratio

21%


  1. (College of Technology and Data, Yantai Nanshan University, Yantai 265713, China {zxy_8204, gxj33490905}@163.com)
  2. (College of Education Science and Technology, Anshan Normal University, Anshan 114056, China yizhifan007@163.com)



Collaborative filtering, Long short-term memory network, Learning resource recommendation, Attention mechanism, Neural matrix factorization, Teaching vertical network

1. Introduction

In the digital age, the development of online learning has not only enriched teaching resources and methods, but also caused the problem of information explosion. Therefore, Learning Resource Recommendation (LRR) is particularly important [1]. LRR refers to recommending learning materials, textbooks, courses, and other resources that are suitable for different learners’ knowledge levels. LRR can help learners learn and acquire knowledge more efficiently. Raj and Renumol reviewed the research on LRR in adaptive scenes to explore the latest research trends in LRR systems. Most studies used learners’ cognitive aspects as labels or reliance [2]. Afsar et al. analyzed the recommendation problem solving methods based on deep reinforcement learning. They also proposed a reinforcement learning-based recommendation system framework. This helped to efficiently solve recommendation problems [3]. Hasan and Ferdous put forward a mixed recommending means that integrated text to image conversion. The alternating least squares means was utilized for enhancing the movie recommendation fidelity. The proposed system’s Root Mean Square Error (RMSE) was 0.8951. The recommendation accuracy was 97% [4]. Yu et al. analyzed how contrastive learning could alleviate the data sparsity. And they proposed a method of using noise-based embedding augmentation to generate contrastive learning views and demonstrated its recommending fidelity and training efficiency [5].

Collaborative Filtering (CF) is a extensively utilized algorithm in recommendation systems that utilizes the similarity between user groups or items to recommend content that may be of interest to users [6]. Wu et al. divided the recommendation modeling work into CF and information rich recommendations, and conducted a synthetic retrospect of neural recommendation means. This helped to provide theoretical references for researchers and practitioners of recommendation systems [7]. Fkih conducted an in-depth review of the similarity metrics utilized in CF-based recommendation systems and tested their performance through experimental research. Improved triangle similarity was the most suitable similarity measure for user-based filtering recommendation systems [8]. Zhou et al. put forward a self-supervised CF means that simplified Siamese networks by enhancing the potential embeddings generated by backbone networks. The proposed framework improved the accuracy by 17.79% and had high recommendation accuracy [9].

As computer technology advances, more deep learning techniques are applied to recommendation systems. In recommendation systems, user behavior data are usually a sequence. Long Short-Term Memory (LSTM) networks validly obtains long-term reliance in sequence data, thereby achieving more accurate recommendations [10]. Duan et al. put forward a new architecture using LSTM, and proposed a “recovery gate” to compensate for the memory deficiency caused by the forgetting mechanism. The proposed method achieved comparable performance to sophisticated methods on challenging sequence recommendation datasets [11]. Saraswat and Srishti adopted a recursive neural network to classify different recommendation items and make recommendations based on categories. The proposed method had an accuracy of 84% and an F1 score of 80%, demonstrating good recommendation performance [12]. Wang et al. put forward a deep learning-based interest point recommending method that considered privacy protection. They combined attention mechanisms to capture users’ long-term and short-term preferences. The recommended time was less than 130 ms [13].

In summary, although there is currently a lot of research on recommendation models, current online LRR methods often only consider learners’ historical learning records. These methods ignore the learners’ actual mastery of knowledge points, lack personalized guidance. In this context, this study proposes a knowledge tracking means on the foundation of LSTM and an LRR means on the foundation of on CF. The research aims to propose an effective online LRR method to further improve learners’ learning outcomes and develop online education. There are two main research innovations. Firstly, a new knowledge tracking means on the foundation of deep learning is proposed. Meanwhile, two gating mechanisms are designed to model learners’ learning and forgetting behaviors during knowledge state updates. The second point is to combine model-based CF and attention mechanisms to address the attribute features processing.

2. Methods and Materials

The study utilizes gating mechanisms in LSTM to update learners’ knowledge states. Attention mechanism is introduced to capture the learner’s current knowledge state. A knowledge tracking means on the foundation of LSTM is built. Meanwhile, a model-based CF is utilized to build an LRR model based on CF.

2.1. Construction of Knowledge Tracking Model based on LSTM

Before conducting LRR, students’ mastery status on different knowledge points is obtained through knowledge tracking. Learners’ learning behavior is usually time-series data. LSTM is proposed to deal with Recurrent Neural Network (RNN)’s long-term reliance. It can analyze inputs using time series and has been proven to have superior performance in processing time-series 14,15]. Therefore, the study adopts LSTM to build a knowledge tracking model. LSTM includes three gating mechanisms. Input gates can control whether input information is updated into the cell unit, represented by Eq. (1).

(1)
{ i t = σ ( W i h t 1 + V i x t + b i ) , c ¯ t = tanh ( W c h t 1 + V c x t + b c ) , c t = f t c t 1 + i t c ¯ t .

In Eq. (1), i t stands for the input gate state. W i means a weight matrix. tanh stands for an activation function. c ¯ t refers to candidate cell units. W c stands for vector units’ weight matrix. c t refers to the memory cell node’s latest state. The forget gate state is represented by Eq. (2) [16].

(2)
f t = σ ( W f h t 1 + V f x t + b f ) .

In Eq. (2), σ stands for the sigmoid activation function. W f stands for the forget gate’s weight matrix. h t 1 stands for the input from the previous moment. V stands for the weight matrix of h t 1 . x t refers to the input value at the current time. b stands for bias term. The output gate is represented by Eq. (3).

(3)
{ o t = σ ( W o h t 1 + V o x t + b o ) , h t = o t tanh ( c t ) .

In Eq. (3), o t refers to the output gate state. W o refers to the weight matrix. h t refers to the output at the current time. In knowledge tracking, a concept matrix M is constructed for all knowledge, with each column representing a knowledge vector M ( i ) . The knowledge point k t and problem q t at time t are transformed into knowledge point embedding e t and problem embedding z t . z t is merged with the problem response r t , represented by Eq. (4).

(4)
z ˜ t = { [ z t 0 ] ,  if  r t = 1 , [ 0 z t ] ,  if  r t = 0.

In Eq. (4), represents the connection operation. The inner product of knowledge point embeddings and each column of knowledge vectors in the concept matrix is calculated. The results are input into the softmax function, represented by Eq. (5) [17].

(5)
β t = softmax [ e t M ( i ) ] .

In Eq. (5), β t represents the problem and various knowledge points’ correlation. Most existing knowledge tracking models judge learners’ knowledge status based on their performance in the problem, which cannot reflect their actual knowledge level [18]. Therefore, the study utilizes the gating mechanism in LSTM and designs two gating mechanisms, learning and forgetting, to update learners’ knowledge states. The learner’s learning benefit is represented by Eq. (6).

(6)
{ L t = tanh ( w 1 T [ y t a t t h t 1 ] + b 1 ) , y t = β t z t .

In Eq. (6), w 1 T and b 1 refer to the parameters that need to be optimized. a t t refers to the time spent embedding in answering question q t . y t refers to embedding problems with authority. A learning gate is designed to control learners’ absorption of knowledge, represented by Eq. (7).

(7)
δ t = σ ( w 2 T [ y t a t t h t 1 ] + b 2 ) .

Learners’ knowledge gain L ˜ t after one learning interaction ( q t , r t ) is represented by Eq. (8).

(8)
L ˜ t = δ t L t .

The forget gate is represented by Eq. (9).

(9)
F t = σ ( w 3 T [ y t i t t h t 1 ] + b 3 ) .

In Eq. (9), i t t refers to the time interval embedding between the learner’s current interaction and the previous interaction on the same problem. Therefore, the knowledge state update of learners after experiencing a learning interaction is represented by Eq. (10).

(10)
h t = L ˜ t + L t h t 1 .

The difficulty of the problem can have an impact on the learner’s response to the question. Therefore, attention mechanisms are introduced to capture how different knowledge concepts and problem difficulty affect knowledge tracking for learners in their current knowledge state. When processing input data, the attention mechanism assigns different weights to enable the model to focus on key parts of the input data, thereby improving processing accuracy and efficiency 19,20]. The attention mechanism is represented by Eq. (11).

(11)
Attention ( Q , K , V ) = softmax ( ( Q K T ) / d k ) V .

In Eq. (11), Q means query. K means key. V means value. d k means the dimensions of Q and K . Learner data are collected and preprocessed. Each knowledge concept’s difficulty d ( k t ) and the problem’s difficulty d ( q t ) are calculated, represented by Eq. (12).

(12)
{ d ( q t ) = c o u n t R q i = 1 / c o u n t q i , d ( k t ) = c o u n t R k i = 1 / c o u n t k i .

In Eq. (12), c o u n t R q i = 1 represents the frequency that question q i has been correctly answered by learners. c o u n t q i means the frequency that the question q i is answered by learners. c o u n t R k i = 1 means the frequency that the knowledge concept k i has been correctly grasped. c o u n t k i means the frequency that knowledge concept k i is answered by learners.

Fig. 1. Framework diagram of knowledge tracking model.

../../Resources/ieie/IEIESPC.2025.14.6.815/fig1.png

The attention network’s output is represented by Eq. (13).

(13)
o u t = Attention ( concat ( h t , d ( q t ) , d ( k t ) ) ) .

Subsequently, the study increases the model nonlinearity through a position Feedforward Neural Network (FNN). FNN consists of multiple levels of nodes that transmit information in specific directions without feedback loops. Therefore, FNN is highly effective in handling static input data and classification tasks [21]. The FNN output is represented by Eq. (14).

(14)
FNN ( o u t ) = ReLU ( o u t w 4 T + b 4 ) w 5 T + b 5 .

The study utilizes the sigmoid function to activate the fully connected layer and minimizes the loss of the prediction layer and problem response r t through a binary cross entropy loss function, as shown in formula (15).

(15)
Prediction = sigmoid ( FNN ( o u t ) w 6 T + b 6 ) .

In summary, the knowledge tracking model constructed in this study mainly includes three modules: knowledge weight calculation, learning and forgetting, and attention. Fig. 1 shows a specific LSTM based knowledge tracking model.

2.2. Construction of the Collaborative Filtering Based-Learning Resource Recommendation Model

To further improve learning outcomes, this study conducts personalized LRR based on the knowledge mastery of different learners. CF mainly includes two types: heuristic based and model-based [22]. Compared to traditional user-based CF, model-based CF has higher flexibility. Therefore, the study adopts model-based CF to build LRR systems. Matrix Factorization (MF) is a method utilized to process high-dimensional data, with the main objective of decomposing a high-dimensional matrix into the product of a low dimensional matrix [23]. Neural Matrix Factorization (NMF) combines Generalized Matrix Factorization (GMF) and Multi-Layer Perceptron (MLP) to better capture users and items’ complicated relationships 24,25]. Fig. 2 shows the specific structure of NMF.

Fig. 2. Structure diagram of neural matrix factorization model.

../../Resources/ieie/IEIESPC.2025.14.6.815/fig2.png

NMF’s input layer is a one-hot encoding in users or items. One-hot representation is embedded into dense vectors in the feature space and merged to obtain the final prediction result. However, one-hot representation cannot reflect any semantic information, which can lead to the appearance of correlation between features. Therefore, the study adopts an embedding layer in the recommendation model and utilizes one-hot input as the table retrieval index. To address the attribute feature processing, attention mechanisms are introduced to reduce the MF irrationality. The model’s attention level λ j for each attribute is represented by Eq. (16).

(16)
{ v j = tanh ( w 1 V j + b 1 ) , λ j = exp ( ReLU ( w 2 v j + b 2 ) ) k = 1 | I u | exp ( ReLU ( w 2 v k + b 2 ) ) .

In Eq. (16), w 1 and w 2 represent weight matrices. V j means the connection vector between the user’s vector q i in the feature space and the user’s attribute a j . b 1 and b 2 mean bias terms. v j means the K-dimensional vector mapped from V j . I u means the user attributes set. The user’s final latent feature vector is represented by Eq. (17).

(17)
p u = ϕ pooling ( p u , λ j a j ) .

In Eq. (17), represents two vectors’ product element by element. To improve the interaction quality between vectors, paired pooling is utilized to encode features’ correlation in the latent feature space. Paired pooling takes into account the interactions within the set of user attribute features, can further improve user vectors’ interaction quality by multiplying them element by element. Paired pooling is represented by Eq. (18).

(18)
p u = ϕ pairwise ( p u , I ˜ u ) = j = 1 | I ˜ u | p u ( λ j a j ) + j = 1 | I ˜ u | j = j + 1 | I ˜ u | ( λ j a j ) ( λ j a j ) .

To make the model highly nonlinear, a fully connected layer was also added. After linear combination through fully connected layers, the final prediction result is represented by Eq. (19).

(19)
y ^ u i = σ L ( σ 1 ( w 1 ( p u q i ) + b 1 ) ) .

In Eq. (19), L represents the quantity of fully connected layers. p u means the user’s ultimate characteristics. σ is the ReLu activation function. To obtain learners’ rating predictions for learning resources, the output layer’s loss function is Mean Squared Error (MSE). Finally, to accelerate the convergence speed of the model, Adam, which has high computational efficiency, is utilized as the optimizer to enhance the loss function. Adam’s calculation is represented by Eq. (20).

(20)
{ m t = β 1 m t 1 + ( 1 β 1 ) g t , v t = β 2 v t 1 + ( 1 β 2 ) g t 2 .

In Eq. (20), m t represents the first-order momentum of the current gradient. β 1 means first-order matrix exponential decay rate. g t means the current gradient value. v t means the second-order momentum of the current gradient. β 2 means second-order matrix exponential decay rate. Since m t and v t are both zero vectors at the beginning, bias correction is required, represented by Eq. (21).

(21)
{ m ^ t = m t / ( 1 β 1 t ) , v ^ t = v t / ( 1 β 2 t ) .

In Eq. (21), m ^ t means gradient weighted average. v ^ t means that there is a bias in gradient weighting. The optimized parameter θ t is represented by Eq. (22).

(22)
θ t = θ 0 ( α / ( 10 8 + v ^ t ) ) m ^ t .

In Eq. (22), θ 0 represents the initial parameter. α means learning rate.

3. Results

The research mainly analyzed from two aspects. Firstly, the LSTM-based knowledge tracking model was analyzed. Then, the LRR model’s recommendation performance based on CF was analyzed.

3.1. Analysis of the Effect of Knowledge Tracking Model Based on LSTM

To validate the knowledge tracking model based on LSTM, the maximum sequence length is 100, the learning rate is 0.001, the quantity of hidden layers is 128, and the batch size is 64. Assisment2012 and RAIEd2020, with 80% as the training set and 20% as the testing set. Assisment2012 contains 46674 learners, 177654 questions, and 708631 interactions. The RAIEd2020 dataset is a large dataset. 1500000 records were selected for the experiment. After preprocessing to remove redundant information, 393656 learners, 13523 questions, and 1459846 interactions were included. The Area Under the Curve (AUC) was utilized as the evaluating metric. The new means was compared with five advanced models: Deep Knowledge Tracking (DKT), Dynamic Key Value Memory Network (DKVMN), Self-Attention Knowledge Tracing (SAKT), Learning Process-consistent Knowledge Tracing (LPKT), and Convolutional Knowledge Tracking (CKT). Fig. 3 shows six models’ Receiver Operating Characteristic (ROC) curves. From Fig. 3(a), this proposed means had the highest AUC of 82.37% on the Assisment2012 dataset. Next was LPKT, with an AUC of 77.68%. DKVMN performed the worst. From Fig. 3(b), on the RAIEd2020 dataset, the proposed model still had the highest AUC of 81.54%. The knowledge tracking means using LSTM demonstrated nice application in AUC and had superiority.

Fig. 3. Comparison of AUC values for six models.

../../Resources/ieie/IEIESPC.2025.14.6.815/fig3.png

To investigate how hidden layers affected this model, experiments were conducted with the quantity of hidden layers set to 50, 75, 100, and 125. Fig. 4 shows the comparison results of AUC indicators for these six models mentioned above under different quantities of hidden layers. From Fig. 4(a), on the Assisment2012 dataset, as the hidden layers increased, all six models’ AUC increased. The proposed model’s AUC remained the highest, reaching 76.27% when the quantity of hidden layers was 125. From Fig. 4(b), on the RAIEd2020 dataset, the proposed model’s AUC was significantly higher than other means, showing an AUC of 77.52% when the quantity of hidden layers was 125. The increase of hidden layers could improve the model performance to a certain extent.

Fig. 4. Comparison of AUC indicators under different numbers of hidden layers.

../../Resources/ieie/IEIESPC.2025.14.6.815/fig4.png

To investigate the proposed model’s improvement effect, ablation experiments were conducted. The AUC of models with removed learning gates (A), removed forgetting gates (B), removed attention mechanisms (C), and the complete model (D) were compared. Fig. 5 shows the ablation experiment results. On both datasets, the complete model performed the best in terms of AUC metrics. The models without attention mechanism had the lowest AUC, which were 72.14% and 70.46%, respectively. Therefore, attention mechanisms could validly enhance model. The proposed improvement strategy could validly enhance the model. The improvement effect of attention mechanism was the most significant.

Fig. 5. Results of ablation experiment.

../../Resources/ieie/IEIESPC.2025.14.6.815/fig5.png

3.2. Performance Analysis of Learning Resource Recommendation Model

To validate the CF-based LRR means, the FrcSub dataset was utilized and divided into training and testing sets in a 7 : 3 ratio. The FrcSub dataset is a small sample set containing 536 scoring data points, with 20 questions and 8 knowledge points examined. For each student in the test set, 40%, 50%, and 60% of the test questions were selected as the recommended test set, with the remaining questions as the training set. The study utilized the PyTorch framework to build a model, with GMF’s embedding layer dimension of 16 and the learning rate of 0.001. This proposed model was compared with traditional NMF, NFM, and Probability Matrix Factorization-Cognitive Diagnosis (PMF-CD) recommendation models that combined probability matrix factorization and cognitive diagnosis. Recommendation precision, recall, and F1 score were utilized as evaluation indicators. Table 1 shows four models’ recommended results on difficult test questions. This new means performed the best in recommending precision, recall, and F1 score. When the proportion of the test set to be recommended was 60%, this new model’s precision, recall, and F1 score were 94.82%, 92.87%, and 93.74%, respectively. The LRR model based on CF showed high recommendation accuracy and performance.

Table 1. Recommended results on difficult test questions.

Model Index Proportion of test question sets to be recommended
40% 50% 60%
NMF Precision/% 88.42 90.89 92.17
Recall/% 82.36 89.48 90.85
F1 score/% 85.31 90.69 91.64
NFM Precision/% 86.95 89.42 90.86
Recall/% 81.29 88.35 89.27
F1 score/% 84.32 89.48 90.16
PMF-CD Precision/% 89.83 90.84 93.27
Recall/% 86.42 81.09 81.52
F1 score/% 87.96 85.74 86.93
Our Precision/% 91.65 92.57 94.82
Recall/% 91.26 94.35 92.87
F1 score/% 91.64 93.59 93.74

The Math dataset contains final exam questions from multiple high schools, covering 4209 students, 415 exercises, and 118 knowledge points. To further validate the proposed model’s superiority, the Math dataset was utilized for testing and separated into training, testing, and validating sets in 7 : 2 : 1. This study utilized accuracy and RMSE as evaluation indicators to compare with three currently advanced models, DKT, DKVMN, and DKT-CF. Fig. 6 shows four models’ test results on the Math dataset. From Fig. 6(a), this new means had the highest recommendation accuracy of 93.45%. Next was the DKT-CF, with a recommendation accuracy of 84.29%. DKT had the lowest recommendation accuracy. From Fig. 6(b), the proposed model had lowest RMSE, at 19.93%. The LRR model based on CF had high recommendation accuracy and certain superiority.

Fig. 6. Test results on the Math dataset.

../../Resources/ieie/IEIESPC.2025.14.6.815/fig6.png

Fig. 7 shows the comparison results of exercise recommendation diversity indicators for four models. On both datasets, the proposed model’s LRR list had the best diversity. The CF-based LRR model could effectively improve the LRR diversity. The larger the dataset, the better the recommendation diversity.

Fig. 7. Comparison of diversity indicators.

../../Resources/ieie/IEIESPC.2025.14.6.815/fig7.png

4. Discussion and Conclusion

In order to provide effective learning resource recommendations, a knowledge tracking means using LSTM and an LRR means using CF are constructed. The results indicate that the proposed knowledge tracking model performed better in precision, recall, and F1 score, with scores of 92.43%, 91.37%, and 92.16%, respectively. On the Assisment2012 dataset, the proposed model had the highest AUC of 82.37%. On the RAIEd2020 dataset, the proposed model still had the highest AUC at 81.54%. On the Assisment2012 dataset, the proposed model consistently had the highest AUC, with an AUC of 76.27% when the quantity of hidden layers was 125. On the RAIEd2020 dataset, when the quantity of hidden layers was 125, the proposed model’s AUC was 77.52%. The complete model performed the best in terms of AUC. The AUC of the models without attention mechanism were the lowest, at 72.14% and 70.46%, respectively. The proposed LRR model performed the best in recommending precision, recall, and F1 score. When the proportion of the test set to be recommended was 60%, this new approach’s precision, recall, and F1 score were 94.82%, 92.87%, and 93.74%, respectively. However, the constructed knowledge tracking model tends to learn embedded representations of problems and knowledge from one-hot. The actual data may contain more information such as images. Therefore, in future research, data from multiple patterns should be further utilized to break through the boundaries of current knowledge tracking methods.

References

1 
Salas-Pilco S. Z., Yang Y., Zhang Z., 2022, Student engagement in online learning in Latin American higher education during the COVID-19 pandemic: A systematic review, British Journal of Educational Technology, Vol. 53, No. 3, pp. 593-619DOI
2 
Raj N. S., Renumol V. G., 2022, A systematic literature review on adaptive content recommenders in personalized learning environments from 2015 to 2020, Journal of Computers in Education, Vol. 9, No. 1, pp. 113-148DOI
3 
Afsar M. M., Crump T., Far B., 2022, Reinforcement learning based recommender systems: A survey, ACM Computing Surveys, Vol. 55, No. 7, pp. 1-38DOI
4 
Hasan M. R., Ferdous J., 2024, Dominance of AI and machine learning techniques in hybrid movie recommendation system applying text-to-number conversion and cosine similarity approaches, Journal of Computer Science and Technology Studies, Vol. 6, No. 1, pp. 94-102DOI
5 
Yu J., Xia X., Chen T., Cui L., Hung N. Q. V., Yin H., 2024, XSimGCL: Towards extremely simple graph contrastive learning for recommendation, IEEE Transactions on Knowledge and Data Engineering, Vol. 36, No. 2, pp. 913-926DOI
6 
Papadakis H., Papagrigoriou A., Panagiotakis C., Kosmas E., Fragopoulou P., 2022, Collaborative filtering recommender systems taxonomy, Knowledge and Information Systems, Vol. 64, No. 1, pp. 35-74DOI
7 
Wu L., He X., Wang X., Zhang K., Wang M., 2023, A survey on accuracy-oriented neural recommendation: From collaborative filtering to information-rich recommendation, IEEE Transactions on Knowledge and Data Engineering, Vol. 35, No. 5, pp. 4425-4445DOI
8 
Fkih F., 2022, Similarity measures for collaborative filtering-based recommender systems: Review and experimental comparison, Journal of King Saud University-Computer and Information Sciences, Vol. 34, No. 9, pp. 7645-7669DOI
9 
Zhou X., Sun A., Liu Y., Zhang J., Miao C., 2023, Selfcf: A simple framework for self-supervised collaborative filtering, ACM Transactions on Recommender Systems, Vol. 1, No. 2, pp. 1-25DOI
10 
Rao A., Plank P., Wild A., Maass W., 2022, A long short-term memory for AI applications in spike-based neuromorphic hardware, Nature Machine Intelligence, Vol. 4, No. 5, pp. 467-479DOI
11 
Duan J., Zhang P., Qiu R., Huang Z., 2023, Long short-term enhanced memory for sequential recommendation, World Wide Web, Vol. 26, No. 2, pp. 561-883DOI
12 
Saraswat M., Srishti , 2022, Leveraging genre classification with RNN for book recommendation, International Journal of Information Technology, Vol. 14, No. 7, pp. 3751-3756DOI
13 
Wang K., Wang X., Lu X., 2023, POI recommendation method using LSTM-attention in LBSN considering privacy protection, Complex & Intelligent Systems, Vol. 9, No. 3, pp. 2801-2812DOI
14 
Liu C., Zhang Y., Sun J., Cui Z., Wang K., 2022, Stacked bidirectional LSTM RNN to evaluate the remaining useful life of supercapacitor, International Journal of Energy Research, Vol. 46, No. 3, pp. 3034-3043DOI
15 
Hunt K. M., Matthews G. R., Pappenberger F., Prudhomme C., 2022, Using a long short-term memory (LSTM) neural network to boost river streamflow forecasts over the western United States, Hydrology and Earth System Sciences, Vol. 26, No. 21, pp. 5449-5472DOI
16 
Liu Y., Chen L., Li C., Liu X., Zhou W., Li K., 2023, Long-term and short-term memory networks based on forgetting memristors, Soft Computing, Vol. 27, No. 23, pp. 18403-18418DOI
17 
Wu J., Wang X., Gao X., Chen J., Fu H., Qiu T., 2024, On the effectiveness of sampled softmax loss for item recommendation, ACM Transactions on Information Systems, Vol. 42, No. 4, pp. 1-26DOI
18 
Terrin É., Triventi M., 2023, The effect of school tracking on student achievement and inequality: A meta-analysis, Review of Educational Research, Vol. 93, No. 2, pp. 236-274DOI
19 
Zhao Q., Zhao H., Zheng K., Wang J., 2022, HyperAttentionDTI: Improving drug–protein interaction prediction by sequence-based deep learning with attention mechanism, Bioinformatics, Vol. 38, No. 3, pp. 655-662DOI
20 
Ramaswamy S. L., Chinnappan J., 2022, RecogNet-LSTM+ CNN: A hybrid network with attention mechanism for aspect categorization and sentiment classification, Journal of Intelligent Information Systems, Vol. 58, No. 2, pp. 379-404DOI
21 
Chen Y., Zhang C., Liu C., Wang Y., Wan X., 2022, Atrial fibrillation detection using a feedforward neural network, Journal of Medical and Biological Engineering, Vol. 42, No. 1, pp. 63-73DOI
22 
Mikhaylov A., Bhatti I. M., Dinăer H., Yüksel S., 2024, Integrated decision recommendation system using iteration-enhanced collaborative filtering, golden cut bipolar for analyzing the risk-based oil market spillovers, Computational Economics, Vol. 63, No. 1, pp. 305-338DOI
23 
Zhang H., Luo F., Wu J., He X., Li Y., 2023, LightFR: Lightweight federated recommendation with privacy-preserving matrix factorization, ACM Transactions on Information Systems, Vol. 41, No. 4, pp. 1-28DOI
24 
Purohit J., Dave R., 2023, Leveraging deep learning techniques to obtain efficacious segmentation results, Archives of Advanced Engineering, Vol. 1, No. 1, pp. 11-26DOI
25 
NoulapeuNgaffo A., Choukair Z., 2022, A deep neural network-based collaborative filtering using a matrix factorization with a twofold regularization, Neural Computing and Applications, Vol. 34, No. 9, pp. 6991-7003DOI

Author

Xiaoying Zhu
../../Resources/ieie/IEIESPC.2025.14.6.815/au1.png

Xiaoying Zhu was born in April 1982, female, from Weihai, Shandong, Han ethnicity. She obtained a Bachelor’s degree in Educational Technology from Tianjin Sport University in 2006 and a Master’s degree in Educational Technology from Jiangsu Normal University in 2010. Research direction: Modern educational technology. From 2010 to 2024, worked as a full-time teacher and lecturer at Yantai Nanshan University. She has published 1 academic paper, edited 2 textbooks, conducted 3 research projects, and obtained 1 invention patent.

Xiaojing Guo
../../Resources/ieie/IEIESPC.2025.14.6.815/au2.png

Xiaojing Guo was born in 1979, female, from Fuxin City, Liaoning Province, Han ethnicity. She obtained a Bachelor’s degree in Computer Science and Technology from Liaoning University of Engineering and Technology in 2003 and a Master’s degree in Computer Technology from Qingdao University in 2009, with a research focus on computer technology. Since 2003, She has been a teacher at Yantai Nanshan University. She has published multiple academic papers, including one core paper at Peking University and one SCI four zone paper. Also served as deputy editor in chief or participated in the compilation and publication of six textbooks, as well as two department level scientific research projects.

Xue Zhang
../../Resources/ieie/IEIESPC.2025.14.6.815/au3.png

Xue Zhang was born in May 1982, female, from Anshan City, Liaoning Province, Han ethnicity. She graduated with a Bachelor’s degree in Educational Technology from Anshan Normal University in July 2005, a Master’s degree in Educational Technology from Jiangsu Normal University in July 2010, and a PhD in Educational Management and Administrative Leadership from Thonburi University in Bangkok in June 2022. She is currently studying. From October 2012 to present, worked as a lecturer at Anshan Normal University, College of Education Science and Technology. She has Published 8 academic papers, published 1 academic work and textbook, conducted 4 research projects, and won 2 academic awards.