Mobile QR Code QR CODE

2024

Acceptance Ratio

21%


  1. (School of Media and Art Design, Guangzhou Huali College, Guangzhou, 510000, China)



Asymptotic regularization, Image restoration, Low-rank matrix recovery, Singular value entropy function, Visual art

1. Introduction

Visual art is an important component of human culture. Art image restoration is a crucial step in protecting and inheriting these cultural heritage sites [1]. As an important carrier of information transmission and aesthetic expression, the integrity and quality of visual art images are crucial. However, many precious visual art images are damaged or missing due to various reasons, such as long history, improper preservation, war damage, etc [2]. Traditional repair methods often rely on manual repair and painting techniques, which are not only costly but also difficult to ensure the repair effect [3]. With the continuous development of artificial intelligence technology, data repair technology has gradually become an effective means to solve this problem. The regularized Low-Rank Matrix (LRM) restoration algorithm is an emerging data restoration technique with powerful capabilities for denoising, filling, and maintaining overall style. In the field of visual arts restoration, regularized low-rank matrix restoration algorithm can be applied to a variety of specific restoration tasks. For example, when restoring an ancient fresco, the color may fade, the pattern may be blurred, or parts of it may be damaged due to age. By dividing the mural image into multiple small blocks, the algorithm can identify similar blocks in the image and form them into a block matrix. In this matrix, each similar block acts as a column vector, and through the similar block search and weight update strategy, the algorithm is able to identify which blocks are visually similar, thus preserving the original style and details of the mural during restoration. The regularized low-rank matrix recovery algorithm is also suitable for a painting that has become blurred due to age. Certain areas of the painting may become unrecognizable due to oxidation or physical damage. By constructing block matrix and using adjacency matrix to describe the similarity between image blocks, the algorithm optimizes the combination of image blocks in the iterative process. In this way, the algorithm can not only restore the clarity of the painting, but also maintain the consistency of its artistic style and color [4,5]. Therefore, a low-rank image restoration algorithm ground on asymptotic regularization is proposed, which classifies image blocks and forms a similar block matrix through similar block search. Then, a low-rank approximation algorithm using asymptotic regularization is introduced to achieve low-rank approximation of the matrix. Finally, the image blocks are re-aggregated to obtain the repaired image. The repair method ground on regularized LRM restoration algorithm can handle a wider range of damage situations, including noise, blur, missing pixels, etc. The research has four parts. The first part is the literature review section, which discusses the current research status of regularized LRM restoration algorithms and visual art image restoration. The second part proposes a low-rank image restoration algorithm based on asymptotic regularization singular value function. Part three verifies its effectiveness and performance through experiments. Part four is the conclusion, which summarizes the research results.

2. Related Works

In recent years, regularized LRM restoration algorithms have received widespread attention in data optimization processing and noise reduction. For LRM recovery and non-convex discontinuous optimization problems, Tao et al. proposed an alternating maximization proximal method with extrapolation to find initial factor pairs with fewer non-zero columns. Subsequently, they utilized smooth non-convex loss for alternating maximization and minimization with extrapolation, thereby obtaining low error and rank solutions [6]. When exploring the rank norm relaxation and model optimization in low-rank regularization, Hu et al. reviewed the latest developments in the two problems faced by all low-rank regularization models. The non-convex relaxation helped alleviate the penalty bias. Compared with convex relaxation, it had more advantages. Different optimization algorithms have their own characteristics. These findings provide important clues for improving the performance of low-rank regularization models [7]. Charisopoulos et al. proposed a standard method ground on non-smooth optimization for recovering LRM from noisy linear measurements. It showed that the non-smooth formula had natural robustness to outliers, which is superior to other methods [8]. When solving the LRM completion iteration, Bauch et al. proposed an iterative least squares method. The sparse least squares problem with lower computational cost in each iteration was solved. This method could recover LRM from few observations, which was stable for additive noise [9]. Zilber et al. built a restoration strategy ground on Gaussian-Newton linearization to address a series of issues related to LRM restoration. This algorithm implicitly maintained an approximate balance of the factor matrix throughout the entire iteration process and performed better than several popular methods in completing uniformly sampled matrices, especially in situations where the observation values were close to the information limit [10].

Recently, influenced by intelligent technology, considerable progress has been made in restoring and preserving fine art. Mei et al. built a novel pyramid attention module to address the insufficient self-similarity utilization of deep convolutional neural networks in image restoration. This module captured remote feature correspondences. The experimental results showed that the algorithm exhibited excellent accuracy and visual quality in image denoising, and mosaic removal [11]. The quality of data obtained from remote sensing imaging sensors is often affected by various types of noise and human factors. Rasti et al. built a restoration strategy based on remote sensing images. Experiments showed that this algorithm had significant effects in restoring remote sensing images [12]. Considering the long time required for image restoration processing, Jiu et al. built a deep learning based image classification and segmentation method. Compared with other advanced methods, this algorithm performed well on conventional datasets, which was better on complex datasets [13]. To overcome the limitations of plug and play image restoration, Zhang et al. incorporated deep denoising priors as modules into iterative algorithms based on semi-quadratic splitting. This plug and play image restoration method based on deep denoising prior was significantly superior to other methods [14]. Traditional image restoration methods based on structural sparse representation often directly fit or transform damaged image data, resulting in poor restoration results. Therefore, Zha et al. built a new mixed structure sparse error model. Experimental results showed that this scheme outperformed other methods [15].

In summary, the regularized LRM restoration algorithm has shown significant advantages in the field of visual art image restoration. Although significant progress has been made in related research, there are still some challenges and issues that need to be further addressed. For example, how to better handle complex image structures and texture information, how to improve restoration efficiency and accuracy, and how to expand and apply to a wider range of visual art image types. Therefore, an LRM recovery method ground on regularized singular values is proposed. Singular value decomposition is a powerful matrix analysis tool, combined with regularization techniques, which can remove noise while maintaining the main information, and improve the recovery effects.

3. The LRM Restoration Algorithm Ground on the Regularized Singular Value Function

Introducing low-rank and sparse constraints can effectively restore the low-rank structure of images and remove noise and outliers. Meanwhile, the regularization term further enhances the robustness and generalization ability, making it better able to adapt to different types of visual art image restoration tasks.

3.1. Low-rank Visual Art Image Restoration Based on Progressive Regularization

LRM restoration has been extensively applied in many fields such as image restoration and deblurring. In an image data matrix, each row or column can be expressed by other rows or columns, meaning that this matrix contains a lot of redundant information [16]. In a sense, image restoration is a typical LRM restoration problem. The image restoration is displayed in Fig. 1.

Fig. 1. Image recovery process.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig1.png

In Fig. 1, the image restoration is to repair and eliminate blurring of the image. Although image restoration techniques have been able to effectively repair or eliminate blurring in images, there is still a significant gap in clarity in the restored and the original image due to various reasons. Therefore, the study introduces a weight update strategy. The weighted kernel norm minimization method is used for matrix filling, as shown in Eq. (1).

(1)
${\mathop{\min }\limits_{X}} \left\| X\right\| _{\omega ,*} s.t.p_{\Omega } (X)=p_{\Omega } (Y). $

In Eq. (1), $\Omega$ represents a binary indicator matrix of the same size as $Y$. $p_{\Omega}(Y)$ represents the Hadamard product of the indicator matrix and the observation matrix $Y$. This limitation indicates that the estimation matrix $X$ matches the lossless information of the observation matrix $Y$. Adjacency matrix is a matrix used to describe the nearest neighbor relationships between nodes. This algorithm takes a one-dimensional array to store all points in the graph, while a two-dimensional array stores the relationships between points [17]. The adjacency matrix is specifically shown in Fig. 2.

Fig. 2. Adjacency matrix diagram.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig2.png

In Fig. 2, $+1$ indicates mutual perception, while blank indicates inability to perceive each other. For an undirected graph, its adjacent matrices must be symmetric, its main diagonal must be 0, and its sub diagonal may not necessarily be 0. In an undirected graph, the degree of any vertex refers to the $i$-th non-zero element quantity. In a directed graph, adjacent matrices may not necessarily be symmetric. Among them, the number of outgoing lines of node $i$ in a directed graph is the non-zero element quantity in row $i$. In a directed graph, the outgoing lines of node $i$ are non-zero elements [18]. LRM recovery takes the rank of the matrix as the sparse measure. The linear measurement is shown in Eq. (2).

(2)
$y=A(X) .$

In Eq. (2), $y$ represents the observed data. $A$ represents the linear operator. $X \in R^{m \times n}$, $y \in R^p$. The definition of the linear operators is specifically shown in Eq. (3).

(3)
$A(X)\buildrel\Delta\over= [<A_{1} ,X>,~...,~<A_{P} ,X>]^{T} . $

In Eq. (3), $\langle A_i, X \rangle$ represents the inner product of the matrix. From a mathematical perspective, the low-rank restoration is reduced to a rank minimization, as displayed in Eq. (4).

(4)
${\mathop{\min }\limits_{X\in R^{m\times n} }} rank(X)~~\text{s.t.}~~y=A(X) .$

In Eq. (4), $rank(X)$ signifies the rank function of matrix $X$. The update process of the weight vector is set to the following form, as shown in Eq. (5).

(5)
$\omega _{i}^{l+1} =C(\sigma _{i} (X_{l} )+\varepsilon _{l} )^{\frac{p}{2} -1} . $

In Eq. (5), $\varepsilon_l$ represents a sequence that gradually decreases in the range of 0 to 1 and approaches 0. Traditional regularization algorithms require $\varepsilon$ to take a small optimal solution and maintain a constant during iteration. This algorithm takes $\varepsilon_l = 1$ as the initial value, increasing from high to low. The maximum value includes some suboptimal solutions first. During the decreasing process, it continuously shrinks to gradually approach the optimal solution. This regularization method can improve the sparsity of rank and achieve better results. A block-based image restoration algorithm is introduced to address the non-local self-similarity characteristics of images. This algorithm mainly includes similar block matching, matrix low-rank approximation, image reconstruction, and iterative diffusion. In the block matching stage, the missing image $Y$ is first divided into $N$ overlapping blocks. For each block, the similarity measure is used to find other blocks that are similar to it. Next, each similar block is transformed into a column vector form, and a block matrix is formed by connecting all column vectors and stacking them. The process of building a block matrix is shown in Fig. 3.

Fig. 3. The process of building a block matrix.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig3.png

In Fig. 3, in the process of building block matrix, the art image is first divided into several small blocks, which are then searched for other similar blocks in the image through similarity search algorithm. Each similar block is converted into column vector form, and these column vectors are connected and stacked to form block matrix. In this process, the adjacency matrix plays an important role, which describes the adjacency relationship between nodes. For image restoration, nodes here can be image blocks, and the adjacency matrix reflects the similarity or correlation between these image blocks. In this way, the adjacency matrix helps the algorithm identify and exploit similar areas in the image, thereby preserving the structure and texture information of the image during the repair process. In image restoration, adjacency matrix not only helps to search and match similar blocks, but also helps to optimize the combination of image blocks in the iterative process, so that the repaired image can better retain the features and styles of the original image.

3.2. Image Restoration Based on Singular Value Entropy Function

Singular value arrangement in image restoration uses singular value decomposition to decompose an image, obtain a set of singular values, and arrange these singular values in descending order. In image restoration, singular value permutation is used to restore or reconstruct damaged or missing image parts. Specifically, singular value decomposition can decompose a complex image matrix into several simple matrices, of which the singular value matrix is an important component. These singular values can reflect important features and structural information. Therefore, the main features can be preserved by retaining larger singular values. For smaller singular values, they can be ignored or set to zero to remove noise, fill missing parts, or perform other image restoration operations [19,20]. Singular value decomposition and arrangement are shown in Fig. 4.

Fig. 4. Singular value decomposition and arrangement.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig4.png

In image processing, scratches are equivalent to collision noise at different positions in the spatial domain, making the disorder of pixel values in the image more severe. On the contrary, due to the random loss of pixel values following a uniform distribution, the confusion caused is not significant. For some systems with high chaos, the entropy minimization method can be effectively improved. To address this issue, an image repair method based on the entropy weight is designed to further optimize scratches. The specific definition of entropy function is shown in Eq. (6).

(6)
$h(x)=-\sum _{i}\frac{\left|x_{i} \right|}{\left\| x\right\| _{1} } \log \frac{\left|x_{i} \right|}{\left\| x\right\| _{1} } . $

In Eq. (6), $x \in R$. Information theory suggests that the maximum entropy of a random variable exists when it has a uniform distribution. In other words, if the numerical value of a random variable is small, its entropy value is lower. That is, its vector $x$ is sparser. Based on the low-rank approximation, the singularity performance of the repaired image can be improved, resulting in better sparsity and more concentrated main information, thereby improving the quality of image restoration. Therefore, the study introduces another rank approximation method, which is the entropy function of singular values. For an LRM $X$, the rank minimization approximation is equal to solving the minimization problem of the corresponding singular value entropy function. This approach can better capture and utilize the low-rank structure of matrices, thereby optimizing the image restoration effect. The rank minimization approximation is specifically shown in Eq. (7).

(7)
${\mathop{\min }\limits_{X}} \lambda h(\sigma (X))~~\text{s.t.}~~Y=X+E .$

Under certain fidelity limitations, the equivalent result of the above problem is displayed in Eq. (8).

(8)
${\mathop{\min }\limits_{X}} \lambda h(\sigma (X))+f(Y,X,E) .$

The entropy function minimization can be transformed into a weighted kernel norm minimization problem. This entropy-based weight is applied to the image restoration framework to perform block processing on the image. Unlike global low-rank, image restoration is achieved by utilizing the low-rank of similar block matrices and solving the weighted kernel norm minimization problem corresponding to entropy gradient weights. It better capture the local structure and maintain the integrity of these structures during the repair process, thereby improving the repair effect. The consensus weight method can cause the repair result to be overly smooth, so the weight value is specifically shown in Eq. (9).

(9)
$w_{j} =\left\{\begin{aligned} 1-\frac{rank(X)}{n} ,rank(X)<n,\\ \frac{1}{n} \quad\quad\quad\quad\quad,rank(X)=n. \end{aligned}\right. $

During the iteration process, an iterative regularization step is introduced to reinforce the improved estimation results by back-projecting the residual image onto the estimated image. Specifically, at the beginning of each iteration, the estimated image output from the previous iteration is updated and then used as the input image for the next iteration. This method can utilize the information from the previous iteration to gradually optimize the image, thereby improving estimation accuracy and stability, as shown in Eq. (10).

(10)
$X_{(k)} =\hat{X}_{(k-1)} +\delta (Y-P_{\Omega } (\hat{X}_{(k-1)} )) .$

In Eq. (10), $\hat{X}$ represents the repair of damaged images. $\delta$ is an adjustment parameter. Peak Signal-to-Noise Ratio (PSNR) is an objective standard used to assess image or video quality. PSNR is frequently used to compare the quality differences between original images and compressed, processed, or reconstructed images, as shown in Eq. (11).

(11)
$PSNR=10\times \log _{10} \left(\frac{MAX^{2} }{\frac{1}{N} \sum _{j=1}^{N}(x_{j} -\hat{x}_{j} )^{2} } \right) .$

Parameters $M$ and $\delta$ directly affect the effectiveness of the algorithm. The impact parameters $M$ and $\delta$ on PSNR is shown in Fig. 5.

Fig. 5. Effect of δ on PSNR.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig5.png

Finally, it is decided to choose $M$ as 5 and $\delta$ as 0.4 to complete the image restoration task. Mean Absolute Error (MAE) and Root Mean Square Error (RMSE) are taken as evaluation metrics to measure the error in the repair results. MAE is a commonly used error measure that represents the average absolute deviation in the repaired and the true. A small MAE indicates that the repair result is approached to the true value and the repair effect is better. The MAE is shown in Eq. (12).

(12)
$MAE=\frac{1}{n} \sum _{i=1}^{n}\left|\hat{y}_{i} -y_{i} \right| .$

RMSE considers the variance and mean of the repair error. Compared with MAE, RMSE is more sensitive to outliers. A small RMSE value indicates better stability of the repair result and better repair effect. The RMSE is shown in Eq. (13).

(13)
$RMSE=\sqrt{\frac{1}{n} \sum _{i=1}^{n}(\hat{y}_{i} -y_{i} ) } . $

The steps of an LRM image restoration method ground on asymptotic regularization singular value function roughly include image partitioning, constructing similar block groups, singular value decomposition, singular value contraction, reconstructing LRM, image block aggregation, etc., as shown in Fig. 6.

Fig. 6. The specific process steps of the algorithm.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig6.png

In Fig. 6, the image to be restored is divided into overlapping or non-overlapping blocks during the image restoration process. For each image block, other similar blocks are searched and combined into a matrix. The matrix composed of each similar block group is subjected to singular value decomposition. Asymptotic regularization singular value function is used to contract singular values. The singular values are adjusted based on their size and preset thresholds or regularization parameters. The adjusted singular values and corresponding left and right singular vectors are used to reconstruct an LRM. All reconstructed low-rank matrices are converted back into image blocks and placed back in their corresponding positions in the image. If the blocks are overlapping, weighted averaging may be necessary to fuse the overlapping regions. The above process is repeated until the image quality no longer significantly improves. The weight updating strategy uses the weighted kernel norm minimization method to fill the matrix, in which the weight vector gradually decreases in the iterative process, the initial value is large to contain part of the suboptimal solution, and then gradually decreases to approximate the optimal solution. This process enhances the sparsity of the rank, so as to obtain better repair effect. The regularization term effectively restores the low-rank structure of the image and removes noise and outliers by introducing low-rank constraints and sparse constraints. At the same time, the robustness and generalization ability of the algorithm are enhanced, so that it can adapt to different types of visual art image restoration tasks. In the process of restoration, these parameters interact with the low-rank components of the image, and the damaged image is decomposed into low-rank components and sparse components through the low-rank approximation model. The low-rank matrix is estimated by the iterative optimization algorithm, and the progressive regularization strategy is used to gradually approximate the original image. Finally, the repaired low-rank components and sparse components are synthesized to obtain the repaired image.

4. Performance and Image Restoration Effect of LRM Algorithm Based on Asymptotic Regularization Singular Value Function

The study first determines the parameters of the experiment, and then conducts experiments to present the effectiveness of the LRM method based on asymptotic regularization singular value function. Then, the LRM based on asymptotic regularization singular value function is applied to image restoration and its application effect is tested.

4.1. Performance of LRM Based on Asymptotic Regularization Singular Value Function

The LRM based on asymptotic regularization singular value function (Algorithm 1) is compared with other image restoration algorithms. Comparative indicators include relative error, running time, stability, precision, recall, and F1 value. The comparison algorithms include image restoration algorithm based on sparse representation (Algorithm 2), image restoration algorithm ground on matched sample blocks (Algorithm 3), the single-step compression perception (Algorithm 4), and the convolutional neural network (Algorithm 5). Table 1 displays the results.

Table 1. Comparison of various indexes of different algorithms.

Algorithm Relative error Total algorithm time (s) Stability (%) Precision (%) Recall (%) F1 value (%)
Algorithm 1 0.001 28.54 88 93.64 92.36 92.51
Algorithm 2 0.009 39.78 76 67.23 79.85 85.37
Algorithm 3 0.02 25.16 81 78.42 67.53 74.93
Algorithm 4 0.11 39.28 83 72.53 73.54 70.29
Algorithm 5 0.25 45.72 75 76.92 77.08 71.461

In Table 1, the relative error of Algorithm 1 was 0.001, the running time was 28.54s, the stability was 88%, the accuracy was 93.64%, the recall was 92.36%, and the F1 value was 92.51. Overall, various indicator parameters are relatively excellent. The relative error, running time, stability, precision, recall, and F1 value of the other four algorithms are lower and longer than the LRM based on asymptotic regularization singular value function. Therefore, it indicates that the LRM based on asymptotic regularization singular value function has better performance. The Receiver Operating Characteristic (ROC) curve depicts the balance in true positive rate and false positive rate under different classification thresholds. The Area Under the Curve (AUC) quantifies the overall performance of the ROC curve, with values closer to 1 indicating better performance. The ROC, AUC values, and curve generation time of the LRM based on asymptotic regularization singular value function are shown in Fig. 7.

Fig. 7. Comparison of ROC, AUC values and specific running time.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig7.png

In Fig. 7(a), the ROC curve of the LRM based on asymptotic regularization singular value function was good, with an AUC value close to 1, indicating good performance. The running time in Fig. 7(b) ranged from 20 to 50 minutes, indicating that the algorithm had a high complexity, a large dataset, and more computing resources, but it was within a controllable range. The RMSE of the LRM ground on asymptotic regularization singular value function is further evaluated. The other parameters remain unchanged, only changing the size of the missing rate to evaluate the performance of Algorithm 1 and Algorithm 2 under different missing rates, as shown in Fig. 8.

Fig. 8. RMSE comparison of two algorithms.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig8.png

In Fig. 8, the minimum RMSE of Algorithm 1 was 0.1082, and the minimum RMSE of Algorithm 2 was 0.1897, indicating that Algorithm 1 had higher accuracy in reconstructing image data. The RMSE of the two algorithms increases as the sample size decreases. Although the sample size is decreasing, Algorithm 1 can still maintain relatively stable performance. This further demonstrates the stability and superiority of the proposed LRM based on asymptotic regularization singular value function.

The LRM algorithm based on asymptotically regularized singular value function is compared with other advanced image restoration algorithms. Comparison algorithms include Wavelet Deep Neural Network (WDNN), Reference Constrained Image Restoration Algorithm, Wavelet deep neural Network (RCIR), Incremental Transformer Algorithm (IT), and Efficient Low-Rank Matrix Factorization (ELRMF). The accuracy and recall rates of the five algorithms are shown in Fig. 9.

Fig. 9. Accuracy and recall of five algorithms.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig9.png

In Fig. 9(a), as iterations increased, the accuracy of the algorithm stabilized above 0.9 after 50 iterations, which showed the algorithm efficiency and its stability. When predicting positive samples, the algorithm mistakenly identifies the repaired area as undamaged area less often, reduces unnecessary modifications, and maintains the original features of the image. Fig. 9(b) further demonstrates the advantage of the algorithm in terms of recall rate. That is, the algorithm is able to identify most of the actual damaged areas, reducing the risk of missing damaged parts of the original image. In visual art image restoration, any omission can cause the image to lose its artistic value and historical value.

4.2. Image Restoration Effect

To assess the effectiveness of LRM based on asymptotic regularization singular value functions in image restoration, this study applies them to images with 50% missing data and analyzes the PSNR of five image restoration algorithms, as shown in Table 2.

Table 2. PSNR for image repair with 50% missing data.

Graphics Algorithm 1 Algorithm 2 Algorithm 3 Algorithm 4 Algorithm 5
1 0.93 0.58 0.70 0.62 0.80
2 0.91 0.62 0.76 0.65 0.79
3 0.88 0.73 0.73 0.66 0.72
4 0.96 0.71 0.75 0.78 0.73
5 0.92 0.82 0.81 0.71 0.71
6 0.95 0.69 0.68 0.74 0.75
7 0.97 0.75 0.74 0.73 0.77
8 0.89 0.66 0.71 0.79 0.58
Mean value 0.93 0.69 0.74 0.71 0.73

In Table 2, Algorithm 1 had a relatively high PSNR on different images, with an average of 0.93. However, Algorithms 2-5 only had 0.69, 0.74, 0.71, and 0.73, respectively. It indicates that Algorithm 1 has better image restoration quality. The difference from the original image is relatively small. Algorithms 1-3 are used to recover artificial data containing 20%, 30%, 40%, and 50% random deletion rates. 30 independent repeated experiments are conducted. Fig. 10 displays the standard deviation of the PSNR obtained.

Fig. 10. Image repair under different deletion rate.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig10.png

In Fig. 10, Figs. 10(a)-10(d) show the random deletion rates of 20%, 30%, 40%, and 50%, respectively. Algorithm 1 showed the highest accuracy in restoring image data, with an average restoration accuracy of over 85%. This means that Algorithm 1 can effectively recover lost information and maintain high accuracy when dealing with artificial datasets containing missing data with different proportions. The study further analyzes the impact of different node numbers on the recognition rate of Algorithm 1 on two different datasets, as shown in Fig. 11.

Fig. 11. The influence of different number of nodes on different datasets.

../../Resources/ieie/IEIESPC.2025.14.6.753/fig11.png

Fig. 11(a) displays the recognition results of the Face Scrub dataset. Fig. 11(b) displays the recognition results of the Tiny Images Dataset dataset. Algorithm 1 performed better on both datasets. As the number of nodes increased, the final recognition rate stabilized at over 80%. The recognition rates of Algorithm 2 and 3 were not high. This once again indicates that the LRM based on asymptotic regularization singular value function is more effective in the image restoration.

5. Conclusion

The image of visual art is often damaged for various reasons, which urgently needs to be restored. This study combined regularized LRM approximation and singular value function to maintain the global structure and texture details. From the research results, the minimum RMSE of the LRM based on asymptotic regularization singular value function was 0.1082, the relative error was 0.001, the running time was 28.54s, the stability was 88%, the precision was 93.64%, and the recall was 92.36%. The LRM algorithm performed better on different datasets. As the number of nodes increased, the final recognition rate stabilized at over 80%. In randomly missing data of 20%, 30%, 40%, and 50%, the LRM algorithm based on asymptotic regularization singular value function showed the highest accuracy in restoring image data, with an average restoration accuracy of over 85%. This algorithm demonstrates good restoration ability and stability in processing visual art images containing complex textures and details, which has higher accuracy in reconstructing image data. When dealing with artificial datasets containing missing data with different proportions, it is possible to effectively recover lost information while maintaining high accuracy. However, when dealing with visual art images with rich texture details, low-rank approximation may smooth out some important local features, resulting in the repaired image lacking the vividness and detail expressiveness of the original image. Future research can further optimize this method and explore how to better preserve the texture details of images in low-rank restoration processes.

References

1 
Chen X., Qin J., 2021, Regularized Kaczmarz algorithms for tensor recovery, SIAM Journal on Imaging Sciences, Vol. 14, No. 4, pp. 1439-1471DOI
2 
Miao J., Kou K. I., 2021, Color image recovery using low-rank quaternion matrix completion algorithm, IEEE Transactions on Image Processing, Vol. 31, No. 1, pp. 190-201DOI
3 
Zhang X., Lu H., Guo D., Lai Z., Ye H., Peng X., 2022, Accelerated MRI reconstruction with separable and enhanced low-rank Hankel regularization, IEEE Transactions on Medical Imaging, Vol. 41, No. 9, pp. 2486-2498DOI
4 
Yu K., Wang X., Dong C., Tang X., Loy C. C., 2021, Path-restore: Learning network path selection for image restoration, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 44, No. 10, pp. 7078-7092DOI
5 
Yedder H. B., Cardoen B., Hamarneh G., 2021, Deep learning for biomedical image reconstruction: A survey, Artificial Intelligence Review, Vol. 54, No. 1, pp. 215-251DOI
6 
Tao T., Qian Y., Pan S., 2022, Column l2,0-norm regularized factorization model of low-rank matrix recovery and its computation, SIAM Journal on Optimization, Vol. 32, No. 2, pp. 959-988DOI
7 
Hu Z., Nie F., Wang R., Li X., 2021, Low rank regularization: A review, Neural Networks, Vol. 136, No. 1, pp. 218-232DOI
8 
Charisopoulos V., Chen Y., Davis D., Díaz M., Ding L., Drusvyatskiy D., 2021, Low-rank matrix recovery with composite optimization: Good conditioning and rapid convergence, Foundations of Computational Mathematics, Vol. 21, No. 6, pp. 1505-1593DOI
9 
Bauch J., Nadler B., Zilber P., 2021, Rank 2r iterative least squares: Efficient recovery of ill-conditioned low rank matrices from few entries, SIAM Journal on Mathematics of Data Science, Vol. 3, No. 1, pp. 439-465DOI
10 
Zilber P., Nadler B., 2022, GNMR: A provable one-line algorithm for low rank matrix recovery, SIAM Journal on Mathematics of Data Science, Vol. 4, No. 2, pp. 909-934DOI
11 
Me Y., Fan Y., Zhang Y., Yu J., Zhou Y., Liu D., Shi H., 2023, Pyramid attention network for image restoration, International Journal of Computer Vision, Vol. 131, No. 12, pp. 3207-3225DOI
12 
Rasti B., Chang Y., Dalsasso E., Denis L., Ghamisi P., 2021, Image restoration for remote sensing: Overview and toolbox, IEEE Geoscience and Remote Sensing Magazine, Vol. 10, No. 2, pp. 201-230DOI
13 
Jiu M., Pustelnik N., 2021, A deep primal-dual proximal network for image restoration, IEEE Journal of Selected Topics in Signal Processing, Vol. 15, No. 2, pp. 190-203DOI
14 
Zhang K., Li Y., Wuo W., Zhang L., Gool L. V., Timofte R., 2021, Plug-and-play image restoration with deep denoiser prior, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 44, No. 10, pp. 6360-6376DOI
15 
Zha Z., Wen B., Yuan X., Zhou J., Zhu C., Kot A. C., 2021, A hybrid structural sparsification error model for image restoration, IEEE Transactions on Neural Networks and Learning Systems, Vol. 33, No. 9, pp. 4451-4465DOI
16 
Wang H., Li W., Hu L., Zhang C., He Q., 2022, Structural smoothness low-rank matrix recovery via outlier estimation for image denoising, Multimedia Systems, Vol. 28, No. 1, pp. 241-255DOI
17 
Pokala P. K., Hemadri R. V., Seelamantula C. S., 2021, Iteratively reweighted minimax-concave penalty minimization for accurate low-rank plus sparse matrix decomposition, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 44, No. 12, pp. 8992-9010DOI
18 
Sharma P., Bisht I., Sur A., 2023, Wavelength-based attributed deep neural network for underwater image restoration, ACM Transactions on Multimedia Computing, Communications and Applications, Vol. 19, No. 1, pp. 1-23DOI
19 
Tong T., Ma C., Chi Y., 2021, Accelerating ill-conditioned low-rank matrix estimation via scaled gradient descent, The Journal of Machine Learning Research, Vol. 22, No. 1, pp. 6639-6701DOI
20 
Choudhuri S., Adeniye S., Sen A., 2023, Distribution alignment using complement entropy objective and adaptive consensus-based label refinement for partial domain adaptation, Artificial Intelligence and Applications, Vol. 1, No. 1, pp. 43-51DOI

Author

Shuping Lin
../../Resources/ieie/IEIESPC.2025.14.6.753/au1.png

Shuping Lin acquired her master’s degree from Guangdong University of Technology in 2016. She presently holds the positions of Deputy Director at the Lingnan Urban and Rural Culture and Art Research Institute, Academic Director of School, and Lecturer in Design at Guangzhou Huali College. Her research endeavors primarily concentrate on Lingnan urban and rural spatial culture, as well as artistic image processing. Lin has spearheaded 5 research projects and been involved in over ten additional ones. She has been honored with a provincial-level award for outstanding teaching cases and has contributed to the development of two textbooks planned under the “14th Five-Year Plan” for higher education in the field of applied fine arts and design. Furthermore, she has authored more than 20 papers published in national and international academic journals. Her design works have garnered over 20 awards in various design competitions, and she has mentored students who have collectively won more than 150 awards in such competitions.