中国科学院大学学报  2024, Vol. 41 Issue (2): 151-164   PDF    
Robust individualized subgroup analysis
ZHANG Xiaoling, REN Mingyang, ZHANG Sanguo     
School of Mathematical Sciences, University of Chinese Academy of Sciences, Beijing 100049, China; Key Laboratory of Big Data Mining and Knowledge Management, Chinese Academy of Sciences, Beijing 100049, China
Abstract: Subgroup analysis of heterogeneous groups is a crucial step in the development of individualized treatment and personalized marketing strategies. Regression-based approaches are one of the main schools of subgroup analysis, a paradigm that divides predictor variables into two parts with heterogeneous and homogeneous effects and divides the sample into subgroups based on the heterogeneous effects. However, most of the existing regression-based subgroup analysis methods have two major limitations: First, they still consider the sample homogeneous within subgroups and do not fully consider individual effects; Second, the common contamination phenomenon of homogeneous effect variables is not taken into account, which will lead to large bias in the model results. To address these challenges, we propose a robust individualized subgroup analysis. We use a multidirectional separation penalty function to achieve individualized effects analysis for the heterogeneous part of the model and use γ-divergence to obtain robust estimates for the contaminated homogeneous part. We also propose an efficient alternating iterative two-step algorithm, combining coordinate descent and alternating direction method of multipliers (ADMM) techniques to implement this process. Our proposed method is further illustrated by simulation studies and analysis of a skin cutaneous melanoma dataset.
Keywords: subgroup analysis    multidirectional separation penalty    robust regression    variable selection    
稳健的个体化亚组分析
张晓灵, 任明旸, 张三国     
中国科学院大学数学科学学院, 北京 100049; 中国科学院大数据挖掘与知识管理重点实验室, 北京 100049
摘要: 异质群体的亚组分析是实现个体化医疗和个性化营销的关键所在。基于回归的方法是亚组分析的主要流派之一, 这种范式将预测变量分为具有异质效应和同质效应的两部分, 并根据异质变量是否相同将样本分为不同的亚组。然而, 现有的基于回归的亚组分析方法大多有两大局限性: 第一, 它们仍然认为亚组内的样本是同质的, 没有充分考虑个体效应; 第二, 没有考虑到同质变量中常见污染现象, 这将导致模型结果出现较大偏差。为应对这些挑战, 提出一种稳健的个体化亚组分析方法。使用多向分离惩罚函数估计模型异质部分的个体化效应, 并使用γ散度得到同质部分的稳健估计。还提出一种高效的交替迭代的两步算法, 这一方法结合了坐标下降法和交替方向乘子法。数值模拟和对皮肤黑色素瘤数据的分析进一步验证了所提方法的有效性。
关键词: 亚组分析    多向分离惩罚    稳健回归    变量选择    

In recent years, there has been a growing demand to explore individualized models, which have a wide range of applications for personalized medicine, personalized education and personalized marketing. In the era of big data, heterogeneous data is one of the key challenges in data analytics, which is to correctly identify subgroups from a heterogeneous population in order to target treatment or marketing for each subgroup. For example, in the fight against diseases such as cancer, the effectiveness of a new medicine to treat a disease is evaluated for the whole population. However, if there is significant heterogeneity in treatment effects due to genetic variation or environmental influences, then new treatments are likely to be particularly effective for some patient subgroups and ineffective or less effective for others. Therefore, subgroups of patients with desired outcomes need to be identified based on appropriate statistical methods.

In order to solve this problem, we need to identify potential subgroup structures. In general, subgroup identification can be achieved by clustering samples. To group different individuals, Hocking et al.[1] and Lindsten et al.[2] used Lp fused penalties, treating clustering as a problem of penalized regression. Pan et al.[3] and Ma and Huang[4] used a non-convex fused penalty to reduce bias. However, the fused penalty approach focuses on subgroups rather than model selection of individual coefficients. In other words, although existing subgroup analysis methods have explored possible heterogeneous effects across samples, they have not adequately considered individual effects. For example, in genetic studies to identify biomarkers associated with a particular disease, a gene may be a relevant biomarker for one individual in the population but not for other individuals. Furthermore, for different genes, the effects on heterogeneous covariates may vary between individuals. Therefore, individualized variable selection is important, as different individuals may have different sets of biomarker genes. In addition, the rise of precision medicine and personalized marketing strategies has driven us to develop more effective personalized treatments and recommendations by selecting unique characteristics for each individual. The rich collection of data information makes the use of individualized models feasible and convincing, as traditional aggregate models cannot incorporate heterogeneous effects across individuals. Tang et al.[5] proposed an effective method for individualized model selection using a multidirectional separation penalty function to select significant covariates for different individuals and to simultaneously identify subgroups based on the effects of heterogeneous covariates.

Although the individualized model proposed by Tang et al.[5] can realize the feature selection of individuals and the subgroup division of heterogeneous covariates, it is only for ideal data. However, in the real data, especially in the data of genes and diseases, there are often more complex situations. For example, in The Cancer Genome Atlas (TCGA) collection data on skin cucumber melanoma (SKCM) data, studies have shown that[6] the environmental variables are heterogeneous, important to some samples and not important to some samples. And the remaining high-dimensional gene expression variables are homogeneous, and some homogeneous variables are completely unimportant, so variable selection is needed. In addition, due to some technical reasons, there are often some measurement errors in gene expression data, resulting in long-tailed distributions or contamination. Therefore, robust methods need to be used to process this part of data. The method of individualized model[5] can not select homogeneous variables, and it is not suitable for dealing with contaminated data.

Outliers in data are often encountered with biomedicine, image processing and other areas. However, traditional linear models require assumptions and expectations of the correct variables. The data may be contaminated due to inadequate access to information, errors in subjective judgement and measurement errors, which will lead to bias in the estimates derived from traditional linear models. Robust estimation methods have thus been developed to reduce the impact of outliers on data analysis. The initially used is the least absolute deviation (LAD), but the calculation is more complex. In recent years, the divergence-based methods have been developed. Two of the more common methods are density power divergence and γ-divergence. The MDPD (minimum density power divergence) method was first proposed by Basu et al.[7], it is mainly used to solve the problem of parameter estimation for density distributions. Many statistical models constructed on the basis of density power divergence have been shown in the literature to exhibit excellent robustness. Fujisawa and Eguchi[8] proposed a robust parameter estimation method for Gaussian mixture models based on density power divergence; Ghosh and Basu[9] and Durio and Isaia[10] have extended the method to regression problems and have shown good robustness; Zang et al.[11]proposed a high-dimensional robust parameter estimation method based on density power divergence and applied it to a high-dimensional linear regression model with multiple response variables. The MDPD method is more robust than the LAD method and can deal well with data contamination and heavy-tailed distribution of residuals. Jones et al.[12] first proposed the γ-divergence method for robust estimation of parameters from a single distribution. It was later extended to robust regression methods for low-dimensional data by Fujisawa and Eguchi[13], so that the estimates obtained have strong robustness, and the potential bias can be sufficiently small even in the case of heavy contamination. None of the other robust methods can achieve these properties, and the estimates are affected by the proportion of contaminated data. It has been shown that the estimation has good statistical and numerical advantages. Kawashima and Fujisawa[14] proposed a robust sparse regression based on γ-divergence to establish robust properties from Pythagorean relations. Hung et al.[15] proposed γ-logistic regression. As γ-logistic regression can ignore the bias caused by the contamination distribution and the proportion of contamination, the probability of the wrong category in the model does not need to be modelled. The MDPD logistic regression was also compared with the γ-logistic regression, showing that the γ-logistic regression has stronger robustness. Ren et al.[16] used γ-divergence on a high-dimensional generalized linear model to deal with multiple types of anomalous responses and rigorously established consistency in variable selection and estimation bounds. However, γ-divergence has not been used for individualized subgroup analysis.

The main contribution of this paper is as follows. First, we propose a robust regression-based individualized subgroup analysis method. Specifically, a multidirectional separation penalty is introduced to analyze heterogeneous individualized effects, and γ-divergence and regularization techniques are introduced to simultaneously achieve variable selection and robust analysis of homogeneous effects with possible contamination data. Second, an effective twostep algorithm with stepwise alternating iterations, combining coordinate descent and ADMM techniques, is proposed to address the difficulties of objective function optimization. Numerical simulations demonstrate the effectiveness of this algorithm. Third, the real data for skin cutaneous melanoma (SKCM) effectively explores the individualized heterogeneous effects of this disease and provides a practical analytical framework for the analysis of such complex diseases.

1 Methodology 1.1 Model settings

We formulate the problem under the heterogeneous regression model. For the ith individual, yi is a response variable, Xi=(xi1, …, xip)T is a p-dimensional vector of predictors with heterogeneous effects, and Zi=(zi1, …, ziq)T is a q-dimensional vector of homogengous effects. The model is denoted as

$ y_i=\boldsymbol{X}_i^{\mathrm{T}} \boldsymbol{\beta}_i+\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}+\boldsymbol{\varepsilon}_i, i=1, \cdots, n, $ (1)

where each individual has a unique heterogeneous effect βi=(βi1, …, βip)T related to some certain variables Xi, and β=(β1, …, βn). The homogeneous effect α=(α1, …, αq)T is related to Zi. The random errors εi are independent and E(εi)=0, Var(εi) < ∞.

The heterogeneous linear model (1) can be decomposed into

$ \left\{\begin{array}{l} y_i^{(1)}=\boldsymbol{X}_i^{\mathrm{T}} \boldsymbol{\beta}_i+\boldsymbol{\varepsilon}_i^{(1)}, \\ y_i^{(2)}=\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}+\boldsymbol{\varepsilon}_i^{(2)}, \\ y_i=y_i^{(1)}+y_i^{(2)}, \\ E\left(\boldsymbol{\varepsilon}_i^{(1)}\right)=0, \operatorname{Var}\left(\boldsymbol{\varepsilon}_i^{(1)}\right)<\infty, \\ E\left(\boldsymbol{\varepsilon}_i^{(2)}\right)=0, \operatorname{Var}\left(\boldsymbol{\varepsilon}_i^{(2)}\right)<\infty . \end{array}\right. $ (2)

We can see that the decomposition of this model (2) combines the existing subgroup analysis heterogeneous model and the classical homogeneous linear model.

1.2 Robust individual estimation based on γ-divergence

We propose a robust individualized subgroup analysis method based on γ-divergence, which can deal with the linear model with contamination in the homogeneous part and has a individual penalty in the heterogeneous part. According to (2), our objective function is

$ Q(\boldsymbol{\alpha}, \boldsymbol{\beta}, \boldsymbol{\tau})=L_1(\boldsymbol{\beta})+S(\boldsymbol{\beta}, \boldsymbol{\tau})+L_2(\boldsymbol{\alpha})+P(\boldsymbol{\alpha}) . $ (3)

The first term L1(β) is the square loss function,

$ L_1(\boldsymbol{\beta})=\frac{1}{2} \sum\limits_{i=1}^n\left(y_i^{(1)}-\boldsymbol{X}_i^{\mathrm{T}} \boldsymbol{\beta}_i\right)^{\mathrm{T}}\left(y_i^{(1)}-\boldsymbol{X}_i^{\mathrm{T}} \boldsymbol{\beta}_i\right) . $ (4)

The second term S(β, τ) is the multidirectional separation penalty (MDSP). The MDSP was first proposed by Tang et al.[5]. We use the MDSP to get the heterogeneity estimate $\hat{\boldsymbol{\beta}}$. We consider different subgrouping with respect to different heterogeneous predictors. We assume that every heterogeneous variable Xi has Bk subgroups,

$ \begin{aligned} & \beta_{i k}= \begin{cases}\tau_k^{(l)}, & \text { if } i \in g_k^{(l)}, \quad l=1, \cdots, B_k-1, \\ 0, & \text { if } i \in g_k^{(0)}, \end{cases} \\ & \;\;\;\;\;\;\;\;\;\;\;\;\text { for } i=1, \cdots, n \text {, } \end{aligned} $ (5)

where τk(l)(l=1, …, Bk-1) is unknow nonzero sub-homogeneous effect shared by individuals with lth subgroup, and each potential subgroup is represented by the index partition set {gk(l)}l=0, 1, …, Bk-1 heterogeneous covariate as an example: one is a subgroup of zero effect(βik=0, igk)and the other is a subgroup of non-zero effect(βik=τk, igkc). And the multidirectional separation (MDSP) function S(τ, β) is defined as

$ S(\boldsymbol{\beta}, \boldsymbol{\tau})=\sum\limits_{i=1}^n \sum\limits_{k=1}^p s_{\lambda_2}\left(\beta_{i k}, \tau_k\right), $ (6)
$ s_{\lambda_2}\left(\beta_{i k}, \tau_k\right)=\lambda_2 \min \left(\left|\beta_{i k}\right|, \left|\beta_{i k}-\tau_k\right|\right), $ (7)

where λ2 is a tuning parameter for individual penalty. Given τk, the MDSP function provides βik another shrinking direction τk except 0, which can reduce the bias and reduce the sparsity of weak signals. The MDSP term $\sum\limits_{i=1}^n s_{\lambda_2}\left(\beta_{i k}, \tau_k\right)$ is a center-based clustering. The center τk of the subgroup and the subgroup members obtained from each contraction direction are iteratively updated. This allows each individual to shrink in the best direction, thus improving the individual model fit, while further adaptively estimating $\hat{\tau}$ to obtain subgroups of individuals. More specific details about MDSP can be found in Tang et al.[5].

The third term L2(α) is the γ-divergence loss function. According to the existing literature[13] that uses γ-divergence to deal with linear regression, they argue that the response variable y will deviate partially from the normal distribution due to the y presence of outliers. And by adjusting the appropriate γ values, robust estimates can be obtained with deviating from the normal distribution. Therefore, we adopted one of the same settings and practices as theirs. According to the second equation of model (2), we can get conditional probability of yi(2)

$ f\left(y_i^{(2)} \mid \boldsymbol{Z}_i ; \boldsymbol{\alpha}\right)=\frac{1}{\sqrt{2 \pi} \sigma} \exp \left\{-\frac{\left(y_i^{(2)}-\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}\right)^2}{2 \sigma^2}\right\} . $ (8)

We use γ-divergence to deal with contamination in homogenerous variables. For two density functions fθ(x) and g(x), the γ-divergence is defined as

$ \begin{aligned} & D_\gamma\left(g(x), f_\theta(x)\right)=\frac{1}{\gamma(\gamma+1)} \\ & \left\{\|g(x)\|_\gamma-\int\left(\frac{f_\theta(x)}{\left\|f_\theta(x)\right\|_{\gamma+1}}\right)^\gamma g(x) \mathrm{d} x\right\}, \gamma>0, \end{aligned} $ (9)

where fθ is the model distribution under p dimensional parameter θ, g is the distribution of data generation, and $\|g(x)\|_\gamma=\left(\int g(x)^\gamma \mathrm{d} x\right)^{\frac{1}{\gamma}}, \left\|f_\theta(x)\right\|_{\gamma+1}=$$\left(\int f_\theta^{\gamma+1}(x) \mathrm{d} x\right)^{\frac{1}{\gamma+1}}$. And the parameter γ balances robustness and efficiency, which means a bigger γ corresponding to more robust but less efficient estimation. Noted that as γ→0, Dγ(g, fθ) is a version of the Kullback-Leibler divergence in the limiting case.

Neglecting the terms independent of the unknown parameters, the empirical version[13] of the γ-divergence loss function is obtained by

$ L_2(\boldsymbol{\alpha})=-\frac{1}{n} \sum\limits_{i=1}^n \frac{f\left(y_i^{(2)} \mid \boldsymbol{Z}_i ; \boldsymbol{\alpha}\right)^\gamma}{\left(\int f\left(y^{(2)} \mid \boldsymbol{Z}_i ; \boldsymbol{\alpha}\right)^{(1+\gamma)} \mathrm{d} y^{(2)}\right)^{\gamma /(1+\gamma)}}, $ (10)

where f(yi(2)|Zi; αi) is the conditional probability of yi giving Zi. Following Fujisawa and Eguchi[13], then we can get the γ-loss function

$ L_2(\boldsymbol{\alpha})=-\frac{1}{n}\left(\frac{1+\gamma}{2 \pi \sigma^2}\right)^{\frac{\gamma}{2(1+\gamma)}} \sum\limits_{i=1}^n \exp \left(-\frac{\gamma}{2 \sigma^2}\left(y_i^{(2)}-\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}\right)^2\right). $ (11)

The last term is the Lasso penalty to select variables,

$ P(\boldsymbol{\alpha})=\lambda_1 \sum\limits_{j=1}^q\left|\alpha_j\right|, $ (12)

where λ1 is a tuning parameter for Lasso penalty.

However, it was found that yi(1) and yi(2) are not recognizable, we can not solve (3) directly. Therefore, we design a two-step algorithm that can give reasonable solutions, which is described in Section 1.3. More discussion on the identifiability of the objective function and the effectiveness of the algorithm is shown in the final remark of Section 1.3.1.

Regarding the theoretical properties of parameter estimation, there is really no discussion of related issues in this article, and this theoretical nature of parameter estimation presented in this article is indeed a great challenge. Based on the theoretical properties of the estimates obtained by γ-divergence, Fujisawa and Eguchi[13] proved in detail the theoretical properties of the estimates obtained by linear regression based on γ-divergence; Hung et al.[15] applied γ-divergence to logistic regression and proved the consistency and robustness of the obtained estimates; Ren et al.[16] further applied γ-divergence to high-dimensional generalized linear regression and gives the theoretical properties of the estimates. Regarding the MDSP penalty term, it was first proposed by Tang et al.[5] to apply MDSP to heterogeneous linear models to achieve individualized variable selection, they use a weighted least squares loss function and prove the large sample theoretical properties of the estimation. We proposed a robust individualized subgroup analysis method based on the above, and the theoretical properties of this estimation will be the subject of further research in the future.

1.3 Computation 1.3.1 Overall two-step algorithm

To solve the problem that yi(1) and yi(2) are not recognizable, we proposed a two-step method. Our idea is to divide the solution of the model (2) into two parts, iterating alternately. First, the initial value of the coefficient α(0), β(0) is obtained by the method in Tang et al.[5]. Next, the two parts of the model are solved alternately. In the first part, we think of β as a known, then we obtain a general linear model, which only contains homogeneous coefficients α. So the objective function in this step is

$ Q(\boldsymbol{\alpha})=L_2(\boldsymbol{\alpha})+P(\boldsymbol{\alpha}), $ (13)

where L2(α) is defined by (11), P(α) is defined by (12). Then we get the estimation $ \hat{\boldsymbol{\alpha}} $ by solving the γ-divergence loss function with Lasso penalty. The optimization algorithm is described in Section 1.3.2. In the second part, we fix α, then we get heterogeneity estimation $\hat{\boldsymbol{\beta}}$ by the following objective function

$ Q(\boldsymbol{\beta}, \boldsymbol{\tau})=L_1(\boldsymbol{\beta})+S(\boldsymbol{\beta}, \boldsymbol{\tau}), $ (14)

where L1(β) is defined by (4), and S(β, τ) is MDSP. The two parts iterate alternately until convergence. The optimization algorithm is described in Section 1.3.3. The detailed procedure is described as follows.

Initialize: Based on ADMM algorithm in Tang et al.[5], the initial values of homogeneous coefficient α(0) and heterogeneous coefficient β(0) are solved.

Step 1: We get the value of α(l) from the previous step, the coefficients of homogeneous covariates α(l) are reestimated based on γ-divergence. We reestimate the coefficients of homogeneous covariates α(l+1) by solving (13).

Step 2: We get the value of β(l) from the previous step, solve the objective function with multidirectional separation penalty, and reestimate the coefficient of heterogeneous covariate β(l+1) by solving (14). Thus, the step 1 and step 2 are repeated until convergence.

Remark 1: We acknowledge that there is a recognizability problem in the design of the underlying true model, and there is indeed a gap between the objective function and the current algorithm, and the solution optimized by the algorithm may not correspond exactly to the objective function, but the result obtained by the algorithm is also a reasonable solution with good numerical results. It is the direction of our future exploration to explore the problem of recognizability of the underlying true model in depth.

Remark 2: To eliminate the problem of recognizability of the current model, another possible model is

$ \begin{aligned} Q(\boldsymbol{\alpha}, \boldsymbol{\beta})= & -\frac{1}{n} \sum\limits_{i=1}^n \frac{f\left(y_i \mid \boldsymbol{Z}_i, \boldsymbol{X}_i ; \boldsymbol{\alpha}, \boldsymbol{\beta}\right)^\gamma}{\left(\int f\left(y \mid \boldsymbol{Z}_i, \boldsymbol{X}_i ; \boldsymbol{\alpha}, \boldsymbol{\beta}\right)^{(1+\gamma)} \mathrm{d} y\right)^{\frac{\gamma}{(1+\gamma)}}}+ \\ & \lambda\left(\sum\limits_{j=1}^q\left|\alpha_j\right|+\sum\limits_{i=1}^p\left|\beta_i\right|\right) . \end{aligned} $ (15)

This objective function (15) can be solved by the classical coordinate descent method, which is omitted here. This scheme we also tried, and the numerical results are not good. The numerical results show that both the estimation of the homogeneous coefficient α and the heterogeneous coefficient β have poorer results compared to our proposed method. And the MSE of the estimated homogeneity coefficients α for our proposed method is significantly lower than the model (15). The lower value of RI obtained by the model (15) suggests that the direct use of γ-divergence to the whole y may be less suitable in the presence of individual effects. Our intuitive interpretation is that if the γ-divergence is added directly to the whole y, because the whole y contains a portion of individual effects, that portion is not normally distributed at all, and such the y will cause the robustness of the γ-divergence to fail completely. On the other hand, we may regard the heterogeneous data as outliers, we can not judge whether the data is contaminated or caused by heterogeneous variables. So we use γ-divergence loss function in the homogeneous part to greatly reduce the not robust results caused by homogeneous data contamination, and use MDSP in the heterogeneous part to realize the individualized selection of heterogeneous covariates.

1.3.2 CD algorithm

To minimize the (13), we use coordinate descent algorithm to solve this problem. According to (11), the derivative of Q(α) to αj is

$ \begin{aligned} \nabla_{\alpha_j} Q= & \frac{\partial Q(\boldsymbol{\alpha})}{\partial \alpha_j} \\ = & -\frac{1}{n}\left(\frac{1+\gamma}{2 \pi \sigma^2}\right)^{\frac{\gamma}{2(1+\gamma)}} \frac{\gamma}{\sigma^2} \sum\limits_{i=1}^n\left(y_i^{(2)}-\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}\right) z_{i j} \\ & \exp \left(-\frac{\gamma}{2 \sigma^2}\left(y_i^{(2)}-\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}\right)^2\right)+\lambda_1 \operatorname{sgn}\left(\alpha_j\right), \end{aligned} $ (16)

then we let Q(α) derivative of σ2, and we let it equals 0

$ \begin{gathered} \frac{\partial Q(\boldsymbol{\alpha})}{\partial \sigma^2}=-\frac{1}{n}\left(\frac{1+\gamma}{2 \pi \sigma^2}\right)^{\frac{\gamma}{2(1+\gamma)}} \frac{\gamma}{2}\left(\frac{1}{\sigma^2}\right)^2 \\ \sum\limits_{i=1}^n\left[\left(y_i^*-\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}\right)^2-\frac{\sigma^2}{1+\gamma}\right] \\ \exp \left(-\frac{\gamma}{2 \sigma^2}\left(y_i^*-\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}\right)^2\right)=0 . \end{gathered} $ (17)

With fixed γ and λ1, we calculate the gradient according to (16), select the step length using the armijo criterion, and update α according to the coordinate descent method, then use the bisection method to estimate σ2.

1.3.3 ADMM algorithm

In order to optimize the objective function (14), the constraint set is introduced and transformed into solving the following constrained optimization problem

$ \min \limits_{\boldsymbol{\beta}, \boldsymbol{\nu}, \boldsymbol{\tau}} L_\gamma(\boldsymbol{\beta})+S_{\lambda_2}(\boldsymbol{\beta}, \boldsymbol{\tau}) \quad \text { s.t. } \quad \boldsymbol{\beta}=\boldsymbol{\nu}. $ (18)

where βnp×1=(βij)1≤in, 1≤jp, νnp×1=(νij)1≤in, 1≤jp. The augmented Lagrange multiplier method is used to solve (18) by introducing Lagrange multiplier Λ, κ.

$ \min\limits _{\boldsymbol{\beta}, \boldsymbol{\nu}, \boldsymbol{\tau}} L_\gamma(\boldsymbol{\beta})+S_{\lambda_n}(\boldsymbol{\beta}, \boldsymbol{\tau}) \quad \text { s.t. } \quad \boldsymbol{\beta}=\boldsymbol{\nu} . $ (19)
$ \begin{aligned} & \boldsymbol{L}_\gamma(\boldsymbol{\beta}, \boldsymbol{\nu}, \boldsymbol{\tau})=L_\gamma(\boldsymbol{\beta})+S_{\lambda_n}(\boldsymbol{\beta}, \boldsymbol{\tau})+ \\ & \mathit{\boldsymbol{ \boldsymbol{\varLambda} }}^{\mathrm{T}}(\boldsymbol{\beta}-\boldsymbol{\nu})+\frac{\boldsymbol{\kappa}}{2}\|\boldsymbol{\beta}-\boldsymbol{\nu}\|_2^2, \end{aligned} $ (20)

where Λnp×1=(Λij)1≤in, 1≤jp, κ is a fixed number. Then we use ADMM algorithm to solve the following optimization problem:

$ \boldsymbol{\beta}_{(l+1)}=\underset{\boldsymbol{\beta}}{\operatorname{argmin}} L_\gamma(\boldsymbol{\beta})+\frac{\kappa}{2}\left\|\boldsymbol{\beta}-\boldsymbol{\nu}_{(l)}+\kappa^{-1} \mathit{\Lambda}^{(l)}\right\|_2^2, $ (21)
$ \begin{gathered} \left\{\boldsymbol{\nu}_{(l+1)}, \boldsymbol{\tau}_{(l+1)}\right\}=\underset{\nu, \tau}{\operatorname{argmin}} S_{\lambda_n}(\boldsymbol{\nu}, \boldsymbol{\tau})+ \\ \frac{\boldsymbol{\kappa}}{2}\left\|\boldsymbol{\beta}_{(l+1)}-\boldsymbol{\nu}+\boldsymbol{\kappa}^{-1} \mathit{\Lambda}^{(l)}\right\|_2^2, \end{gathered} $ (22)
$ \mathit{\Lambda}_{(l+1)}=\mathit{\Lambda}_{(l)}+\boldsymbol{\kappa}\left(\boldsymbol{\beta}_{(l+1)}-\boldsymbol{\nu}_{(l+1)}\right) . $ (23)

For more details about ADMM algorithm for (14), please refer to Tang et al.[5]. The overall algorithm in our approach is described in Table 1.

Table 1 Algorithm 1

Regarding the algorithm for the solution, we use a two-step iterative method. Given the initial values, the first step uses the alternating direction method of multipliers (ADMM) to solve an optimization problem containing the MDSP of the optimization problem. In this step, we use the ADMM algorithm borrowed from the method used by Tang et al.[5] in their solution. Boyd et al.[17] proved that the ADMM algorithm can guarantee the convergence of residuals, objective function, and pairwise variables under the assumption of general. Tang et al.[5] show that ADMM algorithm can converge to a stable point when solving the penalty term with MDSP, and in practice, it can converge to a local minimum by iteration. And most of the individuals are insensitive to the initial values except those near the subgroup boundary. Therefore, using the ADMM algorithm to solve this problem can ensure convergence.

In the second step, we use the coordinate descent method to solve a loss function based on γ-divergence. We use this algorithm by drawing from existing papers on γ-divergence-based regression[14, 16] which use the CD algorithm in solving such problems. The convergence of the algorithm of coordinate descent is a very general framework, for the main term is the likelihood function are applicable, so its convergence can be guaranteed. Both optimization subproblems are convergent, so it is a natural result that the whole algorithm is convergent. For all of our simulated and real data sets, convergence was successfully achieved within 50 overall iterations (mostly within 20 iterations).

Regarding the efficiency of the overall two-step method, we performed numerical simulations on a computer configured with an Apple M1 (ARM64) chip (total number of cores 8, memory 16 GB), and the time for 100 repetitions at a given setting was 1 to 2 hours, with an average time of less than 1 minute for one calculation. The complexity of the algorithm is O(n3), and the relatively high complexity of the algorithm is due to the high complexity of the ADMM algorithm used in solving the objective function containing the MDSP. However, in general, the time required to solve is shorter and the algorithm is more efficient.

1.3.4 Tuning parameter

In this article, we should tune three parameters γ, λ1, λ2. It is worth noting that the first tuning parameter γ balance the estimation efficiency and robustness. However, there is no consistent way to select γ. The second parameter λ1 is Lasso penalty parameter, and it can continuously shrink coeffients toward zero, which really improves prediction ability via the bias variance trade-off. The last parameter λ2 is MDSP parameter, which control the individual variables selection. Bayes Information Criterion (BIC) criteria are able to identify the true model consistently. Here, we use BIC criteria to select optimal parameters λ1, λ2, γ.

Regarding the parameter γ, according to the definition of γ-divergence, and studies related to γ-divergence, it is shown that the value of γ balances the robustness of the model. However, regarding the selection of parameter γ, Basu et al.[7] pointed out that there is no consistent best way to choose a suitable parameter γ. Therefore, there are many ways to choose γ. The first method is to use cross-validation together with other parameters to be optimized to select[18]. The second method does not select the parameters based on the data, but artificially gives the γ value directly[14]. The third approach is to use specified rules (e.g. BIC criterion) for parameter selection together with other parameters that need to be optimized[11, 15-16, 19]. Since cross-validation is slow and relying on artificially given parameter values without data-based selection is unreliable, we draw on the existing literature[11, 16] to use the BIC criterion for parameter γ selection.

2 Simulation 2.1 Model for simulation

We used the simulation model given by

$ \begin{aligned} & y_i=x_{i 1} \beta_{i 1}+x_{i 2} \beta_{i 2}+\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}+e, \\ & e \sim N(0, 1), i=1, \cdots, n . \end{aligned} $

We set the sample size n=180, 120, and the number of homogeneous explanatory variables q=20, 40, respectively. The true coefficients were give by

$ \begin{aligned} & \boldsymbol{\beta}_1=\left(\theta_1, \cdots, \theta_1, \cdots, \theta_1, 0, \cdots, 0\right), \\ & \boldsymbol{\beta}_2=\left(0, \cdots, 0, \theta_2, \cdots, \theta_2, \cdots, \theta_2\right), \\ & \boldsymbol{\alpha}=\left(\theta_3, \cdots, \theta_3, 0, \cdots, 0\right)^{\mathrm{T}}, \end{aligned} $

where θ1=2, θ2=-1, θ3=1. The number of nonzero elements θ1 in β1 is the same as the number of θ2 in β2, which is 2n/3. And the number of θ3 is 10.

The variables are generated by two ways. Let X=(X1, …, Xn)T, Z=(Z1, …, Zn)T. The first way was heterogeneity explanatory variables X and homogeneous explanatory variables Z are independent. And X are generated from a normal distribution N(0, Σ1) with $\Sigma_1=\left(\begin{array}{cc}1 & 0.2 \\ 0.2 & 1\end{array}\right)$. The homogeneous explanatory variable Z are also generated from a normal distribution N(0, Σ2). And we consider two structures of covariance matrix Σ2. The first structure is autoregressive correlation (AR), which is given by Σ2=(ρ|i-j|)1≤i, jp with ρ=0.2. The banded correlation is the second structure. We consider σij=0.5, if |i-j|=1, and 0 otherwise. The second way to generate variables X and Z are dependent. They are generated from a multivariate normal distribution with mean 0 and covariance R(ρ), where R(ρ) is the correlation matrix with AR structure like Σ2, and ρ=0.2.

Outliers in homogeneous variables were incorporated into simulations. We investigated two outlier ratios (r=0.1 and 0.3). The outliers are generated from N(μ, cΣ2), where μ =(1, …, 1)T, c=2.

2.2 Performance measure

The mean squared error (MSE) were examined to verify the predictive performance and fitness of regression coefficient:

$ \begin{aligned} & \operatorname{MSE}(\boldsymbol{\alpha})=\frac{1}{q} \sum\limits_{j=1}^q\left(\alpha_j^*-\hat{\alpha}_j\right)^2, \\ & \operatorname{MSE}(\boldsymbol{\beta})=\frac{1}{n p} \sum\limits_{i=1}^n \sum\limits_{j=1}^p\left(\beta_{i j}^*-\hat{\beta}_{i j}\right)^2, \end{aligned} $

where (xi*, zi*, yi*)(i=1, 2, …, n) is the test sample generated from the simulation model without outliers, βij* and αj* are the true coefficients. The true positive rate (TPR) and true negative rate (TNR) of coefficients were:

$ \begin{aligned} & \operatorname{TPR}(\hat{\boldsymbol{\alpha}})=\frac{\left|\left\{j \in\{1, \cdots, q\}: \hat{\alpha}_j \neq 0 \wedge \alpha_j^* \neq 0\right\}\right|}{\left|\left\{j \in\{1, \cdots, q\}: \alpha_j^* \neq 0\right\}\right|}, \\ & \operatorname{TNR}(\hat{\boldsymbol{\alpha}})=\frac{\left|\left\{j \in\{1, \cdots, q\}: \hat{\alpha}_j=0 \wedge \alpha_j^*=0\right\}\right|}{\left|\left\{j \in\{1, \cdots, q\}: \alpha_j^*=0\right\}\right|}, \\ & \operatorname{TPR}(\hat{\boldsymbol{\beta}})= \\ & \frac{\left|\left\{i \in\{1, \cdots, n\}, j \in\{1, \cdots, p\}: \hat{\beta}_{i j} \neq 0 \wedge \beta_{i j}^* \neq 0\right\}\right|}{\left|\left\{i \in\{1, \cdots, n\}, j \in\{1, \cdots, p\}: \beta_{i j}^* \neq 0\right\}\right|}, \\ & \operatorname{TNR}(\hat{\boldsymbol{\beta}})= \\ & \frac{\left|\left\{i \in\{1, \cdots, n\}, j \in\{1, \cdots, p\}: \hat{\beta}_{i j}=0 \wedge \beta_{i j}^*=0\right\}\right|}{\left|\left\{i \in\{1, \cdots, n\}, j \in\{1, \cdots, p\}: \beta_{i j}^*=0\right\}\right|} . \\ & \end{aligned} $

The rand index (RI) is a measure of the similarity between two data clusterings. It is definated by

$ \mathrm{RI}=1-\frac{I\left(M_1\right)-I\left(M_2\right)}{C_n^2} \text {, } $

where function I(M) is the number of upper triangular nonzero elements in matrix M. Ms(s=1, 2) represents class matrix, its element aij=1 means subject i and j are in the same group, otherwise aij=0. If subject i and subject j have completely identical heterogeneity coefficients, we consider they belong to the same class. M1 is real class matrix, M2 is estimate class matrix.

2.3 Comparative method

We compare the performance of our proposed method γ-MDSP with five subgrouping based variable selection approaches: 1) the original MDSP method; 2) the pairwise fused Lasso with an Lasso penalty (FLPa): $\lambda_1 \sum_{i=1}^n\left\|\beta_i\right\|_1+$$\lambda_2 \sum_{k=1}^p \sum_{i < j}\left|\beta_{i k}-\beta_{j k}\right|$, was sovled by R package penalized (version 0.9-50); 3) fusion and feature selection with a truncated Lasso penalty (FTLP): $\lambda_1 \sum_{k=1}^p \sum_{i=1}^n J_\tau \left(\left|\beta_{i k}\right|\right) +$$\lambda_2 \sum_{k=1}^p \sum_{i < j} J_\tau\left(\left|\beta_{i k}-\beta_{j k}\right|\right)$, where Jτ(a)=$\min \left(\frac{a}{\tau}, 1\right)$, was implemented by R package penalized (version 0.9-50).

In addition, we also focus on the following methods about subgrouping and clustering aspect: 1) clustering based on response variables with a Lasso penalty (CVL). First we cluster based on the response variables, and then we use Lasso penalty on each cluster to realize variables selection; 2) clustering based on residual with an Lasso penalty (CRL). First, under the homogeneity assumption $y_i=\boldsymbol{X}_i^{\mathrm{T}} \boldsymbol{\beta}+\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}+\boldsymbol{\varepsilon}_i$, we use Lasso penalty to select variables, then we cluster them based on residuals $y_i-\boldsymbol{X}_i^{\mathrm{T}} \hat{\boldsymbol{\beta}}-\boldsymbol{Z}_i^{\mathrm{T}} \hat{\boldsymbol{\alpha}}$, and we also apply Lasso penalty on each cluster.

We also compare the γ-divergence with some other robust loss functions: 1) least absolute deviation (LAD) solved by R package L1pack (version 0.38.196); 2) Huber function solved by R package MASS (version 7.3-55).

Moreover, we used MDSP based on known true data and true important homogeneous variables in the first estimator (Oracle1). In the second estimator (Oracle2), we know true important homogeneous variables and original MDSP method is used.

In our method, we need an initial point to obtain the estimate, and in this experiment, we used the estimate of MDPS as an initial point. The tuning parameter γ, MDSP penalty λ1 and Lasso penalty λ2 are selected via line search to minimize Bayesian information criterion.

2.4 Results

Table 2 is the results that heterogeneous variable X is independent of homogeneous variable Z. Table 2 shows the results of that X and Z are dependent. In the main text we only show Table 2 and Table 3. And these tables provide the MSE, TPR, and TNR of coefficients α and β, the last column displays the RI for all methods.

Table 2 Independent X and Z, q=20, r=0.1

Table 3 Dependent X and Z, q=20, r=0.1

The proposed method has the smallest MSE of α in all settings besides Oracle1. The MSE (α) of our method are significantly improved compared to other methods. This is because our method introduce the γ-divergence to increase the model robustness. Also, our method has the smallest MSE(β) besides Oracle1, which has a significant improvement on original MDSP approach. This is because we get more precise $\hat{\boldsymbol{\alpha}}$ and we reduce errors of $\hat{\boldsymbol{\beta}}$ by iteration.

In addition, the results show that our approach has the biggest TPR(α) in all methods besides Oracle1. As for RI, we can find that our method has almost the biggest value in all settings. Although our method has lower TNR than CVL and CRL, CVL and CRL have the worest performance on TPR, which makes the worst result of RI in all approaches. We should notice that the results of TNR(α) of all methods are not too high, and this is a challenge for above approaches.

In general, the proposed method has the closest values to Oracle1. Our method is robust against the contamination data and misspecification of subgroup numbers in terms of the consistently smallest MSE and highest TPR among all approaches.

3 Real data application

In this section, we apply the proposed robust individualized subgroup analysis method to the data of SKCM. Skin cutaneous melanoma is one of the highly lethal and aggressive skin diseases. The identification of biomarkers is of great importance for clinical treatment. But handling high-dimensional data analysis and identifying potential genes on the dataset is challenging Zhang et al.[20]. The data includes 342 individuals, the response variable is Breslow' s depth, and the covariants includes 9 environmental effects and 20189 mRNA expression. The response variable Breslow' s depth in the dataset is considered to be a prognostic factor for melanoma in medicine, that is, the greater the Breslow depth, the lower the survival rate. The heterogeneous effect of environmental factors on individual skin melanoma has been mentioned many times in the field of biomedicine[21]. Therefore, it is reasonable to take environmental variables as heterogeneous variables in genetic and environmental data.

Consider a linear model with heterogeneous variables, $y_i=\boldsymbol{X}_i^{\mathrm{T}} \boldsymbol{\beta}_i+\boldsymbol{Z}_i^{\mathrm{T}} \boldsymbol{\alpha}+\boldsymbol{\varepsilon}_i$, denote yi as the Breslow's depth (log-transformed) of ith-individual, Xi as the environmental variables of ith-individual, and Zi as the mRNA expression. We removed covariates with missing value rates greater than 0.15 and removed other individuals with missing values. The final remaining 294 individuals with 7 environmental covariates. Meanwhile, we used prescreening for the selection of genetic variables, resulting in 50 remaining mRNA variables. To identify environmental factors significantly associated with melanoma in each individual, we used 7 environmental covariates as heterogeneous potential predictors and added genetic covariates as homogeneous variables.

The gene coefficients are shown in the Table 4 and the results indicate that 11 of the 50 genes are significantly associated with melanoma, and the selected genes have been shown in the available literature to be highly correlated with skin cutaneous melanoma. Lambert[22] mentioned the gene CCNK as one of the 20 most significantly expressed genes in squamous cell carcinoma of the skin, and also CCNK is a gene involved in DNA repair and has a better role in patient prognosis[23]. CNOT11, one of the eight subunits of the mammalian CCR4-NOT complex, plays a dual role in the control of tumor progression[24]. In characterizing the transcriptome of Spitzoid neoplasms cutaneous melanocytic proliferations with digital mRNA expression profiles, EIF2B4 is one of the most informative genes[25]. In addition, LSM12 was detected more frequently in male carriers containing preferred partner BRCA1 fusion mutations, and cutaneous melanoma were the frequent tumors demonstrating BRCAm in males[26]. Also, PSMD9 showed significant upregulated expression in primary tumors[27]. Also, the TRIAP1 gene is involved in cell cycle regulation and cellular stress response, regulating P53-mediated apoptosis or cell death in response to stresses such as UV irradiation or DNA damage[28].

Table 4 Coefficients of gene

The three heterogeneous variables AJCC NODES PATHOLOGIC PN, AJCC TUMOR PATHOLOGIC PT, and GENDER were selected for analysis, and subgroups were classified according to whether the coefficients of the heterogeneous variables were zero or not. The 294 individuals were classified into a total of four subgroups, and the values of the non-zero coefficients are shown in Table 5. The four subgroups classified were tested for differences with the rest of the variables and the results showed significant differences in AJCC METASTASIS PATHOLOGIC PM, CLARK LEVEL AT DIAGNOSIS, and SAMPLE TYPE. This suggests that environmental variables have a strong heterogeneous effect on melanoma patients, so that the risk of melanoma is diverse at the individual level and other environmental factors. Therefore, in the treatment and prevention of melanoma, individualized treatment plans should be carried out at the individual level.

Table 5 Information about subgroups
4 Discussion

In this paper, we consider a robust individualized linear model based on γ-divergence. For heterogeneous variables, we use MDSP penalty term to realize individualized penalty. For contaminated homogeneous variables, we use gamma divergence to obtain robust estimation of anti pollution.

We further propose a two-step iterative method to calculate the above process. In addition, we have carried out numerical simulation, and the results show that our method has better effect. In addition, the analysis of SKCM data shows that this method has strong applicability. And in the future work, it can be extended to the generalized linear model to make it more generalized.

References
[1]
Hocking T D, Joulin A, Bach F, et al. Clusterpath an algorithm for clustering using convex fusion penalties[C/OL]//Proceedings of the 28th International Conference on Machine Learning. June 28-July 2, 2011, Bellevue, Washington, USA. ICML, 2011: 1. https://hal.archives-ouvertes.fr/hal-00591630.
[2]
Lindsten F, Ohlsson H, Ljung L. Clustering using sum-of-norms regularization: with application to particle filter output computation[C]//2011 IEEE Statistical Signal Processing Workshop. June 28-30, 2011, Nice, France. IEEE, 2011: 201-204. DOI: 10.1109/SSP.2011.5967659.
[3]
Pan W, Shen X T, Liu B H. Cluster analysis: Unsupervised learning via supervised learning with a non-convex penalty[J]. Journal of Machine Learning Research, 2013, 14(1): 1865.
[4]
Ma S J, Huang J. A concave pairwise fusion approach to subgroup analysis[J]. Journal of the American Statistical Association, 2017, 112(517): 410-423. DOI:10.1080/01621459.2016.1148039
[5]
Tang X W, Xue F, Qu A. Individualized multidirectional variable selection[J]. Journal of the American Statistical Association, 2021, 116(535): 1280-1296. DOI:10.1080/01621459.2019.1705308
[6]
Bhalla S, Kaur H, Dhall A, et al. Prediction and analysis of skin cancer progression using genomics profiles of patients[J]. Scientific Reports, 2019, 9(1): 1-16. DOI:10.1038/s41598-019-52134-4
[7]
Basu A, Harris I R, Hjort N L, et al. Robust and efficient estimation by minimising a density power divergence[J]. Biometrika, 1998, 85(3): 549-559. DOI:10.1093/biomet/85.3.549
[8]
Fujisawa H, Eguchi S. Robust estimation in the normal mixture model[J]. Journal of Statistical Planning and Inference, 2006, 136(11): 3989-4011. DOI:10.1016/j.jspi.2005.03.008
[9]
Ghosh A, Basu A. Robust estimation for independent non-homogeneous observations using density power divergence with applications to linear regression[J]. Electronic Journal of Statistics, 2013, 7: 2420-2456. DOI:10.1214/13-EJS847
[10]
Durio A, Isaia E D. The minimum density power divergence approach in building robust regression models[J]. Informatica, 2011, 22(1): 43-56. DOI:10.15388/Informatica.2011.313
[11]
Zang Y G, Zhao Q, Zhang Q Z, et al. Inferring gene regulatory relationships with a high-dimensional robust approach[J]. Genetic Epidemiology, 2017, 41(5): 437-454. DOI:10.1002/gepi.22047
[12]
Jones M C, Hjort N L, Harris I R, et al. A comparison of related density-based minimum divergence estimators[J]. Biometrika, 2001, 88(3): 865-873. DOI:10.1093/biomet/88.3.865
[13]
Fujisawa H, Eguchi S. Robust parameter estimation with a small bias against heavy contamination[J]. Journal of Multivariate Analysis, 2008, 99(9): 2053-2081. DOI:10.1016/j.jmva.2008.02.004
[14]
Kawashima T, Fujisawa H. Robust and sparse regression via γ-divergence[J]. Entropy, 2017, 19(11): 608. DOI:10.3390/e19110608
[15]
Hung H, Jou Z Y, Huang S Y. Robust mislabel logistic regression without modeling mislabel probabilities[J]. Biometrics, 2018, 74(1): 145-154. DOI:10.1111/biom.12726
[16]
Ren M Y, Zhang S G, Zhang Q Z. Robust high-dimensional regression for data with anomalous responses[J]. Annals of the Institute of Statistical Mathematics, 2021, 73(4): 703-736. DOI:10.1007/s10463-020-00764-1
[17]
Boyd S, Parikh N, Chu E, et al. Distributed optimization and statistical learning via the alternating direction method of multipliers[J]. Foundations and Trends in Machine Learning, 2011, 3(1): 1-122. DOI:10.1561/2200000016
[18]
Smith S A, O'Meara B C. treePL: divergence time estimation using penalized likelihood for large phylogenies[J]. Bioinformatics, 2012, 28(20): 2689-2690. DOI:10.1093/bioinformatics/bts492
[19]
Mollah M N H, Eguchi S, Minami M. Robust prewhitening for ICA by minimizing β-divergence and its application to FastICA[J]. Neural Processing Letters, 2007, 25(2): 91-110. DOI:10.1007/s11063-006-9023-8
[20]
Zhang L, Wang Q, Wang L J, et al. OSskcm: an online survival analysis webserver for skin cutaneous melanoma based on 1085 transcriptomic profiles[J]. Cancer Cell International, 2020, 20: 176. DOI:10.1186/s12935-020-01262-3
[21]
Volkovova K, Bilanicova D, Bartonova A, et al. Associations between environmental factors and incidence of cutaneous melanoma. Review[J]. Environmental Health: A Global Access Science Source, 2012, 11(Suppl 1): S12. DOI:10.1186/1476-069X-11-S1-S12
[22]
Lambert S R. Molecular profiling of cutaneous squamous cell carcinoma[D]. London, Queen Mary University of London, 2010. https://qmro.qmul.ac.uk/xmlui/handle/123456789/564.
[23]
Ramazzotti D, Lal A, Wang B, et al. Multi-omic tumor data reveal diversity of molecular mechanisms that correlate with survival[J]. Nature Communications, 2018, 9: 4453. DOI:10.1038/s41467-018-06921-8
[24]
Bartlam M, Yamamoto T. The structural basis for deadenylation by the CCR4-NOT complex[J]. Protein & Cell, 2010, 1(5): 443-452. DOI:10.1007/s13238-010-0060-8
[25]
Hillen L M, Geybels M S, Spassova I, et al. A digital mRNA expression signature to classify challenging spitzoid melanocytic neoplasms[J]. FEBS Open Bio, 2020, 10(7): 1326-1341. DOI:10.1002/2211-5463.12897
[26]
Sun P, Li Y, Chao X, et al. Clinical characteristics and prognostic implications of BRCA-associated tumors in males: a pan-tumor survey[J]. BMC Cancer, 2020, 20(1): 994. DOI:10.1186/s12885-020-07481-1
[27]
Miñoza J M A, Rico J A, Zamora P R F, et al. Biomarker discovery for meta-classification of melanoma metastatic progression using transfer learning[EB/OL]. (2021-05-27)[2022-04-08]. https://www.preprints.org/manuscript/202105.0670/v1.
[28]
López S, Smith-Zubiaga I, García de Galdeano A, et al. Comparison of the transcriptional profiles of melanocytes from dark and light skinned individuals under basal conditions and following ultraviolet-B irradiation[J]. PLoS One, 2015, 10(8): e0134911. DOI:10.1371/journal.pone.0134911