In the big data era, data related to internet users is characterized by an increasing number of samples and dimensional characteristics. This data plays a pivotal role in diverse fields, including augmented reality, mixed reality, virtual reality, and the anticipated 6G services (Rani et al., 2022). However, as the number of samples and characteristics expands, the complexity of data processing escalates. Hence, utilizing “low-loss” dimensionality reduction (DR) techniques to derive the best lowdimensional data has become a pressing concern in numerous burgeoning domains (Jia et al., 2022). Current DR techniques mainly cater to datasets with many samples. They often directly address the sample’s characteristics, as seen in methods like Principal Component Analysis (PCA), Factor Analysis (FA), and Non-negative Matrix Factorization (NMF) (Roy et al., 2022; Sawant & Bhurchandi, 2022; Yang et al., 2022). These algorithms excel in extracting low-dimensional data from large sample sets. However, in high-dimensional datasets with limited samples, the number of characteristic dimensions exceeding the sample count results in a singular covariance matrix. When applied to high-dimensional data with a smaller sample size, their performance diminishes. Lasso, as a convex optimization method based on the L1 norm, possesses characteristics such as sparse characteristics, data simplification, and reserved subset reduction (Bhadra et al., 2019). It can be applied to extract characteristics from high-dimensional datasets with small samples that exhibit multicollinearity. Though Lasso has found applications in areas such as image recognition and gene analysis (Deutelmoser et al., 2021; Liu et al., 2021), it has no explicit solution. It needs to rely on estimation algorithms for determining regression coefficients diminishes the interpretability of its result.
The Transformer, as a method for extracting characteristic from a global perspective (Vaswani et al., 2017), overcomes the RNN’s limitation of non-parallel computation. Its Attention mechanism also enhances model interpretability. But its proficiency in grasping local details lags behind RNN and CNN. The positional information it uses is an artificial index, lacking flexibility for linear transformations. In summary, current DR techniques often fall short in balancing global information and interpretability, especially in high-dimensional and small-sample datasets. Turning to alternative dimension reduction methods might be the key to rectifying these traditional model limitations.
Clustering stands as a cornerstone tool in data mining applications and research (Ezugwu et al., 2022), occasionally utilized for DR. It’s an unsupervised learning algorithm, that aims to identify natural groupings within a set of patterns, points, or objects (Golalipour et al., 2021). Among the common algorithms, K-Means is computationally efficient but is tailored for convex data (Nasiri et al., 2021). The hierarchical clustering, though more complex, delivers superior accuracy (Tian et al., 2022). In addition, none of these clustering algorithms take the global optimum as the objective function (Golalipour et al., 2021). Each clustering algorithm performs well on suitable data, but for any dataset, selecting the optimal parameters does not always get good performance. For large datasets with outliers, the result obtains by the most complex clustering algorithm do not meet the actual clustering requirements. Therefore, there is an urgent need to develop new characteristic extraction algorithms and validation indicators to produce an accurate and low-complexity model for all realistic data (Oyewole & Thopil, 2023).
In this paper, we construct a new DRMBIP, referring to the four foundational concepts of neighbors, links, objective function, and similarity measure in the ROCK clustering algorithm (Guha et al., 2000). The model takes the correlation of the original data as the similarity measure between the indicators, which solves the problem of solving multidimensional small sample data.
Meanwhile, the model takes maximizing the sum of intra-class correlations, minimizing the sum of inter-class correlations, and minimizing the out-of-bounds penalties as the objective function. It enables the model to find the best neighbors and links in the case of satisfying the inter-class and intra-class correlation threshold.
Specifically, the DRMBIP proposed in this paper has the following four advantages:
More interpretable. In other clustering algorithms such as K-Means, ROCK, DBSCAN, etc., the clustering centers are virtual nodes (Ding & He, 2004; Fahim, 2023; Flegner et al., 2023). However, in the results of the reduced dimensional model in this paper, the class center indicator is real and has accurate practical significance. No need to consider the number of clusters. The number of clusters can have a huge impact on the effectiveness (Song et al., 2022). The DRMBIP solves this problem, and not only obtains the optimal number of clusters, but also makes the model effect globally optimal. Automatic elimination of abnormal indicators. In real data, there are indicators that are difficult to classify. And it is unreasonable to force them into a certain class when using clustering algorithms for dimensionality reduction (Wang et al., 2021). In the DRMBIP of this paper, it is possible to perform dimensionality reduction while eliminating a certain number of indicators, which improves the effect of dimensionality reduction. Easier to view results. The results of the model often face difficulties in visualization. In this paper, the correlation between the indicators is used as the basic data of the model, which can be simpler and more intuitive to display the results of dimensionality reduction in a two-dimensional plane.
Since this paper focuses on linear and global, common existing dimensionality reduction algorithms can be classified into four categories (Espadoto et al., 2019): Linear & Global, Linear & Local, Nonlinear & Global, and Nonlinear & Local.
Linear & Global uses a linear function that preserves the structure of the pairwise distances between all data samples. Representative algorithms are PCA (Abdi & Williams, 2010), FA (Stephenson, 1935), and NMF (Lee & Seung, 1999). PCA is a dimensionality reduction method that transforms the data variables into a set of linearly uncorrelated variables by applying a linear transformation to the data variables. Scholars have proved that PCA is a continuous solution for the clustering membership indicators in K-Means. It suggests PCA potentially performs clustering (Ding & He, 2004). The difference between FA and PCA is that the former uses variance maximization for dimensionality reduction, while the latter performs DR based on the dependencies between variables. However, NMF decomposes the original data matrix into the multiplication of the base matrix W and the coefficient matrix H, i.e., W×H≈V, it uses the coefficient matrix H instead of the original matrix V to achieve dimensionality reduction.
Linear & Local is to use linear function in the process of dimensionality reduction to preserve local features. The representative algorithms are LLE (Roweis & Saul, 2000), LE (Belkin & Niyogi, 2001), and t-SNE (Van der Maaten & Hinton, 2008). The LLE algorithm involves only matrix operations and is easy to implement, but it is not suitable for closed streaming data. The LE algorithm is less computationally and faster in execution compared to the LLE algorithm, but it is less robust to noise and cannot learn the intrinsic geometric structure from noisy data. The t-SNE algorithm improves the computational efficiency compared to the SNE algorithm, but it has no unique optimal solution and cannot be used for prediction. Currently, the corresponding mainstream linear variants of these three common DR algorithms are LPP (Yao et al., 2023), NPE (Zhang et al., 2023), and t-SNLE (Xia et al., 2021). They alleviate the curse of dimensionality and retain the advantages of cluster separation, but it is difficult to maintain the topological connectivity of the data after dimensionality reduction (Yao et al., 2023).
Nonlinear & Global is the use of nonlinear functions to preserve the structure of the pairwise distances between all data samples. Representative algorithms are nonlinear MDS (Kruskal & Wish, 1978) and ISOMAP (Tenenbaum et al., 2000). The MDS algorithm uses Euclidean distances to measure the distances between samples. It is difficult for the algorithm to represent the true geometric structure in low-dimensional space. The effect of the MDS algorithm using nonlinear function optimization is greatly affected by the chosen function. However, it is difficult to fundamentally improve the shortcomings of the MDS algorithm on nonlinear streaming data. The essence of ISOMAP is consistent with MDS. The difference is that ISOMAP uses Geodesic distances instead of Euclidean distances. The Geodesic distance can more realistically reflect the real low-dimensional geometric structure of the flow shape and better measure the distance between samples.
Nonlinear & Local is the use of nonlinear functions to preserve local features during data dimensionality reduction. Representative algorithms are LLE (Roweis & Saul, 2000), LE (Belkin & Niyogi, 2001), t-SNE (Van der Maaten & Hinton, 2008), and UMAP (McInnes et al., 2018). The UMAP algorithm assumes that the sample data is uniformly distributed in the topological space and can be mapped from these finite data samples to a lower dimensional space. Currently, these techniques are more popular for dimensionality reduction and visual exploration of multidimensional datasets (Xia et al., 2021). For example, t-SNE and UMAP are excellent at separation, but slightly inferior at compactness.
Clustering, as an indispensable tool in data mining, can be used in combination with dimensionality reduction such as PCA (Shuai, 2022; Yang et al., 2022; Yunita et al., 2022), t-SNE (Ullah et al., 2022), and UMAP (Can-can et al., 2022). And it can use alone to reduce the dimensionality of the indicator. The most common practice is to use dimensionality reduction methods to compress the data dimensions locally or globally. And the indicators with lower dimensions can be used in data mining work. It is less common to use clustering alone to reduce indicators. It retains indicators that are highly interpretable and interchangeable, providing high flexibility in dimensionality reduction. However, most clustering algorithms have strict requirements, requiring data to satisfy convexity assumptions and are difficult to handle outliers.
Integer programming, as a very powerful tool for solving the problem of reality today, has a wide range of applications (Achterberg et al., 2020). Advances in research related to data sparsity and matrix decomposition have led to a rapid increase in the solution size of commercial solvers such as Gurobi, Cplex, and Mosek (Li et al., 2020). Moreover, the optimal solution can be solved in reasonable time for large-scale cases. Such as urban traffic light scheduling for pedestrian-vehicle mixed-flow networks (Gupta et al., 2022) and efficient operation of distributed generation networks (Cortés et al., 2018). The wide range of applications and the increasing solution scale of integer linear programming provide opportunities for more and more ideas. Many of the concepts in the ROCK clustering algorithm (Guha et al., 2000) can be expressed using integer programming tools, making it possible to construct a dimensionality reduction model.
In this paper, the data of relevant indicators affecting life expectancy is obtained from the Global Health Observatory (GHO). The data are used as the basis for dimensionality reduction analysis. First, we use the DRMBIP to eliminate the abnormal indicators and obtain the filtered indicators. second, we use the correlation data between the indicators to display results. It visualizes the complex relationship between different types of indicators in two-dimensional plane. Finally, we can combine the correlation results of the model with the original categorization between the indicators for further dimensionality reduction. This allows us to find potentially important factors that influence life expectancy.
In this section, we will give a description of the parameters and variables. Then, we will focus on the two main models and their mathematical formulas in this paper. Finally, we establish relevant intra-class and inter-class evaluation indexes and introduce some evaluation indexes of traditional clustering models to evaluate the model results comprehensively.
To build model of DRMBIP, the following symbol must be introduced:
denotes the subscript of the indicator | |
denotes the subscript of alternative class center | |
denotes the set of all subscripts of indicators | |
denotes the set of all subscripts of alternative class centers | |
denotes the number of enabled indicators | |
denotes the number of classes in the result | |
denotes the lower limit parameter of intra-class correlation | |
denotes the upper limit parameter of inter-class correlation | |
denotes the weight of the inter-class correlations in the objective function | |
denotes the weight of intra-class correlations in the objective function | |
denotes the weight of the part above |
|
denotes the weight of the difference of |
|
denotes the absolute value of the Pearson correlation coefficient between the |
|
is similar with |
Define the decision variables as follows:
To solve the problem that the number of clusters is difficult to determine, we propose two DRMBIP with reference to the concepts of ROCK clustering algorithm. The DRMBIP uses the correlation as the basic data, the minimum value of the correlation between the indicator and class center is used to measure the compactness. And the correlation between the indicators of class center is used to measure the separation between the classes. In general, multidimensional data in different situations often have different dimensionality reduction requirements. To meet this demand, we set the threshold for measuring the dimensionality reduction as a parameter in DRMBIP-p to improve flexibility.
Objective function (1): minimize the sum of inter-class correlations, maximize the sum of intra-class correlations, and minimize the sum of out-of-bounds penalties, given the weights
In DRMBIP-p, the optimal number of clusters can be determined by inputting three parameters:
In some practical cases, compactness and separation need to be achieved simultaneously. For this purpose, the parameters in Constraint (6) and Constraint (7) in DRMBIP-p can be transformed into variables then the constraints can be linearized. To facilitate comparisons with other models, constraint must be added to control the number of classes in DRMBIP-v.
Objective function (8): minimize the sum of inter-class correlation, maximize intra-class correlation, and minimize the difference between
The DRMBIP-v differs from DRMBIP-p in that it considers compactness and separation as equally important. Moreover, adding it to the Objective function (8) maximizes
With reference to the traditional clustering model evaluation indexes, this paper introduces five evaluation indexes, namely,
The formula for calculating the
The formula for calculating the
The formula for calculating the
The formula for calculating the
The formula for calculating the
The dimensionality reduction models proposed in this paper use the correlation as the similarity measure between the indicators. Thus, we can construct evaluation indexes based on the correlation in both intra-class and inter-class aspects. Intra-class compactness based on correlation can be measured in two ways. One way is to calculate the correlation of all two indicators in the class, and the other way is to calculate the correlation between the center of the class and the rest of the indicators.
Minimum intra-class correlation is an indicator used to measure compactness within a class. This indicator is chosen after calculating the minimum correlation between two indicators in each class and then selecting the smallest of them.
The formula for calculating the minimum intra-class correlation is as follows:
When adding the constraint of correlation, DRMBIP uses the correlation between the class center and the other indicators. Thus, the intra-class and class center minimum correlation must be calculated to get the actual value under this constraint.
The calculation is the same as the calculation of the minimum intra-class correlation, except that when the minimum correlation is calculated, one of the indicators is changed to the class center.
The formula for calculating the minimum intra-class and class center correlation is as follows:
Since
The formulas for the average intra-class correlation and the average intra-class and class center correlation are as follows:
If fewer or no indicators are filtered, the quality of indicators may result in a single indicator as one class. The effect of extreme values generated in this case needs to be excluded. Therefore, the calculation of
The formula for the average intra-class correlation (not a single class) is as follows:
While considering the intra-class compactness, the inter-class separation of the results also needs to be considered. For this reason, the maximum inter-class correlation is introduced as a measure of the inter-class separation. The maximum inter-class correlation is susceptible to extreme values, which can be excluded by introducing the average inter-class correlation.
The formulas for the maximum inter-class correlation and the average inter-class correlation are as follows:
Life expectancy is one of the three major human development indexes and a key indicator of the progress of human societies (Baumann, 2021; Sen, 1998). Possession of specific disease-causing genes is not the only reason for contracting diseases (Farhud, 2022). The increase in the vintage of the prescription drugs years has increased the average age of death by 1.23 years in 26 countries (Lichtenberg, 2022), assuming the disease is fixed. Thus, the adequate and advanced medical care is an important means of increasing human life expectancy. However, the uneven distribution of healthcare resources in developed countries due to irrational policies has exacerbated the widening of healthcare inequalities (Alexiou et al., 2021). In the specific period of the COVID-19 pandemic, this inequality was even more obvious, which seriously affected the population’s health and economic recovery (Bayati et al., 2022; Currie et al., 2021). Accordingly, deepening the potentially important factors affecting life expectancy plays a very important role in improving human life expectancy. In this section, we will identify the important factors affecting life expectancy and analyze them in depth to find the important factors and make recommendations.
In this paper, we obtained 1464 samples of 141 factors in 183 countries or regions as of March 2023 from the Global Health Observatory (GHO), a public safety and health database established by the World Health Organization (WHO). It includes developed and developing countries such as Asia, Europe, and the United States, as well as poor countries such as Afghanistan, Burundi, and the Central African Republic.
The 141 diseases fall into three main categories or 32 subcategories, the first of which includes subcategories on communicable diseases, maternal, perinatal, and nutritional status. The second main category, non-communicable diseases, is the most numerous with 81 diseases. The third main category includes only the subcategories of unintentional and intentional injuries. Selected indicators of life expectancy for a sample of men in selected countries in 2019 are shown in Table 1.
Selected disease data for a sample of men in selected countries, 2019.
Disease Name | Australia | Canada | USA | Japan | China | Congo | Niger | Somalia |
---|---|---|---|---|---|---|---|---|
Tuberculosis | 0.03 | 0.06 | 0.34 | 1.96 | 25.36 | 1.66 | 2.36 | 5.91 |
Syphilis | 0 | 0 | 0.03 | 0.01 | 0.99 | 0.06 | 0.06 | 0.23 |
HIV/AIDS | 0.05 | 0.12 | 4.12 | 0.04 | 21.23 | 1.68 | 0.63 | 0.34 |
Diarrhoeal Diseases | 0.12 | 0.36 | 4.08 | 1.14 | 4.7 | 0.81 | 8.87 | 8.11 |
Meningitis | 0.02 | 0.04 | 0.33 | 0.17 | 3.46 | 0.16 | 2.73 | 2.97 |
Acute Hepatitis A | 0.01 | 0.01 | 0.24 | 0.05 | 0.05 | 0.01 | 0.08 | 0.42 |
Diabetes Mellitus | 2.46 | 3.38 | 35.64 | 3.07 | 80.63 | 0.54 | 1.14 | 1.75 |
Peptic Ulcer Disease | 0.14 | 0.23 | 1.78 | 1.58 | 24.13 | 0.04 | 0.51 | 0.39 |
Kidney Diseases | 1.52 | 2.24 | 44.01 | 21.37 | 103.69 | 0.31 | 1.63 | 1.27 |
Skin Diseases | 0.27 | 0.28 | 2.42 | 1.01 | 2.62 | 0.02 | 0.11 | 0.07 |
Rheumatoid Arthritis | 0.07 | 0.09 | 0.65 | 0.74 | 4.12 | 0.01 | 0 | 0.02 |
Neural Tube Defects | 0.03 | 0.03 | 0.43 | 0.05 | 1.01 | 0.08 | 0.75 | 0.3 |
Road Injury | 0.91 | 1.35 | 29.18 | 3.07 | 187.4 | 1.02 | 3.93 | 3.01 |
Self-Harm | 2.34 | 3.27 | 40.69 | 13.51 | 72.51 | 0.26 | 0.84 | 0.93 |
Since the interpretability of dimensionality reduction methods such as PCA is poor, the K-Means algorithm, which has stronger interpretability, was chosen as the comparison algorithm (Boutsidis et al., 2014; Maldonado et al., 2015). To ensure the comparability of the models, we use DRMBIP-p and DRMBIP-v to reduce the dimensionality of the indicators. For these two results, we utilize the K-Means algorithm for clustering dimensionality reduction, and the results are labeled as Case 2 and Case 4, respectively. Meanwhile, the K-Means algorithm was utilized to reduce the dimensionality of all the indicators to form a control group, which was noted as Case 5.
The lower limit of the intra-class correlation
Evaluation indexes for dimensionality reduction algorithms.
Case 1 | Case 2 | Case 3 | Case 4 | Case 5 | |
---|---|---|---|---|---|
0.6018 | 0.6018 | 0.0090 | 0.0033 | ||
0.8106 | 0.1345 | 0.0394 | |||
0.6669 | 0.6669 | 0.3578 | 0.0226 | ||
0.9086 | 0.8033 | 0.6845 | |||
0.6669 | 0.6669 | 0.3578 | 0.0226 | ||
0.2807 | 0.8381 | 0.3279 | |||
0.2807 | 0.8381 | 0.3279 | |||
1.0282 | 1.2433 | 1.6834 | |||
2.6207 | 1.3752 | 2.3058 | |||
0.8184 | 0.8184 | 0.3313 | 0.2488 | ||
0.5095 | 0.5095 | 0.2300 | 0.1721 | ||
0.7684 | 1.8128 | 1.4645 |
In terms of indicator filtering effect, comparing Case 2 with Case 5, the K-Means algorithm with indicator filtering gets better dimensionality reduction effect. Its performance on the evaluation indicators is better than the K-Means algorithm without indicator filtering. However, in the comparison between Case 4 and Case 5, the K-Means algorithm with indicator filtering performs generally on the four indexes involving inter-class distances but performs better on all other indexes.
In terms of dimensionality reduction effect, comparing Case 1 with Case 2, the DRMBIP-p filtered indexes performed better. Only
Comparing Case 1 and Case 3, it is not difficult to find that DRMBIP-p performs better in the evaluation index of intra-class correlation. But there is a huge gap between DRMBIP-v and DRMBIP-p in terms of inter-class correlation, which is due to the pursuit of compactness and separation at the same time in DRMBIP-v.
Comparing the three subfigures in the first row of Figure 1, it is easy to find that the dimensionality reduction results of DRMBIP-p have smaller inter-class correlation and larger intra-class correlation than the K-Means dimensionality reduction results after the indicator filter, and the effect is better. Unlike DRMBIP-p, DRMBIP-v must maximize intra-class correlation and minimize inter-class correlation when performing indicator filtering. As a result, the filtered set of indicators has extremely strong class characteristics. This leads to errors in that classes contain indicators with low correlation in the K-Means algorithm.
In summary, DRMBIP obtains good results. DRMBIP-p adopts the form of parameter to set the lower limit of intra-class correlation and the upper limit of interclass correlation. It makes the indicator filtering more flexible, and the selected indicators have better compatibility with other algorithms. On the other hand, DRMBIP-v variables are the lower and upper limit parameters of DRMBIP-p. It makes the results of DRMBIP-v more extreme, sacrificing the compatibility with other models in exchange for more obvious class characteristics.
Among the 141 factors affecting life expectancy, diseases account for 130 indicators, most of which are subdivided into subcategories. After filtering and dimensionality reduction of the indicators, the dimensionality can be further reduced by combining these actual categorizations. To facilitate the analysis, we draw the results of the dimensionality reduction of DRMBIP-p and DRMBIP-v in a corresponding mind map, as shown in Figure 2 below.
The dimensionality reduction results of DRMBIP include subcategories such as tuberculosis, diarrhoeal diseases, hepatitis, digestive diseases, neonatal conditions, and congenital anomalies. In addition, the four subcategories of hepatitis, digestive diseases, neonatal conditions, and congenital anomalies include several diseases with high correlation among the diseases, which may share common predisposing factors and require further traceability. Compared to DRMBIP-v, DRMBIP-p lacked diseases in the categories of genitourinary diseases, skin diseases, sudden infant death syndrome, unintentional injuries, and respiratory diseases. In DRMBIP-v, minimizing the inter-class correlation eliminated these diseases that were not highly correlated with the best class centers, resulting in a more pronounced distinction between classes. These diseases with different class characteristics may not have common causative factors and therefore require separate focus. In contrast, DRMBIP-v is built with more relaxed parameters to find the best indicators while maintaining inter-class and intra-class correlation thresholds. Compared with DRMBIP-p, DRMBIP-v does not include malignant neoplasms, which also indicates that there are some common factors among the different causes of malignant neoplasms. But these causative factors are still quite different from each other, and these common factors should be explored, while the factors inducing different types of malignant neoplasms should be studied separately. Similarly, cardiovascular diseases selected only the disease of stroke in DRMBIP-p. However, it selected two diseases that belonged to two categories in DRMBIP-v. It indicated that there were more complex causative factors in cardiovascular diseases and that it was difficult to find the causes of cardiovascular diseases, which should be monitored separately to find out the causes of the diseases respectively.
Analysis of the 30 indicators analyzed reveals several key factors that influence life expectancy. We have summarized the areas on which countries at all stages of development should focus, in the hope of maximizing the efficiency of the use of health resources and improving life expectancy for all.
Many infectious diseases can be transmitted through the digestive tract, such as hepatitis and diarrhoeal diseases, and protecting water sources is tantamount to blocking the transmission of viruses. In 2016, an outbreak of acute jaundice hepatitis occurred in Chad, Africa, where unsafe water use led to the widespread spread of the disease. Countries should improve the safety of water sources to fundamentally disrupt the transmission of such infectious diseases.
At present, many countries do not pay enough attention to food safety. For example, the artificial sweetener aspartame, which was first classified as “possibly carcinogenic to humans” in July this year, is now widely used by the food industry in various countries. Currently, many countries have lax control over food additives, leading to excessive intake by consumers and resulting in an increase in the prevalence of digestive diseases and cancers every year.
Conservative estimates put the global average incidence of malformed fetuses at 2%-3%, with the incidence of severely malformed fetuses at 0.5%-1%. For high-risk pregnant women with a family history of serious genetic defects, infectious diseases in early pregnancy, alcohol, drug addiction, those who have received radiation therapy, as well as those who are malnourished and weak, the risk of malformed fetuses will be even higher. And these high-risk pregnant women should be screened as much as possible.
In this paper, we propose two DRMBIP to find the most dominant geometric structures from high-dimensional complex data and reveal their influence patterns. Taking the life expectancy data of 184 countries in the world as an example, we calculate the correlation between 141 relevant disease indicators and use DRMBIP to eliminate abnormal indicators and filter indicators. We find 30 important factors affecting life expectancy. It automatically clustered into the optimal two classes based on the correlation. Taking the central indicators of each class as the starting point, we analyzed the correlation indicators and disease categories. We found that there are four main categories of disease factors affecting life expectancy, namely hepatitis, digestive diseases, congenital anomalies, and malignant neoplasms.
We compare the effect of the proposed model with the traditional K-Means algorithm on indicator dimensionality reduction, and find that the algorithm in this paper has the following advantages:
More interpretable The class center, as an important entry point for clustering analysis, should have strong interpretability. When traditional clustering algorithms are used for indicator dimensionality reduction, the class center is not real, from the virtual class center into the analysis may lead to a certain bias. The models proposed in this paper are based on dimensionality reduction on the factors affecting life expectancy, and the class center indicators affecting life expectancy are obtained. DRMBIP-p: preterm birth complications; trachea, bronchus, lung cancers. DRMBIP-v: cardiomyopathy, myocarditis, endocarditis; other congenital anomalies. From the center of the class, we mined out the three most likely common influences and gave suggestions for prevention and control. No need to consider the number of clusters In the traditional clustering model, some models use the number of clusters as a parameter, and the number of clusters seriously affects the model. To obtain the best result, it is necessary to consider the factors such as the compactness and the separation. And it needs to test the parameter to choose the best value in the set of the number of clusters. DRMBIP variables the parameter number of clusters, so the model can find the number of clusters with the best clustering effect. In selecting the 30 important indicators influencing life expectancy, the optimal number of clusters can be found. Eventually, the relevant indicators were classified into two classes with obvious class characteristics. Automatic elimination of abnormal indicators There are many indicators that are difficult to classify, and it is not appropriate to force them into a particular class. By analyzing the relevant indicators affecting life expectancy, we found that the average value of the correlation is 0.4693. It indicates that many indicators are difficult to classify and are in low-density areas in the indicator space. To select the important factors from the indicators affecting life expectancy, this paper sets the number of class centers as 30. It obtains better results which not only shows excellent clustering and dimensionality reduction effect, but also provides a compatible low-dimensional indicator system for other clustering models. Easier to view results In multidimensional data, the geometric structure between data is often difficult to display in the two-dimensional plane. DRMBIP adopts correlation to measure the distance between indicators, it makes the visualization display easy. We can mine the main geometric structures and special details in the intra-class and inter-class indicators based on the indicator relationship maps of the 30 important indicators.
At present, the research on DRMBIP is still in the exploratory stage. In the future, we will further validate the dimensionality reduction effect under different sizes of datasets. At the same time, we will develop parallel dimensionality reduction algorithms under big data. Eventually, a complete set of accurate dimensionality reduction algorithms for big data will be formed for data mining personnel.