Scalable Fuzzy Clustering With Anchor Graph
Scalable Fuzzy Clustering With Anchor Graph
Abstract—Fuzzy clustering algorithms have been widely used to reveal the possible hidden structure of data. However, with the
increasing of data amount, large scale data has brought genuine challenges for fuzzy clustering. Most fuzzy clustering algorithms suffer
from the long time-consumption problem since a large amount of distance calculations are involved to update the solution per iteration.
To address this problem, we introduce the popular anchor graph technique into fuzzy clustering and propose a scalable fuzzy clustering
algorithm referred to as Scalable Fuzzy Clustering with Anchor Graph (SFCAG). The main characteristic of SFCAG is that it addresses
the scalability issue plaguing fuzzy clustering from two perspectives: anchor graph construction and membership matrix learning.
Specifically, we select a small number of anchors and construct a sparse anchor graph, which is beneficial to reduce the computational
complexity. We then formulate a trace ratio model, which is parameter-free, to learn the membership matrix of anchors to speed up the
clustering procedure. In addition, the proposed method enjoys linear time complexity with the data size. Extensive experiments
performed on both synthetic and real world datasets demonstrate the superiority (both effectiveness and scalability) of the proposed
method over some representative large scale clustering methods.
2.3 Fast Fuzzy Clustering Based on Anchor Graph only the nearest anchor can be the neighbor of xi with
Anchor graph model is a recently proposed graph based similarity 1.
learning model for large scale problems, which constructs a In practice, we prefer to learn a sparse bi , i.e., only the k
similarity matrix to measure the relationship between data nearest neighbors of xi have chance to connect to xi , to
points and anchors by capturing the intrinsic structure of achieve better performance and alleviate computation bur-
data distribution. Then, clustering is performed on this sub- den. Then, according to the method P proposed in [42], g can
tle similarity matrix. be set as g ¼ k2 kxi akþ1 k22 12 kh¼1 kxi ah k22 . Thus, the
Inspired by the anchor graph, Nie et al. [40] propose a solution to problem (3) is
Fast Fuzzy Clustering algorithm based on Anchor Graph 8 2 2
< kxi akþ1 k2 kxi aj k2
(FFCAG) to reduce the time-consumption and gives intui- P jk
bij ¼ kkxi akþ1 k22 kh¼1 kxi ah k22 ; : (4)
tive interpretations for their model. Specifically, FFCAG sol- :
ves the following optimization problem 0; others
It can be easily observed that the constructed anchor graph
max Tr UT BT B g 1 I g 2 BT 11T B U ; (2)
U1¼1;U0 B 2 Rnm is sparse. It has much less spurious connections
between dissimilar points and tends to exhibit high quality.
where B is the data-to-anchor similarity matrix and U is the It is also worth noting that the construction method is
membership matrix of anchors. g 1 and g 2 are the regulariza- extremely efficient, having a linear computational complexity
tion parameters. By utilizing the anchor graph model, with data size OðndmÞ, which is helpful to reduce the compu-
FFCAG can able to improve the computational efficiency. tational burden and obtain better clustering performance.
However, it requires more parameters to tune for better per-
formance, which is impractical since the value of parame- 3.2 Membership Matrix Learning
ters may be an arbitrary number. The computational cost of fuzzy clustering mainly comes
Therefore, in this paper, we propose a parameter-free from the full-size membership matrix learning model, since
fuzzy clustering algorithm named scalable fuzzy clustering the number of data size is huge in large scale problems and
with anchor graph, which is simple and scalable, having lin- learning a full-size membership matrix is inefficient. Accord-
ear time complexity with data size. ing to [43], [44], once the membership matrix associated with
anchors can be referred, the membership matrix of data
3 METHODOLOGY points can be obtained by a simple linear combination. There-
fore, we propose to construct a scalable fuzzy clustering
In this section, we elaborate the proposed scalable fuzzy
model by coupling anchor graph and membership matrix of
clustering with anchor graph (SFCAG) for large scale data
anchors.
in detail, which attempts to address the scalability issue
Define U 2 Rmc as the membership matrix of anchors.
plaguing fuzzy clustering from two perspectives: anchor
For an anchor, to achieve a clear clustering division, the dif-
graph construction and membership matrix learning.
ference between membership values belonging to different
clusters should be diverse. Therefore, we obtain the mem-
3.1 Anchor Graph Construction bership matrix of anchors via solving the following problem
In our work, we address the scalability issue through a
small number of representative anchors, which can ade- max Tr UT BT BU : (5)
quately cover the intrinsic manifold structure of data points. U1¼1;U0
Using Eq. (6) as a regularization term, we have the over- Define fðt Þ
all optimization problem written as
T T f ðt Þ ¼ Tr UTtþ1 BT B t 11T Utþ1 : (15)
TrðU B BUÞ
max : (8)
U1¼1;U0 TrðUT 11T UÞ
where Utþ1 can be efficiently calculated according to Algo-
rithm 2 described in the next part.
This economical membership matrix learning model Then, we can get the slope of fðÞ at point t
indeed mitigates the computational burden of the full-sized
data points model. After obtaining U, the membership
f 0 ðt Þ ¼ Tr UTtþ1 11T Utþ1 0 : (16)
matrix of data points can be expressed as F ¼ BU accord-
ing to the similarities between data points and anchors.
We use a liner function gðÞ to approximate the linear
function fðÞ, such that
3.3 Optimization
To solve the challenging formulation in Eq. (8), we propose gðÞ ¼ f 0 ðt Þð t Þ þ f ðt Þ
to first transform it into an equivalent form for easily solv-
ing. Then an iterative optimization algorithm is developed ¼ Tr UTtþ1 BT B 11T Utþ1 (17)
to solve the equivalent model.
Suppose that problem (8) obtains maximum value if
Let gðtþ1 Þ ¼ 0, we have
U ¼ U , that is,
Tr UT BT BU Tr UTtþ1 BT BUtþ1
T T ¼ ; (9) tþ1 ¼ : (18)
Tr U 11 U Tr UTtþ1 11T Utþ1
and
Since gðÞ is an approximation of fðÞ, we can update t
Tr UT BT BU
T T ; 8U : (10) by tþ1 . Thus, we can find the root of fðÞ ¼ 0 and the opti-
Tr U 11 U mal solution in problem (8).
max Tr UT BT B 11T U ¼ 0 : (12) Define M ¼ 11T BT B, problem (19) can be written as
U
Define the function min Tr UT MU : (20)
U1¼1;U0
f ðÞ ¼ max Tr UT BT B11T U : (13)
U Problem (20) is a Quadratic Programming problem,
which can be efficiently solved by the Augmented Lagrang-
Obviously, fð Þ ¼ 0. Therefore, we can obtain the opti- ian Multiplier (ALM) method [45], [46]. According to ALM,
mal by finding the root of equation fðÞ ¼ 0. In the next a slack variable P is introduced and then problem (20) is
two parts, we develop an iterative algorithm to efficiently equivalently transformed as
find the optimal solution . The detailed procedure of the
proposed SFCAG is described in Algorithm 1. min Tr UT MP : (21)
U1¼1;U0;U¼P
Taking the derivative of Eq. (23) w.r.t. P, we can obtain That is, tþ1 t , is monotonically increasing in
each iteration. Therefore, Algorithm 1 monotonically
1 increases the objective function value of problem (8) in
P¼ mU þ S MT U : (24)
m each iteration until converge. u
t
Fig. 1. Visualization of the clustering results of the proposed SFCAG on the two-moon dataset.
TABLE 3
Clustering Performance Comparisons of Different Methods on the Synthetic Dataset
Methods KM FCM RSFCM MSFCM PLRSC DFKM LSC FSC FFCAG SFCAG
ACC 0.8300 0.8200 0.8500 0.8200 0.9400 0.8900 0.9480 0.9775 1 1
NMI 0.3425 0.3199 0.3904 0.3199 0.7024 0.5006 0.7716 0.8871 1 1
ARI 0.4328 0.4066 0.4874 0.4066 0.7733 0.6064 0.8147 0.9148 1 1
F-score 0.7150 0.7018 0.7425 0.7018 0.8860 0.7967 0.9085 0.9554 1 1
cost but also obtains better performance. Since these robustness of the algorithm. Therefore, the proposed
datasets are from different scenarios, the clustering method can achieve better performance.
results clearly reveal that the proposed SFCAG is an 4) The anchor graph based clustering methods can
effective and promising clustering algorithm with always obtain better performance than others. This
good scalability in different domains. phenomenon illustrates that the anchor graph can
2) SFCAG obtains best performance on most datasets. well capture the essence structure of complicates
For Covertype dataset, it is an extremely unbalanced data, which tends to enhance the clustering perfor-
dataset, the clustering results of most comparison mance. It also shows that it is effective and feasible
methods will have empty clusters and harm the clus- to introduce anchor graph technique into fuzzy clus-
tering quality. Nevertheless, FFCAG introduces a tering to improve its scalability.
balanced regularization to constraint the size of each 5) Deep clustering methods have poor performance on
cluster and obtains the best clustering performance. small datasets, which may be caused by over-fitting.
Even so, the proposed method still achieves the sec- For medium and large datasets, DFKM performs bet-
ond best clustering result. ter than PLRSC, since DFKM conducts deep features
3) For anchor graph based methods, FSC and LSC, do extraction and clustering simultaneously, while
not perform as well as the proposed model, since PLRSC first learns a low-ran presentation by deep
they need a post-processing to obtain the clustering encoder and then adopts LSC to segment the repre-
assignment, which may case information loss and sentation. Each independent stage inevitably produ-
poor clustering performance. The clustering results ces information loss and thus unreliability of
of FFCAG are often only next to the proposed SFCAG. clustering assignment.
The reason may be that both FFCAG and the pro- 6) KM and FCM always have poor clustering perfor-
posed method adopt anchor graph technique to bal- mance since they are dependent on the initialization
ance clustering performance and computation cost. of the clustering centers or fuzzy membership matrix
Whereas FFCAG has two parameters that need to be and only can be performed on convex samples. while
manually tuned, the proposed SFCAG is parameter- MSFCM performs better due to that it introduces the
free. It avoids not only the laborious parameter tun- triangle inequality to scale the membership degrees
ing process but also the influence of parameter on the of the selected samples to enhance the clustering
Authorized licensed use limited to: Universita degli Studi di Roma Tor Vergata. Downloaded on March 06,2024 at 16:15:15 UTC from IEEE Xplore. Restrictions apply.
8510 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 35, NO. 8, AUGUST 2023
TABLE 4
Clustering Performance Comparison on Three Small Datasets
TABLE 5
Clustering Performance Comparison on Three Medium Datasets
TABLE 6
Clustering Performance Comparison on two Large Datasets
quality. RSFCM adopts l1 -norm instead of squared Waveform, SensIT and Covertype datasets. The experimen-
Frobenius norm as loss function to enhance robust- tal results are presented in Fig. 2. Since FCM, RSFCM and
ness to the outliers. MSFCM have no parameter to tune, their ACC and running
time remain constant, which are indicated by the green, yel-
4.5 Parameter Sensitivity Analysis low and magenta dashed lines, respectively.
In this subsection, we report a detail parameter analysis From the Fig. 2, we can observe that SFCAG can achieve
about the number of anchors m in the range of ½25 ; 26 ; the best performance in terms of ACC when the number of
27 ; 28 ; 29 ; 210 from the aspects of ACC and running time on anchors increases. When the number of anchors is small, the
Authorized licensed use limited to: Universita degli Studi di Roma Tor Vergata. Downloaded on March 06,2024 at 16:15:15 UTC from IEEE Xplore. Restrictions apply.
LIU ET AL.: SCALABLE FUZZY CLUSTERING WITH ANCHOR GRAPH 8511
Fig. 2. Clustering ACC and running time versus the number of selected anchors on Waveform, SensIT and Covertype datasets.
performance of all anchor graph based methods is poor since The running time of SFCAG is more advantageous
the selected anchors cannot effectively represent the intrinsic than compared methods on medium and large data-
structure of data points. When the number of anchors is 27 , sets. Specifically, the time cost of SFCAG is much
SFCAG can obtain better or comparable ACC compared with less than that of fuzzy clustering methods (FCM,
other methods. In addition, as a general trend, it can be seen RSFCM, and MSFCM), while the time cost of SFCAG
that the running time of the graph-based methods become is comparable to that of anchor graph based methods
larger and larger as the number of anchors increases. Even so, (LSC, FSC, and FFCAG). Although the running time
the running time of the proposed method on large dataset is of the proposed method is larger on small datasets, it
much lower than other fuzzy clustering methods. From these can obtain better clustering performance.
findings, we can conclude that the proposed method SFCAG For small datasets, we can observe that fuzzy cluster-
can achieve an elegant balance between computational bur- ing methods often perform more efficiently, as
den and clustering performance. shown in Figs. 3a, 3b, and 3c. The reason may be that
the distance calculation amount is small since data
4.6 Computational Efficiency points involved in updating the solution per itera-
We now evaluate the computational efficiency of the pro- tion are small, which leads to a high efficiency.
posed SFCAG in terms of running time and scalability. For medium and large datasets, anchor graph based
methods are much faster than fuzzy clustering meth-
ods, as shown in Figs. 3g and 3h. The reason is that the
4.6.1 Running Time number of selected anchors is far less than data points,
The running time of different methods on all datasets is which greatly reduces the computational cost and
illustrated in Fig. 3. From this Figure, we have the following accelerates the subsequent clustering procedure. We
conclusions: can see that the SFCAG achieves the best performance
Authorized licensed use limited to: Universita degli Studi di Roma Tor Vergata. Downloaded on March 06,2024 at 16:15:15 UTC from IEEE Xplore. Restrictions apply.
8512 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 35, NO. 8, AUGUST 2023
Fig. 5. The convergence curves of our method on real datasets. In general, the objective function of each dataset converges less than 5 iterations.
Authorized licensed use limited to: Universita degli Studi di Roma Tor Vergata. Downloaded on March 06,2024 at 16:15:15 UTC from IEEE Xplore. Restrictions apply.
LIU ET AL.: SCALABLE FUZZY CLUSTERING WITH ANCHOR GRAPH 8513
Several questions remain to be investigated in the future. [22] H. Wei, L. Chen, and L. Guo, “Kl divergence-based fuzzy cluster
ensemble for image segmentation,” Entropy, vol. 20, no. 4, 2018,
How to generate anchors to effectively capture the Art. no. 273.
[23] M.-S. Yang and Y. Nataliani, “A feature-reduction fuzzy cluster-
manifold structure of data points is important. There- ing algorithm based on feature-weighted entropy,” IEEE Trans.
fore, more elegant methods need to be designed to Fuzzy Syst., vol. 26, no. 2, pp. 817–835, Apr. 2018.
select the optimal representative anchors in the future. [24] S. Eschrich, J. Ke, L. O. Hall, and D. B. Goldgof, “Fast accurate
fuzzy clustering through data reduction,” IEEE Trans. Fuzzy Syst.,
Designing a good anchor graph to more accurately vol. 11, no. 2, pp. 262–270, Apr. 2003.
represent the similarities between data points and [25] M. B. Al-Zoubi, A. Hudaib, and B. Al-Shboul, “A fast fuzzy clus-
anchors is also a good direction of future work. tering algorithm,” in Proc. 6th WSEAS Int. Conf. Artif. Intell. Knowl.
Eng. Data Bases, 2007, pp. 28–32.
[26] J. K. Parker and L. O. Hall, “Accelerating fuzzy-c means using an
REFERENCES estimated subsample size,” IEEE Trans. Fuzzy Syst., vol. 22, no. 5,
[1] J. Gu, L. Jiao, S. Yang, and F. Liu, “Fuzzy double c-means cluster- pp. 1229–1244, Oct. 2014.
ing based on sparse self-representation,” IEEE Trans. Fuzzy Syst., [27] R. J. Hathaway and Y. Hu, “Density-weighted fuzzy c-means
vol. 26, no. 2, pp. 612–626, Apr. 2017. clustering,” IEEE Trans. Fuzzy Syst., vol. 17, no. 1, pp. 243–252,
[2] T. Lei, X. Jia, Y. Zhang, L. He, H. Meng, and A. K. Nandi, Feb. 2009.
“Significantly fast and robust fuzzy C-means clustering algorithm [28] I. A. Atiyah, A. Mohammadpour, and S. M. Taheri, “KC-means: A
based on morphological reconstruction and membership filter- fast fuzzy clustering,” Adv. Fuzzy Syst., vol. 2018, pp. 1–8, 2018.
ing,” IEEE Trans. Fuzzy Syst., vol. 26, no. 5, pp. 3027–3041, Oct. [29] K. Zou, Z. Wang, and M. Hu, “An new initialization method for
2018. fuzzy c-means algorithm,” Fuzzy Optim. Decis. Mak., vol. 7, no. 4,
[3] I.-J. Chiang, C. C.-H. Liu, Y.-H. Tsai, and A. Kumar, “Discovering pp. 409–416, 2008.
latent semantics in web documents using fuzzy clustering,” IEEE [30] Q. Yang, D. Zhang, and F. Tian, “An initialization method for
Trans. Fuzzy Syst., vol. 23, no. 6, pp. 2122–2134, Dec. 2015. fuzzy c-means algorithm using subtractive clustering,” in Proc.
[4] J.-P. Mei, Y. Wang, L. Chen, and C. Miao, “Large scale document 3th Int. Conf. Intell. Netw. Intell. Syst., 2010, pp. 393–396.
categorization with fuzzy clustering,” IEEE Trans. Fuzzy Syst., [31] Z. Cebeci, “Initialization of membership degree matrix for fast
vol. 25, no. 5, pp. 1239–1251, Oct. 2017. convergence of fuzzy c-means clustering,” in Proc. Int. Conf. on
[5] L. Hu and K. C. Chan, “Fuzzy clustering in a complex network Artif. Intell. Data Process., 2018, pp. 1–5.
based on content relevance and link structures,” IEEE Trans. Fuzzy [32] Z. Cebeci and C. Cebeci, “A fast algorithm to initialize cluster
Syst., vol. 24, no. 2, pp. 456–470, Apr. 2015. centroids in fuzzy clustering applications,” Information, vol. 11,
[6] A. Pister, P. Buono, J.-D. Fekete, C. Plaisant, and P. Valdivia, no. 9, 2020, Art. no. 446.
“Integrating prior knowledge in mixed-initiative social network [33] Y. Shen, W. Pedrycz, Y. Chen, X. Wang, and A. Gacek,
clustering,” IEEE Trans. Vis. Comput. Graphics, vol. 27, no. 2, “Hyperplane division in fuzzy c-means: Clustering big data,”
pp. 1775–1785, Feb. 2021. IEEE Trans. Fuzzy Syst., vol. 28, no. 11, pp. 3032–3046, Nov. 2020.
[7] J. MacQueen et al., “Some methods for classification and analysis [34] S. Zhou, D. Li, Z. Zhang, and R. Ping, “A new membership scaling
of multivariate observations,” in Proc. 5th Berkeley Symp. Math. fuzzy C-means clustering algorithm,” IEEE Trans. Fuzzy Syst.,
Statist. Probability, 1967, pp. 281–297. vol. 29, no. 9, pp. 2810–2818, Sep. 2020.
[8] J. Ye, Z. Zhao, and M. Wu, “Discriminative K-means for clustering,” [35] M. Wang, W. Fu, S. Hao, D. Tao, and X. Wu, “Scalable semi-super-
in Proc. Adv. Neural Informat. Process. Syst., 2007, pp. 1649–1656. vised learning by efficient anchor graph regularization,” IEEE
[9] C. Boutsidis, A. Zouzias, M. W. Mahoney, and P. Drineas, Trans. Knowl. Data Eng., vol. 28, no. 7, pp. 1864–1877, Jul. 2016.
“Randomized dimensionality reduction for K-means clustering,” [36] H. Hu, K. Wang, C. Lv, J. Wu, and Z. Yang, “Semi-supervised
IEEE Trans. Inf. Theory, vol. 61, no. 2, pp. 1045–1062, Feb. 2015. metric learning-based anchor graph hashing for large-scale image
[10] X. Shen, W. Liu, I. Tsang, F. Shen, and Q.-S. Sun, “Compressed K- retrieval,” IEEE Trans. Image Process., vol. 28, no. 2, pp. 739–754,
means for large-scale clustering,” in Proc. 31th AAAI Conf. Artif. Feb. 2019.
Intell., 2017, pp. 2527–2533. [37] D. Cai and X. Chen, “Large scale spectral clustering via landmark-
[11] U. Von Luxburg, “A tutorial on spectral clustering,” Statist. Com- based sparse representation,” IEEE Trans. Cybern., vol. 45, no. 8,
put., vol. 17, no. 4, pp. 395–416, 2007. pp. 1669–1680, Aug. 2014.
[12] Y. Pang, J. Xie, F. Nie, and X. Li, “Spectral clustering by joint spec- [38] W. Zhu, F. Nie, and X. Li, “Fast spectral clustering with efficient
tral embedding and spectral rotation,” IEEE Trans. Cybern., large graph construction,” in Proc. IEEE Int. Conf. Acoust. Speech
vol. 50, no. 1, pp. 247–258, Jan. 2020. Signal Process., 2017, pp. 2492–2496.
[13] Z. Wang, Z. Li, R. Wang, F. Nie, and X. Li, “Large graph clustering [39] H. Hu, R. Wang, F. Nie, X. Yang, and W. Yu, “Fast unsupervised fea-
with simultaneous spectral embedding and discretization,” IEEE ture selection with anchor graph and l_2, 1-norm regularization,”
Trans. Pattern Anal. Mach. Intell., vol. 43, no. 12, pp. 4426–4440, Multimedia Tools Appl., vol. 77, no. 17, pp. 22099–22E113, 2018.
Dec. 2020. [40] F. Nie, C. Liu, R. Wang, Z. Wang, and X. Li, “Fast fuzzy clustering
[14] L. Zadeh, “Fuzzy sets,” Inf. Control, vol. 8, no. 3, pp. 338–353, 1965. based on anchor graph,” IEEE Trans. Fuzzy Syst., vol. 30, no. 7,
[15] J. C. Bezdek, R. Ehrlich, and W. Full, “FCM: The fuzzy c-means clus- pp. 2375–2387, Jul. 2022.
tering algorithm,” Comput. Geosci.s, vol. 10, no. 2/3, pp. 191–203, [41] A. Guillon, M.-J. Lesot, and C. Marsala, “Laplacian regularization
1984. for fuzzy subspace clustering,” in Proc. IEEE Int. Conf. Fuzzy Syst.,
[16] S. Krinidis and V. Chatzis, “A robust fuzzy local information c- 2017, pp. 1–6.
means clustering algorithm,” IEEE Trans. Image Process., vol. 19, [42] F. Nie, X. Wang, and H. Huang, “Clustering and projected cluster-
no. 5, pp. 1328–1337, May 2010. ing with adaptive neighbors,” in Proc. 20th ACM SIGKDD Int.
[17] J. Xu, J. Han, K. Xiong, and F. Nie, “Robust and sparse fuzzy k- Conf. Knowl. Discov. Data Mining, 2014, pp. 977–986.
means clustering,” in Proc. 25th Int. Joint Conf. Artif. Intell., 2016, [43] X. Zhu and J. Lafferty, “Harmonic mixtures: Combining mixture
pp. 2224–2230. models and graph-based methods for inductive and scalable
[18] Z. Bian, H. Ishibuchi, and S. Wang, “Joint learning of spectral clus- semi-supervised learning,” in Proc. 22nd Int. Conf. Mach. Learn.,
tering structure and fuzzy similarity matrix of data,” IEEE Trans. 2005, pp. 1052–1059.
Fuzzy Syst., vol. 27, no. 1, pp. 31–44, Jan. 2019. [44] W. Liu, J. He, and S.-F. Chang, “Large graph construction for scal-
[19] L. Guo, L. Chen, X. Lu, and C. P. Chen, “Membership affinity able semi-supervised learning,” in Proc. 27th Int. Conf. Mach.
lasso for fuzzy clustering,” IEEE Trans. Fuzzy Syst., vol. 28, no. 2, Learn., 2010, pp. 679–686.
pp. 294–307, Feb. 2020. [45] M. R. Hestenes, “Multiplier and gradient methods,” J. Optim. The-
[20] R. Zhang, X. Li, H. Zhang, and F. Nie, “Deep fuzzy k-means with ory Appl., vol. 4, no. 5, pp. 303–320, 1969.
adaptive loss and entropy regularization,” IEEE Trans. Fuzzy [46] D. P. Bertsekas, Constrained optimization and Lagrange Multiplier
Syst., vol. 28, no. 11, pp. 2814–2824, Nov. 2020. Methods. New York, NY, USA: Academic Press, 2014.
[21] L. Chen, C. P. Chen, and M. Lu, “A multiple-kernel fuzzy c-means [47] J. Huang, F. Nie, and H. Huang, “A new simplex sparse learning
algorithm for image segmentation,” IEEE Trans. Syst., Man, model to measure data similarity for clustering,” in Proc. 24th Int.
Cybern., vol. 41, no. 5, pp. 1263–1274, Oct. 2011. Joint Conf. Artif. Intell., 2015, pp. 3569–3575.
Authorized licensed use limited to: Universita degli Studi di Roma Tor Vergata. Downloaded on March 06,2024 at 16:15:15 UTC from IEEE Xplore. Restrictions apply.
8514 IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 35, NO. 8, AUGUST 2023
[48] L. Deng, “The MNIST database of handwritten digit images for Rong Wang received the BS degree in informa-
machine learning research [best of the web],” IEEE Signal Process. tion engineering, the MS degree in signal and
Mag., vol. 29, no. 6, pp. 141–142, Nov. 2012. information processing, and the PhD degree in
[49] G. Cohen, S. Afshar, J. Tapson, and A. Van Schaik, “EMNIST: computer science from Xian Research Institute of
Extending MNIST to handwritten letters,” in Proc. Int. Joint Conf. Hi-Tech, Xian, China, in 2004, 2007 and 2013,
Neural Netw., 2017, pp. 2921–2926. respectively. During 2007 and 2013, he received
[50] J. Li and H. Liu, “Projective low-rank subspace clustering via the PhD degree in the Department of Automation,
learning deep encoder,” in Proc. 26th Int. Joint Conf. Artif. Intell., Tsinghua University, Beijing, China. He is cur-
2017, pp. 2145–2151. rently an associate professor at the School of
[51] K. Zhan, F. Nie, J. Wang, and Y. Yang, “Multiview consensus graph Cybersecurity and School of Artificial Intelligence,
clustering,” IEEE Trans. Image Process., vol. 28, no. 3, pp. 1261–1270, Optics and Electronics (iOPEN), Northwestern
Mar. 2019. Polytechnical University, Xian, China. His research interests focus on
machine learning and its applications.
Chaodie Liu is currently working toward the PhD
degree in the School of Computer Science and
School of Artificial Intelligence, Optics and Electron-
ics (iOPEN), Northwestern Polytechnical University, Xuelong Li (Fellow, IEEE) is a full professor with
Xi’an, Shaanxi, China. Her research interests include the School of Artificial Intelligence, Optics and
machine learning and pattern recognition. Electronics (iOPEN), Northwestern Polytechnical
University, Xi’an, China.
Authorized licensed use limited to: Universita degli Studi di Roma Tor Vergata. Downloaded on March 06,2024 at 16:15:15 UTC from IEEE Xplore. Restrictions apply.