Clustering is a widely deployed unsupervised learning tool. Given data in the Euclidean spoace, K-means clustering is one of the most commonly used clustering methods, which minimize the distance between each point to the centroid of its assigned cluster. Among the popular clustering methods, SDP clustering enjoys the strongest statistical guarantees under the standard Gaussian mixture models in that it achieves an information-theoretic bound for exact recovery. However, the original SDP method is limited to isotropic covariance matrices for Gaussians, and it has prohibitively high costs of solving the SDP optimization problem. This project wants to develop algorithms to improve the computational efficiency and extend the results to more general cases in the following aspects: Extend the algorithms and results to heterogeneous data as well as other types of data like distributions or measures; develop algorithms to enhance the computational performance for SDP or to efficiently solve the SDP for clustering; propose 1-st order and 2-nd order methods to solve general non-negative SDP optimization problems with minimal assumptions, which can be applied to various hidden community detection tasks.
Use this login method if you
don't
have an
@illinois.edu
email address.
(Oops, I do have one)
IDEALS migrated to a new platform on June 23, 2022. If you created
your account prior to this date, you will have to reset your password
using the forgot-password link below.