##
**A clustering ensemble framework based on elite selection of weighted clusters.**
*(English)*
Zbl 1267.62076

Summary: Each clustering algorithm usually optimizes a qualification metric during its progress. The qualification metric in conventional clustering algorithms considers all the features equally important; in other words each feature participates in the clustering process equivalently. It is obvious that some features have more information than others in a data set. So it is highly likely that some features should have lower importance degrees during a clustering or a classification algorithm; due to their lower information or their higher variances, etc. So it is always a desire for all artificial intelligence communities to enforce the weighting mechanism in any task that identically uses a number of features to make a decision. But there is always a certain problem of how the features can be participated in the clustering process (in any algorithm, but especially in clustering algorithm) in a weighted manner. Recently, this problem is dealt with by locally adaptive clustering (LAC). However, like its traditional competitors the LAC suffers from inefficiency in data with imbalanced clusters. This paper solves the problem by proposing a weighted locally adaptive clustering (WLAC) algorithm that is based on the LAC algorithm. However, WLAC algorithm suffers from sensitivity to its two parameters that should be tuned manually. The performance of WLAC algorithm is affected by well-tuning of its parameters. This paper proposes two solutions. The first is based on a simple clustering ensemble framework to examine the sensitivity of the WLAC algorithm to its manual well-tuning. The second is based on cluster selection method.

### MSC:

62H30 | Classification and discrimination; cluster analysis (statistical aspects) |

65C60 | Computational problems in statistics (MSC2010) |

### Software:

UCI-ml
PDF
BibTeX
XML
Cite

\textit{H. Parvin} and \textit{B. Minaei-Bidgoli}, Adv. Data Anal. Classif., ADAC 7, No. 2, 181--208 (2013; Zbl 1267.62076)

Full Text:
DOI

### References:

[1] | Agrawal R, Gehrke J, Gunopulos D, Raghavan P (1998) Automatic subspace clustering of high dimensional data for data mining applications. In: Proceedings of the 1998 ACM SIGMOD international conference on Management of data, pp 94–105 |

[2] | Alizadeh H, Minaei-Bidgoli B, Parvin H (2011a) A new criterion for clusters validation. Artificial Intelligence Applications and innovations, IFIP advances in information and communication technology, vol 364, pp 110–115 |

[3] | Alizadeh H, Minaei-Bidgoli B, Parvin H, Moshki M (2011b) An asymmetric criterion for cluster validation. Developing concepts in applied intelligence, Studies in computational intelligence, vol 363, pp 1–14 |

[4] | Blum A, Rivest R (1992) Training a 3-node neural network is NP-complete. Neural Netw 5:117–127 |

[5] | Chang JW, Jin DS (2002) A new cell-based clustering method for large-high dimensional data in data mining applications. In: Proceedings of the ACM symposium on Applied computing, pp 503–507 |

[6] | Cheng CH, Fu AW, Zhang Y (1999) Entropy-based subspace clustering for mining numerical data. In: Proceedings of the fifth ACM SIGKDD international conference on knowledge discovery and data mining, pp 84–93 |

[7] | Domeniconi C, Al-Razgan M (2009) Weighted cluster ensembles: methods and analysis. ACM Trans Knowl Discov Data. doi: 10.1145/1460797.1460800 |

[8] | Domeniconi C, Gunopulos D, Ma S, Yan B, Al-Razgan M, Papadopoulos D (2007) Locally adaptive metrics for clustering high dimensional data. Data Min Knowl Discov 14:63–97 · Zbl 05146452 |

[9] | Dudoit S, Fridlyand J (2003) Bagging to improve the accuracy of a clustering procedure. Bioinformatics 19:1090–1099 |

[10] | Faceli K, Marcilio CP, Souto D (2006) Multi-objective clustering ensemble. In: Proceedings of the sixth international conference on hybrid intelligent systems (HIS’06) |

[11] | Fred A (2001) Finding consistent clusters in data partitions. In: Second international workshop on multiple classifier systems, pp 309–318 · Zbl 0980.68682 |

[12] | Fred A, Jain AK (2002a) Data clustering using evidence accumulation. In: Proceedings of the 16th international conference on pattern recognition, pp 276–280 |

[13] | Fred A, Jain AK (2002b) Evidence accumulation clustering based on the k-means algorithm. In: Joint IAPR international workshops on structural, syntactic, and statistical pattern recognition, pp 442–451 · Zbl 1073.68743 |

[14] | Fred A, Jain AK (2005) Combining multiple clusterings using evidence accumulation. IEEE Trans Pattern Anal Mach Intell 27:835–850 · Zbl 05111568 |

[15] | Fern XZ, Lin W (2008) Cluster ensemble selection. In: SIAM international conference on data mining, pp 787–797 |

[16] | Jain AK, Dubes RC (1998) Algorithms for clustering data. Prentice Hall, Englewood Cliffs |

[17] | Kohavi R, John RG (1997) Wrappers for feature subset selection. Artif Intell 97:273–324 · Zbl 0904.68143 |

[18] | Liu B, Xia Y, Yu PS (2000) Clustering through decision tree construction. In: Proceedings of the ninth international conference on information and knowledge management, pp 20–29 |

[19] | Miller R, Yang Y (1997) Association rules over interval data. In: Proceedings of ACM SIGMOD international conference on management of data, pp 452–461 |

[20] | Mirzaei A, Rahmati M, Ahmadi M (2008) A new method for hierarchical clustering combination. Intell Data Anal 12:549–571 |

[21] | Minaei-Bidgoli B, Parvin H, Alinejad H, Alizadeh H, Punch W (2011) Effects of resampling method and adaptation on clustering ensemble efficacy. Artif Intell Rev. doi: 10.1007/s10462-011-9295-x |

[22] | Munkres J (1957) Algorithms for the assignment and transportation problems. J Soc Indus Appl Math 5:32–38 · Zbl 0083.15302 |

[23] | Newman CBDJ, Hettich S, Merz C (1998) UCI repository of machine learning databases. http://www.ics.uci.edu/\(\sim\)mlearn/MLSummary.html |

[24] | Parsons L, Haque E, Liu H (2004) Subspace clustering for high dimensional data: a review. ACM SIGKDD Explor Newsl 6:90–105 · Zbl 05442873 |

[25] | Parvin H, Beigi A, Mozayani N (2012a) A clustering ensemble learning method based on the ant colony clustering algorithm. Int J Appl Comput Math 11:286–302 |

[26] | Parvin H, Minaei-Bidgoli B, Parvin S, Alinejad H (2012b) A new classifier ensemble methodology based on subspace learning. J Exp Theor Artif Intell. doi: 10.1080/0952813X.2012.715683 · Zbl 1267.62076 |

[27] | Procopiuc CM, Jones M, Agarwal PK, Murali TM (2002) A Monte Carlo algorithm for fast projective clustering. In: Proceedings of the ACM SIGMOD conference on management of data, pp 418–427 |

[28] | Srikant R, Agrawal R (1996) Mining quantitative association rules in large relational tables. In: Proceedings of the ACM SIGMOD conference on management of data |

[29] | Strehl A, Ghosh J (2002) Cluster ensembles–a knowledge reuse framework for combining multiple partitions. J Mach Learn Res 3:583–617 · Zbl 1084.68759 |

This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. It attempts to reflect the references listed in the original paper as accurately as possible without claiming the completeness or perfect precision of the matching.