. Gan G., Ma Ch,. Wu J. (2007), Data clustering. Theory, algorithms, and applications. Philadelphia, Pennsilvania: SIAM. 455 p.
2. Xu R., Wunsch D.C. (2008), Clustering. Hoboken. John Wiley & Sons. 358 p.
3. Pelleg D., Moore A. (2000), X-means: Extending K-means with efficient estimation of the number of clusters. Proc. of the 17th int. conf. on Machine Learning. San Francisco: Morgan Kaufmann. pp. 727-730.
4. Ishioka T. (2000), Extended K-means with an efficient estimation of the number of clusters. Proc. of second int. conf. Intelligent Data Engineering and Automated Learning IDEAL. 2000. Hong Kong, China. pp.17-22.
5. Ishioka T. (2005), An expansion of X-means for automatically determining the optimal number of clusters. Proc. of the 4th IASTED Int. conf. Computational intelligence. Calgary, Alberta, Canada. pp. 91-96.
6. Schwarz G. (1978), Estimation the dimension of a model. The Annals of Statistics, Vol. 6(2). pp. 461-464.
7. Bozdogan H. (1987), Model selection and Akaike’s information criterion (AIC): The general theory and its analytical extensions. Psychametrica. Vol. 52. pp. 345-370.
8. Jolion J.M., Meer P., Bataouche S.(1991) Robust clustering with applications in computer vision. IEEE Trans.on Pattern Analysis and Machine Intelligence. Vol 13 pp. 291 - 802.
9. Krishnapuram R., Freg C.P. (1992),Fitting and unknown numbers of lines and planes to image to image data through compatible cluster merging. Pattern Recognition. Vol. 25. pp. 385-400.
10. Kass B., Wasserman L. (1995), A reference Bayesian test for nested hypotheses and its relationship to the Schwarz criterion. J. of the
Amer. Statistical Association. Vol. 90. pp.773-795.