Show simple item record

dc.contributor.authorBandyapadhyay, Sayan
dc.contributor.authorFomin, Fedor
dc.contributor.authorGolovach, Petr
dc.contributor.authorLochet, William Alexandre
dc.contributor.authorPurohit, Nidhi
dc.contributor.authorSimonov, Kirill
dc.date.accessioned2024-04-17T11:20:15Z
dc.date.available2024-04-17T11:20:15Z
dc.date.created2023-06-15T18:06:03Z
dc.date.issued2023
dc.identifier.issn0004-3702
dc.identifier.urihttps://hdl.handle.net/11250/3127010
dc.description.abstractk-means and k-median clustering are powerful unsupervised machine learning techniques. However, due to complicated dependencies on all the features, it is challenging to interpret the resulting cluster assignments. Moshkovitz, Dasgupta, Rashtchian, and Frost proposed an elegant model of explainable k-means and k-median clustering in ICML 2020. In this model, a decision tree with k leaves provides a straightforward characterization of the data set into clusters. We study two natural algorithmic questions about explainable clustering. (1) For a given clustering, how to find the “best explanation” by using a decision tree with k leaves? (2) For a given set of points, how to find a decision tree with k leaves minimizing the k-means/median objective of the resulting explainable clustering? To address the first question, we introduce a new model of explainable clustering. Our model, inspired by the notion of outliers in robust statistics, is the following. We are seeking a small number of points (outliers) whose removal makes the existing clustering well-explainable. For addressing the second question, we initiate the study of the model of Moshkovitz et al. from the perspective of multivariate complexity. Our rigorous algorithmic analysis sheds some light on the influence of parameters like the input size, dimension of the data, the number of outliers, the number of clusters, and the approximation ratio, on the computational complexity of explainable clustering.en_US
dc.language.isoengen_US
dc.publisherElsevieren_US
dc.rightsNavngivelse 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/deed.no*
dc.titleHow to find a good explanation for clustering?en_US
dc.typeJournal articleen_US
dc.typePeer revieweden_US
dc.description.versionpublishedVersionen_US
dc.rights.holderCopyright 2023 The Author(s)en_US
dc.source.articlenumber103948en_US
cristin.ispublishedtrue
cristin.fulltextoriginal
cristin.qualitycode2
dc.identifier.doi10.1016/j.artint.2023.103948
dc.identifier.cristin2155058
dc.source.journalArtificial Intelligenceen_US
dc.relation.projectNorges forskningsråd: 314528en_US
dc.identifier.citationArtificial Intelligence. 2023, 322, 103948.en_US
dc.source.volume322en_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Navngivelse 4.0 Internasjonal
Except where otherwise noted, this item's license is described as Navngivelse 4.0 Internasjonal