<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wikimd.org/index.php?action=history&amp;feed=atom&amp;title=Hierarchical_clustering</id>
	<title>Hierarchical clustering - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://wikimd.org/index.php?action=history&amp;feed=atom&amp;title=Hierarchical_clustering"/>
	<link rel="alternate" type="text/html" href="https://wikimd.org/index.php?title=Hierarchical_clustering&amp;action=history"/>
	<updated>2026-04-27T02:19:13Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.44.2</generator>
	<entry>
		<id>https://wikimd.org/index.php?title=Hierarchical_clustering&amp;diff=5653135&amp;oldid=prev</id>
		<title>Prab: CSV import</title>
		<link rel="alternate" type="text/html" href="https://wikimd.org/index.php?title=Hierarchical_clustering&amp;diff=5653135&amp;oldid=prev"/>
		<updated>2024-04-24T01:57:28Z</updated>

		<summary type="html">&lt;p&gt;CSV import&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;[[Image:Clusters.svg|Clusters|thumb]] &amp;#039;&amp;#039;&amp;#039;Hierarchical clustering&amp;#039;&amp;#039;&amp;#039; is a method of [[cluster analysis]] which seeks to build a hierarchy of [[clusters]]. Strategies for hierarchical clustering generally fall into two types: *Agglomerative*: This is a &amp;quot;bottom-up&amp;quot; approach where each observation starts in its own cluster, and pairs of clusters are merged as one moves up the hierarchy. *Divisive*: This is a &amp;quot;top-down&amp;quot; approach where all observations start in one cluster, and splits are performed recursively as one moves down the hierarchy. The results of hierarchical clustering are usually presented in a [[dendrogram]].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
In the field of [[data analysis]], hierarchical clustering is a powerful tool that allows the analyst to identify the natural groupings or structures within a dataset. Unlike [[k-means clustering]], which requires the analyst to specify the number of clusters beforehand, hierarchical clustering does not require the number of clusters to be specified in advance, making it particularly useful for exploratory data analysis.&lt;br /&gt;
&lt;br /&gt;
== Algorithm ==&lt;br /&gt;
The algorithm for hierarchical clustering can be described as follows:&lt;br /&gt;
&lt;br /&gt;
=== Agglomerative Clustering ===&lt;br /&gt;
# Start by treating each [[data point]] as a single cluster.&lt;br /&gt;
# Find the closest (most similar) pair of clusters and merge them into a single cluster.&lt;br /&gt;
# Compute distances (similarities) between the new cluster and each of the old clusters.&lt;br /&gt;
# Repeat steps 2 and 3 until all items are clustered into a single cluster of size &amp;#039;&amp;#039;n&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
=== Divisive Clustering ===&lt;br /&gt;
# Start with all observations in a single cluster.&lt;br /&gt;
# Find the cluster to split and how to split it.&lt;br /&gt;
# Perform the split to create two new clusters.&lt;br /&gt;
# Repeat steps 2 and 3 until each observation is in its own cluster.&lt;br /&gt;
&lt;br /&gt;
== Distance Measures ==&lt;br /&gt;
The choice of distance measure is a critical step in clustering. It defines how the similarity of two elements is calculated and it will influence the shape of the clusters. The most common distance measures used in hierarchical clustering are:&lt;br /&gt;
&lt;br /&gt;
* [[Euclidean distance]]: The standard distance measure also known as straight-line distance.&lt;br /&gt;
* [[Manhattan distance]]: Sum of the absolute differences of their Cartesian coordinates also known as city block distance.&lt;br /&gt;
* [[Cosine similarity]]: Measures the cosine of the angle between two vectors.&lt;br /&gt;
&lt;br /&gt;
== Applications ==&lt;br /&gt;
Hierarchical clustering is widely used in various fields such as:&lt;br /&gt;
&lt;br /&gt;
* [[Biology]], for constructing phylogenetic trees.&lt;br /&gt;
* [[Information retrieval]], for document clustering.&lt;br /&gt;
* [[Social sciences]], for clustering individuals based on their characteristics.&lt;br /&gt;
* [[Market research]], for customer segmentation.&lt;br /&gt;
&lt;br /&gt;
== Advantages and Disadvantages ==&lt;br /&gt;
=== Advantages ===&lt;br /&gt;
* Does not require the number of clusters to be specified in advance.&lt;br /&gt;
* Easy to implement and provides hierarchical relationships among the observations.&lt;br /&gt;
&lt;br /&gt;
=== Disadvantages ===&lt;br /&gt;
* Can be computationally expensive, especially for large datasets.&lt;br /&gt;
* The results can be sensitive to the choice of distance measure and linkage criteria.&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
* [[Cluster analysis]]&lt;br /&gt;
* [[K-means clustering]]&lt;br /&gt;
* [[Dendrogram]]&lt;br /&gt;
* [[Data analysis]]&lt;br /&gt;
&lt;br /&gt;
[[Category:Cluster analysis]]&lt;br /&gt;
[[Category:Data mining]]&lt;br /&gt;
[[Category:Machine learning]]&lt;br /&gt;
&lt;br /&gt;
{{stb}}&lt;/div&gt;</summary>
		<author><name>Prab</name></author>
	</entry>
</feed>