100% satisfaction guarantee Immediately available after payment Both online and in PDF No strings attached
logo-home
Summary Ecological Methods: Applied Statistics (visual) $6.68   Add to cart

Summary

Summary Ecological Methods: Applied Statistics (visual)

 8 views  1 purchase
  • Course
  • Institution

This is a summary of the third part of Ecological Methods (WEC31806), namely Applied Statistics. The summary shows all the lectures in the applied part and has been made extra visual in order to understand the material quickly and properly. In addition, with sufficient examples.

Preview 4 out of 35  pages

  • October 3, 2024
  • 35
  • 2024/2025
  • Summary
avatar-seller
16. Cluster analysis
Clustering: grouping data points based on similarity
Data points within a cluster are similar to each other, and dissimilar do data points in other clusters
→ useful to find groups that are assumed to exist in reality (e.g. vegetation type, animal behaviour)

Clustering = partitioning (same term)

- Clustering is not about revealing gradients
→ ordination is about revealing gradients
→ clustering is about detecting discrete groups with small differences between members
- Clustering is not the same as classification
→ classification is about creating groups based on known labels
Similarity and dissimilarity are essential components of clustering analysis
Distance between pairs of:
→ points
→ cluster of points

Types of clustering:

- Flat clustering (K-means clustering): creates a flat set of clusters without any structure
- Hierarchical clustering: creates a hierarchy of clusters (thus within internal structure)
Flat clustering (K-means)
K-means: the simplest clustering algorithm, where we must define a target number K, which refers to
the number of means (centers) we want our dataset to partition around.
→ Each observation is assigned to the cluster with the nearest mean
→ Only deals with difference between clusters and not within clusters

Steps:

1. Randomly locates initial cluster centers
2. Assign records to nearest cluster mean
3. Compute new cluster means
4. Repeats 2 & 3 a few iterations
→ new data points can be assigned to the cluster
with the nearest center
→ disadvantage: number of clusters is assigned by
eye



Learning algorithm: algorithm that learns; tries a
few times and then knows a definite outcome.
→ does not necessarily result in exactly the same outcome when the analyses is repeated

,Hierarchical clustering
Hierarchical clustering does not require us to pre-
specify the number of clusters to be generated,
and results in a dendrogram

Dendrogram: tree-like diagram that records the
sequences of merges or splits




Root node: upper node where all samples belong to
Leaf (terminal node): cluster with only one sample

→ similarity of two observations is bases on the height where branches containing those two
observations first are fused
→ we cannot use the proximity of two observations along the horizontal axis for similarity

Types of hierarchical clustering:

- Agglomerative clustering (merges): builds nested clusters by merging smaller cluster with a
bottom-up approach
- Divisive clustering (splits): builds nested cluster by merging smaller clusters with a top-down
approach




Disadvantage: when a new datapoint is
added, the entire dendrogram needs to
be recalculated

,Similarity and dissimilarity
Distance between pairs
Euclidean How the crow flies




Manhattan How the taxi drives
→ distance along the axis




Jaccard Intersection/union: relative similarity
→ for binary data




Jaccard distance:




0.67 → 4 out of 6 species differ between the sites

, If variables differ in measure (e.g. temperature and weight), scale the columns to mean = 0, sd = 1
→ same scal



Linkage: how we quantify the dissimilarity between clusters:

- Single: minimum distance between clusters
- Often leading to clusters with different size
- Shape of clusters can become elongated
- Complete: maximum distance between clusters
- Size of clusters become more compact
- Average: average between clusters - Handles outliers and noise well
- Ward: minimum variance method - Lead to more uniformly sized clusters
- More difficult to compute, thus slower for
large datasets

The benefits of buying summaries with Stuvia:

Guaranteed quality through customer reviews

Guaranteed quality through customer reviews

Stuvia customers have reviewed more than 700,000 summaries. This how you know that you are buying the best documents.

Quick and easy check-out

Quick and easy check-out

You can quickly pay through credit card or Stuvia-credit for the summaries. There is no membership needed.

Focus on what matters

Focus on what matters

Your fellow students write the study notes themselves, which is why the documents are always reliable and up-to-date. This ensures you quickly get to the core!

Frequently asked questions

What do I get when I buy this document?

You get a PDF, available immediately after your purchase. The purchased document is accessible anytime, anywhere and indefinitely through your profile.

Satisfaction guarantee: how does it work?

Our satisfaction guarantee ensures that you always find a study document that suits you well. You fill out a form, and our customer service team takes care of the rest.

Who am I buying these notes from?

Stuvia is a marketplace, so you are not buying this document from us, but from seller niekvandeven. Stuvia facilitates payment to the seller.

Will I be stuck with a subscription?

No, you only buy these notes for $6.68. You're not tied to anything after your purchase.

Can Stuvia be trusted?

4.6 stars on Google & Trustpilot (+1000 reviews)

84190 documents were sold in the last 30 days

Founded in 2010, the go-to place to buy study notes for 14 years now

Start selling
$6.68  1x  sold
  • (0)
  Add to cart