Editing
Unsupervised Learning and Clustering
(section)
Jump to navigation
Jump to search
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== <span style="color: #FFFFFF;">Remembering</span> == * '''Unsupervised learning''' β Learning patterns from data without labels or target outputs. * '''Clustering''' β Partitioning data into groups (clusters) where items within a group are more similar to each other than to items in other groups. * '''K-means''' β A centroid-based clustering algorithm that iteratively assigns points to the nearest of K cluster centers and updates centers. * '''Hierarchical clustering''' β Builds a tree (dendrogram) of nested clusters; agglomerative (bottom-up) or divisive (top-down). * '''DBSCAN (Density-Based Spatial Clustering of Applications with Noise)''' β Finds clusters as dense regions separated by low-density areas; can discover arbitrarily shaped clusters. * '''Gaussian Mixture Model (GMM)''' β A probabilistic model representing the data as a mixture of K Gaussian distributions; fit by EM algorithm. * '''Expectation-Maximization (EM)''' β An iterative algorithm for fitting latent variable models; alternates between expectation (compute responsibility) and maximization (update parameters). * '''PCA (Principal Component Analysis)''' β Linear dimensionality reduction that finds the directions of maximum variance. * '''t-SNE''' β A non-linear dimensionality reduction for visualization; preserves local neighborhood structure. * '''UMAP''' β Faster, more scalable alternative to t-SNE for visualization; better preserves global structure. * '''Autoencoder''' β A neural network trained to reconstruct its input through a bottleneck; the bottleneck gives a compressed representation. * '''Silhouette score''' β A clustering quality metric measuring how similar a point is to its own cluster vs. other clusters; range [-1, 1]. * '''Elbow method''' β Heuristic for choosing K in K-means: plot inertia vs. K, choose the "elbow" where improvement slows. * '''Inertia''' β Within-cluster sum of squared distances from each point to its cluster center; K-means minimizes this. </div> <div style="background-color: #006400; color: #FFFFFF; padding: 20px; border-radius: 8px; margin-bottom: 15px;">
Summary:
Please note that all contributions to BloomWiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
BloomWiki:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Navigation menu
Personal tools
Not logged in
Talk
Contributions
Create account
Log in
Namespaces
Page
Discussion
English
Views
Read
Edit
View history
More
Search
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Tools
What links here
Related changes
Special pages
Page information