A guide on how to use OpenAI embeddings and clustering techniques to analyze survey data and extract meaningful topics and actionable insights from the responses.
The process involves transforming textual survey responses into embeddings, grouping similar responses through clustering, and then identifying key themes or topics to aid in business improvement.
A visual representation of papers on ArXiv using UMAP and nomic-embed.
An article discussing the use of embeddings in natural language processing, focusing on comparing open source and closed source embedding models for semantic search, including techniques like clustering and re-ranking.
ASCVIT V1 aims to make data analysis easier by automating statistical calculations, visualizations, and interpretations.
Includes descriptive statistics, hypothesis tests, regression, time series analysis, clustering, and LLM-powered data interpretation.
- Accepts CSV or Excel files. Provides a data overview including summary statistics, variable types, and data points.
- Histograms, boxplots, pairplots, correlation matrices.
- t-tests, ANOVA, chi-square test.
- Linear, logistic, and multivariate regression.
- Time series analysis.
- k-means, hierarchical clustering, DBSCAN.
Integrates with an LLM (large language model) via Ollama for automated interpretation of statistical results.
This article provides a beginner-friendly introduction to HDBSCAN, a powerful hierarchical clustering algorithm that extends the capabilities of DBSCAN by handling varying densities more effectively. It compares HDBSCAN to DBSCAN and KMeans, highlighting the advantages of HDBSCAN in handling clusters of different shapes and sizes.
An overview of clustering algorithms, including centroid-based (K-Means, K-Means++), density-based (DBSCAN), hierarchical, and distribution-based clustering. The article explains how each type works, its pros and cons, provides code examples, and discusses use cases.
A simple and intuitive explanation of DBSCAN (Density-Based Spatial Clustering of Applications with Noise), a clustering algorithm that can identify outliers, extract new features, compress data, and perform novelty detection. The article provides a fast implementation of DBSCAN in Python.
This article introduces interpretable clustering, a field that aims to provide insights into the characteristics of clusters formed by clustering algorithms. It discusses the limitations of traditional clustering methods and highlights the benefits of interpretable clustering in understanding data patterns.
Discusses reasons why clustering in data science might not produce desired results and how to address these issues.
This article discusses a method for automatically curating high-quality datasets for self-supervised pre-training of machine learning systems. The method involves successive and hierarchical applications of k-means on a large and diverse data repository to obtain clusters that distribute uniformly among data concepts, followed by a hierarchical, balanced sampling step from these clusters. The experiments on three different data domains show that features trained on the automatically curated datasets outperform those trained on uncurated data while being on par or better than ones trained on manually curated data.