Vector Database as a Big Data Analysis Tool for AI Agents
Workflows from the webinar "Build production-ready AI Agents with Qdrant and n8n".
This series of workflows shows how to build big data analysis tools for production-ready AI agents with the help of vector databases. These pipelines are adaptable to any dataset of images, hence, many production use cases.
- Uploading (image) datasets to Qdrant
- Set up meta-variables for anomaly detection in Qdrant
- Anomaly detection tool
- KNN classifier tool
For anomaly detection
- The first pipeline to upload an image dataset to Qdrant.
2. This is the second pipeline to set up cluster (class) centres & cluster (class) threshold scores needed for anomaly detection.
- The third is the anomaly detection tool, which takes any image as input and uses all preparatory work done with Qdrant to detect if it's an anomaly to the uploaded dataset.
For KNN (k nearest neighbours) classification
- The first pipeline to upload an image dataset to Qdrant.
- The second is the KNN classifier tool, which takes any image as input and classifies it on the uploaded to Qdrant dataset.
To recreate both
You'll have to upload crops and lands datasets from Kaggle to your own Google Storage bucket, and re-create APIs/connections to Qdrant Cloud (you can use Free Tier cluster), Voyage AI API & Google Cloud Storage.
[This workflow] Setting Up Cluster (Class) Centres & Cluster (Class) Threshold Scores for Anomaly Detection
Preparatory workflow to set cluster centres and cluster threshold scores so anomalies can be detected based on these thresholds.
Here, we're using two approaches to set up these centres: the "distance matrix approach" and the "multimodal embedding model approach".