If you spend some time browsing, there are some examples already available for Python SIFT/SURF bag of words (BoW) classifier in the internet. They use clustering (usually K-Means) to build dictionary of visual vocabularies (usually with sklearn or cv2 clustering library) of SIFT/SURF features. However, most of the sample codes that I found can’t properly handle big number(> 100) of vocabularies/clusters, while some papers (such as this one) shows best result are achieved using 2000+ clusters.
Building visual dictionary using cv2.BOWKMeansTrainer is super slow when using > 100 clusters. While using sklearn.cluster.KMeans solves the speed issue, it requires huge amount of memory (8 GB of RAM is still insufficient to handle > 400 clusters). That’s where klearn.cluster.MiniBatchKMeans comes into picture.
import cv2
import numpy as np
import progressbar
from sklearn.cluster import MiniBatchKMeans
...
def build_dictionary(xfeatures2d, dir_names, file_paths, dictionary_size):
print('Computing descriptors..')
desc_list = []
num_files = len(file_paths)
bar = progressbar.ProgressBar(maxval=num_files).start()
for i in range(num_files):
p = file_paths[i]
image = cv2.imread(p)
gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)
kp, dsc = xfeatures2d.detectAndCompute(gray, None)
desc_list.extend(dsc)
bar.update(i)
bar.finish()
print('Creating BoW dictionary using K-Means clustering with k={}..'.format(dictionary_size))
dictionary = MiniBatchKMeans(n_clusters=dictionary_size, batch_size=100, verbose=1)
dictionary.fit(np.array(desc_list))
return dictionary
...
# usage example
sift = cv2.xfeatures2d.SIFT_create()
dir_names = ['class1', 'class2', ...]
file_paths = ['/data/class1/1.jpg', '/data/class1/2.jpg', ..., '/data/class2/1.jpg', '/data/class2/2.jpg', ...]
dictionary_size = 2800
dictionary = build_dictionary(sift, dir_names, file_paths, dictionary_size)
Using the code above, I was able to complete whole process of raw data (2000+ images) preprocessing, building SIFT dictionary, cross-validating 6 classifiers within 52 minutes (still quite long, but acceptable ?). While using SURF, it takes around 45 minutes. The complete main files for both experiments can be found in here and here.
Getting verified SSL information with Python (3.x) is very easy. Code examples for it are…
By default, Spring Data Couchbase implements single-bucket configuration. In this default implementation, all POJO (Plain…
Last year, Google released Firebase Auth Emulator as a new component in Firebase Emulator. In…
One of the authentication protocol that is supported by most of Google Cloud services is…
If you need to to add a spatial information querying in your application, PostGIS is…
Amazon Web Service Transcribe provides API to automatically convert an audio speech file (mp3/wav) into…