一. 传统的关键点检测和特征描述算法有:
1. SIFT
import numpy as np
import cv2
imgname1 = './555.jpg'
imgname2 = './666.jpg'
sift = cv2.xfeatures2d.SIFT_create()
img1 = cv2.imread(imgname1)
gray1 = cv2.cvtColor(img1, cv2.COLOR_BGR2GRAY) #灰度处理图像
kp1, des1 = sift.detectAndCompute(img1,None) #des是描述子
img2 = cv2.imread(imgname2)
gray2 = cv2.cvtColor(img2, cv2.COLOR_BGR2GRAY)#灰度处理图像
kp2, des2 = sift.detectAndCompute(img2,None) #des是描述子
# hmerge = np.hstack((gray1, gray2)) #水平拼接
# cv2.imshow("gray", hmerge) #拼接显示为gray
# cv2.waitKey(0)
img3 = cv2.drawKeypoints(img1,kp1,img1,color=(255,0,255)) #画出特征点,并显示为红色圆圈
img4 = cv2.drawKeypoints(img2,kp2,img2,color=(255,0,255)) #画出特征点,并显示为红色圆圈
cv2.imwrite('./555_sift.png', img3)
cv2.imwrite('./666_sift.png', img4)
# hmerge = np.hstack((img3, img4)) #水平拼接
# cv2.imshow("point", hmerge) #拼接显示为gray
# cv2.waitKey(0)
# BFMatcher解决匹配
bf = cv2.BFMatcher()
matches = bf.knnMatch(des1,des2, k=2)
img5 = cv2.drawMatchesKnn(img1,kp1,img2,kp2,matches,None,flags=2)
cv2.imwrite('./555_666_siftmatch.png', img5)
cv2.imshow("BFmatch", img5)
cv2.waitKey(0)
cv2.destroyAllWindows()
2. SURF
3. ORB
import cv2
import numpy as np
#from pylsd.lsd import lsd
import math
img1 = cv2.imread('./333.jpg',0)#导入灰度图像
img2 = cv2.imread('./444.jpg',0)
def drawMatches(img1, kp1, img2, kp2, matches):
rows1 = img1.shape[0]
cols1 = img1.shape[1]
rows2 = img2.shape[0]
cols2 = img2.shape[1]
out = np.zeros((max([rows1,rows2]),cols1 + cols2, 3),dtype = 'uint8')
#拼接图像
out[:rows1, :cols1] = np.dstack([img1, img1,img1])
out[:rows2, cols1:] = np.dstack([img2, img2,img2])
for mat in matches:
img1_idx = mat.queryIdx
img2_idx = mat.trainIdx
(x1,y1) = kp1[img1_idx].pt
(x2,y2) = kp2[img2_idx].pt
#绘制匹配点
cv2.circle(out, (int(x1),int(y1)),4,(255,255,0),1)
cv2.circle(out,(int(x2)+cols1,int(y2)),4,(0,255,255),1)
cv2.line(out,(int(x1),int(y1)),(int(x2)+cols1,int(y2)),(255,0,0),1)
return out
detector = cv2.ORB_create()
kp1 = detector.detect(img1,None)
# for p in kp1:
# x_ = p.pt[0]
# y_ = p.pt[1]
# print(x_, y_)
kp2 = detector.detect(img2,None)
kp1,des1 = detector.compute(img1,kp1)
kp2,des2 = detector.compute(img2,kp2)
bf = cv2.BFMatcher(cv2.NORM_HAMMING,crossCheck = True)
matches = bf.match(des1,des2)
img3 = drawMatches(img1,kp1,img2,kp2,matches[:50])
# img3 = cv2.drawKeypoints(img1,kp1,None,color = (0,255,0),flags = 0)
cv2.imwrite("333_444_orb.jpg",img3)
cv2.imshow('orbTest',img3)
cv2.waitKey(0)
4. AKAZE
二. 基于深度学习方法
1. 特征提取
(1)Discriminative learning of deep convolutional feature point descriptors
(2)LIFT: Learned Invariant Feature Transform 2016
2. 特征匹配
(1)MatchNet: “MatchNet: Unifying Feature and Metric Learning for Patch-Based Matching”, CVPR 2015
paper: http://www.cs.unc.edu/~xufeng/cs/papers/cvpr15-matchnet.pdf
github:https://github.com/hanxf/matchnet
(2)UCN: Universal Correspondence Network NIPS 2016
(3)DGC-Net: “DGC-Net: Dense Geometric Correspondence Network”, CVPR 2019