不同旋转下的 Harris 角点检测器

Ben*_*sAI 2 python opencv computer-vision corner-detection

我在 python 中使用哈里斯角检测的 opencv 实现。我的问题是关于下面 gif 中显示的行为 - 当图像旋转时,角停止被检测到(在各种旋转时)。完整代码:

import cv2

image_path = 'image1.jpg'

original_image = cv2.imread(image_path)

def play(video, name='video', wait=60, key='q'):
    for f in video:
        cv2.imshow(name, f)
        if cv2.waitKey(wait) == ord(key):
            return

def rotate(image, theta, point=(0,0)):
    M = cv2.getRotationMatrix2D((point[1], point[0]), theta, 1)
    return cv2.warpAffine(image, M, (image.shape[1], image.shape[0]))

def rotate_detect(image):
    for theta in range(0, 360):
        img = rotate(image, theta, (original_image.shape[0] / 2, original_image.shape[1] / 2))

        gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)

        dst = cv2.cornerHarris(gray,9,13,0.04)

        #result is dilated for marking the corners, not important
        dst = cv2.dilate(dst,None)

        # Threshold for an optimal value, it may vary depending on the image.
        threshold = 0.005
        img[dst>threshold*dst.max()]=[0,0,255]
        yield img

play(rotate_detect(original_image), wait=60)

cv2.destroyAllWindows()
Run Code Online (Sandbox Code Playgroud)

基于

在此处输入图片说明

使用的图像可以在这里找到。从 gif 中可能不太清楚(如果您运行代码会更清楚),但是当网格线水平/垂直时会检测到角点。

如果blocksize参数增加,我们可以获得所需的行为(以检测所有旋转的角点)。

问题 - 如何解释 gif 中显示的行为?

Cat*_*ree 5

更新:

使用goodFeaturesToTrack()而不是使用手动代码从 Harris 响应图中检索角。

如果您想使用 OpenCV 获得正确的结果,则必须根据您的用例调整blockSizequalityLevel

左边是 DIPlib,右边是 OpenCV 结果。

在此处输入图片说明


DIPlib 代码:

from __future__ import print_function
from timeit import default_timer as timer
import PyDIP as dip
import numpy as np
import imageio
import os

img = imageio.imread('https://i.stack.imgur.com/hkbFD.jpg')
img = dip.Image(img)
img.SetColorSpace('rgb')
img = dip.ColorSpaceManager.Convert(img,'gray')

animation = []
times = []

for angle in range(0,360):
    img_rot = dip.Rotation2D(img, np.radians(angle), '3-cubic', 'add zeros')
    img_rot = img_rot.Pad(np.maximum(img.Sizes(), img_rot.Sizes()))
    img_rot.Crop(img.Sizes())

    start = timer()
    harris = dip.HarrisCornerDetector(img_rot, sigmas=[3.0])
    end = timer()
    times.append(end - start)
    harris *= dip.Maxima(harris, connectivity=2)
    harris = dip.Dilation(harris > 1, 5)
    harris = dip.Overlay(img_rot, harris)
    harris.Convert('UINT8')
    animation.append(harris)

print('Mean computation time: {:f}s'.format(np.mean(times)))
print('Median computation time: {:f}s'.format(np.median(times)))
print('Std computation time: {:f}s'.format(np.std(times)))

save_folder = 'DIPlib'
if not os.path.exists(save_folder):
    os.mkdir(save_folder)
for idx, img in enumerate(animation):
    imageio.imsave('{}/Harris_DIPlib_{:03d}.png'.format(save_folder, idx), img)
Run Code Online (Sandbox Code Playgroud)

OpenCV 代码:

from __future__ import print_function
from __future__ import division
from timeit import default_timer as timer
import argparse
import numpy as np
import cv2 as cv
import os

parser = argparse.ArgumentParser(description='Test Harris corners rotation invariance.')
parser.add_argument('--input', default='', type=str, help='Input image path')
parser.add_argument('--save', default=False, type=bool, help='Save results')
parser.add_argument('--maxCorners', default=500, type=int, help='Maximum number of corners')
parser.add_argument('--qualityLevel', default=0.03, type=float, help='Minimal accepted quality of image corners')
parser.add_argument('--minDistance', default=11, type=int, help='Minimum possible Euclidean distance between the returned corners')
parser.add_argument('--blockSize', default=11, type=int, help='Size of an average block for computing a derivative covariation matrix over each pixel neighborhood')
args = parser.parse_args()

harris_params = dict(maxCorners = args.maxCorners,
                     qualityLevel = args.qualityLevel,
                     minDistance = args.minDistance,
                     blockSize = args.blockSize,
                     useHarrisDetector = True)
print('harris_params:\n', harris_params)

image_path = 'hkbFD.jpg'
original_image = cv.imread(image_path)

def play(video, name='video', wait=60, key='q'):
    idx = 0
    directory = 'OpenCV'
    if args.save and not os.path.exists(directory):
        os.makedirs(directory)

    times = []
    for f, elapsed_time in video:
        times.append(elapsed_time)
        cv.imshow(name, f)

        if args.save:
            filename = directory + '/Harris_OpenCV_%03d.png' % idx
            cv.imwrite(filename, f)
            idx += 1

        if cv.waitKey(wait) == ord(key):
            return

    print('Mean computation time: {:f}s'.format(np.mean(times)))
    print('Median computation time: {:f}s'.format(np.median(times)))
    print('Std computation time: {:f}s'.format(np.std(times)))

def rotate(image, theta, point=(0,0)):
    M = cv.getRotationMatrix2D((point[1], point[0]), theta, 1)
    return cv.warpAffine(image, M, (image.shape[1], image.shape[0]))

def rotate_detect(image):
    for theta in range(0, 360):
        img = rotate(image, -theta, (original_image.shape[0] / 2, original_image.shape[1] / 2))

        gray = cv.cvtColor(img, cv.COLOR_BGR2GRAY)

        start = timer()
        harris_cv = cv.goodFeaturesToTrack(gray, **harris_params)
        elapsed_time = timer() - start

        for c in harris_cv:
            cv.circle(img, (int(c[0,0]), int(c[0,1])), 8, (0,0,255))

        yield (img, elapsed_time)

play(rotate_detect(original_image), wait=60)
Run Code Online (Sandbox Code Playgroud)

老的:

算法说明

本文分析了Harris角点检测器的实现。

@文章{ipol.2018.229,

title   = {{An Analysis and Implementation of the Harris Corner Detector}},
author  = {Sánchez, Javier and Monzón, Nelson and Salgado, Agustín},
journal = {{Image Processing On Line}},
volume  = {8},
pages   = {305--328},
year    = {2018},
doi     = {10.5201/ipol.2018.229},
Run Code Online (Sandbox Code Playgroud)

}

算法如下:

在此处输入图片说明

步骤 1 未包含在原始论文中


OpenCV Harris 角点检测实现


Scikit-image 角点检测实现

代码很简单:

    Axx, Axy, Ayy = structure_tensor(image, sigma)

    # determinant
    detA = Axx * Ayy - Axy ** 2
    # trace
    traceA = Axx + Ayy

    if method == 'k':
        response = detA - k * traceA ** 2
    else:
        response = 2 * detA / (traceA + eps)

    return response
Run Code Online (Sandbox Code Playgroud)

structure_tensor()

    image = _prepare_grayscale_input_2D(image)

    imx, imy = _compute_derivatives(image, mode=mode, cval=cval)

    # structure tensore
    Axx = ndi.gaussian_filter(imx * imx, sigma, mode=mode, cval=cval)
    Axy = ndi.gaussian_filter(imx * imy, sigma, mode=mode, cval=cval)
    Ayy = ndi.gaussian_filter(imy * imy, sigma, mode=mode, cval=cval)

    return Axx, Axy, Ayy
Run Code Online (Sandbox Code Playgroud)

_compute_derivatives()

    imy = ndi.sobel(image, axis=0, mode=mode, cval=cval)
    imx = ndi.sobel(image, axis=1, mode=mode, cval=cval)

    return imx, imy
Run Code Online (Sandbox Code Playgroud)

OpenVX Harris 角点检测规范

可以在这里找到。

根据这些规范,供应商可以提供针对其平台优化的自定义实现。例如,ARM 计算库


OpenCV 和 Scikit-image 之间的比较。

  • 使用的版本:
OpenCV: 4.2.0-dev
Numpy: 1.18.1
scikit-image: 0.16.2
Run Code Online (Sandbox Code Playgroud)
  • 哈里斯参数 k=0.04
  • 修改后的 OpenCV Harris 函数参数,实验的默认值为:blockSize=3ksize=1
  • 修改后的 Scikit-image Harris 函数参数,实验默认值: sigma=1
  • 左:OpenCV 结果;右:Scikit 图像结果

数独图片

  • 默认参数:

在此处输入图片说明

  • OpenCV: blockSize=7, apertureSize=3; Scikit-image: sigma=5:

在此处输入图片说明

可以计算重复率,但不确定我的代码是否完全正确:

在此处输入图片说明

距离阈值为 5。

从我的实验来看,看起来 Scikit-image 产生了更好的角定位精度。

Blox 图像

  • 默认参数:

在此处输入图片说明

  • OpenCV: blockSize=3, apertureSize=3; Scikit-image: sigma=3:

在此处输入图片说明

OP 数独图像

  • 默认参数:

在此处输入图片说明

  • OpenCV: blockSize=7, apertureSize=3; Scikit-image: sigma=7:

在此处输入图片说明

校准器图像

  • 默认参数:

在此处输入图片说明


代码在这里


Cri*_*ngo 5

这个答案只是为了表明哈里斯角点检测器的正确实现应该是完全旋转不变的。

正如动画图像所示,有一些检测(在背景中和紧密的簇内)以随机角度出现或消失。这些是由不完美的插值和不可避免的浮点舍入误差引起的。但大多数检测在所有角度都是一致的。[注意:GIF 太大,无法上传到 SO,我在此处显示为帧,请随意运行脚本来查看动画。]

0度 20度

使用的实现是DIPlib中的实现。它使用高斯梯度来计算导数,并使用高斯低通滤波来进行局部平均。高斯滤波器是完全各向同性滤波器。该实现始终使用单精度浮点运算,以最大限度地减少舍入误差。你可以在这里找到它。DIPlib 旨在精确量化,因此试图避免不必要的数值错误。[披露:我是一名作家。]

import diplib as dip
import numpy as np
import imageio

img = imageio.imread('https://i.stack.imgur.com/hkbFD.jpg')
img = dip.Image(img)
img.SetColorSpace('rgb')
img = dip.ColorSpaceManager.Convert(img,'gray')

animation = []

for angle in range(0,180,4):
   img_rot = dip.Rotation2D(img, angle/180*3.14159, '3-cubic', 'add zeros')
   img_rot = img_rot.Pad(np.maximum(img.Sizes(), img_rot.Sizes()))
   img_rot.Crop(img.Sizes())
   harris = dip.HarrisCornerDetector(img_rot, sigmas=[3.0])
   harris *= dip.Maxima(harris, connectivity=2)
   harris = dip.Dilation(harris > 1, 5)
   harris = dip.Overlay(img_rot, harris)
   harris.Convert('UINT8')
   animation.append(harris)

imageio.mimsave('./so.gif', animation, fps=4) # This file is too large for SO

imageio.imsave('./so_00.png', animation[0]) # Frame at 0 degrees
imageio.imsave('./so_20.png', animation[5]) # Frame at 20 degrees
Run Code Online (Sandbox Code Playgroud)