I have 2 test images here. My is question is, how to map the square in first image to the quadrilateral in the second image without cropping the image.
Image 1:
Image 2:
Here is my current code using openCV warpPerspective function.
import cv2 import numpy as np img1_square_corners = np.float32([[253,211], [563,211], [563,519],[253,519]]) img2_quad_corners = np.float32([[234,197], [520,169], [715,483], [81,472]]) h, mask = cv2.findHomography(img1_square_corners, img2_quad_corners) im = cv2.imread("image1.png") out = cv2.warpPerspective(im, h, (800,800)) cv2.imwrite("result.png", out)
Result:
As you can see, because of dsize=(800,800) parameter in the warpPerspective function, I can't get full view of image 1. If I adjust the dsize, the square won't map properly. Is there any way to resize the output image so that I can get whole picture of image 1?
OpenCV program in python to demonstrate warpPerspective() function to read the given image and align the given image and then fit the size of the aligned image to the size of the original image using warpPerspective() function: #importing the module cv2 and numpy.
In OpenCV, images are represented as 3-dimensional Numpy arrays. An image consists of rows of pixels, and each pixel is represented by an array of values representing its color. Where [72 99 143] , etc., are the blue, green, and red (BGR) values of that one pixel.
My solution is to calculate the result image size, and then do a translation.
def warpTwoImages(img1, img2, H): '''warp img2 to img1 with homograph H''' h1,w1 = img1.shape[:2] h2,w2 = img2.shape[:2] pts1 = float32([[0,0],[0,h1],[w1,h1],[w1,0]]).reshape(-1,1,2) pts2 = float32([[0,0],[0,h2],[w2,h2],[w2,0]]).reshape(-1,1,2) pts2_ = cv2.perspectiveTransform(pts2, H) pts = concatenate((pts1, pts2_), axis=0) [xmin, ymin] = int32(pts.min(axis=0).ravel() - 0.5) [xmax, ymax] = int32(pts.max(axis=0).ravel() + 0.5) t = [-xmin,-ymin] Ht = array([[1,0,t[0]],[0,1,t[1]],[0,0,1]]) # translate result = cv2.warpPerspective(img2, Ht.dot(H), (xmax-xmin, ymax-ymin)) result[t[1]:h1+t[1],t[0]:w1+t[0]] = img1 return result dst_pts = float32([kp1[m.queryIdx].pt for m in good]).reshape(-1,1,2) src_pts = float32([kp2[m.trainIdx].pt for m in good]).reshape(-1,1,2) M, mask = cv2.findHomography(src_pts, dst_pts, cv2.RANSAC, 5.0) result = warpTwoImages(img1_color, img2_color, M)
Yes, but you should realise that the output image might be very large. I quickly wrote the following Python code, but even a 3000 x 3000 image could not fit the output, it is just way too big due to the transformation. Although, here is my code, I hope it will be of use to you.
import cv2 import numpy as np import cv #the old cv interface img1_square_corners = np.float32([[253,211], [563,211], [563,519],[253,519]]) img2_quad_corners = np.float32([[234,197], [520,169], [715,483], [81,472]]) h, mask = cv2.findHomography(img1_square_corners, img2_quad_corners) im = cv2.imread("image1.png")
Create an output image here, I used (3000, 3000) as an example.
out_2 = cv.fromarray(np.zeros((3000,3000,3),np.uint8))
By using the old cv
interface, I wrote directly to the output, and so it does not get cropped. I tried this using the cv2
interface, but for some reason it did not work... Maybe someone can shed some light on that?
cv.WarpPerspective(cv.fromarray(im), out_2, cv.fromarray(h)) cv.ShowImage("test", out_2) cv.SaveImage("result.png", out_2) cv2.waitKey()
Anyway, this gives a very large image, that contains your original image 1, warped. The entire image will be visible if you specify the output image to be large enough. (Which might be very large indeed!)
I hope that this code may help you.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With