I'm currently doing a project based on the methodology described in this paper: Camera calibration from a single night sky image
As a beginner in computer vision, I do not quite understand how I can implement the method used in the paper to find the centre of all the bright spots (luminaries) within an image, particular on the paragraph in section 4.1:
The surrounding patch of size 15 × 15 pixels (Figure 1(a)), is upsampled by a given factor (Figure 1(c)) and the corresponding gradient map is calculated (Figure 1(d)). Starting from the brightest region, the gray value threshold is decreased, until an energy function is maximized. The energy function is defined as the sum of the border gradients and normalized by the border length (Figure 1(e)). This leads to a segmented star image shown in Figure 1(f). The segmentation ensures that the weighted centre of gravity algorithm [11] gives a robust estimation.
From my understanding, I think I can do a Laplacian / Sobel gradient function on the upsampled image, but after that I'm not too sure how I can perform the energy function part and produce the segmented image. Also I would also want to understand how implement the weighted centre of gravity algorithm to find the centre of the bright spot using openCV or other python library.
Much appreciated if any of you can provide some lights on this.
Thanks and regards.
To find the center of an image, the first step is to convert the original image into grayscale. We can use the cvtColor() method of cv2 as we did before. We read the image and convert it to a grayscale image. The new image is stored in gray_img.
The susceptible method to finding the brightest spot in an image is to use the cv2. minMaxLoc function without any pre-processing. This function requires a single argument, which is our grayscale image.
The main thing to take away is energy function
used in this context is any function that is used for a maximization problem. Here, the energy function is the sum of gradients/derivatives/differences (i.e. "detected borders likelihood" in this case).
Since you seem to have a non-algorithmic background, I suggest you to read on breadth-first search (remember an image is a very specific type of graph, where every edge is a pixel, connected to adjacent ones), recursion, and floodfill.
grad_img
= max_per_pixel(sobel_horiz,sobel_vert).seed
of the star.region
that consists of the seed
. Keep adding adjacent pixels to that region
(recommend breadth-first traversal). Calculate the energy by the sum of pixel values in grad_img
with pixel coordinates the borders of the region
. If the energy is higher than the energy of the previous iteration, the new pixel is added to the region
. If not, the pixel is rejected.My solution is a bit different than their solution. They actually run a floodfill algorithm that fills all pixels of brightness [threshold;255], calculating energy func, decreasing the threshold, rinse and repeat, stopping when they maximize the energy function. Note that their algorithm is very inefficient as they are making effectively up to 255 floodfills for every pre-detected star compared to 1 floodfill in my proposal, and that could be a performance issue in practice.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With