• 검색 결과가 없습니다.

Improvement of Optical Flow Estimation

N/A
N/A
Protected

Academic year: 2023

Share "Improvement of Optical Flow Estimation"

Copied!
7
0
0

로드 중.... (전체 텍스트 보기)

전체 글

(1)

Improvement of Optical Flow Estimation

by Using the Hampel Filter for Low-End Embedded Systems

Ji-il Park1, Yeongseok Lee1, Eungyo Suh1, Hyunyong Jeon2, Kuk-Jin Yoon1∗ and Kyung-soo Kim1∗

Abstract—Owing to the recent advances in the field of deep- learning-based approaches, state-of-the-art performance has been achieved for optical flow estimation. However, non-deep-learning- based improvement in the optical flow estimation performance is still required because many platforms, such as small UGVs and drones, involve constraints that make mounting GPUs difficult.

Thus, in this study, we do not apply deep learning methods;

rather, we improve the accuracy of the optical flow pipeline by optimizing only its fundamental parameters. The optical flow estimation performance is influenced by the number of coarse-to- fine estimation pyramids, the filter applied to each pyramid level, the window size, and the graduated nonconvexity (GNC) step number. No significant differences from the previous research were achieved by optimizing the parameters heuristically because they have already been optimized. Therefore, we decided to change the filter applied in each pyramid, which is the most important factor in determining the optical flow estimation performance. As a result of verification, the optical flow with the median filter did not show good optical flow estimation performance due to oversmoothing of the image boundary, and the optical flow with the weighted median filter proposed to overcome this drawback could not well address the complexity of the equation and deal with the large computation cost. The proposed Hampel filter shows better performance by minimizing the loss of the original image in the image smoothing process and reduced computational complexity compared to the weighted median filter. The principle of the Hampel filter is similar to that of the median filter, except that if the reference pixel is statistically close to the pixel median in the window, it works by preserving the original pixel value. This is usually used for filtering outliers in a 1D signal but has never been applied for 2D image filters.

This study expanded the application of the existing Hampel filter to include 2D images. Compared to the conventional median filter, this filter demonstrates relatively good performance, as it can considerably reduce the loss in the image data and improve the excessive edge smoothing effect such that it can better identify motion. The performance-related improvements achieved by this research were verified using the KITTI Vision Benchmark Suite, and the simulation results demonstrate that the Hampel filter can replace the existing filter and that it has strengths related to multi-object detection, background recognition, and the identification of moving objects at the edges of images.

Index Terms—Computer vision for automation, visual track-

Manuscript received: February 23, 2021; Revised: June 2, 2021; Accepted:

June 30, 2021. This paper was recommended for publication by Associate Editor Okamura, Allison and Editor Cadena Lerma, Cesar upon evaluation of the Associate Editor and Reviewers’ comments. This work was supported by the Institute of Civil Military Technology Cooperation funded by the Defense Acquisition Program Administration and Ministry of Trade, Industry and En- ergy of the Korean government under grant no.UM19303RD3. (Corresponding author: Kuk-Jin Yoon, Kyung-soo Kim.)

Ji-il Park, Yeongseok Lee, Eungyo Suh, Kuk-Jin Yoon, and Kyung-soo Kim are with the Department of Mechanical Engineering, Korea Advanced Institute of Science and Technology (KAIST), 291, Daehak-ro, Yuseong-gu, Daejeon, Republic of Korea. (e-mail: tinz64@kaist.ac.kr; yeongseok94@kaist.ac.kr;

bomtorazek@kaist.ac.kr; kjyoon@kaist.ac.kr; kyungsookim@kaist.ac.kr).

Hyunyong Jeon is with the Department of Mechanical Engineering, Korea Military Academy (KMA), 574, Hwarang-Ro, Seoul, Republic of Korea. (e- mail: junhy425@kaist.ac.kr).

Digital Object Identifier (DOI): see top of this page.

ing, recognition

I. INTRODUCTION

A

UTONOMOUS driving consists of perception, localization and mapping, path planning, prediction, decision making, and control. Currently, as deep learning technology is applied to each of these areas, the development of autonomous driving technology is rapidly progressing. Among the abovementioned areas, deep learning is actively being applied in the area of perception because it is the foundation of autonomous driving and thus considerably influences the other factors. In addition, perception is viewed as brain activity based on human vision, so the need for deep learning in the area of perception is greater than that in any other area. However, in the case of small unmanned platforms with difficulties in terms of GPU mounting, deep-learning-based perception is difficult to apply because the current deep learning methods are limited in terms of implementation without GPUs. The same is true in the case of optical flow, which is widely used for perception. In this context, this study is focused on improving the optical flow performance by modifying a basic algorithm rather than by applying deep learning.

Most previously used methods, such as those used by Horn and Schunck (HS) [1] and Lucas and Kanade (LK) [2], aimed at optimizing the original algorithm. Attempts to improve the optical flow performance through the application of deep learn- ing are common. However, without optimizing the underlying algorithm, performance improvement through deep learning alone is limited. In particular, fundamental optimization of the optical flow algorithms can lead to better performance than that achievable through application of deep learning alone.

Thus, in this study, we examined how the performance can be optimized without applying deep learning by fundamentally improving the optical flow algorithm, which is an essential element for the implementation of perception.

As expected, after analyzing the top-ranked papers of the KITTI Vision Benchmark Suite Scene Flow Evaluation 2015 [3], we found that all papers applied deep learning. This result can be attributed to the fact that complex densification procedures aimed at achieving reliable optical flow estimation in all regions require deep learning. Thus, most researchers are actively studying the application of deep learning to compensate for the limitations of the existing algorithms.

First, we reviewed the papers aimed at improving the optical flow performance by applying deep learning methods. Ren et al.obtained an optical flow based on multiple images and lowered the end point error (EPE) by employing flow fusion.

In this work, two neural networks were used to obtain and fuse the optical flow [4]. Zhang et al. suggested the use of

(2)

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/LRA.2021.3095927, IEEE Robotics and Automation Letters

2 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. JUNE, 2021

general multiobjective optimization based on the consistency and distinctiveness of the examined features and applied a convolutional neural network (CNN) to identify the good features through dense matching, as such features can help achieve impressive results [5].

In addition, ConvNets applied deep learning to process only meaningful operations without processing probabilisti- cally unnecessary calculations, thus reducing the high com- putational cost of optical flow calculations [6]. In addition, LiteFlowNet [7], an improved version of FlowNet2 [8] that is considerably lighter with the same level of performance, was designed to cascade the correction of both the flow based on feature correlation and the subpixel units within each pyra- mid level. Furthermore, feature-specific local convolution was implemented to achieve further correction per pyramid level after cascading. PWC-net [9] combined domain information from the classical optical flow techniques and deep learning to create an algorithm that was more compact and faster than FlowNet2 [8]. Similar to in the traditional coarse-to-fine approach, current optical flow estimation was used to warp the features of the second image in the CNN, and a cost volume was calculated based on these warped features and the features of the first image. This volume was then used to predict the optical flow by using the CNN.

For SelFlow [10], PWC-net [9] was developed in the form of multiframe flow estimation, and this approach employed self-supervised learning. In this method, a CNN is first trained without considering occlusion. After the occlusion is predicted separately, the CNN is trained using a new cost function that reflects this factor. Similarly, most studies improved the optical flow performance by implementing deep learning rather than through fundamental improvements. In addition, there are studies that implement sequential adversarial learning for visual odometry using optical flow [11] and transfer learning- based visual tracking with gaussian processes regression using sequential image information similar to optical flow [12].

Additionally, most papers that did not adopt deep learning were ranked lower than those that employed deep learning.

Yin et al. [13] added the total variation (data term) and an L-1 norm term (regularization term) and regularized these terms to the frequency domain through a Fourier transform. The main differences between this algorithm and the other algorithms are that noise is eliminated and the energy cost is lowered by setting the optimized boundary. This algorithm is simply an advanced form of the HS algorithm because it is a classical optical flow estimation method that uses a median filter at each pyramid level, similar to the conventional approach. Chen et al. [14] interpreted the regularization term in terms of filtering and proposed a new optical flow estimation algorithm that included a framework to guide regularizers according to the given filtering method.

As demonstrated by the abovementioned studies, various approaches have been adopted to improve the optical flow per- formance. However, research on improving the optimization of the tuning parameters or on the filtering method for improv- ing the fundamental optical flow estimation performance is lacking. Therefore, in this study, we have examined how to improve the optical flow performance by enhancing only the

optical flow algorithm without applying deep learning.

In Section 2, we choose a target algorithm that can be applied with both the Hampel filter and a median filter to compare the performance of each; additionally, we identify the tuning parameters that can improve the optical flow performance. In Section 3, the principle of the Hampel filter we self-programed to expand to 2D space and its simulation method are explained. Finally, in Section 4, the performance and features of the Hampel filter are discussed in detail based on the simulation results.

II. RESEARCH APPROACH A. Analysis of the Target Algorithm

Among various papers on optical flow without deep learn- ing, the paper “Secrets of optical flow estimation and their principles“, in which the Hampel filter can be applied, was selected as the target paper for comparison. In that study, the authors defined a baseline algorithm called “classical“, which can be considered a descendant of the original HS formulation, and established a simple model by applying various techniques [15]. Additionally, the authors achieved statistically significant improvements through a small number of keys and examined the principles pertaining to their success.

Similar to others, that study achieved good results by applying a median filter, which is a nonlinear digital filter used to remove noise from images and signals; however, interestingly, the energy function was high. This is because the objective function calculated based on the difference between the horizontal and vertical component (u, v) values of successive images has increased while passing through the median filter. This result indicates that the optical flow estimation performance can be improved by changing the filter.

In addition, a median filter is effective in removing impul- sive noise; to accomplish this, it uses medium values after arranging the windowed values in order of their magnitude.

However, this approach encountered an oversmoothing prob- lem, which occurred where surrounding pixels were more prevalent than target pixels in a corner or for a slender object.

Accordingly, this paper proposed a weighted median filter that included nonlocal terms [15]. However, this type of weighted median filter also presents a disadvantage. The Classic+NL- Full method, which has a weighted median filter of 15×15 neighborhoods on every pixel, should take three times as long as other methods to achieve a specified level of performance improvement.

The Classic+NL method uses a weighted median filter of 15×15 neighborhoods at the boundary and a median filter of 5×5 neighborhoods at the nonboundary, and this method is faster than the Classic+NL-Full method; however, addi- tional calculation time is required to identify the boundary when this method is adopted. The Classic+NL-Fast method, which adopts a two-stage graduated nonconvexity (GNC) [17]

process to perform fewer computations than the Classic+NL method and reduce the number of warps to three, can reduce the computation time while incurring a slight performance penalty. However, this method is effective only when the examined movement is small. In conclusion, all three methods

(3)

Fig. 1. Visualization of the effects of the Hampel filter on an image, including a comparison of a corrupted grayscale image [16] with different types of filters and tuning parameters. (a), (b), (c), and (d) show images filtered by the median (t = 0), Hampel (t = 1), Hampel (t = 1.5), and Hampel (t = 2) filters with K = 2, respectively. (e), (f), (g), and (h) show images filtered by the median (t = 0), Hampel (t = 1), Hampel (t = 1.5), and Hampel (t = 2) filters with K = 7, respectively. In contrast to the median filter, the Hampel filter removes noise while preserving the data of the original image. This tendency is more pronounced in cases with higher K and t values. The calculated SSIM and MSE values of the filtered images based on the reference image show that the larger the error bound is, the higher the SSIM index and the lower the MSE. This result not only shows that the image filtering effect is improved when the Hamper filter with a large error bound is applied but also means that the optical flow estimation performance can be improved by minimizing the over-smoothing problem of the median filter.

have the disadvantage of requiring considerable time to per- form complex mathematical operations and solve complicated formulas.

To solve the abovementioned problems, a small number of keys were selected, and the selected tuning parameters included the filtering method, window size (K), error bound (t), number of steps in the GNC process, and number of iterations at the pyramid level.

First, median filtering was used to perform coarse-to-fine estimation, which is commonly performed in this context, and warping was applied to increase the accuracy at the pyramid level, which is a step of increasing resolution; subsequently, a median filter was used. However, median filtering involves two limitations. First, this approach cannot effectively remove the noise generated in succession. Second, because the pixel values are arranged in order of their size, the pixel values may be changed even if they are not particularly problematic.

B. Introduction of the Hampel filter

The first problem can be solved by setting the window size larger than the noise in succession. However, the second problem conflicts with the solution to the first problem because a larger window size leads to a stronger image filtering effect, which may cause blurring in the original image, and the boundary or corners of the image, in particular, may not be preserved. Therefore, setting the appropriate window size is important. For these reasons, we decided to apply the new filtering method instead of median filtering.

To address the abovementioned issues, the Hampel filter [18] was proposed to overcome the problems of oversmoothing on the boundary portions of an image, which is encountered when using a median filter, complexity of the equation, and long computation time. The Hampel filter is a decision-based filter that replaces the middle value of a data window with the median value when the middle value is determined to be an outlier that lies more than a certain distance from the median value. The Hampel filter exhibits a strong filtering effect and

(4)

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/LRA.2021.3095927, IEEE Robotics and Automation Letters

4 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. JUNE, 2021

Fig. 2. Graphical description of the 2D Hampel filter. While a median filter always replaces the pixel value of a window with the median of the corresponding window, the Hampel filter replaces this value only when the pixel value of the center of the window is greater than t · S(x, y).

reduces the image data loss by more than that achievable using a median filter; thus, this filter can reduce the error rates when applied to optical flow, thereby increasing the reliability.

III. 2D HAMPEL FILTER A. Extension of the Hampel Filter to 2D Space

Since the original Hampel filter was developed by filtering only the outliers of a 1-dimensional curve, we program the Hampel filter to be used for 2D images. Furthermore, accord- ing to the tuning parameters described above, the threshold of the standard deviation is selected as a tuning parameter in addition to K (window size) and t (error bound). In other words, the code of this filter is designed to apply various deviations. If t is set as 0, then the filter is identical to a median filter, so careful consideration is required.

The median absolute deviation (MAD) is applied to min- imize the distortion caused by outliers. Additionally, the MAD factor or constant scale factor κ, which depends on the considered distribution, is calculated using the following equation:

κ = 1

√2er f c−11/2≈ 1.4826 (1) where er f c−1is the inverse function of the complementary error function. Then, the standard deviation of the window can be estimated by substituting the MAD as follows: ˆσ = 1.4826MAD.

The equation of the MAD scale estimate S(x, y) is as follows. For the filter response, the center pixel value I(x, y) is applied when the standard deviation is within t standard deviations, whereas the median value M(x, y) is applied when it is outside t standard deviations.

S(x, y) = κ · mediani, j∈[−K,K]|I(x + i, y + j) − M(x, y)| (2)

Fig. I. Optical flow estimation methodologies to be compared and their settings. ‘L’ and ‘NL’ denote local and nonlocal, respectively. The proposed method has only t 6= 0.

Parameters Classic+NL-Fast BA Classic+Hampel (Proposed)

GNC steps 2 2 2

Maximum iterations

per pyramid level 3 3 3

K(Window size) 2(L) / 7(NL) 2 2

t(Error threshold) 0 0 1

Interpolation method bicubic cubic bicubic Weighting function Generalized

Charbonnier Lorentzian Generalized Charbonnier

If=

(I(x, y), if |I(x + i, y + j) − M(x, y)| ≤ t · S(x, y) M(x, y), if |I(x + i, y + j) − M(x, y)| > t · S(x, y) (3)

B. Effects of the 2D Hampel Filter as an Image Filter To observe the effects of the proposed 2D Hampel filter when used as an image filter, a simple test using an arbitrary 8-bit grayscale image from the web [16] was employed. A corrupted image was generated using the Image Processing Toolbox of MATLAB R2018a [19] with Gaussian noise of σ = 1 in pixel value and salt and pepper noise of density 0.04.

Then, the corrupted image was filtered using a simple median filter and a Hampel filter with K = 2, 7 and t = 0, 1, 1.5, 2.

The similarities between the filtered and original images were evaluated with the structural similarity (SSIM) index [20] and mean square error (MSE) of the pixel value.

The visualization of the results of this analysis is pre- sented in Figure 1. Both the median filter and Hampel filter demonstrated a strong ability in terms of eliminating outliers.

Therefore, both filters were shown to be effective in filtering out salt and pepper noise. Additionally, since the median filter always replaces the pixel value with the median value of the corresponding window, this filter can hide the details of the image features. In contrast, the Hampel filter does not always filter out the details of the image features.

The results shown in Figure 1 show that the filtered image exhibits increasing similarity with the original image with increasing t. The characteristics of the image with respect to K and t can vary based on the noise type and amount. However, these results show that when proper values are used for K and t, the preservation of the characteristics of the original image can be significantly improved.

C. Simulation Methods

Further work is carried out to analyze the differences in the performance with variation of the parameters K and t, which determine the characteristics of the Hampel filter. The values used for K and t are (2, 7) and (1, 2, 3), respectively. We use

“the ratio of pixels with a flow error greater than 3 pixels“ as the metric, which is the standard metric used by the KITTI 2015 Vision Benchmark Suite.

(5)

Fig. 3. Results of the optical flow estimation: (a), (b), and (c) show the optical flow estimations obtained using the Classic+NL-Fast, BA, and Classic+Hampel methods, respectively. Our proposed Classic+Hampel method exhibits relatively good performance in terms of multi-object detection, occluded object detection, and object identification at the edge of the images. Considering that most optical flow estimation methods encounter problems at motion boundaries, our proposed method can be considered to show meaningful results.

In this study, three methods are compared and analyzed. The first method is the Classic+NL-Fast method using a weighted median filter. As previously mentioned, a weighted median filter estimates whether the neighboring pixels in an image are in the same plane and assigns more weight to these pixels based on where they are supposed to be in the plane. In this method, the number of GNC steps is limited to 2, and the number of iterations per pyramid level is limited to 3 to facilitate the calculation.

The second method is the BA method, which is the method adopted in the target paper. This method uses a simple median filter. To reduce the required computation time and ensure that it is equivalent to the time needed by the Classic+NL- Fast method, thereby enabling comparison under identical conditions, we modify the number of GNC steps adopted in this method as well as the number of iterations. The third

Fig. II. Comparison of EPE with parameter variation. The difference with respect to changes in K and t was small, but K= 2 and t = 1 yielded the optimal result.

EPE K= 2 K= 7

t= 1 t= 2 t= 3 t= 1 t= 2 t= 3 Average EPE 43.6743 43.7345 43.9019 43.7548 43.8436 43.8328 Std. Dev. EPE 24.0219 24.0840 24.0217 24.2848 24.0259 24.0760

method is the proposed method, the Classic+Hampel method, which replaces the median filter at each pyramid level with a Hampel filter. Moreover, we modify the number of GNC steps and number of iterations so that the methods are equally comparable.

IV. SIMULATION RESULTS A. Parameter Verification Results

We simulated various tuning parameters, including window size K, error threshold t, the number of GNC steps, and the number of iterations within each pyramid level. Among these parameters, optimizing the number of GNC steps and number of iterations at each pyramid level did not result in considerable variation in the performance. Therefore, we compared the performance of only these methods with various values of K and t. Surprisingly, in the comparison of the methodologies, K = 2 and t = 1 corresponded to the best performance. This phenomenon occurred because the other parameters of our baseline code had already been optimized with K = 2 and t = 0, which correspond to a 5×5 median filter. Thus, the final comparison considered only the different filtering methodologies and not the variations in K and t.

(6)

This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/LRA.2021.3095927, IEEE Robotics and Automation Letters

6 IEEE ROBOTICS AND AUTOMATION LETTERS. PREPRINT VERSION. JUNE, 2021

Fig. III. Overall statistics of the % of pixels with >3 px flow errors. The proposed method exhibits marginal performance in terms of the average EPE and consistent results among the input images. The Hampel filter is written in MATLAB code, so its processing time is not comparable to that of the other filters written in C or C++.

EPE Classic+NL-Fast BA Classic+Hampel

(Proposed)

Average EPE 33.79% 42.38% 34.70%

Std. Dev. EPE 21.17% 23.79% 20.49%

Processing time 77.41s 34.08s 93.34s

B. Application Results of the Hampel Filter

To verify the effects of the Hampel filter, we compared the Classic+NL-fast, BA, and Classic+Hampel methods by employing the KITTI 2015 Vision Benchmark; the results are as follows. Overall, the percentage of pixels with more than 3-pixel errors in flow magnitude is slightly greater in the case of the Classic+Hampel method than in the case of the Classic+NL-Fast method. However, the standard deviation of the percentage of erroneous pixels is lower in the case of the Classic+Hampel method. Thus, the Classic+Hampel method provides consistent performance regardless of the images used as input.

However, the EPE is not always proportional to an al- gorithm’s absolute performance. Although the percentage of erroneous pixels may be higher for a particular method, the flow error of a moving object with a small area for that method could be lower; additionally, the background features with a wide area could have a larger flow error. In this case, the examined algorithm may exhibit better performance in terms of detecting moving objects or flow boundaries.

Therefore, we compared the flow images directly while considering the error percentage level because one of the most important indices of the performance of an algorithm is how well it distinguishes between moving objects and flow boundaries. The flow output of image 14 of the KITTI 2015 dataset when the Classic+Hampel method was used could effectively distinguish the flow boundaries of the three moving vehicles in the image. In particular, image 112 resulted in better performance because the occluded objects moving in opposite directions could be accurately detected. Additionally, the Classic+Hampel method could accurately distinguish the background and motion boundaries of the vehicle in image 26.

Thus, our proposed Classic+Hampel method is relatively robust in terms of detecting multiple and occluded objects and objects on the boundaries of images. Considering that most optical flow estimation methods encounter problems at motion boundaries where the assumption of spatial smoothness is violated [21], our proposed method can be considered to show meaningful results.

The Hampel filter does not smooth the large flow area, which decreases its numerical performance. Additionally, our algorithm does not consider local and nonlocal areas, so this property is applied globally. Nevertheless, by using the Hampel filter, we can replace the median filter and its related

Fig. 4. Performance comparison result with the state-of-the-art methods using the KITTI 2012 dataset. Of course, the EPE is not always proportional to an algorithm’s absolute perfor- mance, but it shows an optical flow estimation performance that is not far behind that of state-of-the-art methods.

complex algorithm with a simple decision-based filter that results in marginal performance. In summary, the results show that the Hampel filter helps preserve the optical flow and boundaries of small objects with relative motion with respect to the background. In addition, it is meaningful that further applications of the 2D Hampel filter in terms of optical flow estimation have been demonstrated.

C. Comparison with State-of-the-art Methods

Additionally, we compared state-of-the-art optical flow es- timation with non-deep -learning methods to prove the per- formance of the proposed method. For the comparison, the most recent algorithms were selected among the algorithms ranked high on the KITTI benchmark. As mentioned above, since many optical flow estimation studies have used deep learning from recent years, the most recent studies that did not apply deep learning were found to be published between 2015 and 2017. Ultimately, OSF [22], DIS-FAST [23], MirrorFlow [24], and PPM [25] were selected. For the dataset performance comparison, the KITTI 2012 dataset was used instead of the previously used KITTI 2015 dataset for sufficient performance verification.

Figure 4 shows that the OSF performance is the best and the DIS-FAST performance is the worst based on the EPE value, and our proposed algorithm is ranked next to the PPM.

As mentioned above, we can see that our proposed algorithm shows similar performance to the latest algorithm by applying Hampel filter instead of the median filter and its related complex algorithm without considering local and non-local filters. This means that it is possible to further improve the optical flow estimation performance to which the Hampel filter is applied—for example, the weighted Hampel filter to which the local and non-local filters are applied.

(7)

V. CONCLUSIONS

In conclusion, our analysis of the existing studies demon- strated that optimization of the tuning parameters was per- formed to a certain extent in the previous research. Accord- ingly, we can confirm that the performance can be improved by changing the filter, which is different from these methods.

In particular, we found that when our proposed Hampel filter was applied, a similar level of performance could be achieved without changing the weighting or complicated localizing of a weighted median filter. In terms of obtaining an image with a smaller error, the object identification of the Hampel filter was confirmed to be superior to that of the Classic+NL-Fast method. Our proposed Classic+Hampel method significantly reduced the average number of erroneous pixels compared to the BA method; however, the average number of erroneous pixels increased slightly compared to the Classic+NL-Fast method. However, as mentioned above, this was an erroneous value, as unnecessary pixels resulted from checking the actual image. Therefore, our method is more robust in terms of performance consistency since the error variance is smaller.

Additionally, even when we employed various input images, our algorithm produced more consistent error variance and performance than the Classic+NL-Fast and BA methods.

This robustness of the Hampel filter can be particularly advantageous when applied to small unmanned robots for in- telligence, surveillance and reconnaissance (ISR) because one of the most important factors in the practical use of unmanned platforms is that they possess a level of robustness that enables them to run stably and detect moving objects accurately within various environments and in the presence of various variables such as weather changes, day and night changes, and the appearance of sudden objects. In this sense, this study is meaningful because the Hampel filter demonstrates not only good performance as an image filter but also applicability in the context of optical flow estimation. In future research, we will aim to achieve optimal performance in terms of optical flow estimation by applying the Hampel filter to various optical flow algorithms. We also expect that our proposed Hampel filter will exhibit faster and less error-prone performance if a weighting technique is also applied, such as weighted median filtering. We also plan to conduct additional research to thor- oughly verify numerically the robustness of detecting multiple and obscured objects and objects at the image boundary, which is the strength of the proposed algorithm.

REFERENCES

[1] B. K. Horn and B. G. Schunck, “Determining optical flow,” Artificial intelligence, vol. 17, no. 1-3, pp. 185–203, 1981.

[2] B. D. Lucas, T. Kanade et al., “An iterative image registration technique with an application to stereo vision,” 1981.

[3] A. Geiger, P. Lenz, C. Stiller, and R. Urtasun, “The kitti vision benchmark suite,” URL http://www. cvlibs. net/datasets/kitti, 2015.

[4] Z. Ren, O. Gallo, D. Sun, M.-H. Yang, E. Sudderth, and J. Kautz, “A fusion approach for multi-frame optical flow estimation,” in 2019 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE, 2019, pp. 2077–2086.

[5] F. Zhang and B. W. Wah, “Fundamental principles on learning new features for effective dense matching,” IEEE Transactions on Image Processing, vol. 27, no. 2, pp. 822–836, 2017.

[6] A. Ranjan and M. J. Black, “Optical flow estimation using a spatial pyramid network,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 2720–2729.

[7] T.-W. Hui, X. Tang, and C. Change Loy, “Liteflownet: A lightweight convolutional neural network for optical flow estimation,” in Pro- ceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 8981–8989.

[8] E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, and T. Brox,

“Flownet 2.0: Evolution of optical flow estimation with deep networks,”

in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 2462–2470.

[9] D. Sun, X. Yang, M.-Y. Liu, and J. Kautz, “Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 8934–8943.

[10] P. Liu, M. Lyu, I. King, and J. Xu, “Selflow: Self-supervised learning of optical flow,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 4571–4580.

[11] S. Li, F. Xue, W. Xin, Z. Yan, and H. Zha, “Sequential adversarial learning for self-supervised deep visual odometry,” 10 2019, pp. 2851–

2860.

[12] J. Gao, H. Ling, W. Hu, and J. Xing, “Transfer learning based visual tracking with gaussian processes regression,” in Computer Vision – ECCV 2014, D. Fleet, T. Pajdla, B. Schiele, and T. Tuytelaars, Eds.

Cham: Springer International Publishing, 2014, pp. 188–203.

[13] Z. Yin, T. Darrell, and F. Yu, “Hierarchical discrete distribution decom- position for match density estimation,” 12 2018.

[14] J. Chen, Z. Cai, J. Lai, and X. Xie, “A filtering-based framework for optical flow estimation,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 29, no. 5, pp. 1350–1364, 2018.

[15] D. Sun, S. Roth, and M. J. Black, “Secrets of optical flow estimation and their principles,” in 2010 IEEE computer society conference on computer vision and pattern recognition. IEEE, 2010, pp. 2432–2439.

[16] A. Whitehead. Rose - grayscale. [Online].

Available: https://www.publicdomainpictures.net/en/view- image.php?image=201528&picture=rose-grayscale

[17] A. Blake and A. Zisserman, Visual reconstruction. MIT press, 1987.

[18] R. K. Pearson, Y. Neuvo, J. Astola, and M. Gabbouj, “Generalized hampel filters,” EURASIP Journal on Advances in Signal Processing, vol. 2016, no. 1, p. 87, 2016.

[19] MATLAB, version 9.2 (R2018a). Natick, Massachusetts: The Math- Works Inc., 2018.

[20] Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli et al., “Image quality assessment: from error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004.

[21] D. Sun, S. Roth, J. P. Lewis, and M. J. Black, “Learning optical flow,” in Computer Vision – ECCV 2008, D. Forsyth, P. Torr, and A. Zisserman, Eds. Berlin, Heidelberg: Springer Berlin Heidelberg, 2008, pp. 83–97.

[22] M. Menze and A. Geiger, “Object scene flow for autonomous vehicles,”

in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2015.

[23] T. Kroeger, R. Timofte, D. Dai, and L. Van Gool, “Fast optical flow using dense inverse search,” in Computer Vision – ECCV 2016, B. Leibe, J. Matas, N. Sebe, and M. Welling, Eds. Cham: Springer International Publishing, 2016, pp. 471–488.

[24] J. Hur and S. Roth, “Mirrorflow: Exploiting symmetries in joint optical flow and occlusion estimation,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), Oct 2017.

[25] F. Kuang, “Patchmatch algorithms for motion estimation and stereo reconstruction,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), December 2017.

참조

관련 문서

Empirical results of this study indicate that the incremental explanatory power of earning is greater than that of cash flow in the mature stage of the

The purpose of this study is to examine the effects of flow experience through the dance sports media, a dance sports image, and identification on

The innovation of the mass media not only has changed the delivery process of a product but also has expanded into the image industry such as

- The Flow Rate is not Sales (which was $514,405) because inventory is measured in the cost to purchase goods, not in the sales revenue that may be earned from the

This will involve the solution of (n+1) linear systems.. Write a routine that determines the flux at each region and currents at the external boundary. Generate the edit shown

• In any flow of a barotropic inviscid fluid, the circulation about any closed path does not vary with time if the contour is imagined to move with the fluid, that is, always to

With these results, we can suggest the hypothesis that increased cardiac output in the patients who administered ketamine increased muscle blood flow, and this is the

These results confirm that the 12-week swimming exercise based on the swimming program applied in this study has a positive effect on improvement of