Lane Detection For Automatic Cars
Abstract
The first stage in developing an autonomous car is the lane detection system. To help us identify lanes, we've borrowed a pair of ready-made models. As a rule, these two models are very time-consuming and expensive to compute. To lessen the burden on the computer, we developed a technique called the "row anchor based" approach. The computational burden is reduced, and the no-visual-clue issue is addressed by using this technique. It is exceedingly challenging to identify lanes when we are unable to see them clearly, as occurs in inclement weather, when water is on the lanes, or when the lanes are not designated. No-visual-clue is the term for this kind of issue. ResNet-18, which is used for pretrained models, has been utilized. Because of this, velocity will rise. ResNet-34 is another option, but it is too resource-intensive for this particular project. Road detection from one image is used to locate the road in a picture so it can be used as a district in the automation of the driving system within the vehicles for moving the vehicle on the correct road given a picture captured from a camera attached to a vehicle moving on a road, which road may or may not be level, have clearly described edges, or have some previous acknowledged patterns thereon. Here, we apply techniques for vanishing point identification, Hough Transformation Space, area of interest detection, edge detection, and canny edge detection for road recognition to locate the road inside the picture acquired by the vehicle. To train our model to recognize the road in the fresh image processed by the car, we typically use hundreds of images of roads from different locations.
References
Garnett, N., Cohen, R., Pe’er, T., Lahav, R., Levi, D.: 3d-lanenet: End-to-end 3d multiple lane detection. In: Proceedings of the IEEE International Conference on Computer Vision. pp. 2921–2930 (2019)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 770–778 (2016)
L. Li, K. Ota, and M. Dong, “Humanlike driving: Empirical decisionmaking system for autonomous vehicles,” IEEE Transactions on Vehicular Technology, vol. 67, no. 8, pp. 6814–6823, 20.
Q. Zou, L. Ni, Q. Wang, Q. Li, and S. Wang, “Robust gait recognition by integrating inertial and rgbd sensors,” IEEE Transactions on Cybernetics , vol. 48, no. 4, pp. 1136–1150, 2018.
J. R. Medel and A. Savakis, “Anomaly detection in video using predictive convolutional long short-term memory networks,” arXiv preprint arXiv:1612.00390, 2016.
P. Tokmakov, K. Alahari, and C. Schmid, “Learning video object segmentation with visual memory,” 2017 IEEE International Conference on Computer Vision (ICCV), pp. 4491–4500, 2017.
Refbacks
- There are currently no refbacks.
Copyright © 2012 - 2023, All rights reserved.| ijitr.com
International Journal of Innovative Technology and Research is licensed under a Creative Commons Attribution 3.0 Unported License.Based on a work at IJITR , Permissions beyond the scope of this license may be available at http://creativecommons.org/licenses/by/3.0/deed.en_GB.