• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 348
  • 42
  • 20
  • 13
  • 10
  • 8
  • 5
  • 4
  • 3
  • 3
  • 2
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 541
  • 541
  • 253
  • 210
  • 173
  • 134
  • 113
  • 111
  • 108
  • 89
  • 87
  • 80
  • 75
  • 74
  • 73
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
211

Accelerating Multi-target Visual Tracking on Smart Edge Devices

Nalaie, Keivan January 2023 (has links)
\prefacesection{Abstract} Multi-object tracking (MOT) is a key building block in video analytics and finds extensive use in surveillance, search and rescue, and autonomous driving applications. Object detection, a crucial stage in MOT, dominates in the overall tracking inference time due to its reliance on Deep Neural Networks (DNNs). Despite the superior performance of cutting-edge object detectors, their extensive computational demands limit their real-time application on embedded devices that possess constrained processing capabilities. Hence, we aim to reduce the computational burdens of object detection while maintaining tracking performance. As the first approach, we adapt frame resolutions to reduce computational complexity. During inference, frame resolutions can be tuned according to the complexity of visual scenes. We present DeepScale, a model-agnostic frame resolution selection approach that operates on top of existing fully convolutional network-based trackers. By analyzing the effect of frame resolution on detection performance, DeepScale strikes good trade-offs between detection accuracy and processing speed by adapting frame resolutions on-the-fly. Our second approach focuses on enhancing the efficiency of a tracker by model adaptation. We introduce AttTrack to expedite tracking by interleaving the execution of object detectors of different model sizes in inference. A sophisticated network (teacher) runs for keyframes only while, for non-keyframe, knowledge is transferred from the teacher to a smaller network (student) to improve the latter’s performance. Our third contribution involves exploiting temporal-spatial redundancies to enable real-time multi-camera tracking. We propose the MVSparse pipeline which consists of a central processing unit that aggregates information from multiple cameras (on an edge server or in the cloud) and distributed lightweight Reinforcement Learning (RL) agents running on individual cameras that predict the informative blocks in the current frame based on past frames on the same camera and detection results from other cameras. / Thesis / Doctor of Science (PhD)
212

Real time Optical  Character Recognition  in  steel  bars  using YOLOV5

Gattupalli, Monica January 2023 (has links)
Background.Identifying the quality of the products in the manufacturing industry is a challenging task. Manufacturers use needles to print unique numbers on the products to differentiate between good and bad quality products. However, identi- fying these needle printed characters can be difficult. Hence, new technologies like deep learning and optical character recognition (OCR) are used to identify these characters. Objective.The primary ob jective of this thesis is to identify the needle-printed characters on steel bars. This ob jective is divided into two sub-ob jectives. The first sub-ob jective is to identify the region of interest on the steel bars and extract it from the images. The second sub-ob jective is to identify the characters on the steel bars from the extracted images. The YOLOV5 and YOLOV5-obb ob ject detection algorithms are used to achieve these ob jectives. Method. Literature review was performed at first to select the algorithms, then the research was to collect the dataset, which was provided by OVAKO. The dataset included 1000 old images and 3000 new images of steel bars. To answer the RQ2, at first existing OCR techniques were used on the old images which had low accuracy levels. So, the YOLOV5 algorithm was used on old images to detect the region of interest. Different rotation techniques are applied to the cropped images(cropped after the bounding box is detected) no promising result is observed so YOLOV5 at the character level is used in identifying the characters, the results are unsatisfactory. To achieve this, YOLOV5-obb was used on the new images, which resulted in good accuracy levels. Results. Accuracy and mAP are used to assess the performance of OCRs and selected ob ject detection algorithms. The current study proved Existing OCR was also used in the extraction, however, it had an accuracy of 0%, which implies it failed to identify characters. With a mAP of 0.95, YOLOV5 is good at extracting cropped images but fails to identify the characters. When YOLOV5-obb is used for attaining orientation, it achieves a mAP of 0.93. Due to time constraint, the last part of the thesis was not implemented. Conclusion. The present research employed YOLOV5 and YOLOV5-obb ob ject detection algorithms to identify needle-printed characters on steel bars. By first se- lecting the region of interest and then extracting images, the study ob jectives were met. Finally, character-level identification was performed on the old images using the YOLOV5 technique and on the new images using the YOLOV5-obb algorithm, with promising results
213

Object detection and sensor data processing for off-road autonomous vehicles

Foster, Timothy 30 April 2021 (has links)
Autonomous vehicles require intelligent systems to perceive and navigate unstructured envi- ronments. The scope of this project is to improve and develop algorithms and methods to support autonomy in the off-road problem space. This work explores computer vision architectures to support real-time object detection. Furthermore, this project explores multimodal deep fusion and sensor processing for off-road object detection. The networks are compared to and based off of the SqueezeSeg architecture. The MAVS simulator was utilized for data collection and semantic ground truth. The results indicate improvements from the SqueezeSeg performance metrics.
214

STUDENT ATTENTIVENESS CLASSIFICATION USING GEOMETRIC MOMENTS AIDED POSTURE ESTIMATION

Gowri Kurthkoti Sridhara Rao (14191886) 30 November 2022 (has links)
<p> Body Posture provides enough information regarding the current state of mind of a person. This idea is used to implement a system that provides feedback to lecturers on how engaging the class has been by identifying the attentive levels of students. This is carried out using the posture information extracted with the help of Mediapipe. A novel method of extracting features are from the key points returned by Mediapipe is proposed. Geometric moments aided features classification performs better than the general distances and angles features classification. In order to extend the single person pose classification to multi person pose classification object detection is implemented. Feedback is generated regarding the entire lecture and provided as the output of the system. </p>
215

Low-Observable Object Detection and Tracking Using Advanced Image Processing Techniques

Li, Meng 21 August 2014 (has links)
No description available.
216

Empirical Evaluation of AdaBoost Method in Detecting Transparent and Occluded Objects

Tamang, Sujan 29 May 2018 (has links)
No description available.
217

Histogram-based template matching object detection in images with varying brightness and contrast

Schrider, Christina Da-Wann 16 October 2008 (has links)
No description available.
218

Simultaneous object detection and segmentation using top-down and bottom-up processing

Sharma, Vinay 07 January 2008 (has links)
No description available.
219

Detecting Successful Throws

Almousa, Sami, Morad, Gorgis January 2023 (has links)
This project aims to create a robot system that can accurately figure out if the throws are successful. This can help make various industrial tasks more efficient. The system uses implemented methods to process data from fisheye camera data and depth sensor data, to check the quality of the throws. The main goal is to find out if the thrown object reaches its target or not, with more advanced tasks including predicting its path when frames are lost or not tracked properly.To put the system together the Robot Operating System (ROS) was used for handling data and processing, as well as different tools and techniques, like bag files and OpenCV. A variety of methods and algorithms were used to apply background subtraction, clustering, curve fitting, marking objects and drawing the path they take in the air. The depth sensor data processing is included to make up for the limitations of 2D camera data, providing more accurate and reliable tracking of thrown objects.
220

Object Detection in Paddy Field for Robotic Combine Harvester Based on Semantic Segmentation / セマンティックセグメンテーションに基づくロボットコンバインのための物体検出

Zhu, Jiajun 25 September 2023 (has links)
京都大学 / 新制・課程博士 / 博士(農学) / 甲第24913号 / 農博第2576号 / 新制||農||1103(附属図書館) / 京都大学大学院農学研究科地域環境科学専攻 / (主査)教授 飯田 訓久, 教授 近藤 直, 教授 野口 良造 / 学位規則第4条第1項該当 / Doctor of Agricultural Science / Kyoto University / DFAM

Page generated in 0.0858 seconds