• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 91
  • 23
  • 17
  • 15
  • 13
  • 12
  • 5
  • 4
  • 3
  • 1
  • 1
  • 1
  • Tagged with
  • 220
  • 220
  • 72
  • 62
  • 60
  • 54
  • 41
  • 37
  • 36
  • 32
  • 28
  • 27
  • 27
  • 23
  • 22
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
41

Human Motion Detection and Visual Augmentation of Chopin’s Etudes

Kerr, David Philip 25 November 2014 (has links)
Chopin’s Etudes are difficult musical compositions for advanced piano students. Helmut Brauss, a professional pianist and educator, has created a number of videos to teach students motion patterns that will help them perfect the Etudes. The subtleties of motion shown in the videos are not apparently obvious to students, and in our research, we have developed four markerless based approaches to visually augment the videos: Predictive Optical Flow, Historical Optical Flow, Predictive Hand Tracking and Historical Hand Tracking. A survey of students learning the Etudes was conducted, and it was determined that the participants found the Historical techniques to be the most useful. No difference could be found between the usefulness of the Optical Flow and Hand Tracking augmentations. / Graduate
42

Caminhamento fotogramétrico utilizando o fluxo óptico filtrado

Barbosa, Ricardo Luís [UNESP] January 2006 (has links) (PDF)
Made available in DSpace on 2014-06-11T19:30:31Z (GMT). No. of bitstreams: 0 Previous issue date: 2006Bitstream added on 2014-06-13T19:40:16Z : No. of bitstreams: 1 barbosa_rl_dr_prud.pdf: 1353454 bytes, checksum: 109a85af0d056d18c721a2c6af70ce93 (MD5) / Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES) / Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP) / Em certas condições, os sensores de orientação e posicionamento (INS e GPS) de um Sistema Móvel de Mapeamento Terrestre (SMMT) ficam indisponíveis por algum intervalo de tempo casionando a perda da orientação e do posicionamento das imagens capturadas neste intervalo. Neste trabalho, é proposta uma solução baseada apenas nas imagens sem a utilização de sensores ou informações externas às mesmas, através do fluxo óptico. Um sistema móvel com um par de vídeo câmaras, denominado Unidade Móvel de Mapeamento Digital (UMMD), foi utilizado para testar a metodologia proposta em uma via plana. As câmaras são fixadas em uma base com um afastamento entre as câmaras de 0,94m e paralelas ao eixo de deslocamento (Y). A velocidade do veículo é estimada, inicialmente, com base no fluxo óptico denso. Em seguida, a estimação da velocidade é melhorada após uma filtragem, que consiste em: utilizar os vetores que apresentam comportamento radial na metade inferior das imagens e que foram detectados pelo algoritmo de Canny, acrescida uma segunda etapa na estimação da velocidade com eliminação de erros grosseiros. Com a velocidade estimada e sabendo-se o tempo de amostragem do vídeo, o deslocamento de cada imagem é determinado e esta informação é utilizada como aproximação inicial para o posicionamento das câmaras. Os resultados mostraram que a velocidade estimada ficou próxima da velocidade verdadeira e a qualidade do ajustamento se mostrou razoável, considerando-se a não utilização de sensores externos e de pontos de apoio. / Under certain conditions the positioning and orientation sensors such as INS and GPS of a land-based mobile mapping system may fail for a certain time interval. The consequence is that the images captured during this time interval may be misoriented or even may have no orientation. This thesis proposes a solution to orient the images based only on image processing and a photogrammetric technique without any external sensors in order to overcome the lack of external orientation. A land-based mobile mapping system with a pair of video cameras and a GPS receiver was used to test the proposed methodology on an urban flat road. The video cameras were mounted on the roof of the vehicle with both optical axes parallel to the main road axis (Y). The methodology is based on the velocity estimation of the vehicle, which is done in two steps. Initially, the dense optical flow is computed then the velocity estimation is obtained through a filtering strategy that consists of using radial vectors in the low parts of the images. These radial vectors are detected by the Canny algorithm. The vehicle velocity is re-estimated after eliminating the optical flow outliers. With the reestimated velocity and with the video sampling time the spatial displacement of each image (with respect to the previous one of the sequence) is determined. The results show that the estimated velocity is pretty close to the true one and the quality of the least square adjustment is quite acceptable, considering that no external sensors were used.
43

Video Deinterlacing using Control Grid Interpolation Frameworks

January 2012 (has links)
abstract: Video deinterlacing is a key technique in digital video processing, particularly with the widespread usage of LCD and plasma TVs. This thesis proposes a novel spatio-temporal, non-linear video deinterlacing technique that adaptively chooses between the results from one dimensional control grid interpolation (1DCGI), vertical temporal filter (VTF) and temporal line averaging (LA). The proposed method performs better than several popular benchmarking methods in terms of both visual quality and peak signal to noise ratio (PSNR). The algorithm performs better than existing approaches like edge-based line averaging (ELA) and spatio-temporal edge-based median filtering (STELA) on fine moving edges and semi-static regions of videos, which are recognized as particularly challenging deinterlacing cases. The proposed approach also performs better than the state-of-the-art content adaptive vertical temporal filtering (CAVTF) approach. Along with the main approach several spin-off approaches are also proposed each with its own characteristics. / Dissertation/Thesis / M.S. Electrical Engineering 2012
44

The Potential of Visual Features : to Improve Voice Recognition Systems in Vehicles Noisy Environment

Jafari Moghadamfard, Ramtin, Payvar, Saeid January 2014 (has links)
Multimodal biometric systems have been subject of study in recent decades, theirunique characteristic of Anti spoofing and liveness detection plus ability to deal withaudio noise made them technology candidates for improving current systems such asvoice recognition, verification and identification systems.In this work we studied feasibility of incorporating audio-visual voice recognitionsystem for dealing with audio noise in the truck cab environment. Speech recognitionsystems suffer from excessive noise from the engine and road traffic and cars stereosystem. To deal with this noise different techniques including active and passive noisecancelling have been studied.Our results showed that although audio-only systems are performing better in noisefree environment their performance drops significantly by increase in the level of noisein truck cabins, which by contrast does not affect the performance of visual features.Final fused system comprising both visual and audio cues, proved to be superior toboth audio-only and video-only systems.
45

Walking Assistant – A Mobile Aid for the Visually-Impaired

Miller, Adin T 01 June 2014 (has links) (PDF)
The most common navigation aid visually-impaired people employ is a white cane, but, recently, technology has given rise to a varied set of sophisticated navigation aids. While these new aids can provide more assistance to a visually-impaired person than a white cane, they tend to be expensive due to a small market segment, which in turn can reduce their accessibility. In an effort to produce a technologically-advanced yet accessible navigation aid, an Android application is proposed that detects and notifies users about obstacles within their path through the use of a smartphone's camera. While the smartphone is mounted on a harness worn by the user, the Walking Assistant application operates by capturing images as the user walks, finding features of objects within each frame, and determining how the features have moved from image to image. If it is discovered that an object is moving towards the user, the Walking Assistant will activate the smartphone's vibration mode to alert the user to the object's presence. Additionally, the user can control the Walking Assistant through the use of either touch or voice commands. By conducting real-world tests, it was determined that the Walking Assistant can correctly identify obstacles 42.1% of the time, while generating false positive obstacle identifications only 15.0% of the time. The accuracy of the Walking Assistant can be further improved by implementing additional features, such as a fuzzy-decision-based thresholding system or image stabilization.
46

4D-CT Lung Registration and its Application for Lung Radiation Therapy

Min, Yugang 01 January 2012 (has links)
Radiation therapy has been successful in treating lung cancer patients, but its efficacy is limited by the inability to account for the respiratory motion during treatment planning and radiation dose delivery. Physics-based lung deformation models facilitate the motion computation of both tumor and local lung tissue during radiation therapy. In this dissertation, a novel method is discussed to accurately register 3D lungs across the respiratory phases from 4D-CT datasets, which facilitates the estimation of the volumetric lung deformation models. This method uses multi-level and multi-resolution optical flow registration coupled with thin plate splines (TPS), to address registration issue of inconsistent intensity across respiratory phases. It achieves higher accuracy as compared to multi-resolution optical flow registration and other commonly used registration methods. Results of validation show that the lung registration is computed with 3 mm Target Registration Error (TRE) and approximately 3 mm Inverse Consistency Error (ICE). This registration method is further implemented in GPU based real time dose delivery simulation to assist radiation therapy planning.
47

A NEURAL METHOD OF COMPUTING OPTICAL FLOW BASED ON GEOMETRIC CONSTRAINTS

KAIMAL, VINOD GOPALKRISHNA January 2002 (has links)
No description available.
48

INFLUENCE OF EDGE RATE, GLOBAL OPTICAL FLOW RATE, ANGLE, AND EXPANSION RATE ON BRAKING BEHAVIOR

Russell, Sheldon M. 27 October 2010 (has links)
No description available.
49

Image Segmentation and Range Estimation Using a Moving-aperture Lens

Subramanian, Anbumani 07 May 2001 (has links)
Given 2D images, it still remains a big challenge in the field of computer vision to group the image points into logical objects (segmentation) and to determine the locations in the scene (range estimation). Despite the decades of research, a single solution is yet to be found. Through our research we have demonstrated that a possible solution is to use moving aperture lens. This lens has the effect of introducing small, repeating movements of the camera center so that objects appear to translate in the image, by an amount that depends on distance from the plane of focus. Our novel method employs optical flow techniques to an image sequence, captured using a video camera with a moving aperture lens. For a stationary scene, optical flow magnitudes and direction are directly related to the three-dimensional object distance and location from the observer. Exploiting this information, we have successfully extracted objects at different depths and estimated the locations of objects in the scene, with respect to the plane of focus. Our work therefore demonstrates an ability for passive range estimation, without emitting any energy in an environment. Other potential applications include video compression, 3D video broadcast, teleconferencing and autonomous vehicle navigation. / Master of Science
50

Sub-frame synchronisation and motion interpolation for panoramic video stitching / Synkronisering och Interpolering av Videodata för Panoramagenerering

Remì, Chierchia January 2022 (has links)
This study was carried out in collaboration with Tracab, a brand leader in real-time digital sports data. As a result, the application field is centred on sports analytics. The technology, for instance, consists of multiple cameras that capture a football pitch in a panoramic setup. The alignment of two or more cameras in both a spatial and temporal manner is referred to as sub-frame synchronisation. Because the cameras are already in the same geometric coordinates, only temporal synchronisation will be addressed in this project. The main method for retrieving the desynchronisation information that affects the cameras is based on optical flow. The off-sync cameras' spacial information is then synthesised to the time required by the synchronisation constraint using motion interpolation. In addition, the created system is compared to a real-time intermediate flow interpolation approach. The latter method relies on machine learning techniques, whereas this study focuses on more traditional methods. The metrics Peak Signal-to-Noise Ratio and Structural Similarity Index Measure are used to address the quality criteria required by this subject of study. Furthermore, visually perceived quality is examined to identify differences between measured and perceived quality. The results reveal that in every realistic situation investigated, temporal synchronisation can be addressed by an error measure of less than 1ms. The frame synthesis stage, on the other hand, fails to accurately estimate complicated scenarios, while the machine learning approach stands out. The implemented approach, on the other hand, addresses fast-moving objects with greater precision. Furthermore, the machine learning approach is unable to interpolate intermediate frames in arbitrary time steps, which is critical for the project's application. Finally, considering the lack of real-time computational speed and the quality achieved by machine learning approaches, more research is required in these directions. / Denna studie genomfördes i samarbete med Tracab, en marknadsledare inom digital sportdata levererad i realtid. Studiens applikationsområde kommer där av centreras kring sportdata där två eller flera kameror filmar en fotbollsplan i ett videopanorama. Kamerasynkroniseringen måste ske både spatialt och temporalt. Eftersom kamerorna har samma position kommer endast den temporala synkronisering tas upp i detta projekt. Den övergripande metoden för att göra detta är baserat på optiskt flöde. Data från en ej synkroniserad kamera syntetiseras via en synkroniseringkonstant mha. rörelseinterpolering. Detta jämförs även mot ett tillvägagångssätt som bygger på maskininlärning medan man i denna studie fokuserar på en mer traditionell lösningsmetod. Mätvärdena Peak Signal-to-Noise Ratio och Structural Similarity Index Measure används som kvalitetskriteria. Även visuellt upplevd kvalitet undersöks för att identifiera skillnaden mellan mätt och upplevd kvalitet. Resultatet visar att vid realistiska situationer kan den temporala synkroniseringen beräknas till under 1ms. Den syntetiserade datan lyckas dock inte estimera komplicerade situationer, medan maskininlärningsmetoden presterar bra. Dock så klarar studiens lösningsmetod att bättre generera objekt i snabb rörelse. Vidare så kan inte maskininlärningsmetoden generera video med en godtycklig tidförskjutning, något som är avgörande för projektets tillämpningsområde. Slutligen, med tanke på svårigheter i realtidsberäkning kontra kvaliteten hos maskin- inlärningsmetoder krävs därför mer forskning inom området.

Page generated in 0.075 seconds