• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 43
  • 17
  • 7
  • 4
  • 1
  • 1
  • Tagged with
  • 82
  • 32
  • 19
  • 15
  • 14
  • 13
  • 13
  • 12
  • 12
  • 12
  • 10
  • 10
  • 9
  • 9
  • 8
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
71

Návrh řízení všesměrového mobilního robotu O3-X / Design of omni directional mobile robot (O3-X) control

Olša, Petr January 2010 (has links)
This thesis deals with the design of a three-wheeled omni-directional robot control. The model of control is designed for robot´s omni-directional platform driven by maxon motor with the intelligent positioning controller EPOS. The design of control contains: - installation of the coordinated systems and transformation from one of them into another - design of system´s kinematical model - creation of classes for control and communication with EPOS - creation of the simulative program - planning of the mobile robot´s path - verification that the system is working The solution was based on continuous accelerated motion and the maximal acceleration of wheels was concerned, so that the slip would be suppressed. The function of the model was partly verified.
72

Geometric model of a dual-fisheye system composed of hyper-hemispherical lenses /

Castanheiro, Letícia Ferrari January 2020 (has links)
Orientador: Antonio Maria Garcia Tommaselli / Resumo: A combinação de duas lentes com FOV hiper-hemisférico em posição opostas pode gerar um sistema omnidirecional (FOV 360°) leve, compacto e de baixo custo, como Ricoh Theta S e GoPro Fusion. Entretanto, apenas algumas técnicas e modelos matemáticos para a calibração um sistema com duas lentes hiper-hemisféricas são apresentadas na literatura. Nesta pesquisa, é avaliado e definido um modelo geométrico para calibração de sistemas omnidirecionais compostos por duas lentes hiper-hemisféricas e apresenta-se algumas aplicações com esse tipo de sistema. A calibração das câmaras foi realizada no programa CMC (calibração de múltiplas câmeras) utilizando imagens obtidas a partir de vídeos feitos com a câmara Ricoh Theta S no campo de calibração 360°. A câmara Ricoh Theta S é composto por duas lentes hiper-hemisféricas fisheye que cobrem 190° cada uma. Com o objetivo de avaliar as melhorias na utilização de pontos em comum entre as imagens, dois conjuntos de dados de pontos foram considerados: (1) apenas pontos no campo hemisférico, e (2) pontos em todo o campo de imagem (isto é, adicionar pontos no campo de imagem hiper-hemisférica). Primeiramente, os modelos ângulo equisólido, equidistante, estereográfico e ortogonal combinados com o modelo de distorção Conrady-Brown foram testados para a calibração de um sensor da câmara Ricoh Theta S. Os modelos de ângulo-equisólido e estereográfico apresentaram resultados melhores do que os outros modelos. Portanto, esses dois modelos de projeção for... (Resumo completo, clicar acesso eletrônico abaixo) / Abstract: The arrangement of two hyper-hemispherical fisheye lenses in opposite position can design a light weight, small and low-cost omnidirectional system (360° FOV), e.g. Ricoh Theta S and GoPro Fusion. However, only a few techniques are presented in the literature to calibrate a dual-fisheye system. In this research, a geometric model for dual-fisheye system calibration was evaluated, and some applications with this type of system are presented. The calibrating bundle adjustment was performed in CMC (calibration of multiple cameras) software by using the Ricoh Theta video frames of the 360° calibration field. The Ricoh Theta S system is composed of two hyper-hemispherical fisheye lenses with 190° FOV each one. In order to evaluate the improvement in applying points in the hyper-hemispherical image field, two data set of points were considered: (1) observations that are only in the hemispherical field, and (2) points in all image field, i.e. adding points in the hyper-hemispherical image field. First, one sensor of the Ricoh Theta S system was calibrated in a bundle adjustment based on the equidistant, equisolid-angle, stereographic and orthogonal models combined with Conrady-Brown distortion model. Results showed that the equisolid-angle and stereographic models can provide better solutions than those of the others projection models. Therefore, these two projection models were implemented in a simultaneous camera calibration, in which the both Ricoh Theta sensors were considered i... (Complete abstract click electronic access below) / Mestre
73

Omnidirectional Phase Matching In Zero-Index Media

Gagnon, Justin 22 April 2021 (has links)
Since its inception, the field of nonlinear optics has only increased in importance as a result of a growing number of applications. The efficiency of all parametric nonlinear optical processes is limited by challenges associated with phase-matching requirements. To address this constraint, a variety of approaches, such as quasi-phase-matching, birefringent phase matching, and higher-order-mode phase matching have historically been used to phase-match interactions. However, the methods demonstrated to date suffer from the inconvenience of only being phase-matched for one specific arrangement of beams, typically co-propagating along the same axis. This stringency of the phase-matching requirement results in cumbersome optical configurations and large footprints for integrated devices. In this thesis, we show that phase-matching requirements in parametric nonlinear optical processes may be satisfied for all orientations of input and output beams when using zero-index media: a condition of omnidirectional phase matching. To validate this theory, we perform experimental demonstrations of phase matching for five separate FWM beam configurations to confirm this phenomenon. Our measurements constitute the first experimental observation of the simultaneous generation of a forward- and backward-propagating signal with respect to the pump beams in a medium longer than a free-space optical wavelength, allowing us to determine the coherence length of our four-wave-mixing process. Our demonstration includes nonlinear signal generation from spectrally distinct counter-propagating pump and probe beams, as well as the excitation of a parametric process with the probe beam's wave vector orthogonal to the wave vector of the pump beam. By sampling all of these beam configurations, our results explicitly demonstrate that the unique properties of zero-index media relax traditional phase-matching constraints, and provide strong experimental evidence for the existence of omnidirectional phase matching in zero-index media. This property can be exploited to facilitate nonlinear interactions and miniaturize nonlinear devices, and adds to the established exceptional properties of low-index materials.
74

A Comparative Study of Omnidirectional and Differential Drive Systems for Mobile Manipulator Robots : A Performance Review of Strengths and Weaknesses / En jämförande studie om omnidirektionell drift och differentialdrift för mobila manipulatorer : En prestationsrecension av styrkor och svagheter

Vestman, Rebecka January 2023 (has links)
This thesis investigates the strengths and weaknesses of omnidirectional drive and differential drive systems on mobile manipulator robots. Based on a literature study, a hypothetical use case, and identified Key Performace Indicators the drive system’s effects on the performance of the mobile manipulator are evaluated. A qualitative approach was used for evaluation. The research methodology involved analyzing the wheel characteristics of each drive system, identifying parameters affecting the performance, and assessing the two drive system characteristics within the context of the hypothetical use case. Six Key Performance Indicators such as pose accuracy, space utilization, and manipulability were formulated and examined to determine the comparative strengths and weaknesses of the drive systems. The results confirmed that omnidirectional drive systems exhibit greater maneuverability and agility while differential drive systems are less complex and often more durable in rough conditions. However, the results also show that in many cases the answer on what drive system to use will depend on many factors and that these factors can affect the overall performance of the mobile manipulator. This study provides insights into performance-affecting parameters and relevant performance aspects by examining the strengths and weaknesses of omnidirectional and differential drive systems. While acknowledging the need for caution in generalizing the findings and assuming validity in real-world applications, the results obtained serve as a starting point for further investigations. / Detta examensarbete undersöker styrkor och svagheter hos omnidirektionella drivsystem och differentialdrivsystem på mobila manipulatorer. En kvalitativ metod, baserat på en litteraturstudie, ett hypotetiskt användningsscenario och identifierade prestationsindikatorer, användes för att utvärdera drivsystemets effekter på den mobila manipulatorns prestanda. Forskningsmetodiken innebar att analysera hjulegenskaperna för varje drivsystem, identifiera parametrar som påverkar prestandan och bedöma de två drivsystemens egenskaper inom ramen för det hypotetiska användningsfallet. Viktiga prestationsindikatorer som poserings noggrannhet, utrymmesutnyttjande och manipulerbarhet undersöktes för att fastställa och jämförande styrkorna och svagheterna hos drivsystemen. Resultaten bekräftade att omnidirektionella drivsystem uppvisar större manövrerbarhet och smidighet medan differentialdrivsystem ofta är mindre komplexa och mer hållbara under tuffa förhållanden. Dock visar resultaten även att svaret på vilket drivsystem som ska användas i många fall beror på flertalet faktorer och att dessa faktorer kan påverka den mobila manipulatorns totala prestanda. Denna studie ger insikter i prestandapåverkande parametrar och relevanta prestandaaspekter genom att undersöka styrkorna och svagheterna hos omnidirektionella och differentiella drivsystem. Samtidigt som man erkänner behovet av försiktighet med att generalisera resultaten och anta giltighet i verkliga tillämpningar, tjänar de erhållna resultaten som en utgångspunkt för ytterligare undersökningar.
75

A Low-cost Omni-directional Visual Bearing Only Localization System

hart, charles 21 February 2014 (has links)
No description available.
76

Appearance-based mapping and localization using feature stability histograms for mobile robot navigation

Bacca Cortés, Eval Bladimir 20 June 2012 (has links)
This work proposes an appearance-based SLAM method whose main contribution is the Feature Stability Histogram (FSH). The FSH is built using a voting schema, if the feature is re-observed, it will be promoted; otherwise it progressively decreases its corresponding FSH value. The FSH is based on the human memory model to deal with changing environments and long-term SLAM. This model introduces concepts of Short-Term memory (STM), which retains information long enough to use it, and Long-Term memory (LTM), which retains information for longer periods of time. If the entries in the STM are rehearsed, they become part of the LTM (i.e. they become more stable). However, this work proposes a different memory model, allowing to any input be part of the STM or LTM considering the input strength. The most stable features are only used for SLAM. This innovative feature management approach is able to cope with changing environments, and long-term SLAM. / Este trabajo propone un método de SLAM basado en apariencia cuya principal contribución es el Histograma de Estabilidad de Características (FSH). El FSH es construido por votación, si una característica es re-observada, ésta será promovida; de lo contrario su valor FSH progresivamente es reducido. El FSH es basado en el modelo de memoria humana para ocuparse de ambientes cambiantes y SLAM a largo término. Este modelo introduce conceptos como memoria a corto plazo (STM) y largo plazo (LTM), las cuales retienen información por cortos y largos periodos de tiempo. Si una entrada a la STM es reforzada, ésta hará parte de la LTM (i.e. es más estable). Sin embargo, este trabajo propone un modelo de memoria diferente, permitiendo a cualquier entrada ser parte de la STM o LTM considerando su intensidad. Las características más estables son solamente usadas en SLAM. Esta innovadora estrategia de manejo de características es capaz de hacer frente a ambientes cambiantes y SLAM de largo término.
77

A Novel Approach for Spherical Stereo Vision / Ein Neuer Ansatz für Sphärisches Stereo Vision

Findeisen, Michel 27 April 2015 (has links) (PDF)
The Professorship of Digital Signal Processing and Circuit Technology of Chemnitz University of Technology conducts research in the field of three-dimensional space measurement with optical sensors. In recent years this field has made major progress. For example innovative, active techniques such as the “structured light“-principle are able to measure even homogeneous surfaces and find its way into the consumer electronic market in terms of Microsoft’s Kinect® at the present time. Furthermore, high-resolution optical sensors establish powerful, passive stereo vision systems in the field of indoor surveillance. Thereby they induce new application domains such as security and assistance systems for domestic environments. However, the constraint field of view can be still considered as an essential characteristic of all these technologies. For instance, in order to measure a volume in size of a living space, two to three deployed 3D sensors have to be applied nowadays. This is due to the fact that the commonly utilized perspective projection principle constrains the visible area to a field of view of approximately 120°. On the contrary, novel fish-eye lenses allow the realization of omnidirectional projection models. Therewith, the visible field of view can be enlarged up to more than 180°. In combination with a 3D measurement approach, thus, the number of required sensors for entire room coverage can be reduced considerably. Motivated by the requirements of the field of indoor surveillance, the present work focuses on the combination of the established stereo vision principle and omnidirectional projection methods. The entire 3D measurement of a living space by means of one single sensor can be considered as major objective. As a starting point for this thesis chapter 1 discusses the underlying requirement, referring to various relevant fields of application. Based on this, the distinct purpose for the present work is stated. The necessary mathematical foundations of computer vision are reflected in Chapter 2 subsequently. Based on the geometry of the optical imaging process, the projection characteristics of relevant principles are discussed and a generic method for modeling fish-eye cameras is selected. Chapter 3 deals with the extraction of depth information using classical (perceptively imaging) binocular stereo vision configurations. In addition to a complete recap of the processing chain, especially occurring measurement uncertainties are investigated. In the following, Chapter 4 addresses special methods to convert different projection models. The example of mapping an omnidirectional to a perspective projection is employed, in order to develop a method for accelerating this process and, hereby, for reducing the computational load associated therewith. Any errors that occur, as well as the necessary adjustment of image resolution, are an integral part of the investigation. As a practical example, an application for person tracking is utilized in order to demonstrate to which extend the usage of “virtual views“ can increase the recognition rate for people detectors in the context of omnidirectional monitoring. Subsequently, an extensive search with respect to omnidirectional imaging stereo vision techniques is conducted in chapter 5. It turns out that the complete 3D capture of a room is achievable by the generation of a hemispherical depth map. Therefore, three cameras have to be combined in order to form a trinocular stereo vision system. As a basis for further research, a known trinocular stereo vision method is selected. Furthermore, it is hypothesized that, applying a modified geometric constellation of cameras, more precisely in the form of an equilateral triangle, and using an alternative method to determine the depth map, the performance can be increased considerably. A novel method is presented, which shall require fewer operations to calculate the distance information and which is to avoid a computational costly step for depth map fusion as necessary in the comparative method. In order to evaluate the presented approach as well as the hypotheses, a hemispherical depth map is generated in Chapter 6 by means of the new method. Simulation results, based on artificially generated 3D space information and realistic system parameters, are presented and subjected to a subsequent error estimate. A demonstrator for generating real measurement information is introduced in Chapter 7. In addition, the methods that are applied for calibrating the system intrinsically as well as extrinsically are explained. It turns out that the calibration procedure utilized cannot estimate the extrinsic parameters sufficiently. Initial measurements present a hemispherical depth map and thus con.rm the operativeness of the concept, but also identify the drawbacks of the calibration used. The current implementation of the algorithm shows almost real-time behaviour. Finally, Chapter 8 summarizes the results obtained along the studies and discusses them in the context of comparable binocular and trinocular stereo vision approaches. For example the results of the simulations carried out produced a saving of up to 30% in terms of stereo correspondence operations in comparison with a referred trinocular method. Furthermore, the concept introduced allows the avoidance of a weighted averaging step for depth map fusion based on precision values that have to be calculated costly. The achievable accuracy is still comparable for both trinocular approaches. In summary, it can be stated that, in the context of the present thesis, a measurement system has been developed, which has great potential for future application fields in industry, security in public spaces as well as home environments.
78

Robot navigation in sensor space

Keeratipranon, Narongdech January 2009 (has links)
This thesis investigates the problem of robot navigation using only landmark bearings. The proposed system allows a robot to move to a ground target location specified by the sensor values observed at this ground target posi- tion. The control actions are computed based on the difference between the current landmark bearings and the target landmark bearings. No Cartesian coordinates with respect to the ground are computed by the control system. The robot navigates using solely information from the bearing sensor space. Most existing robot navigation systems require a ground frame (2D Cartesian coordinate system) in order to navigate from a ground point A to a ground point B. The commonly used sensors such as laser range scanner, sonar, infrared, and vision do not directly provide the 2D ground coordi- nates of the robot. The existing systems use the sensor measurements to localise the robot with respect to a map, a set of 2D coordinates of the objects of interest. It is more natural to navigate between the points in the sensor space corresponding to A and B without requiring the Cartesian map and the localisation process. Research on animals has revealed how insects are able to exploit very limited computational and memory resources to successfully navigate to a desired destination without computing Cartesian positions. For example, a honeybee balances the left and right optical flows to navigate in a nar- row corridor. Unlike many other ants, Cataglyphis bicolor does not secrete pheromone trails in order to find its way home but instead uses the sun as a compass to keep track of its home direction vector. The home vector can be inaccurate, so the ant also uses landmark recognition. More precisely, it takes snapshots and compass headings of some landmarks. To return home, the ant tries to line up the landmarks exactly as they were before it started wandering. This thesis introduces a navigation method based on reflex actions in sensor space. The sensor vector is made of the bearings of some landmarks, and the reflex action is a gradient descent with respect to the distance in sensor space between the current sensor vector and the target sensor vec- tor. Our theoretical analysis shows that except for some fully characterized pathological cases, any point is reachable from any other point by reflex action in the bearing sensor space provided the environment contains three landmarks and is free of obstacles. The trajectories of a robot using reflex navigation, like other image- based visual control strategies, do not correspond necessarily to the shortest paths on the ground, because the sensor error is minimized, not the moving distance on the ground. However, we show that the use of a sequence of waypoints in sensor space can address this problem. In order to identify relevant waypoints, we train a Self Organising Map (SOM) from a set of observations uniformly distributed with respect to the ground. This SOM provides a sense of location to the robot, and allows a form of path planning in sensor space. The navigation proposed system is analysed theoretically, and evaluated both in simulation and with experiments on a real robot.
79

Robot s autonomním audio-vizuálním řízením / Robot with autonomous audio-video control

Dvořáček, Štěpán January 2019 (has links)
This thesis describes the design and realization of a mobile robot with autonomous audio-visual control. This robot is able of movement based on sensors consisting of camera and microphone. The mechanical part consists of components made with 3D print technology and omnidirectional Mecanum wheels. Software utilizes OpenCV library for image processing and computes MFCC a DTW for voice command detection.
80

Spatio-Temporal Networks for Human Activity Recognition based on Optical Flow in Omnidirectional Image Scenes

Seidel, Roman 29 February 2024 (has links)
The ability of human beings to perceive the environment around them with their visual system is called motion perception. This means that the attention of our visual system is primarily focused on those objects that are moving. The property of human motion perception is used in this dissertation to infer human activity from data using artificial neural networks. One of the main aims of this thesis is to discover which modalities, namely RGB images, optical flow and human keypoints, are best suited for HAR in omnidirectional data. Since these modalities are not yet available for omnidirectional cameras, they are synthetically generated and captured with an omnidirectional camera. During data generation, a distinction is made between synthetically generated omnidirectional data and a real omnidirectional dataset that was recorded in a Living Lab at Chemnitz University of Technology and subsequently annotated by hand. The synthetically generated dataset, called OmniFlow, consists of RGB images, optical flow in forward and backward directions, segmentation masks, bounding boxes for the class people, as well as human keypoints. The real-world dataset, OmniLab, contains RGB images from two top-view scenes as well as manually annotated human keypoints and estimated forward optical flow. In this thesis, the generation of the synthetic and real-world datasets is explained. The OmniFlow dataset is generated using the 3D rendering engine Blender, in which a fully configurable 3D indoor environment is created with artificially textured rooms, human activities, objects and different lighting scenarios. A randomly placed virtual camera following the omnidirectional camera model renders the RGB images, all other modalities and 15 predefined activities. The result of modelling the 3D indoor environment is the OmniFlow dataset. Due to the lack of omnidirectional optical flow data, the OmniFlow dataset is validated using Test-Time Augmentation (TTA). Compared to the baseline, which contains Recurrent All-Pairs Field Transforms (RAFT) trained on the FlyingChairs and FlyingThings3D datasets, it was found that only about 1000 images need to be used for fine-tuning to obtain a very low End-point Error (EE). Furthermore, it was shown that the influence of TTA on the test dataset of OmniFlow affects EE by about a factor of three. As a basis for generating artificial keypoints on OmniFlow with action labels, the Carnegie Mellon University motion capture database is used with a large number of sports and household activities as skeletal data defined in the BVH format. From the BVH-skeletal data, the skeletal points of the people performing the activities can be directly derived or extrapolated by projecting these points from the 3D world into an omnidirectional 2D image. The real-world dataset, OmniLab, was recorded in two rooms of the Living Lab with five different people mimicking the 15 actions of OmniFlow. Human keypoint annotations were added manually in two iterations to reduce the error rate of incorrect annotations. The activity-level evaluation was investigated using a TSN and a PoseC3D network. The TSN consists of two CNNs, a spatial component trained on RGB images and a temporal component trained on the dense optical flow fields of OmniFlow. The PoseC3D network, an approach to skeleton-based activity recognition, uses a heatmap stack of keypoints in combination with 3D convolution, making the network more effective at learning spatio-temporal features than methods based on 2D convolution. In the first step, the networks were trained and validated on the synthetically generated dataset OmniFlow. In the second step, the training was performed on OmniFlow and the validation on the real-world dataset OmniLab. For both networks, TSN and PoseC3D, three hyperparameters were varied and the top-1, top-5 and mean accuracy given. First, the learning rate of the stochastic gradient descent (Stochastic Gradient Descent (SGD)) was varied. Secondly, the clip length, which indicates the number of consecutive frames for learning the network, was varied, and thirdly, the spatial resolution of the input data was varied. For the spatial resolution variation, five different image sizes were generated from the original dataset by cropping from the original dataset of OmniFlow and OmniLab. It was found that keypoint-based HAR with PoseC3D performed best compared to human activity classification based on optical flow and RGB images. This means that the top-1 accuracy was 0.3636, the top-5 accuracy was 0.7273 and the mean accuracy was 0.3750, showing that the most appropriate output resolution is 128px × 128px and the clip length is at least 24 consecutive frames. The best results could be achieved with a learning rate of PoseC3D of 10-3. In addition, confusion matrices indicating the class-wise accuracy of the 15 activity classes have been given for the modalities RGB images, optical flow and human keypoints. The confusion matrix for the modality RGB images shows the best classification result of the TSN for the action walk with an accuracy of 1.00, but almost all other actions are also classified as walking in real-world data. The classification of human actions based on optical flow works best on the action sit in chair and stand up with an accuracy of 1.00 and walk with 0.50. Furthermore, it is noticeable that almost all actions are classified as sit in chair and stand up, which indicates that the intra-class variance is low, so that the TSN is not able to distinguish between the selected action classes. Validated on real-world data for the modality keypoint the actions rugpull (1.00) and cleaning windows (0.75) performs best. Therefore, the PoseC3D network on a time-series of human keypoints is less sensitive to variations in the image angle between the synthetic and real-world data than for the modalities RGB images and optical flow. The pipeline for the generation of synthetic data with regard to a more uniform distribution of the motion magnitudes needs to be investigated in future work. Random placement of the person and other objects is not sufficient for a complete coverage of all movement magnitudes. An additional improvement of the synthetic data could be the rotation of the person around their own axis, so that the person moves in a different direction while performing the activity and thus the movement magnitudes contain more variance. Furthermore, the domain transition between synthetic and real-world data should be considered further in terms of viewpoint invariance and augmentation methods. It may be necessary to generate a new synthetic dataset with only top-view data and re-train the TSN and PoseC3D. As an augmentation method, for example, the Fourier Domain Adaption (FDA) could reduce the domain gap between the synthetically generated and the real-world dataset.:1 Introduction 2 Theoretical Background 3 Related Work 4 Omnidirectional Synthetic Human Optical Flow 5 Human Keypoints for Pose in Omnidirectional Images 6 Human Activity Recognition in Indoor Scenarios 7 Conclusion and Future Work A Chapter 4: Flow Dataset Statistics B Chapter 5: 3D Rotation Matrices C Chapter 6: Network Training Parameters

Page generated in 0.1175 seconds