1 |
Multi-Modal Control: From Motion Description Languages to Optimal ControlDelmotte, Florent 16 November 2006 (has links)
The goal of the proposed research is to provide efficient methods for defining, selecting and encoding multi-modal control programs. To this end, modes are recovered from system observations, i.e. quantized input-output strings are converted into consistent mode sequences within the Motion Description Language (MDL) framework. The design of such modes can help identify and predict the behaviors of complex systems (e.g. biological systems such as insects) and inspire the design and control of robust semi-autonomous systems (e.g. navigating robots).
In this work, the efficiency of a method will be defined by the complexity and expressiveness of specific control programs. The insistence on low-complexity programs is originally motivated by communication constraints on the computer control of semi-autonomous systems, but also by our belief that, as complex as they may look, natural systems indeed use short motion schemes with few basic behaviors. The attention is first focused on the design of such short-length, few-distinct-modes mode sequences within the MDL framework. Optimal control problems are then addressed. In particular, given a mode sequence, the question of deciding when the system should switch from one mode to another in order to achieve some reachability requirements is studied. Finally, we propose to investigate how sampling strategies affect complexity and reachability, and how an acceptable trade-off between these conflicting entities can be reached.
|
2 |
Motion description languages: from specification to executionMartin, Patrick J. 24 March 2010 (has links)
Many emerging controls applications have seen increased operational complexity due to the deployment of embedded, networked systems that must interact with the physical environment. In order to manage this complexity, we design different control modes for each system and use motion description languages (MDL) to specify a sequence of these controllers to execute at run-time.
Unfortunately, current MDL frameworks lose some of the important details (i.e. power, spatial, or communication capabilities) that affect the execution of the control modes.
This work presents several computational tools that work towards
closing MDL's specification-to-execution gap, which can result in undesirable behavior of complex systems at run-time. First, we develop the notion of an MDL compiler for control specifications with spatial, energy, and temporal constraints. We define a new MDL for networked systems and develop an algorithm that automatically generates a supervisor to prevent incorrect execution of the multi-agent MDL program. Additionally, we derive conditions for checking if an MDL program satisfies actuator constraints and develop an algorithm to insert new control modes that maintain actuator bounds during the execution of the MDL program.
Finally, we design and implement a software architecture that facilitates the development of control applications for systems with power, actuator, sensing, and communication constraints.
|
3 |
Video motion description based on histograms of sparse trajectoriesOliveira, Fábio Luiz Marinho de 05 September 2016 (has links)
Submitted by Renata Lopes (renatasil82@gmail.com) on 2017-06-06T19:12:19Z
No. of bitstreams: 1
fabioluizmarinhodeoliveira.pdf: 1410854 bytes, checksum: cb71ee666cda7d462ce0dd33963a988c (MD5) / Approved for entry into archive by Adriana Oliveira (adriana.oliveira@ufjf.edu.br) on 2017-06-07T13:33:08Z (GMT) No. of bitstreams: 1
fabioluizmarinhodeoliveira.pdf: 1410854 bytes, checksum: cb71ee666cda7d462ce0dd33963a988c (MD5) / Made available in DSpace on 2017-06-07T13:33:08Z (GMT). No. of bitstreams: 1
fabioluizmarinhodeoliveira.pdf: 1410854 bytes, checksum: cb71ee666cda7d462ce0dd33963a988c (MD5)
Previous issue date: 2016-09-05 / Descrição de movimento tem sido um tema desafiador e popular há muitos anos em
visão computacional e processamento de sinais, mas também intimamente relacionado a
aprendizado de máquina e reconhecimento de padrões. Frequentemente, para realizar essa
tarefa, informação de movimento é extraída e codificada em um descritor. Este trabalho
apresenta um método simples e de rápida computação para extrair essa informação e
codificá-la em descritores baseados em histogramas de deslocamentos relativos. Nossos
descritores são compactos, globais, que agregam informação de quadros inteiros, e o que
chamamos de auto-descritor, que não depende de informações de sequências senão aquela
que pretendemos descrever. Para validar estes descritores e compará-los com outros tra
balhos, os utilizamos no contexto de Reconhecimento de Ações Humanas, no qual cenas
são classificadas de acordo com as ações nelas exibidas. Nessa validação, obtemos resul
tados comparáveis aos do estado-da-arte para a base de dados KTH. Também avaliamos
nosso método utilizando as bases UCF11 e Hollywood2, com menores taxas de reconhe
cimento, considerando suas maiores complexidades. Nossa abordagem é promissora, pelas
razoáveis taxas de reconhecimento obtidas com um método muito menos complexo que os
do estado-da-arte, em termos de velocidade de computação e compacidade dos descritores
obtidos. Adicionalmente, experimentamos com o uso de Aprendizado de Métrica para a
classificação de nossos descritores, com o intuito de melhorar a separabilidade e a com
pacidade dos descritores. Os resultados com Aprendizado de Métrica apresentam taxas
de reconhecimento inferiores, mas grande melhoria na compacidade dos descritores. / Motion description has been a challenging and popular theme over many years within
computer vision and signal processing, but also very closely related to machine learn
ing and pattern recognition. Very frequently, to address this task, one extracts motion
information from image sequences and encodes this information into a descriptor. This
work presents a simple and fast computing method to extract this information and en
code it into descriptors based on histograms of relative displacements. Our descriptors
are compact, global, meaning it aggregates information from whole frames, and what we
call self-descriptors, meaning they do not depend on information from sequences other
than the one we want to describe. To validate these descriptors and compare them to
other works, we use them in the context of Human Action Recognition, where scenes are
classified according to the action portrayed. In this validation, we achieve results that are
comparable to those in the state-of-the-art for the KTH dataset. We also evaluate our
method on the UCF11 and Hollywood2 datasets, with lower recognition rates, considering
their higher complexity. Our approach is a promising one, due to the fairly good recogni
tion rates we obtain with a much less complex method than those of the state-of-the-art,
in terms of speed of computation and final descriptor compactness. Additionally, we ex
periment with the use of Metric Learning in the classification of our descriptors, aiming
to improve the separability and compactness of the descriptors. Our results for Metric
Learning show inferior recognition rates, but great improvement for the compactness of
the descriptors.
|
Page generated in 0.1256 seconds