81 |
Concurrent Error Detection in Finite Field Arithmetic OperationsBayat Sarmadi, Siavash January 2007 (has links)
With significant advances in wired and wireless technologies and also increased shrinking in the size of VLSI circuits, many devices have become very large because they need to contain several large units. This large number of gates and in turn large number of transistors causes the devices to be more prone to faults. These faults specially in sensitive and critical applications may cause serious failures and hence should be avoided.
On the other hand, some critical applications such as cryptosystems may also be prone to deliberately injected faults by malicious attackers. Some of these faults can produce erroneous results that can reveal some important secret information of the cryptosystems. Furthermore, yield factor improvement is always an important issue in VLSI design and fabrication processes. Digital systems such as cryptosystems and digital signal processors usually contain finite field operations. Therefore, error detection and correction of such operations have become an important issue recently.
In most of the work reported so far, error detection and correction are applied using redundancies in space (hardware), time, and/or information (coding theory). In this work, schemes based on these redundancies are presented to detect errors in important finite field arithmetic operations resulting from hardware faults. Finite fields are used in a number of practical cryptosystems and channel encoders/decoders. The schemes presented here can detect errors in arithmetic operations of finite fields represented in different bases, including polynomial, dual and/or normal basis, and implemented in various architectures, including bit-serial, bit-parallel and/or systolic arrays.
|
82 |
Analysis of cross-system porting and porting errors in software projectsRay, Baishakhi 11 November 2013 (has links)
Software forking---creating a variant product by copying and modifying an existing project---is often considered an ad hoc, low cost alternative to principled product line development. To maintain forked projects, developers need to manually port existing features or bug-fixes from one project to another. Such manual porting is not only tedious but also error-prone. When the contexts of the ported code vary, developers often have to adapt the ported code to fit its surroundings. Faulty adaptations or inconsistent updates of the ported code could potentially introduce subtle inconsistencies in the codebase. To build a deeper understanding to cross-system porting and porting related errors, this dissertation investigates: (1) How can we identify ported code from software version histories? (2) What is the overhead of cross-system porting required to maintain forked projects? (3) What is the extent and characteristics of porting errors that occur in practice? and (4) How can we detect and characterize potential porting errors? As a first step towards assessing the overhead of cross-system porting, we implement REPERTOIRE, a tool to analyze repeated work of cross-system porting across peer projects. REPERTOIRE can detect ported edits between program patches with high accuracy of 94% precision and 84% recall. Using REPERTOIRE, we study the temporal, spatial, and developer dimensions of cross-system porting using 18 years of parallel evolution history of the BSD product family. Our study finds that cross-system porting happens periodically and the porting rate does not necessarily decrease over time. The upkeep work of porting changes from peer projects is significant and currently, porting practice seems to heavily depend on developers doing their porting job on time. Analyzing version histories of Linux and FreeBSD, we derive five categories of porting errors, including incorrect control- and data-flow, code redundancy, and inconsistent identifier and token renamings. Leveraging this categorization, we design a static control- and data-dependence analysis technique, SPA, to detect and characterize porting inconsistencies. SPA detects porting inconsistencies with 65% to 73% precision and 90% recall, and identify inconsistency types with 58% to 63% precision and 92% recall on average. In a comparison with two existing error detection tools, SPA outperforms them with 14% to 17% better precision. / text
|
83 |
The interaction between speech perception and speech production: implications for speakers with dysarthriaSchaefer, Martina Christina Marion January 2013 (has links)
The purpose of the research presented here was to systematically investigate the role of speech perception on speech production in speakers of different ages and those with PD and hypokinetic dysarthria. For this, the experimental designs of auditory perturbation and mimicry were chosen. The initial research phase established that the magnitude of
compensation to auditory vowel perturbation was reduced in 54 speakers of New Zealand English (NZE) when compared to previous studies conducted with speakers of American (AE) and Canadian English (CE). A number of factors were studied to determine possible predictors of compensation and distinguish between potential changes associated with ageing.
However, no predictors of compensation were found for the overall group. Post-hoc analyses established an increased variability in response patterns in NZE when compared to previous studies of AE and CE. Subsequent follow-up analyses focused on the response-dependent categories of (1) big compensators, (2) compensators, (3) big followers, and (4) followers. Linear mixed-effect modelling revealed that in big compensators, the magnitude of compensation was greater in speakers who exhibited larger F1 baseline standard deviation
and greater F1 vowel distances of HEAD relative to HEED and HAD. F1 baseline standard deviation was found to have a similar predictive value for the group of compensators. No
predictors of compensation were found for the other two subgroups. Phase two was set up as a continuation of phase one and examined whether a subset of 16 speakers classified as big compensators adapted to auditory vowel perturbation. Linear mixed-effect modelling revealed that in the absence of auditory feedback alterations, big compensators maintained
their revised speech motor commands for a short period of time until a process of de-adaptation was initiated. No predictors of adaptation were found for the group. Due to the
unexpected results from the first two research phases indicating a dominant weighting of somatosensory feedback in NZE compared to auditory-perceptual influences, a different experimental paradigm was selected for phase three - mimicry. The purpose of this study was to determine whether eight speakers with PD and dysarthria and eight age-matched healthy controls (HC) are able to effectively integrate speech perception and speech production when attempting to match an acoustic target. Results revealed that all speakers were able to modify their speech production to approximate the model speaker but the acoustic dimensions of their speech did not move significantly closer to the target over the three mimicry attempts. Although speakers with moderate levels of dysarthria exhibited greater acoustic distances (except for the dimension of pitch variation), neither the perceptual nor the acoustic analyses found significant differences in mimicry behaviour across the two groups. Overall, these findings were considered preliminary evidence that speech perception and speech production can at least to some extent be effectively integrated to induce error-correction mechanisms and subsequent speech motor learning in these speakers with PD and dysarthria.
|
84 |
Proposal of two solutions to cope with the faulty behavior of circuits in future technologiesRhod, Eduardo Luis January 2007 (has links)
A diminuição no tamanho dos dispositivos nas tecnologias do futuro traz consigo um grande aumento na taxa de erros dos circuitos, na lógica combinacional e seqüencial. Apesar de algumas potenciais soluções começarem a ser investigadas pela comunidade, a busca por circuitos tolerantes a erros induzidos por radiação, sem penalidades no desempenho, área ou potência, ainda é um assunto de pesquisa em aberto. Este trabalho propõe duas soluções para lidar com este comportamento imprevisível das tecnologias futuras: a primeira solução, chamada MemProc, é uma arquitetura baseada em memória que propõe reduzir a taxa de falhas de aplicações embarcadas micro-controladas. Esta solução baseia-se no uso de memórias magnéticas, que são tolerantes a falhas induzidas por radiação, e área de circuito combinacional reduzida para melhorar a confiabilidade ao processar quaisquer aplicações. A segunda solução proposta aqui é uma implementação de um IP de infra-estrutura para o processador MIPS indicada para sistemas em chip confiáveis, devido a sua adaptação rápida e por permitir diferentes níveis de robustez para a aplicação. A segunda solução é também indicada para sistemas em que nem o hardware nem o software podem ser modificados. Os resultados dos experimentos mostram que ambas as soluções melhoram a confiabilidade do sistema que fazem parte com custos aceitáveis e até, no caso da MemProc, melhora o desempenho da aplicação. / Device scaling in new and future technologies brings along severe increase in the soft error rate of circuits, for combinational and sequential logic. Although potential solutions are being investigated by the community, the search for circuits tolerant to radiation induced errors, without performance, area, or power penalties, is still an open research issue. This work proposes two solutions to cope with this unpredictable behavior of future technologies: the first solution, called MemProc, is a memory based architecture proposed to reduce the fault rate of embedded microcontrolled applications. This solution relies in the use magnetic memories, which are tolerant to radiation induced failures, and reduced combinational circuit area to improve the reliability when processing any application. The second solution proposed here is an infrastructure IP implementation for the MIPS architecture indicated for reliable systems-on-chip due to its fast adaptation and different levels of application hardening that are allowed. The second solution is also indicated for systems where neither the hardware nor the software can be modified. The experimental results show that both solutions improve the reliability of the system they take part with affordable overheads and even, as in the case of the MemProc solution, improving the performance results.
|
85 |
Quaternary CLB a falul tolerant quaternary FPGARhod, Eduardo Luis January 2012 (has links)
A diminuição no tamanho dos transistores vem aumentando cada vez mais o número de funções que os dispositivos eletrônicos podem realizar. Apesar da diminuição do tamanho mínimo dos transistores, a velocidade máxima dos circuitos não consegue seguir a mesma taxa de aumento. Um dos grandes culpados apontados pelos pesquisadores são as interconexões entre os transistores e também entre os componentes. O aumento no número de interconexões dos circuitos traz consigo um significativo aumento do cosumo de energia, aumento do atraso de propagação dos sinais, além de um aumento da complexidade e custo do projeto dos circuitos integrados. Como uma possível solução a este problema é proposta a utilização de lógica multivalorada, mais especificamente, a lógica quaternária. Os dispositivos FPGAs são caracterizados principalmente pela grande flexibilidade que oferecem aos projetistas de sistemas digitais. Entretanto, com o avanço nas tecnologias de fabricação de circuitos integrados e diminuição das dimensões de fabricação, os problemas relacionados ao grande número de interconexões são uma preocupação para as próximas tecnologias de FPGAs. As tecnologias menores que 90nm possuem um grande aumento na taxa de erros dos circuitos, na lógica combinacional e sequencial. Apesar de algumas potenciais soluções começara a ser investigadas pela comunidade, a busca por circuitos tolerantes a erros induzidos por radiação, sem penalidades no desempenho, área ou potência, ainda é um assunto de pesquisa em aberto. Este trabalho propõe o uso de circuitos quaternários com modificações para tolerar falhas provenientes de eventos transientes. Como principal contribuição deste trabalho destaca-se o desenvolvimento de uma CLB (do inglês Configurable Logic Block) quaternária capaz de suportar eventos transientes e, na possibilidade de um erro, evitá-lo ou corrigi-lo. / The decrease in transistor size is increasing the number of functions that can be performed by the electronic devices. Despite this reduction in the transistors minimum size, the circuit’s speed does not follow the same rate. One of the major reasons pointed out by researchers are the interconnections between the transistors and between the components. The increase in the number of circuit interconnections brings a significant increase in energy consumption, propagation delay of signals, and an increase in the complexity and cost of new technologies IC designs. As a possible solution to this problem the use of multivalued logic is being proposed, more specifically, the quaternary logic. FPGA devices are characterized mainly by offering greater flexibility to designers of digital systems. However, with the advance in IC manufacturing technologies and the reduced size of the minimum fabricated dimensions, the problems related to the large number of interconnections are a concern for future technologies of FPGAs. The sub 90nm technologies have a large increase in the error rate of its functions for the combinational and sequential logic. Although potential solutions are being investigated by the community, the search for circuits tolerant to radiation induced errors, without performance, area, or power penalties, is still an open research issue. This work proposes the use of quaternary circuits with modifications to tolerate faults from transient events. The main contribution of this work is the development of a quaternary CLB (Configurable Logic Block) able to withstand transient events and the occurrence of soft errors.
|
86 |
Error detection in wastewater treatment plants using mass balancesKarlsson, Maja January 2018 (has links)
Process data from wastewater treatment plants are often corrupted by errors. These data provide a basis for operating the plant, therefore effort should be made to improve the data quality. Currently, Stockholm Vatten och Avfall uses a method where they quantitatively verify water flow measurement data by comparing it to water level measurements. In this thesis, an alternative approach based on mass balancing to detect errors was evaluated. The aim was to find, implement and evaluate a mass balance based method to detect and locate errors. The objective was to use this method to corroborate the flow verification method used by Stockholm Vatten och Avfall, and to improve flow data from Bromma Wastewater treatment plant. The chosen method consisted of two major steps, gross error detection and data reconciliation. A case study was performed where the method was tested on both simulated data with known added errors, real process data and finally a case where the suggested method was compared to the flow verification method. The results showed that this method was efficient in detecting a gross error when only one flow measurement was erroneous and that the estimation of the error magnitude was good. However, the suggested method was not useful for corroboration of the flow verification method. With the flow verification method, the flow in one filter basin at the time was examined. The suggested method required the combined flow in all 24 filter basins, which made it difficult to compare the two methods. The method has potential to be valuable for error detection in wastewater treatment plants, and to be used as a live tool to detect gross errors.
|
87 |
Proposal of two solutions to cope with the faulty behavior of circuits in future technologiesRhod, Eduardo Luis January 2007 (has links)
A diminuição no tamanho dos dispositivos nas tecnologias do futuro traz consigo um grande aumento na taxa de erros dos circuitos, na lógica combinacional e seqüencial. Apesar de algumas potenciais soluções começarem a ser investigadas pela comunidade, a busca por circuitos tolerantes a erros induzidos por radiação, sem penalidades no desempenho, área ou potência, ainda é um assunto de pesquisa em aberto. Este trabalho propõe duas soluções para lidar com este comportamento imprevisível das tecnologias futuras: a primeira solução, chamada MemProc, é uma arquitetura baseada em memória que propõe reduzir a taxa de falhas de aplicações embarcadas micro-controladas. Esta solução baseia-se no uso de memórias magnéticas, que são tolerantes a falhas induzidas por radiação, e área de circuito combinacional reduzida para melhorar a confiabilidade ao processar quaisquer aplicações. A segunda solução proposta aqui é uma implementação de um IP de infra-estrutura para o processador MIPS indicada para sistemas em chip confiáveis, devido a sua adaptação rápida e por permitir diferentes níveis de robustez para a aplicação. A segunda solução é também indicada para sistemas em que nem o hardware nem o software podem ser modificados. Os resultados dos experimentos mostram que ambas as soluções melhoram a confiabilidade do sistema que fazem parte com custos aceitáveis e até, no caso da MemProc, melhora o desempenho da aplicação. / Device scaling in new and future technologies brings along severe increase in the soft error rate of circuits, for combinational and sequential logic. Although potential solutions are being investigated by the community, the search for circuits tolerant to radiation induced errors, without performance, area, or power penalties, is still an open research issue. This work proposes two solutions to cope with this unpredictable behavior of future technologies: the first solution, called MemProc, is a memory based architecture proposed to reduce the fault rate of embedded microcontrolled applications. This solution relies in the use magnetic memories, which are tolerant to radiation induced failures, and reduced combinational circuit area to improve the reliability when processing any application. The second solution proposed here is an infrastructure IP implementation for the MIPS architecture indicated for reliable systems-on-chip due to its fast adaptation and different levels of application hardening that are allowed. The second solution is also indicated for systems where neither the hardware nor the software can be modified. The experimental results show that both solutions improve the reliability of the system they take part with affordable overheads and even, as in the case of the MemProc solution, improving the performance results.
|
88 |
Quaternary CLB a falul tolerant quaternary FPGARhod, Eduardo Luis January 2012 (has links)
A diminuição no tamanho dos transistores vem aumentando cada vez mais o número de funções que os dispositivos eletrônicos podem realizar. Apesar da diminuição do tamanho mínimo dos transistores, a velocidade máxima dos circuitos não consegue seguir a mesma taxa de aumento. Um dos grandes culpados apontados pelos pesquisadores são as interconexões entre os transistores e também entre os componentes. O aumento no número de interconexões dos circuitos traz consigo um significativo aumento do cosumo de energia, aumento do atraso de propagação dos sinais, além de um aumento da complexidade e custo do projeto dos circuitos integrados. Como uma possível solução a este problema é proposta a utilização de lógica multivalorada, mais especificamente, a lógica quaternária. Os dispositivos FPGAs são caracterizados principalmente pela grande flexibilidade que oferecem aos projetistas de sistemas digitais. Entretanto, com o avanço nas tecnologias de fabricação de circuitos integrados e diminuição das dimensões de fabricação, os problemas relacionados ao grande número de interconexões são uma preocupação para as próximas tecnologias de FPGAs. As tecnologias menores que 90nm possuem um grande aumento na taxa de erros dos circuitos, na lógica combinacional e sequencial. Apesar de algumas potenciais soluções começara a ser investigadas pela comunidade, a busca por circuitos tolerantes a erros induzidos por radiação, sem penalidades no desempenho, área ou potência, ainda é um assunto de pesquisa em aberto. Este trabalho propõe o uso de circuitos quaternários com modificações para tolerar falhas provenientes de eventos transientes. Como principal contribuição deste trabalho destaca-se o desenvolvimento de uma CLB (do inglês Configurable Logic Block) quaternária capaz de suportar eventos transientes e, na possibilidade de um erro, evitá-lo ou corrigi-lo. / The decrease in transistor size is increasing the number of functions that can be performed by the electronic devices. Despite this reduction in the transistors minimum size, the circuit’s speed does not follow the same rate. One of the major reasons pointed out by researchers are the interconnections between the transistors and between the components. The increase in the number of circuit interconnections brings a significant increase in energy consumption, propagation delay of signals, and an increase in the complexity and cost of new technologies IC designs. As a possible solution to this problem the use of multivalued logic is being proposed, more specifically, the quaternary logic. FPGA devices are characterized mainly by offering greater flexibility to designers of digital systems. However, with the advance in IC manufacturing technologies and the reduced size of the minimum fabricated dimensions, the problems related to the large number of interconnections are a concern for future technologies of FPGAs. The sub 90nm technologies have a large increase in the error rate of its functions for the combinational and sequential logic. Although potential solutions are being investigated by the community, the search for circuits tolerant to radiation induced errors, without performance, area, or power penalties, is still an open research issue. This work proposes the use of quaternary circuits with modifications to tolerate faults from transient events. The main contribution of this work is the development of a quaternary CLB (Configurable Logic Block) able to withstand transient events and the occurrence of soft errors.
|
89 |
Proposal of two solutions to cope with the faulty behavior of circuits in future technologiesRhod, Eduardo Luis January 2007 (has links)
A diminuição no tamanho dos dispositivos nas tecnologias do futuro traz consigo um grande aumento na taxa de erros dos circuitos, na lógica combinacional e seqüencial. Apesar de algumas potenciais soluções começarem a ser investigadas pela comunidade, a busca por circuitos tolerantes a erros induzidos por radiação, sem penalidades no desempenho, área ou potência, ainda é um assunto de pesquisa em aberto. Este trabalho propõe duas soluções para lidar com este comportamento imprevisível das tecnologias futuras: a primeira solução, chamada MemProc, é uma arquitetura baseada em memória que propõe reduzir a taxa de falhas de aplicações embarcadas micro-controladas. Esta solução baseia-se no uso de memórias magnéticas, que são tolerantes a falhas induzidas por radiação, e área de circuito combinacional reduzida para melhorar a confiabilidade ao processar quaisquer aplicações. A segunda solução proposta aqui é uma implementação de um IP de infra-estrutura para o processador MIPS indicada para sistemas em chip confiáveis, devido a sua adaptação rápida e por permitir diferentes níveis de robustez para a aplicação. A segunda solução é também indicada para sistemas em que nem o hardware nem o software podem ser modificados. Os resultados dos experimentos mostram que ambas as soluções melhoram a confiabilidade do sistema que fazem parte com custos aceitáveis e até, no caso da MemProc, melhora o desempenho da aplicação. / Device scaling in new and future technologies brings along severe increase in the soft error rate of circuits, for combinational and sequential logic. Although potential solutions are being investigated by the community, the search for circuits tolerant to radiation induced errors, without performance, area, or power penalties, is still an open research issue. This work proposes two solutions to cope with this unpredictable behavior of future technologies: the first solution, called MemProc, is a memory based architecture proposed to reduce the fault rate of embedded microcontrolled applications. This solution relies in the use magnetic memories, which are tolerant to radiation induced failures, and reduced combinational circuit area to improve the reliability when processing any application. The second solution proposed here is an infrastructure IP implementation for the MIPS architecture indicated for reliable systems-on-chip due to its fast adaptation and different levels of application hardening that are allowed. The second solution is also indicated for systems where neither the hardware nor the software can be modified. The experimental results show that both solutions improve the reliability of the system they take part with affordable overheads and even, as in the case of the MemProc solution, improving the performance results.
|
90 |
Desenvolvimento de um software para reconciliação de dados de processos quimicos e petroquimicos / Development of software for data reconciliation of chemical and petrochemical processesBarbosa, Agremis Guinho 11 June 2003 (has links)
Orientador: Rubens Maciel Filho / Dissertação (mestrado) - Universidade Estadual de Campinas, Faculdade de Engenharia Quimica / Made available in DSpace on 2018-08-10T21:51:34Z (GMT). No. of bitstreams: 1
Barbosa_AgremisGuinho_M.pdf: 1501070 bytes, checksum: c20fd373ba5e239e2b783608aebbc7f2 (MD5)
Previous issue date: 2003 / Resumo: O objetivo deste trabalho é o desenvolvimento de rotinas computacionais para o condicionamento de dados provenientes de um processo químico, de modo que estes sejam consistentes para a representação do comportamento do processo. A descrição adequada do comportamento de um processo é a base fundamental de qualquer sistema de controle e/ou otimização, uma vez que será em resposta às medições deste processo (sua descrição) que os referidos sistemas atuarão. Desta forma o tratamento e correção dos erros de medição, especificamente, e a estimativa de parâmetros, de um modo mais geral, constituem uma etapa que não deve ser negligenciada no controle e otimização de processos. O condicionamento de dados estudado neste trabalho é a reconciliação de dados, que tem como característica principal o uso de um modelo de restrições para condicionar a informação. Geralmente os modelos de restrição são balanços de massa e energia e os somatórios das frações mássicas e molares, mas outros modelos também podem ser usados. Matematicamente, a reconciliação de dados é um problema de otimização sujeito a restrições. Neste trabalho, a formulação do problema de reconciliação é a dos mínimos quadrados ponderados sujeito a restrições e a abordagem para a sua solução é a fatoração QR. Objetiva-se também reunir as rotinas desenvolvidas em uma única ferramenta computacional para a descrição, resolução e análise dos resultados do problema de reconciliação de dados, constituindo-se em um software de fácil utilização e que tenha ainda um mecanismo de comunicação com banco de dados, conferindo-lhe interatividade em tempo real com sistemas de aquisição de dados de processo / Abstract: The purpose of this work is the development of computational routines for conditioning chemical process data in order to represent the process behavior as reliable as possible. Reliable process description is fundamental for any control or optimization system development, since they respond to the process measurements (its description). Thus, data conditioning and correction of process measurement errors, and parameter estimation are a step that should not be neglected in process control and optimization.
The data conditioning considered in this work is data reconciliation which has as the main characteristic the use of a constraint model. In general constraint models are mass and energy balances and mass and molar fraction summation, but other models may be used.
Under a mathematical point of view, data reconciliation is an optimization subject to constraints. In this work, it is used the formulation of weighed least squares subject to constraints and QR factorization approach to solve the problem. The additional objective of this work is to accommodate the developed routines in such a way to build up an integrated computational tool characterized by its easy to use structure, capability to solve and perform data reconciliation. Its structure takes into account the interaction with data bank, giving it real time interactiveness with process data acquisition systems / Mestrado / Desenvolvimento de Processos Químicos / Mestre em Engenharia Química
|
Page generated in 0.0421 seconds