Spelling suggestions: "subject:"data integrity"" "subject:"data ntegrity""
11 |
Data integrity in RFID systemsAlchazidis, Nikolaos. 09 1900 (has links)
One of the main problems that affect the data integrity of passive RFID systems is the collision between the tags. A popular anticollision algorithm which dominates the standards in HF and UHF passive RFID systems is Framed Slotted Aloha (FSA) and some variations of FSA. Throughput and Average time delay of the RFID system which determines the performance/efficiency of the system are reduced rapidly when the number of tags inside the interrogation zone is increased. Using larger frame sizes is not always the solution. This thesis discusses and compares the existing protocols, and proposes a variation of FSA, called Progressing Scanning (PS) algorithm. The PS algorithm divides the tags in the interrogation zone into smaller groups, and gives the ability to the reader to communicate each time with one of them. For performance analysis, the PS algorithm was evaluated with the parameters of a typical passive RFID system at 2.45 GHz . The results showed that the PS algorithm can improve the efficiency of the RFID system and provide a reliable solution for cases with a high density of tags in the area (over 800 tags ).
|
12 |
Projekt N10 : ProjektsrapportSimeon, Nika January 2007 (has links)
<p>DUE TO COPYRIGHT-RESTRICTIONS THIS PAPER IS NOT AVAILABLE FOR DOWNLOAD!</p><p>The thesis describes a system which communicates in real time with data loggers. The system has been streamlined and integrated with existing application so that each user can get a graphical presentation in real time on what has been sent to and from the units. The user interface and communication has been designed to be robust, user friendly, secure and offer functionality that yields the users of the system added value. The system is flexible from the design perspective and is low maintenance.</p>
|
13 |
Projekt N10 : ProjektsrapportSimeon, Nika January 2007 (has links)
DUE TO COPYRIGHT-RESTRICTIONS THIS PAPER IS NOT AVAILABLE FOR DOWNLOAD! The thesis describes a system which communicates in real time with data loggers. The system has been streamlined and integrated with existing application so that each user can get a graphical presentation in real time on what has been sent to and from the units. The user interface and communication has been designed to be robust, user friendly, secure and offer functionality that yields the users of the system added value. The system is flexible from the design perspective and is low maintenance.
|
14 |
Duomenų vientisumo apribojimų realizavimo strategijos didelėje įmonėje tyrimas / Investigation of strategy for implementation of data integrity constraints in large application systemPreibys, Justas 28 January 2008 (has links)
Darbe apžvelgiami duomenų vientisumo apribojimų tipai ir jų realizavimo būdai, analizuojami jų privalumai ir trūkumai. Kiekvienam apribojimo tipui, pagal veikimo trukmės ir įgyvendinimo sudėtingumo charakteristikas, parinktas labiausiai efektyvus realizavimo būdas bei eksperimentiškai patvirtintos analizės metu iškeltos hipotezės. Remiantis eksperimentinio tyrimo rezultatais, sudaryta ir aprašyta duomenų vientisumo apribojimų realizacijos strategijos parinkimo metodika, kuri galėtų padėti didelėms įmonėms efektyviai įgyvendinti duomenų vientisumo apribojimus, sunaudojant kuo mažiau laiko, darbo ir sistemos resursų, bet tuo pačiu užtikrinant duomenų vientisumą ir korektiškumą. / This thesis describes types of data integrity constraints, the ways for implementation these constraints, and their analyzed advantages and disadvantages. It is chosen most effective way for implementation every type of data integrity constraint, according to their run duration and complexity of implementation characteristics, and experimentally approved hypotheses, which were raised during analysis. With reference to the results of the experiment, is made and described the methodology of implementation strategy for data integrity constraints, which could help effectively implement data integrity constraints for large application systems, using as less as it is possible time, workforce and system resources, and at the same time ensuring data integrity and correctness. Read more
|
15 |
A study regarding the effectiveness of game play as part of an information security awareness program for novicesLabuschagne, William Aubrey 09 1900 (has links)
Technology has become intertwined into society daily life which is not only limited to personal life but also extending into the business world. Availability, integrity and confidentiality are critical information security factors to consider when interacting with technology. Conversely many unsuspecting users have fallen prey to cyber criminals. The
majority of threats encountered could have been prevented by the victims if they had sufficient knowledge to first identify and then mitigate the threat. The use of information security awareness programs provides a platform whereby users are informed about such
threats. The success of these programs is significantly reduced if the content is not transferred in the most effective method to improve understanding and result in a change of behaviour. This dissertation addresses the effectiveness of using a gaming platform within an
information security awareness program. The use of games allows for the users to apply knowledge within a potential scenario as seen with pilots using flight simulators. End users who have no information security background should have a safe platform where threats can be identified and methods taught to mitigate the threats. A wide selection of security awareness frameworks exist, the most appropriate framework should be considered first. The different phases of the framework would be applied within the dissertation with the main objective to ultimately determine the effectiveness of games within security awareness programs. Data was collected during the implemented information security awareness program using quantitative instruments. These included questionnaires and a developed online game designed from the literature reviewed during the study. The analysed data highlighted the effects of extrinsic motivation on knowledge transfer and validated the positive impact of
game play. / Computing / M. Tech. (Information Technology) Read more
|
16 |
Development of Methods for Improved Data Integrity and Efficient Testing of Wind Tunnel Models for Dynamic Test Conditions in Unsteady and Nonlinear Flight RegimesHeim, Eugene Henry DeWendt 05 February 2004 (has links)
Today's high performance aircraft are operating in expanded flight envelopes, often maneuvering at high angular rates at high angles-of-attack, even above maximum lift. Current aerodynamic models are inadequate in predicting flight characteristics in the expanded envelope, such as rapid aircraft departures and other unusual motions. Unsteady flows of aircraft are of real concern. The ability to accurately measure aerodynamic loads directly impacts the ability to accurately model and predict flight. Current wind tunnel testing techniques do not adequately address the data fidelity of a test point under the influence of fluctuating loads and moments. Additionally, forced oscillation test techniques, one of the primary tools used to develop dynamic models, do not currently provide estimates of the uncertainty of the results during an oscillation cycle. Further, in testing models across a range of flight conditions, there are frequently parts of the envelope which are well behaved and require few data points to arrive at a sound answer, and other parts of the envelope where the responses are much more active and require a large sample of data to arrive at an answer with statistical significance. Currently, test methods do not factor changes of flow physics into data acquisition schemes, so in many cases data are obtained over more iterations than required, or insufficient data may be obtained to determine a valid estimate. Methods of providing a measure of data integrity for static and forced oscillation test techniques are presented with examples. A method for optimizing required forced oscillation cycles based on decay of uncertainty gradients and balance tolerances is also presented. / Master of Science Read more
|
17 |
Understanding the Impacts of Data Integrity Attacks in the Context of Transactive Control SystemsBiswas, Shuchismita January 2018 (has links)
The rapid growth of internet-connected smart devices capable of exchanging energy price information and adaptively controlling the consumption of connected loads, has paved the way for transactive control to make inroads in the modern grid. Transactive control frameworks integrate the wholesale and retail energy markets, and enable active participation of end users, thereby playing a key role in managing the rising number of distributed assets.However, the use of internet for the communication of data among the building, distribution,and transmission levels makes the system susceptible to external intrusions. A skilled adversary can potentially manipulate the exchanged data with the intention to inflict damage to the system or increase financial gains. In this thesis, the effect of such data integrity attacks on information exchanged between the distribution systems operator and end-users is investigated. Impact on grid operations is evaluated using different categories like operational, financial, user comfort and reliability parameters. It is shown that attack impact depends on a number of factors like attack duration, time of attack, penetration rate etc besides the attack magnitude. The effect of an attack continues to persist for some time after its removal and hence effective detection and mitigation strategies will be required to ensure system resilience and robustness. / Master of Science / Transactive energy is a framework where price-responsive loads adjust their energy consumption at a certain time according to the real-time energy price sent by the utility. Field demonstrations in recent years have shown that transactive control can effectively manage grid objectives and also monetarily benefit both the electric utility and end-users. Therefore, transactive energy is expected to make inroads into conventional grid operations in the next few years. As successful operation of such a market depends on the information exchanged among different stakeholders, a malicious adversary may try to inject false data and affect system operations. This thesis investigates how manipulating data in the transactive energy platform affects system operations and financial gains of different stakeholders. Understanding system behavior under attack conditions will help in formulating effective detection and mitigation strategies and enhancing system resilience. Read more
|
18 |
Comparative Analysis and Implementation of High Data Rate Wireless Sensor Network Simulation FrameworksLaguduva Rajaram, Madhupreetha 12 1900 (has links)
This thesis focuses on developing a high data rate wireless sensor network framework that could be integrated with hardware prototypes to monitor structural health of buildings. In order to better understand the wireless sensor network architecture and its consideration in structural health monitoring, a detailed literature review on wireless sensor networks has been carried out. Through research, it was found that there are numerous simulation software packages available for wireless sensor network simulation. One suitable software was selected for modelling the framework. Research showed that Matlab/Simulink was the most suitable environment, and as a result, a wireless sensor network framework was designed in Matlab/Simulink. Further, the thesis illustrates modeling of a simple accelerometer sensor, such as those used in wireless sensor networks in Matlab/Simulink using a mathematical description. Finally, the framework operation is demonstrated with 10 nodes, and data integrity is analyzed with cyclic redundancy check and transmission error rate calculations.
|
19 |
An investigation into data management as a strategic information tool and its importance at the Durban University of TechnologyFrancis, Ramani 26 November 2012 (has links)
Submitted in fulfilment of the requirements for the Degree of Master Technologiae: Public Management, Durban University of Technology, 2012. / The purpose of this study was to investigate data management as a strategic
information tool and its importance at the Durban University of Technology. The
problem revolved around, inter alia, data management and data accuracy as
structured interventions impacting on sound decision making. There were many
challenges identified with regard to data management at the Durban University
of Technology.
The research design adopted a quantitative methodological approach that was
used for collecting data through the use of a precoded self administered
questionnaire. The empirical component involved a survey method considering
that in was an in-house investigation and the target population equated to only
174 respondents. A significant response rate of 74% was obtained using the
personal method for the data collection. Several hypotheses were formulated
relating to data quality initiatives, data owners and their responsibility and
frequency of data analysis in order to determine accuracy. These were tested
using the Pearson chi-square test as well as data that was analyzed to
determine frequencies and percentages of responses. The data was analyzed
using the computerized Statistical Program for Social Sciences (SPSS)
program. There were some significant findings that emerged from the empirical
analysis.
A highly significant finding was that 95.31% of the respondents
strongly agreed that data management and integrity is of utmost importance at
the Durban University of Technology.
One of the recommendation suggest that an imperative for the Durban
University of Technology to manage its data as an asset, a policy on data
integrity and integration policy should be developed and implemented. Another
recommendation highlighted and staff should strive to attain proper classification
on the database, considering that this directly impacts on the accuracy of the
HEMIS submissions to the Ministry of Education for the state allocated subsidy.
The study concludes with directions for further research as well. Read more
|
20 |
A practical framework for harmonising welfare and quality of data output in the laboratory-housed dogHall, Laura E. January 2014 (has links)
In the UK, laboratory-housed dogs are primarily used as a non-rodent species in the safety testing of new medicines and other chemical entities. The use of animals in research is governed by the Animals (Scientific Procedures) Act (1986, amended 2012) and legislation is underpinned by the principles of humane experimental technique: Replacement, Reduction and Refinement. A link between animal welfare and the quality of data produced has been shown in other species (e.g. rodents, nonhuman primates), however, no established, integrated methodology for identifying or monitoring welfare and quality of data output previously existed in the laboratory-housed dog. In order to investigate the effects of planned Refinements to various aspects of husbandry and regulated procedures, this project sought to integrate behavioural, physiological and other measures (e.g. cognitive bias, mechanical pressure threshold) and to provide a means for staff to monitor welfare whilst also establishing the relationship between welfare and quality of data output. Affective state was identified using an established method of cognitive bias testing, before measuring welfare at ‘baseline’ using measures of behaviour and physiology. Dogs then underwent ‘positive’ and ‘negative’ behavioural challenges to identify the measures most sensitive to changing welfare and most suitable for use in a framework. The resulting Welfare Assessment Framework, developed in three groups of dogs from contrasting backgrounds within the facility, found a consistent pattern of behaviour, cardiovascular function, affect and mechanical pressure threshold (MPT). Dogs with a negative affective state had higher blood pressure at baseline than those with positive affective states, and the magnitude of the effect of negative welfare suggests that welfare may act as a confound in the interpretation of cardiovascular data. The responses to restraint included increases in blood pressure and heart rate measures which approached ceiling levels, potentially reducing the sensitivity of measurement. If maintained over time this response could potentially have a negative health impact on other organ systems and affecting the data obtained from those. Dogs with a negative welfare state also had a lower mechanical pressure threshold, meaning they potentially experienced greater stimulation from unpleasant physical stimuli. Taken together with the behaviours associated with a negative welfare state (predominantly vigilant or stereotypic behaviours) the data suggest that dogs with a negative welfare state have a greater behavioural and physiological response to stimuli in their environment; as such, data obtained from their use is different from that obtained from dogs with a positive welfare state. This was confirmed by examining the effect size (Cohen’s d ) resulting from the analysis of affective state on cardiovascular data. An increase in variance, particularly in the small dog numbers typical of safety assessment studies, means a reduction in the power of the study to detect the effect under observation; a decrease in variation has the potential to reduce the number of dogs use, in line with the principle of Reduction and good scientific practice. The development of the framework also identified areas of the laboratory environment suitable for Refinement (e.g. restriction to single-housing and restraint) and other easily-implemented Refinements (e.g. feeding toy and human interaction) which could be used to improve welfare. As a result of this, a Welfare Monitoring Tool (WMT) in the form of a tick sheet was developed for technical and scientific staff to identify those dogs at risk of reduced welfare and producing poor quality data, as well as to monitor the effects of Refinements to protocols. Oral gavage is a common regulated procedure, known to be potentially aversive and was identified as an area in need of Refinement. A program of desensitisation and positive reinforcement training was implemented in a study also comparing the effects of a sham dose condition versus a control, no-training, condition. A number of the measures used, including home pen behaviour, behaviour during dosing, MPT and the WMT showed significant benefits to the dogs in the Refined condition. Conversely, dogs in the sham dose condition showed more signs of distress and took longer to dose than dogs in the control condition. The welfare of control dogs was intermediate to sham dose and Refined protocol dogs. This project identified a positive relationship between positive welfare and higher quality of data output. It developed and validated a practical and feasible means of measuring welfare in the laboratory environment in the Welfare Assessment Framework, identified areas in need of Refinement and developed practical ways to implement such Refinements to husbandry and regulated procedures. As such it should have wide implications for the pharmaceutical industry and other users of dogs in scientific research. Read more
|
Page generated in 0.0769 seconds