• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 59
  • 15
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 89
  • 89
  • 42
  • 27
  • 16
  • 13
  • 11
  • 11
  • 11
  • 10
  • 10
  • 9
  • 9
  • 9
  • 9
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
31

Architectural level risk assessment

Hassan, Ahmed E. January 1900 (has links)
Thesis (Ph. D.)--West Virginia University, 2004. / Title from document title page. Document formatted into pages; contains xx, 157, [12] p. : ill. (some col.). Includes abstract. Includes bibliographical references (p. 143-157).
32

Proposal to develop enhancements and extensions of formal models for risk assessment in software projects /

Murrah, Michael R. January 2002 (has links) (PDF)
Thesis (M.S. in Software Engineering)--Naval Postgraduate School, September 2002. / Thesis advisor(s): Luqi, Nabendu Chaki. Includes bibliographical references (p. 115-116). Also available online.
33

Validating cohesion metrics by mining open source software data with association rules

Singh, Pariksha January 2008 (has links)
Dissertation submitted for the fulfillment of the requirement for the degree of Masters in Information Technology, Department of Information Technology, Faculty of Accounting and Informatics, Durban University of Technology, 2008. / Competitive pressure on the software industry encourages organizations to examine the effectiveness of their software development and evolutionary processes. Therefore it is important that software is measured in order to improve the quality. The question is not whether we should measure software but how it should be measured. Software measurement has been in existence for over three decades and it is still in the process of becoming a mature science. The many influences of new software development technologies have led to a diverse growth in software measurement technologies which have resulted in various definitions and validation techniques. An important aspect of software measurement is the measurement of the design, which nowadays often means the measurement of object oriented design. Chidamer and Kemerer (1994) designed a metric suite for object oriented design, which has provided a new foundation for metrics and acts as a starting point for further development of the software measurement science. This study documents theoretical object oriented cohesion metrics and calculates those metrics for classes extracted from a sample of open source software packages. For each open source software package, the following data is recorded: software size, age, domain, number of developers, number of bugs, support requests, feature requests, etc. The study then tests by means of association rules which theoretical cohesion metrics are validated hypothesis: that older software is more cohesive than younger software, bigger packages is less cohesive than smaller packages, and the smaller the software program the more maintainable it is. This study attempts to validate existing theoretical object oriented cohesion metrics by mining open source software data with association rules.
34

Performance measurement as a tool for software engineering

Van Aardt, Jan Markus 22 July 2005 (has links)
Some software development teams regard software performance measurement as a mere luxury. When it happens, it often tends to be infrequent, insufficient and subjective. Countless software projects were sent into an uncontrollable spiral of poor management and unsatisfactory results. By revisiting old ideas and policies, many companies have turned themselves around. To ensure that software engineering does the same, technologies and procedures have to be reevaluated. The fact that many companies have decided to cut costs on technology expenditure necessitates software development teams to look for alternative options for deploying high performance software systems. As many companies are moving into the electronic era and evolving to the next stage of evolution, electronic commerce, the more important it has become to apply these concepts on Internet development projects and procedures. The Internet market has shown that two software providers are aiming for worldwide domination of Internet server deployment, being Microsoft and Apache. Currently, the Apache web server is the most commonly used server on the Internet today (60%), with Microsoft's Internet Information Server (25%) in a strong second place. The need for higher throughput and better services is getting more with each passing day. It increases the pressure on these two software vendors to provide the best architecture for their clients' needs. This study intends to provide the reader with an objective view of a basic performance comparison between these two products and tries to find a correlation between the performance tests and the products' popularity standings. The tests for this study were performed on identical hardware architectures with one difference, being the operating system. By comparing the costly proprietary Microsoft solution with its cheaper open source rival, Linux, certain opinions were tested. Would a product developed by a software company that invests millions of dollars in their products perform better than this free-for-all solution, or would the selfless inputs of hundreds of developers all over the world finally payoff through the creation of the world's best Internet server? The results of these tests were evaluated through formal statistical methods, providing overall comparisons of several common uses of web servers. These results were implemented in a small field test to prove the findings in practice with some interesting outcomes in terms of supportive technologies, new rapid application development (RAD) tools and data access models. This research in itself will not change the mind of any Internet programmer. What it hopes to achieve is to demonstrate software engineers that current processes and methods of developing software are not always the right way of doing things. Furthermore, it highlights many important factors often ignored or overlooked while managing software projects. Change management, process re-engineering and risk management form crucial elements of software development projects. By not adhering to certain critical elements of software development, software projects stand the chance of not reaching their goals and could even fail completely. Performance measurement acts as a tool for software engineering, providing guidelines for technological decisions, project management and ultimately, project success. / Dissertation (MSc (Computer Science))--University of Pretoria, 2005. / Computer Science / unrestricted
35

MEASUREMENT AND ITS HISTORICAL CONTEXT

Gedela, Naga Venkata Praveen babu 12 November 2008 (has links)
No description available.
36

Assessment of software measurement

Berry, Michael, CSE, UNSW January 2006 (has links)
Background and purpose. This thesis documents a program of five studies concerned with the assessment of software measurement. The goal of this program is to assist the software industry to improve the information support for managers, analysts and software engineers by providing evidence of where opportunities for improving measurement and analysis exist. Methods. The first study examined the assessment of software measurement frameworks using models of best practice based on performance/success factors. The software measurement frameworks of thirteen organisations were surveyed. The association between a factor and the outcome experienced with the organisations' frameworks was then evaluated. The subsequent studies were more info-centric and investigated using models of information quality to assess the support provided for software processes. For these studies, information quality models targeting specific software processes were developed using practitioner focus groups. The models were instantiated in survey instruments and the responses were analysed to identify opportunities to improve the information support provided. The final study compared the use of two different information quality models for the assessing and improving information support. Assessments of the same quantum of information were made using a targeted model and a generic model. The assessments were then evaluated by an expert panel in order to identify which information quality model was more effective for improvement purposes. Results. The study of performance factors for software measurement frameworks confirmed the association of some factors with success and quantified that association. In particular, it demonstrated the importance of evaluating contextual factors. The conclusion is that factor-based models may be appropriately used for risk analysis and for identifying constraints on measurement performance. Note, however, that a follow-up study showed that some initially successful frameworks subsequently failed. This implied an instability in the dependent variable, success, that could reduce the value of factor-based models for predicting success. The studies of targeted information quality models demonstrated the effectiveness of targeted assessments for identifying improvement opportunities and suggest that they are likely to be more effective for improvement purposes than using generic information quality models. The studies also showed the effectiveness of importance-performance analysis for prioritizing improvement opportunities.
37

Performance metrics for IT projects success within a financial organisation.

Makhubele, K. M. January 2016 (has links)
M. Tech. Business Information Systems / The technological world is emerging and advancing rapidly. Organisations aspire to meet their business goals and strategic objectives. IT projects have the ability to enable organisations meet their business goals and objectives hence organisations invest a lot of time and money on implementing them. The increasing need for organisations to use technology has made organisations implement IT projects. IT projects are implemented to achieve efficiency and effectiveness that is crucial in achieving business goals and strategic objectives. Other benefits include, improved and increased customer satisfaction, improved growth and development within the teams and competitive advantage. Despite an increase in studies on the performance metrics for IT project success in organisations, IT projects still fail. Many IT Projects are considered failed when they are not delivered on time, not delivered within budget, and when the delivered solution does not meet the business requirements. This research study aimed at determining the performance metrics for IT project success within a financial organisation
38

A Tool for Measuring the Size, Structure and Complexity of Software

Versaw, Larry 05 1900 (has links)
The problem addressed by this thesis is the need for a software measurement tool that enforces a uniform measurement algorithm on several programming languages. The introductory chapter discusses the concern for software measurement and provides background for the specific models and metrics that are studied. A multilingual software measurement tool is then introduced, that analyzes programs written in Ada, C, Pascal, or PL/I, and quantifies over thirty different program attributes. Metrics computed by the program include McCabe's measure of cyclomatic complexity and Halstead's software science metrics. Some results and conclusions of preliminary data analysis, using the tool, are also given. The appendices contain exhaustive counting algorithms for obtaining the metrics in each language.
39

[en] AN APPROACH TO EVOLUTION IN SOFTWARE MEASUREMENT REPOSITORIES: THE CLAIRVOYANT SYSTEM / [pt] UMA ABORDAGEM PARA A EVOLUÇÃO TRANSPARENTE EM REPOSITÓRIOS DE MEDIÇÃO DE SOFTWARE: O SISTEMA CLAIRVOYANT

BERNARDO ARRAES VINHOSA 02 October 2007 (has links)
[pt] O sistema Clairvoyant é um protótipo de repositório de medições de software cujo diferencial é oferecer a capacidade de evolução transparente do modelo de medição. Isso quer dizer que as alterações no modelo de medição podem ser feitas sem expor a estrutura interna do armazenamento de medições que torna possível essa evolução. A importância desse tema advém do fato das necessidades de informação que geram as medições em ambientes de engenharia de software estarem em constante mudança. Para que o sistema Clairvoyant comportasse a evolução transparente do modelo de medição, foi necessário dotálo de um meta-modelo de medições e um modelo de consultas adequados a este fim. Na dissertação são explicados esses modelos e a influência deles nos macroprocessos operacionais do repositório de medição (manutenção do modelo de medições, importação de dados, consulta aos dados e exportação de dados). / [en] The Clairvoyant system is a software measurement repository prototype which stands out for allowing transparent evolution in its measurement model. This means that changes can be made to its measurement model without revealing the underlying storage structure which makes possible this evolution. This is an important concern due to the fact that the information needs to which the measurements respond constantly evolve. The Clairvoyant system was designed based on a measurement meta-model and a measurement query model to make it possible to transparently evolve its measurement model. This work explains these models and studies their influence on the repository´s operational macro-processes (measurement model maintenance, measurement data importing, measurement data querying and measurement data exporting).
40

Object-oriented software development effort prediction using design patterns from object interaction analysis

Adekile, Olusegun 15 May 2009 (has links)
Software project management is arguably the most important activity in modern software development projects. In the absence of realistic and objective management, the software development process cannot be managed in an effective way. Software development effort estimation is one of the most challenging and researched problems in project management. With the advent of object-oriented development, there have been studies to transpose some of the existing effort estimation methodologies to the new development paradigm. However, there is not in existence a holistic approach to estimation that allows for the refinement of an initial estimate produced in the requirements gathering phase through to the design phase. A SysML point methodology is proposed that is based on a common, structured and comprehensive modeling language (OMG SysML) that factors in the models that correspond to the primary phases of object-oriented development into producing an effort estimate. This dissertation presents a Function Point-like approach, named Pattern Point, which was conceived to estimate the size of object-oriented products using the design patterns found in object interaction modeling from the late OO analysis phase. In particular, two measures are proposed (PP1 and PP2) that are theoretically validated showing that they satisfy wellknown properties necessary for size measures. An initial empirical validation is performed that is meant to assess the usefulness and effectiveness of the proposed measures in predicting the development effort of object-oriented systems. Moreover, a comparative analysis is carried out; taking into account several other size measures. The experimental results show that the Pattern Point measure can be effectively used during the OOA phase to predict the effort values with a high degree of confidence. The PP2 metric yielded the best results with an aggregate PRED (0.25) = 0.874.

Page generated in 0.0957 seconds