• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 169
  • 144
  • 49
  • 21
  • 14
  • 8
  • 7
  • 5
  • 5
  • 4
  • 4
  • 3
  • 3
  • 3
  • 2
  • Tagged with
  • 472
  • 472
  • 144
  • 139
  • 130
  • 88
  • 86
  • 82
  • 68
  • 64
  • 63
  • 59
  • 54
  • 52
  • 52
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
231

Testing the Internet state management mechanism

Tappenden, Andrew 06 1900 (has links)
This thesis presents an extensive survey of 100,000 websites as the basis for understanding the deployment of cookies across the Internet. The survey indicates cookie deployment on the Internet is approaching universal levels. The survey identifies the presence of P3P policies and dynamic web technologies as major predictors of cookie usage, and a number of significant relationships are established between the origin of the web application and cookie deployment. Large associations are identified between third-party persistent cookie usage and a countrys e-business environment. Cookie collection testing (CCT), a strategy for testing web applications, is presented. Cookies maintained in a browser are explored in light of anti random testing techniques, culminating in the definition of seeding vectors as the basis for a scalable test suite. Essentially CCT seeks to verify web application robustness against the modificationintentional or otherwiseof an application's internal state variables. Automation of CCT is outlined through the definition of test oracles and evaluation criterion. Evolutionary adaptive random (eAR) testing is proposed for application to the cookie collection testing strategy. A simulation study is undertaken to evaluate eAR against the current state-of-the-art in adaptive random testingfixed size candidate set, restricted random testing, quasi-random testing, and random testing. eAR is demonstrated to be superior to the other techniques for block pattern simulations. For fault patterns of increased complexity, eAR is shown to be comparable to the other methods. An empirical investigation of CCT is undertaken. CCT is demonstrated to reveal defects within web applications, and is found to have a substantial fault-triggering rate. Furthermore, CCT is demonstrated to interact with the underlying application, not just the technological platform upon which an application is implemented. Both seeding and generated vectors are found to be useful in triggering defects. A synergetic relationship is found to exist between the seeding and generated vectors with respect to distinct fault detection. Finally, a large significant relationship is established between structural and content similarity measures of web application responses, with a composite of the two similarity measures observed to be superior in the detection of faults. / Software Engineering and Intelligent Systems
232

Introduction to a mobile development methods investigation to Scania IT AB : Case: Rundtursbuss application

Moreira, Juliana January 2012 (has links)
The purpose of this work was both to explore the variety of methods to develop mobile applications and to create a requirement documentation and solution proposal to the case study Rundtursbuss. During the 10 weeks work-experience, I had the possibility to investigate a range of methods to develop mobile applications and apply it to a reference case. The questions regarding which method to choose when developing mobile application are not easy to answer, once there is not only one solution that can fit the needs of all the applications that will be developed at Scania. Each application needs to be evaluated not only from a technical perspective, but also from a user-centered way. One of the best ways to quickly evaluate a project is asking what that makes the project mobile The analysis of each criteria aided in the evaluation of the application requirements and in reaching the conclusion that the most promising methods to develop the application Rundtursbuss are either native or cross-compilation ones. In conclusion, it is crucial to emphasize that while Scania does not have a procedure on how to develop, maintain and coordinate the development of mobile applications, single initiatives will continue to emerge, which increase the cost and affect the quality of their IT products and services.
233

A Web-based Database For Experimental Structural Engineering Research

Turer, Oncel Umut 01 May 2010 (has links) (PDF)
This study presents a web-based database application for storing experimental data and related documents at a single location and sharing them among researchers and engineers from all over the world. The database application, accessible from a website was developed for experimental structural engineering researches, and initially tested at Structural and Earthquake Engineering Laboratory of Civil Engineering Department, METU. The application is composed of two parts. The first part is the database that stores information about projects, specimens, experiments, experimental data, documentation, site members, and member groups at the server side. The second part is the website that provides a functional user interface for easy use of application and providing accessibility from everywhere via internet. After the development of the database and the website, these two parts were attached to each other and application functionalities that enabled users to create, modify, search, and delete projects, specimens and experiments / allowed users to upload/download documentation and experimental data / provided abilities to users to plot test data and share their opinions were ensured. ASP.NET framework and C# programming language was utilized for the web application development. Functionality and usability of the database application was then tested by uploading and sharing various experimental results.
234

Žiniatinklio portalų kūrimo ir projektavimo metodika / Web application development methology

Žilinskas, Tadas 25 May 2004 (has links)
The nature of Web system development is significantly different from conventional software development. Good method for system development is needed. Analysis show, that main known system development methods are good at describing static web application aspects, but does not provide a good description for modeling web application behavioral characteristics. This work uses model driven approach as one, suitable for dealing with web application development problems. The key aspect of model driven approach is specifying system domain model independent of any particular technology (J2EE, Microsoft .NET, etc.) and then generating platform specific model or its implementation. The UML is user-friendly, easy to use and is useful for describing system effectively. Specification expressed in terms of UML can be rendered into an XML document using the OMG’s XMI DTD for UML, which makes UML suitable for model driven approach implementation. In this work, method for specifying application behavior design using UML was presented. This method gives rules for specifying web application behavior model, suitable for its implementation code generation. The main components used in today’s web application which implementation could be generated from the model were also described. It was shown, that it is possible to generate much of implementation, just using such a model representation in XMI. The main idea presented in this work is that web application development should be model driven as... [to full text]
235

Apsaugos nuo SQL injekcijų el.verslo svetainėse metodikos sudarymas ir tyrimas / Development and research of method of protection against SQL injections in e-commerce websites

Ramoška, Aidas 04 November 2013 (has links)
SQL injekcijos atakos taikinys – interaktyvios interneto programos, kurios naudoja duomenų bazės serverius. Šios programos leidžia vartotojams įvesti informaciją ir ją įvedus formuojamos SQL užklausos, kurios siunčiamos į duomenų bazės serverį. Darydamas SQL injekcijos ataką, atakuotojas per įvesties laukus suformuoja kenksmingą SQL užklausos segmentą, kuris modifikuoja buvusią užklausą. Naudodamas SQL injekcijos ataką, atakuotojas gali prieiti prie konfidencialios informacijos, ją modifikuoti ar, apeidamas autorizacijos scenarijų, prisijungti prie sistemos nežinodamas slaptažodžio. Šiame darbe pasiūlytas saugos modulis perima visą vartotojo įvedamą informaciją, pritaiko saugumo taisykles ir taip padidina saugumą apsisaugant nuo SQL injekcijų el. verslo žiniatinklio programose bei registruoja potencialius bandymus sutrikdyti normalų sistemos darbą. Norint įdiegti pasiūlytą saugos modulį, nereikia konfigūruoti serverio ar jo programinės įrangos – modulio diegimo metu keičiasi tik žiniatinklio programos failai. Darbui atlikti pasirinkta PHP programavimo kalba ir MySQL duomenų bazė. Tyrimo metu atlikti testavimo rezultatai parodo, kokius saugos modulio konfigūravimo parametrus reikia taikyti norint užtikrinti maksimalų saugumo lygį. / The target of SQL injection attack – interactive web programs, which use database servers. Those programs allow users to input information and as it is imputed, it forms SQL queries, which are sent into database server. With SQL injection help, the attacker using input fields forms harmful section of SQL query, which modifies previous query. Exploiting attack of SQL injection, the attacker may learn confidential information, modify it or connect to system without knowing the password by authorisation bypass. In this research-paper the proposed security model takes over all information inputted by user, adjusts the safety rules and that way it improves the safety in order to guard from SQL injections at electronic business web systems as well as it register potential attempts to disrupt normal work of the system. In order to install the proposed safety model there is no need to configure the server or its software because in the moment of installation it changes only files of website programs. For purpose of executing this work, we use PHP programming language and MySQL database. During the analysis, the received test results show what configuration parameters of safety model we need to use in order to guarantee the maximum level of safety.
236

Testing the Internet state management mechanism

Tappenden, Andrew Unknown Date
No description available.
237

Ställ en fråga, få ett svar : En vidareutveckling av uDecide’s mobilapplikation / Ask a question, get an answer : Development of uDecide’s mobile phone application

Olsson von Koch, Ebba, Juskova, Aleksandra January 2014 (has links)
Denna rapport presenterar en utveckling av uDecide, som är ett sorts hjälpmedel för att lättare kunna ta beslut. Ett verktyg som man kan använda för att ställa frågor till andra människor med hjälp av en bild och svarsalternativ för att få ett personligt svar eller åsikt. Syftet med detta projekt är att utveckla en användbar hi-fi prototyp som i sin tur ska vidareutvecklas av webbyrån Roxbury och uDecidegrundarna. Arbetet utgår från en befintlig uDecide iOS mobilapplikation, och fokus har lagts på att förbättra användarupplevelsen och implementera nya funktioner. Med hjälp av förstudier och användartester har vi utvecklat en webbaserad, responsiv hi fi-prototyp. Lösningen presenterar ny utveckling av navigering, struktur på inlägg och ny meny. Ett kommentarsfält har implementerats i hi-fi prototypen vilket kommer att ge användarna möjlighet att interagera med varandra och ge mer personliga åsikter. / This report describes a development of the mobile application uDecide. uDecide is a tool which you can use to ask other people a question including a picture and alternative answers to get an individual answer or opinion. The objective of this project is to develop an usable hi-fi prototype, which in turn will be further developed by the web designers Roxbury and the founders of uDecide. The development originates from a current uDecide IOS mobile application and focus is to improve user satisfaction and to implement new functions. Supported by pre-development research and user tests we have developed a webbased, responsive and interactive hi-fi prototype. The solution presents a new development in navigation, structure of additions and responses, including a new menu. A commentary function has been implemented in the hi-fi prototype, which will give the users the option to interact and provide more personal opinions.
238

Smart task logging : Prediction of tasks for timesheets with machine learning

Bengtsson, Emil, Mattsson, Emil January 2018 (has links)
Every day most people are using applications and services that are utilising machine learning, in some way, without even knowing it. Some of these applications and services could, for example, be Google’s search engine, Netflix’s recommendations, or Spotify’s music tips. For machine learning to work it needs data, and often a large amount of it. Roughly 2,5 quintillion bytes of data are created every day in the modern information society. This huge amount of data can be utilised to make applications and systems smarter and automated. Time logging systems today are usually not smart since users of these systems still must enter data manually. This bachelor thesis will explore the possibility of applying machine learning to task logging systems, to make it smarter and automated. The machine learning algorithm that is used to predict the user’s task, is called multiclass logistic regression, which is categorical. When a small amount of training data was used in the machine learning process the predictions of a task had a success rate of about 91%.
239

Live updates in High-availability (HA) clouds

Sanagari, Vivek January 2018 (has links)
Background. High-availability (HA) is a cloud’s ability to keep functioning after one or more hardware or software components fail. Its purpose is to minimize the system downtime and data loss. Many service providers guarantee a Service Level Agreement including uptime percentage of the computing service, which is calculated based on the available time and system downtime excluding the planned outage time. The aim of the thesis is to perform the update of the virtual machines running in the cloud without causing any interruptions to the user by redirecting the resources/services running on them to an alternative virtual machine before the original VM is updated. Objectives. The objectives for the above aim include. • The first objective is to investigate existing solutions for high-availability and, if possible, adapt them to our aim. The alternative is to design our own solution. • The second objective is to implement the solution in an Open Stack environment. As an alternative, we can try a smaller scale implementation under a virtualization platform such as Virtual Box. • The final objective is to run experiments to quantify the effectiveness of our solution in terms of overhead and degree of seamlessness to the users. Methods. An environment with multiple virtual machines may be created to represent multiple virtual servers in the cloud. The state of service provided by the primary virtual machine is saved to persistent storage and the client is redirected to an alternate virtual machine. At that point the primary virtual machine may reboot for an update or any other issues. Results. In the case of CPU Utilization, the mean CPU utilization on Server and Host in scenario 1 are 0.34% and 3.2% respectively. The mean CPU utilization on Primary server and Host in scenario 2 during the failover cycle are 2.0% and 9.7% respectively. The mean CPU utilization on Secondary server and Host in scenario 2 during failover cycle are 0.99% and 8.0% respectively. For the Memory Utilization, the mean Memory usage on server in scenario 1 is 16%. The mean Memory usage on primary server and secondary server in scenario 2 during failover cycle are 37% and 48% respectively. The Time for failover of the high availability environment remains for 6.8 seconds and the time for the off-line node to rejoin the cluster as on-line when told would take 1.5 seconds. The network traffic is measured in Kilobits per second, it is 1.2 Kilobits per second on port 80 in scenario 2 and is 1.4 Kilobits per second between the client and the server in scenario 1. In addition, data traffic on ports 5405, 2224 and 7788 are captured where port 5405 (Pacemaker/Corosync) contains UDP traffic, port 2224 (Pcsd) contains TCP traffic and port 7788 (DRBD) contains TCP traffic. The traffic captured on these ports represent network overhead due to HA. During failover cycle an additional traffic of 45Kb/s, 1.2Kb/s. 7.0Kb/s flow on 5405, 2224 and 7788 ports respectively. Conclusions. From our experiment results we can say that the overhead to handle live updates on high availability environment is approximately 1.1 - 1.7 % of CPU higher in HA mode than when a stand-alone server is used. The overhead is around 21 - 32 % higher in terms of memory utilization for the live updates on the HA system than for the standard server. The network traffic overhead induced by the ports used by high availability environment (5405, 2224, 7788) is approximately 53 Kilobits /Second while the minimum overhead is approximately 16 Kilobits / Second. The Final and the important metric is the Failover time which tells the seamlessness of the service as the environment needs to provide the services uninterrupted to the users. The failover time of the HA model is about just 6.8 seconds leaving the environment highly available. However, the user may notice slight interruption for the requests made during this span.
240

Understanding and automating application-level caching / Entendendo e automatizando cache a nível de aplicação

Mertz, Jhonny Marcos Acordi January 2017 (has links)
O custo de serviços na Internet tem encorajado o uso de cache a nível de aplicação para suprir as demandas dos usuários e melhorar a escalabilidade e disponibilidade de aplicações. Cache a nível de aplicação, onde desenvolvedores manualmente controlam o conteúdo cacheado, tem sido adotada quando soluções tradicionais de cache não são capazes de atender aos requisitos de desempenho desejados. Apesar de sua crescente popularidade, este tipo de cache é tipicamente endereçado de maneira ad-hoc, uma vez que depende de detalhes específicos da aplicação para ser desenvolvida. Dessa forma, tal cache consiste em uma tarefa que requer tempo e esforço, além de ser altamente suscetível a erros. Esta dissertação avança o trabalho relacionado a cache a nível de aplicação provendo uma compreensão de seu estado de prática e automatizando a identificação de conteúdo cacheável, fornecendo assim suporte substancial aos desenvolvedores para o projeto, implementação e manutenção de soluções de caching. Mais especificamente, este trabalho apresenta três contribuições: a estruturação de conhecimento sobre caching derivado de um estudo qualitativo, um levantamento do estado da arte em abordagens de cache estáticas e adaptativas, e uma técnica que automatiza a difícil tarefa de identificar oportunidades de cache O estudo qualitativo, que envolveu a investigação de dez aplicações web (código aberto e comercial) com características diferentes, permitiu-nos determinar o estado de prática de cache a nível de aplicação, juntamente com orientações práticas aos desenvolvedores na forma de padrões e diretrizes. Com base nesses padrões e diretrizes derivados, também propomos uma abordagem para automatizar a identificação de métodos cacheáveis, que é geralmente realizado manualmente por desenvolvedores. Tal abordagem foi implementada como um framework, que pode ser integrado em aplicações web para identificar automaticamente oportunidades de cache em tempo de execução, com base na monitoração da execução do sistema e gerenciamento adaptativo das decisões de cache. Nós avaliamos a abordagem empiricamente com três aplicações web de código aberto, e os resultados indicam que a abordagem é capaz de identificar oportunidades de cache adequadas, melhorando o desempenho das aplicações em até 12,16%. / Latency and cost of Internet-based services are encouraging the use of application-level caching to continue satisfying users’ demands, and improve the scalability and availability of origin servers. Application-level caching, in which developers manually control cached content, has been adopted when traditional forms of caching are insufficient to meet such requirements. Despite its popularity, this level of caching is typically addressed in an adhoc way, given that it depends on specific details of the application. Furthermore, it forces application developers to reason about a crosscutting concern, which is unrelated to the application business logic. As a result, application-level caching is a time-consuming and error-prone task, becoming a common source of bugs. This dissertation advances work on application-level caching by providing an understanding of its state-of-practice and automating the decision regarding cacheable content, thus providing developers with substantial support to design, implement and maintain application-level caching solutions. More specifically, we provide three key contributions: structured knowledge derived from a qualitative study, a survey of the state-of-the-art on static and adaptive caching approaches, and a technique and framework that automate the challenging task of identifying cache opportunities The qualitative study, which involved the investigation of ten web applications (open-source and commercial) with different characteristics, allowed us to determine the state-of-practice of application-level caching, along with practical guidance to developers as patterns and guidelines to be followed. Based on such patterns and guidelines derived, we also propose an approach to automate the identification of cacheable methods, which is often manually done and is not supported by existing approaches to implement application-level caching. We implemented a caching framework that can be seamlessly integrated into web applications to automatically identify and cache opportunities at runtime, by monitoring system execution and adaptively managing caching decisions. We evaluated our approach empirically with three open-source web applications, and results indicate that we can identify adequate caching opportunities by improving application throughput up to 12.16%. Furthermore, our approach can prevent code tangling and raise the abstraction level of caching.

Page generated in 0.3579 seconds