• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 589
  • 119
  • 109
  • 75
  • 40
  • 40
  • 27
  • 22
  • 19
  • 10
  • 7
  • 7
  • 6
  • 6
  • 5
  • Tagged with
  • 1222
  • 1222
  • 179
  • 169
  • 163
  • 156
  • 150
  • 149
  • 148
  • 129
  • 112
  • 110
  • 109
  • 109
  • 107
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
171

Användningsområden för Big data inom analytisk CRM

Nilsson, Per January 2014 (has links)
Customer Relationship Management (CRM) är ett ofta använt konceptför verksamheter att hantera sina kundkontakter. En viktig del av CRMär användningen av tekniska lösningar för att lagra och analysera informationom kunder, till exempel genom data mining för att upptäckamönster hos kunders beteende. Idag produceras allt större mängderdata genom människors användning av informations- och kommunikationsteknik.Traditionell teknik klarar ej av att hantera den variation ochmängd data som existerar, vilket lett till utvecklingen av nya tekniskalösningar för dessa uppgifter. Begreppet Big data brukar användas föratt beskriva stora datamängder. Syftet med denna studie har varit att geökad förståelse för hur Big data kan användas inom CRM. För att uppnådetta har studien undersökt om Big data kan uppfylla behoven för framtidensdata mining inom CRM. Den kvalitativa studien har genomförtsmed en litteraturstudie kring CRM och Big data, och därefter har semistruktureradeintervjuer med svenska IT-konsulter genomförts. Resultatenantyder att teknik för Big data kan vara en möjlig lösning på debehov kring framtidens data mining som identifierats, bland dessa möjlighetenatt använda ett utökat antal datakällor och hantera stora datamängder.Resultaten antyder även att det finns problemområden sommåste beaktas. Ett problem med Big data och användningen av externdata är den osäkerhet kring informationens tillförlitlighet som finns. Detförekommer också diskussion kring den personliga integriteten och vilkaetiska problem som hantering av personlig data kan medföra.
172

Estructura espacial urbana de movilidad desde datos masivos de transporte público en Santiago de Chile

Hernández Godoy, Felipe Andrés January 2017 (has links)
Magíster en Ciencias, Mención Computación. Ingeniero Civil en Computación / La estructura espacial urbana se refiere a la disposición del espacio en la ciudad producto de su forma actual junto con las relaciones subyacentes entre estas. Estas interacciones son generadas por el movimiento de personas, mercancías o información entre un origen y destino, y enmarcadas en un concepto de ciudad entendido como una colección de componentes interrelacionados donde se destacan: el sistema de actividades, sistema de transporte y las relaciones que se generan entre ellos. En este trabajo se intenta caracterizar la estructura espacial para Santiago de Chile mediante tres indicadores: centros (zonas de la ciudad con la capacidad de atraer o concentrar personas), centros de pasada (zonas de la ciudad que sirven para conectar un par de zonas, funcionan como puentes espaciales) y comunidades (zonas de la ciudad que presentan un fuerte nivel de interacción interno). La metodología utiliza análisis de redes y análisis espacial sobre datos de tarjetas inteligentes de pago de transporte público. Los datos son obtenidos a partir de las validaciones hechas con tarjetas inteligentes generadas en parte del sistema de transporte público de la ciudad de Santiago (Transantiago) entre el 14 al 21 de abril de 2013 (una semana). Con las validaciones y un proceso de estimación de bajada es posible estimar etapas y destino en cerca del 80% de los viajes. Con el origen, etapas y destino de cada viaje se construye una red dirigida con pesos en donde se relaciona cada indicador con una métrica de la red. Los centros son asociados al PageRank, los centros de pasada al Betweenness y las comunidades a Infomap. Los resultados muestran que la ciudad continúa manteniendo su estructura con un distrito comercial central (CBD por sus siglas en inglés) en el sector centro-oriente, los centros de pasada son fuertemente influenciados por la red de metro y las comunidades presentan una fuerte unión espacial con una expresión de todas ellas en el CBD, mostrando que el centro de la ciudad es un territorio que les pertenece a todos. Lo anterior muestra una estructura de la ciudad más cercana al monocentrismo al compararla con Singapur.
173

Parallelization of Push-based System for Molecular Simulation Data Analysis with GPU

Akhmedov, Iliiazbek 19 October 2016 (has links)
Modern simulation systems generate big amount of data, which consequently has to be analyzed in a timely fashion. Traditional database management systems follow principle of pulling the needed data, processing it, and then returning the results. This approach is then optimized by means of caching, storing in different structures, or doing some sacrifices on precision of the results to make it faster. When it comes to the point of doing various queries that require analysis of the whole data, this design has the following disadvantages: considerable overhead on traditional disk random I/O framework while reading from the simulation output files and low throughput of the data that consequently results in long latency, and, if there was any indexing to optimize selections, overhead of storing those becomes too big, too. Beside it, indexing will also cause delay during write operations and since most of the queries work with the entire data sets, indexing loses its point. There is a new approach to this problem – Push-based System for Molecular Simulation Data Analysis for processing network of queries proposed in the previous paper and its primary steps are: i) it uses traditional scan-based I/O framework to load the data from files to the main memory and then ii) the data is pushed through a network of queries which consequently filters the data and collect all the needed information which increases efficiency and data throughput. It has a considerable advantage in analysis of molecular simulation data, because it normally involves all the data sets to be processed by the queries. In this paper, we propose improved version of Push-based System for Molecular Simulation Data Analysis. Its major difference with the previous design is usage of GPU for the actual processing part of the data flow. Using the same scan-based I/O framework the data is pushed through the network of queries which are processed by GPU, and due to the nature of science simulation data, this gives a big advantage for processing it faster and easier (it will be explained more in later sections). In the old approach there were some custom data structures such as quad-tree for calculation of histograms to make the processing faster and those involved loss of data and some expectations from the data nature, too. In the new approach due to high performance of GPU processing and its nature, custom data structures were not even needed much, though it didn’t bear any loss in precision and performance.
174

Scaling Big Data Cleansing

Khayyat, Zuhair 31 July 2017 (has links)
Data cleansing approaches have usually focused on detecting and fixing errors with little attention to big data scaling. This presents a serious impediment since identify- ing and repairing dirty data often involves processing huge input datasets, handling sophisticated error discovery approaches and managing huge arbitrary errors. With large datasets, error detection becomes overly expensive and complicated especially when considering user-defined functions. Furthermore, a distinctive algorithm is de- sired to optimize inequality joins in sophisticated error discovery rather than na ̈ıvely parallelizing them. Also, when repairing large errors, their skewed distribution may obstruct effective error repairs. In this dissertation, I present solutions to overcome the above three problems in scaling data cleansing. First, I present BigDansing as a general system to tackle efficiency, scalability, and ease-of-use issues in data cleansing for Big Data. It automatically parallelizes the user’s code on top of general-purpose distributed platforms. Its programming inter- face allows users to express data quality rules independently from the requirements of parallel and distributed environments. Without sacrificing their quality, BigDans- ing also enables parallel execution of serial repair algorithms by exploiting the graph representation of discovered errors. The experimental results show that BigDansing outperforms existing baselines up to more than two orders of magnitude. Although BigDansing scales cleansing jobs, it still lacks the ability to handle sophisticated error discovery requiring inequality joins. Therefore, I developed IEJoin as an algorithm for fast inequality joins. It is based on sorted arrays and space efficient bit-arrays to reduce the problem’s search space. By comparing IEJoin against well- known optimizations, I show that it is more scalable, and several orders of magnitude faster. BigDansing depends on vertex-centric graph systems, i.e., Pregel, to efficiently store and process discovered errors. Although Pregel scales general-purpose graph computations, it is not able to handle skewed workloads efficiently. Therefore, I introduce Mizan, a Pregel system that balances the workload transparently during runtime to adapt for changes in computing needs. Mizan is general; it does not assume any a priori knowledge of the graph structure or the algorithm behavior. Through extensive evaluations, I show that Mizan provides up to 84% improvement over techniques leveraging static graph pre-partitioning.
175

Interpretable and Scalable Bayesian Models for Advertising and Text

Bischof, Jonathan Michael 04 June 2016 (has links)
In the era of "big data", scalable statistical inference is necessary to learn from new and growing sources of quantitative information. However, many commercial and scientific applications also require models to be interpretable to end users in order to generate actionable insights about quantities of interest. We present three case studies of Bayesian hierarchical models that improve the interpretability of existing models while also maintaining or improving the efficiency of inference. The first paper is an application to online advertising that presents an augmented regression model interpretable in terms of the amount of revenue a customer is expected to generate over his or her entire relationship with the company---even if complete histories are never observed. The resulting Poisson Process Regression employs a marginal inference strategy that avoids specifying customer-level latent variables used in previous work that complicate inference and interpretability. The second and third papers are applications to the analysis of text data that propose improved summaries of topic components discovered by these mixture models. While the current practice is to summarize topics in terms of their most frequent words, we show significantly greater interpretability in online experiments with human evaluators by using words that are also relatively exclusive to the topic of interest. In the process we develop a new class of topic models that directly regularize the differential usage of words across topics in order to produce stable estimates of the combined frequency-exclusivity metric as well as proposing efficient and parallelizable MCMC inference strategies. / Statistics
176

Využití dat ze sociálních sítí pro BI / The utilisation of social network data in BI

Linhart, Ondřej January 2014 (has links)
The thesis deals with the topic of social networks, particularly with the opportunities the utilisation of social network data can provide to an enterprise. The thesis is divided into two parts: The theoretical part contains definitions of the terms of data, information and knowledge, followed by descriptions of Business Intelligence and Big Data -- the two means of data analysis in an enterprise, and later by describing social networks themselves. The practical part contains an analysis of the data provided by social networks Facebook and Twitter, and at the same time defines the process of data extraction. The outcome of the analysis is a set of data that may possibly be obtained by the enterprise. This data is then used to determine the possible ways in which enterprises can leverage the data for their business. Finally data provided by Czech e shop is used to provide an example of how an entity can utilise social network data.
177

Porovnanie metód machine learningu pre analýzu kreditného rizika / Comparison of machine learning methods for credit risk analysis

Bušo, Bohumír January 2015 (has links)
Recently, machine learning has been put into connection with a field called ,,Big Data'' more and more. Usually, in this field, a lot of data is available and we need to gather useful information based on this data. Nowadays, when still more and more data is generated by use of mobile phones, credit cards, etc., a need for high-performance methods is serious. In this work, we describe six different methods that serve this purpose. These are logistic regression, neural networks and deep neural networks, bagging, boosting and stacking. Last three methods compose a group called Ensemble Learning. We apply all six methods on real data, which were generously provided by one of the loan providers. These methods can help them to distinguish between good and bad potential takers of loans, when the decision about the loan is being made. Lastly, the results of particular methods are compared and we also briefly outline possible ways of interpretation.
178

Vliv vývojových trendů na řešení projektu BI / The influence of trends in BI project

Kapitán, Lukáš January 2012 (has links)
The aim of this these is to analyse the trends occurring in Business intelligence. It does examine, summarise and judge each of the trends from the point of their usability in the real world, their influence and modification of each phase of the implementation of Bussiness intelligence. It is clear that each of these trends has its positives and negatives which can influence the statements in the evaluation. These factors are taken into consideration and analysed as well. The advantages and disadvantages of the trends are occurring especially in the areas of economical demand and technical difficultness. The main aim is to compare the methods of implementation of Bussiness intelligence with actual trends in BI. In order to achieve this a few crucial points were set: to investigate recent trends in the BI and to define the methods of implementation in the broadest terms. The awaited benefit of this these is already mentioned investigation and analysis of trends in the area of Bussiness intelligence and its use in implementation.
179

Advanced Analytics in Retail Banking in the Czech Republic / Prediktívna analytika v retailovom bankovníctve v Českej republike

Búza, Ján January 2014 (has links)
Advanced analytics and big data allow a more complete picture of customers' preferences and demands. Through this deeper understanding, organizations of all types are finding new ways to engage with existing or potential customers. Research shows that companies using big data and advanced analytics in their operations have productivity and profitability rates that are 5 to 6 percent higher compared to their peers. At the same time it is almost impossible to find a banking institution in the Czech Republic exploiting potential of data analytics to its full extent. This thesis will therefore focus on exploring opportunities for banks applicable in the local context, taking into account technological and financial limitations as well as the market situation. Author will conduct interviews with bank managers and management consultants familiar with the topic in order to evaluate theoretical concepts and the best practices from around the world from the point of Czech market environment, to assess capability of local banks to exploit them and identify the main obstacles that stand in the way. Based on that a general framework for bank managers, who would like to use advanced analytics, will be proposed.
180

Implicações do fenômeno big data na análise para inteligência estratégica

Nesello, Priscila 10 April 2014 (has links)
Uma grande quantidade de dados é produzida diariamente por operações comerciais e financeiras, mídias sociais e dispositivos móveis, sensores e outros equipamentos inseridos no mundo físico. Este fenômeno deu origem ao termo big data, e seus efeitos podem ser percebidos por empresas, ciência e governo. Entretanto, é a inteligência estratégica, não a informação, que auxilia gerentes a extrair valor dos grandes volumes de dados. Para isto, é necessário transformar a informação dispersa no ambiente em conhecimento estruturado e útil à tomada de decisão nas organizações. Este é um processo complexo, pois apesar das ferramentas e técnicas disponíveis é indispensável que o profissional em inteligência saiba lidar com a complexidade cognitiva inerente ao processo de análise. Neste contexto, o objetivo deste trabalho foi o de examinar como o fenômeno big data afeta o processo de análise na atividade de inteligência estratégica. A pesquisa tratou de como o fenômeno big data é percebido pelos entrevistados em suas atividades analíticas em inteligência estratégica e propôs uma análise acerca de suas implicações. Para isso, foi realizado um estudo exploratório qualitativo. Foram entrevistados profissionais brasileiros, residentes nos Estados do Rio Grande do Sul, Rio de Janeiro, Distrito Federal e São Paulo. Estes entrevistados foram selecionados por meio de agentes com atuação, conhecimento e trânsito nos campos de inteligência estratégica e/ou big data. O roteiro que serviu de base para estas entrevistas foi estruturado a partir das dimensões do fenômeno big data e seus efeitos nas atividades analíticas no processo de inteligência estratégica. A técnica utilizada para análise dos dados foi análise de conteúdo. Os resultados indicam que o volume de big data contribui para a compreensão dos métodos de coleta, entretanto prejudica o domínio da matéria. Outras descobertas revelam que para alguns entrevistados big data já integra as práticas profissionais na realização de análises mais elaboradas e no desenvolvimento de projetos específicos. Entretanto para outros, big data ainda não é uma realidade, não sendo percebida a necessidade de utilizar grandes volumes de dados nas análises. Este fato também denota um paradoxo entre a caracterização da produção de conhecimento no campo do big data e o seu uso no campo profissional de inteligência estratégica. Por um lado o maior volume de trabalhos sobre o big data está no campo profissional das organizações produtivas e não na academia e por outro os profissionais em inteligência ainda não percebem o valor do fenômeno para sua atuação profissional. / A considerable amount of data is daily produced by business and financial operations, social media, mobile devices, sensors as well as other gadgets available in the world. This phenomenon gave rise to the big data term whose effects can be perceived by companies, science and governments. However, the strategic intelligence, not the information itself, supports managers eliciting values from big volume of data. For this purpose, transforming the dispersed information in the environment into structure knowledge is necessary and useful for organizations’ decisions. This is a complex process, and despite the tools and available techniques, it is indispensable that the professional in intelligence knows how to deal with inherit cognitive complexity in the courses of analyses. In this context, the objective of the present work was examining how the phenomenon big data affected the course of analyses in the strategic intelligence activity. The research was addressed on how the phenomenon big data was perceived by the interviewers in their analytics activities in strategic intelligence. It also proposed analysis based on its implications. In order to achieve this, a qualitative exploratory study was conducted. Several Brazilian professionals were interviewed, including residents in states such as Rio Grande do Sul, Rio de Janeiro, Distrito Federal and São Paulo. Those interviewed were chosen by agents with experience and knowledge in the strategic intelligence field and/or big data. The guide used for the interviews was structured from dimensions of big data phenomenon and its effects on the course of analyses in the strategic intelligence activity. The technique used for analyzing data was through content review. Results indicate that the volume of big data contributes to the comprehension of collection methods even though it eventually might debilitate the ability to grasp the topic. Other discoveries show that for some of those interviewed, big data has already integrated professional practices on not only performing a more detailed analyses but also developing specific projects. Nevertheless, big data is not a reality yet for others since the necessity of utilizing big volume of data for analysis is not really being perceived. This fact also denotes a paradox between the characterization of production knowledge within big data field and its use in the professional area of strategic intelligence. On the one hand, the great workload about big data is located in the professional area of productive organizations. Not in the academy, though. After all, intelligence professionals have not realized yet the real value of big data phenomenon for their professional performance.

Page generated in 0.0881 seconds