Spelling suggestions: "subject:"ddd"" "subject:"3dd""
151 |
Competitive Intelligence / Competitive IntelligenceMikuš, Ondřej January 2011 (has links)
The diploma thesis focuses on the practical use of competitive intelligence, a method increasingly used to support decision making nowadays. The essential key principle is based on collection of precise data and their thourough analysis. The obtained information gives a comprehensive overview of the competitor analysis. The acquired knowledge of the competition is important for company's strategic decision-making processes. At the beginning the thesis gives a theoretical background on different means of data collecting including the ethic kodex. Folowing is the identification of the modern sources of information. Further, various methods and approaches of competitive intelligence are described. And the theoretical background on competitive strategies and their impact on the market wraps up the theoretical part of the thesis. The method is applied to create departmet of competitive inteligence for the company. Then the various options for further go to market activities are described and the scope of cooperation with selected customers is evaluted. The thesis describes the competitive intelligence advantages and the possible ways of aplications for the competitive strategy. At the end, the recommendations for the management of the company are summed up.
|
152 |
Segmentovaná diskrétní waveletová transformace / Segmentwise Discrete Wavelet TransformPrůša, Zdeněk January 2012 (has links)
Dizertační práce se zabývá algoritmy SegDWT pro segmentový výpočet Diskrétní Waveletové Transformace – DWT jedno i vícedimenzionálních dat. Segmentovým výpočtem se rozumí způsob výpočtu waveletové analýzy a syntézy po nezávislých segmentech (blocích) s určitým překryvem tak, že nevznikají blokové artefakty. Analyzující část algoritmu pracuje na principu odstranění přesahu a produkuje vždy část waveletových koeficientů z waveletové transformace celého signálu, které mohou být následně libovolně zpracovány a podrobeny zpětné transformaci. Rekonstruované segmenty jsou pak skládány podle principu přičtení přesahu. Algoritmus SegDWT, ze kterého tato práce vychází, není v současné podobně přímo použitelný pro vícerozměrné signály. Tato práce obsahuje několik jeho modifikací a následné zobecnění pro vícerozměrné signály pomocí principu separability. Kromě toho je v práci představen algoritmus SegLWT, který myšlenku SegDWT přenáší na výpočet waveletové transformace pomocí nekauzálních struktur filtrů typu lifting.
|
153 |
The Association between Childhood Maltreatment, Substance Use Frequency, and Physical Intimate Partner Violence: A Gene-Environment StudyAura Ankita Mishra (8905460) 15 June 2020 (has links)
<p>This
dissertation evaluated the complex inter-relatedness between co-occurring
childhood maltreatment exposures, physical intimate partner violence
(perpetration and victimization), substance use frequency, and molecular genetics
for substance use, utilizing appropriate developmental models and theoretical
approaches. Three studies were proposed within this dissertation. Data for the
three studies come from a national longitudinal panel study: The National Longitudinal Study
of Adolescent to Adult Health (Add Health; Harris,
2013).
Across studies, latent profile analysis was used to evaluate co-occurring
childhood maltreatment exposures based on type and severity of exposures, which
resulted in three homogenous sub-groups. The first sub-group was composed of
individuals that had high levels of physical abuse exposure and moderate levels
of childhood neglect and emotional abuse exposures (high physical abuse
sub-group). The second sub-group (high sexual abuse sub-group) included
individuals with high severity of sexual abuse exposure and moderate severity
of all other childhood maltreatment types (i.e., physical abuse, emotional
abuse, and neglect). This second sub-group was, therefore, the most vulnerable
in terms of their childhood maltreatment exposure. A final normative sub-group
was also found that included a majority of individuals with low severity of
childhood maltreatment exposure across types. Additionally, across all three
studies, a probabilistic multifaceted genetic risk score (i.e., polygenic risk
score) was created to evaluate substance use related genetic risk. The first
study evaluated the role of co-occurring childhood maltreatment exposure on
substance use development from adolescence to young adulthood while evaluating substance
use related genetic moderation. Generalized estimating equations were used to
test the proposed model in study 1. Findings suggest that the high physical
abuse sub-group was more susceptible to genetic risk and had increases in
substance use frequency only at high levels of genetic risk. In contrast, for
the high sexual abuse sub-group, childhood maltreatment and environmental
exposures were more ubiquitous for substance use development from adolescence
to young adulthood. To elaborate, the high sexual abuse sub-group demonstrated
increases in substance use from adolescence to young adulthood irrespective of
genetic risk. In study 2, substance use frequency in young adulthood was tested
as a mechanism between childhood maltreatment sub-groups and subsequent
physical intimate partner violence perpetration in adulthood. Once again,
genetic moderation for the direct association between childhood maltreatment
sub-groups and substance use frequency in young adulthood was tested within the
larger mediation model. In study 3, physical partner violence victimization in
young adulthood was tested as a mediator of the association between childhood
maltreatment sub-groups and substance use frequency in adulthood. In study 3,
in addition to the above-mentioned genetic risk score, an additional substance
use related dopamine polygenic risk score was also tested. Specifically, in
study 3, genetic moderation by both genetic risk scores was tested on 1) the
direct pathway from childhood maltreatment sub-groups to substance use
frequency in adulthood, and 2) the direct pathway from physical intimate
partner violence victimization in young adulthood to substance use frequency in
adulthood. In both studies 2 and 3, product of co-efficient method was used to
estimate mediation hypothesis, and moderated-mediation models were used to test
for genetic moderation within the mediation model. Research aims for studies 2
and 3 were largely not supported. However, supplementary models indicate that substance
use frequency may not be a causal mechanism but may be a contextual factor
exacerbating the association between childhood maltreatment exposures and
physical intimate partner violence perpetration. Implications for findings are
discussed in detail. </p>
|
154 |
Phenomenological studies of dimensional deconstructionHällgren, Tomas January 2005 (has links)
In this thesis, two applications of dimensional deconstruction are studied. The first application is a model for neutrino oscillations in the presence of a large decon- structed extra dimension. In the second application, Kaluza{Klein dark matter from a latticized universal extra dimension is studied. The goal of these projects have been twofold. First, to see whether it is possible to reproduce the relevant features of the higher-dimensional continuum theory, and second, to examine the effect of the latticization in experiments. In addition, an introduction to the the- ory of dimensional deconstruction as well as to the theory of continuous extra dimensions is given. Furthermore, the various higher-dimensional models, such as Arkani-Hamed{Dvali{Dimopolous (ADD) models and models with universal extra dimensions, that have been intensively studied in recent years, are discussed. / QC 20101202
|
155 |
Reedukace motorických obtíží u dětí mladšího školního věku / Re-education of children with motor difficulties of primary school ageTichý, Pavel January 2016 (has links)
BIBLIOGRAPHIC IDENTIFICATION Name and surname of the author: Mgr. Pavel Tichý Name of the dissertation: Re-education of children with motor difficulties of primary school age Workplace: FTVS UK, Department of Sport Games Supervisor of the work: Doc. PhDr. Vladimír Süss, Ph.D. Year of presentation: 2016 ABSTRACT Objective: The aim was to find out the possibility of the re-education of motor difficulties in children of primary school age and create a re-education intervention program, the aim of which is to reduce these difficulties. Subsequently, to verify whether the intervention program had an effect on the motor difficulties of the children. Method: The research was designed as a prospective intervention study where there were two combined orientations ofeducational research: quantative and qualitative. To solve the problem a research design was used, called Crossover Design (Wilmore et al., 2008, Thomas et al., 2005), which was conceived as an experiment with an intervention re-education program aimed at improving the motor difficulties of children. The quantative part of the research was focused on the assessment of motor skills and to detect motor problems in children of primary school age. MABC-2 (Henderson et al., 2007) battery of motor tests were used for this purpose. The qualitative part of the...
|
156 |
Early Effects of the Tomatis Listening Method in Children with Attention DeficitSacarin, Liliana 26 July 2013 (has links)
No description available.
|
157 |
Attention-Deficit/Hyperactivity Disorder: Women's Accounts of Personal Identity and Social SupportRudd, Melissa Felice 22 December 2022 (has links)
No description available.
|
158 |
Higher Radix Floating-Point Representations for FPGA-Based ArithmeticCatanzaro, Bryan Christopher 22 April 2005 (has links) (PDF)
Field Programmable Gate Arrays (FPGAs) are increasingly being used for high-throughput floating-point computation. It is forecasted that by 2009, FPGAs will provide an order of magnitude greater sustained floating-point throughput than conventional processors. FPGA implementations of floating-point operators have historically been designed to use binary floating-point representations, as do general purpose processors. Binary representations were chosen as the standard over three decades ago because they provide maximal numerical accuracy per bit of floating-point data. However, the unique nature of FPGA-based computation makes numerical accuracy per unit of FPGA resources a more important measure of the usefulness of a given floating-point representation. From this viewpoint, higher radix floating-point representations are well suited to FPGA-based computations, especially high precision calculations which require the support of denormalized numbers. This work shows that higher radix representations lead to more efficient use of FPGA resources. For example, a hexadecimal floating-point adder provides a 30% lower Area-Time product than its binary counterpart, and a hexadecimal floating-point multiplier has a 13% lower Area-Time product than its binary counterpart. This savings occurs while still delivering equal worst-case and better average-case numerical accuracy. This work presents a family of higher radix floating-point representations that are designed specifically to interoperate with standard IEEE floating-point, allowing the creation of floating-point datapaths which operate on standard binary floating-point data, yet use higher radix representations internally. Such datapaths provide higher performance by any measure: they are more accurate numerically, consume less FPGA resources and have shorter latencies. When taking into consideration the unique nature of FPGA-based computing systems, this work shows that binary floating-point representations are not optimal for most FPGA-based arithmetic computations. Higher radix representations can therefore be a useful tool for building efficient custom floating-point datapaths on FPGAs.
|
159 |
Confining Mastery: Understanding the Influence of Parental Incarceration on Mastery in Young AdulthoodShaw-Smith, Unique R. 25 April 2014 (has links)
No description available.
|
160 |
Essays in Industrial Organization and Behavioral EconomicsYermukanova, Binur 16 September 2024 (has links)
In the first chapter of my thesis, I examine a domestic downstream producer that acquires a foreign input supplier with the aim of manipulating the transfer price and shifting profits from a high-tax to a low-tax country. The United States, with a corporate tax rate of 21%, is the leading example of profit shifting behavior, where multinational enterprises like Google, Apple and Facebook shift part of their profits to tax havens like Ireland, the Netherlands, or Switzerland. In our model, the multinational enterprise faces a trade-off: insourcing the input at a high transfer price reduces the corporate tax burden but at the cost of increasing the input cost of its downstream division, which reduces domestic sales and profits. The optimal transfer price balances this trade-off. In an extension that we develop in the model, the regulation under the arm’s length principle is shown to reduce the transfer price, which reduces distortion in the domestic production and expands the region in which vertical FDI benefits consumers and social welfare in the home country. We also show that promoting downstream competition could help align private and social incentives, reinforcing the positive effects of regulation. In the second chapter, I study the effect of consumer learning on shrouding (hiding information about the product price) and pricing strategies of the firms in add-on markets. The presence of consumer myopia in add-on markets often allows firms to exploit consumers who have limited information about the market, even in competitive ones. I formally develop a dynamic theoretical model of add-on markets where consumers learn about the shrouding strategies of firms after they buy and experience expensive add-on products in a previous period. Despite high levels of consumer myopia, I find that firms choose to unshroud the add-on product and sell it at a low price as long as consumer experience is high enough in the market. I also find co-existence of shrouding and unshrouding equilibria with high add-on prices when the level of consumer experience is low enough and the level of consumer myopia is high enough in the market. The results combine the contrasting results in the literature for add-on markets (Gabaix and Laibson (2006) and Dahremoller (2013)), and are robust to an infinite-period extension of the model. In the third chapter, I formally develop a theoretical model of the market for quacks, where market consists of both firms which provide a useless service and some experts. Apart from the naive consumers who sample firms based on anecdotal reasoning, I introduce some consumers who use Bayesian updating based on the distribution of firms in the market (sophistication), and hence can infer some information about firms based on the signals they receive, and study its effect on the behavior of firms in the market. In this setting, I find a Bayesian mixed-strategy Nash equilibrium, in which firms choose to randomize over low prices and sell their services to naive and sophisticated consumers, as long as there is a significant fraction of consumers who update their information based on the distribution of the firms in the market. The higher is the quality of service provided by the expert, the larger is the region in which this Nash equilibrium prevails. In contrast, if the fraction of consumers with incorrect beliefs is high enough, firms sell their services at high prices in equilibrium and only to the latter type of consumers. I extend the benchmark model to include more than one quack (and more than one expert), and the results are shown to be robust to the extensions. When there are two quacks and two experts, I find that firms may ‘split’ the market, and sell their services at different prices.
|
Page generated in 0.0415 seconds