• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 14
  • 13
  • 1
  • 1
  • Tagged with
  • 29
  • 29
  • 11
  • 8
  • 8
  • 7
  • 6
  • 6
  • 5
  • 4
  • 4
  • 4
  • 4
  • 4
  • 4
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Will Mortality Rate of HIV-Infected Patients Decrease After Starting Antiretroviral Therapy (ART)?

Bahakeem, Shaher 07 1900 (has links)
Indiana University-Purdue University Indianapolis (IUPUI) / Background: Many authors have indicated that HIV-infected patients mortality risk is higher immediately following the start of Antiretroviral Therapy. However, mortality rate of HIV-infected patients is expected to decrease after starting Antiretroviral Therapy (ART) potentially complicating accurate statistical estimation of patient survival and, more generally, effective monitoring of the evolution of the worldwide epidemic. Method: In this thesis, we determine if mortality of HIV-patients increases or decreases after the initiation of ART therapy using flexible survival modelling techniques. To achieve this objective, this study uses semi-parametric statistical models for fitting and estimating survival time using different covariates. A combination of the Weibull distribution with splines is compared to the usual Weibull, exponential, and gamma distribution parametric models, and the Cox semi-parametric model. The objective of this study is to compare these models to find the best fitting model so that it can then be used to improve modeling of the survival time and explore the pattern of change in mortality rates for a cohort of HIV-infected patients recruited in a care and treatment program in Uganda. Results: The analysis shows that flexible survival Weibull models are better than usualoff-parametric and semi-parametric model fitting according to the AIC criterion. Conclusion: The mortality of HIV-patients is high right after the initiation of ART therapy and decreases rapidly subsequently.
2

Comparison of proportional hazards and accelerated failure time models

Qi, Jiezhi 30 March 2009
The field of survival analysis has experienced tremendous growth during the latter half of the 20th century. The methodological developments of survival analysis that have had the most profound impact are the Kaplan-Meier method for estimating the survival function, the log-rank test for comparing the equality of two or more survival distributions, and the Cox proportional hazards (PH) model for examining the covariate effects on the hazard function. The accelerated failure time (AFT) model was proposed but seldom used. In this thesis, we present the basic concepts, nonparametric methods (the Kaplan-Meier method and the log-rank test), semiparametric methods (the Cox PH model, and Cox model with time-dependent covariates) and parametric methods (Parametric PH model and the AFT model) for analyzing survival data.<p> We apply these methods to a randomized placebo-controlled trial to prevent Tuberculosis (TB) in Ugandan adults infected with Human Immunodificiency Virus (HIV). The objective of the analysis is to determine whether TB preventive therapies affect the rate of AIDS progression and survival in HIV-infected adults. Our conclusion is that TB preventive therapies appear to have no effect on AIDS progression, death and combined event of AIDS progression and death. The major goal of this paper is to support an argument for the consideration of the AFT model as an alternative to the PH model in the analysis of some survival data by means of this real dataset. We critique the PH model and assess the lack of fit. To overcome the violation of proportional hazards, we use the Cox model with time-dependent covariates, the piecewise exponential model and the accelerated failure time model. After comparison of all the models and the assessment of goodness-of-fit, we find that the log-logistic AFT model fits better for this data set. We have seen that the AFT model is a more valuable and realistic alternative to the PH model in some situations. It can provide the predicted hazard functions, predicted survival functions, median survival times and time ratios. The AFT model can easily interpret the results into the effect upon the expected median duration of illness for a patient in a clinical setting. We suggest that the PH model may not be appropriate in some situations and that the AFT model could provide a more appropriate description of the data.
3

Comparison of proportional hazards and accelerated failure time models

Qi, Jiezhi 30 March 2009 (has links)
The field of survival analysis has experienced tremendous growth during the latter half of the 20th century. The methodological developments of survival analysis that have had the most profound impact are the Kaplan-Meier method for estimating the survival function, the log-rank test for comparing the equality of two or more survival distributions, and the Cox proportional hazards (PH) model for examining the covariate effects on the hazard function. The accelerated failure time (AFT) model was proposed but seldom used. In this thesis, we present the basic concepts, nonparametric methods (the Kaplan-Meier method and the log-rank test), semiparametric methods (the Cox PH model, and Cox model with time-dependent covariates) and parametric methods (Parametric PH model and the AFT model) for analyzing survival data.<p> We apply these methods to a randomized placebo-controlled trial to prevent Tuberculosis (TB) in Ugandan adults infected with Human Immunodificiency Virus (HIV). The objective of the analysis is to determine whether TB preventive therapies affect the rate of AIDS progression and survival in HIV-infected adults. Our conclusion is that TB preventive therapies appear to have no effect on AIDS progression, death and combined event of AIDS progression and death. The major goal of this paper is to support an argument for the consideration of the AFT model as an alternative to the PH model in the analysis of some survival data by means of this real dataset. We critique the PH model and assess the lack of fit. To overcome the violation of proportional hazards, we use the Cox model with time-dependent covariates, the piecewise exponential model and the accelerated failure time model. After comparison of all the models and the assessment of goodness-of-fit, we find that the log-logistic AFT model fits better for this data set. We have seen that the AFT model is a more valuable and realistic alternative to the PH model in some situations. It can provide the predicted hazard functions, predicted survival functions, median survival times and time ratios. The AFT model can easily interpret the results into the effect upon the expected median duration of illness for a patient in a clinical setting. We suggest that the PH model may not be appropriate in some situations and that the AFT model could provide a more appropriate description of the data.
4

Análise dos modelos não paramétricos de avaliação de eficiência e a performance dos bancos comerciais brasileiros

Silva, Tarcio Lopes da January 2006 (has links)
O principal objetivo deste trabalho é descrever e analisar do ponto de vista teórico os principais métodos não paramétricos de avaliação de eficiência e avaliar empiricamente os resultados gerados por esses diferentes métodos aplicados a um mesmo conjunto de dados. Com essa finalidade, realizou-se diversas simulações teóricas para fins comparativos dos tradicionais estimadores DEA e FDH, inclusive de seus procedimentos para inferência e correção de viés, e dos novos estimadores de ordem m e quantil. Empiricamente, utilizamos uma amostra de 184 bancos comerciais brasileiros no período de Junho/1995 à Junho/2004. Os resultados mostraram que as diferentes suposições impostas ao conjunto de produção pelos estimadores DEA e FDH afetam sensivelmente os índices de eficiência de várias firmas. Apesar disso, o uso de mais de um estimador mostrou-se um bom artifício para identificação das unidades com pior desempenho. Os procedimentos disponíveis para correção de viés e inferência, entretanto, mostraram-se deficientes principalmente para as firmas localizadas ao longo da fronteira estimada. Por outro lado, a importância da utilização dos novos estimadores não paramétricos quantil e de ordem m ficou evidente devido a presença de observações consideradas valores extremos, que distorcem os índices de eficiência estimados de outras observações. O uso de tais estimadores, mais robustos a valores extremos e outliers, gerou resultados mais confiáveis. Finalmente, procurou-se investigar se o controle de capital, o segmento de atuação e o porte dos bancos afetam sua eficiência, além de investigar o comportamento da performance do setor durante o período de análise. / The purpose of this piece of work has been to analyse and describe the main non parametric efficiency evaluation methods from a theoretical point of view and to empirically analyse the efficiency scores generated by such methods using the same data sample. For that purpose, theoretical simulations were used in order to compare traditional DEA and FDH estimators, their inference and bias correction procedures included. The new order m and quantil estimators were also used. We empirically used a sample of 184 Brazilian commercial banks spanning the June/1995 to June/2004 period. The results show that several banks’ efficiency scores were noticeably affected by the different assumptions made by the DEA and FDH estimators towards the production set. Nevertheless, the use of more than one estimator proved to be an effective way of identifying the units with the worst performance levels. The available procedures for bias correction and inference, however, proved ineffective for frontier firms. On the other hand, the importance of utilising the new order m and quantil estimators was evidenced. Several observations from our sample were detected as extreme values, which affected the estimated efficiency scores from other observations. The use of the aforementioned estimators, more robust to extreme values and outliers, generated more trustworthy results. Finally, an attempt was made to investigate whether or not capital control, business segment and banks’ size affect their efficiency, besides investigating the sector’s performance behaviour during the target period.
5

Análise dos modelos não paramétricos de avaliação de eficiência e a performance dos bancos comerciais brasileiros

Silva, Tarcio Lopes da January 2006 (has links)
O principal objetivo deste trabalho é descrever e analisar do ponto de vista teórico os principais métodos não paramétricos de avaliação de eficiência e avaliar empiricamente os resultados gerados por esses diferentes métodos aplicados a um mesmo conjunto de dados. Com essa finalidade, realizou-se diversas simulações teóricas para fins comparativos dos tradicionais estimadores DEA e FDH, inclusive de seus procedimentos para inferência e correção de viés, e dos novos estimadores de ordem m e quantil. Empiricamente, utilizamos uma amostra de 184 bancos comerciais brasileiros no período de Junho/1995 à Junho/2004. Os resultados mostraram que as diferentes suposições impostas ao conjunto de produção pelos estimadores DEA e FDH afetam sensivelmente os índices de eficiência de várias firmas. Apesar disso, o uso de mais de um estimador mostrou-se um bom artifício para identificação das unidades com pior desempenho. Os procedimentos disponíveis para correção de viés e inferência, entretanto, mostraram-se deficientes principalmente para as firmas localizadas ao longo da fronteira estimada. Por outro lado, a importância da utilização dos novos estimadores não paramétricos quantil e de ordem m ficou evidente devido a presença de observações consideradas valores extremos, que distorcem os índices de eficiência estimados de outras observações. O uso de tais estimadores, mais robustos a valores extremos e outliers, gerou resultados mais confiáveis. Finalmente, procurou-se investigar se o controle de capital, o segmento de atuação e o porte dos bancos afetam sua eficiência, além de investigar o comportamento da performance do setor durante o período de análise. / The purpose of this piece of work has been to analyse and describe the main non parametric efficiency evaluation methods from a theoretical point of view and to empirically analyse the efficiency scores generated by such methods using the same data sample. For that purpose, theoretical simulations were used in order to compare traditional DEA and FDH estimators, their inference and bias correction procedures included. The new order m and quantil estimators were also used. We empirically used a sample of 184 Brazilian commercial banks spanning the June/1995 to June/2004 period. The results show that several banks’ efficiency scores were noticeably affected by the different assumptions made by the DEA and FDH estimators towards the production set. Nevertheless, the use of more than one estimator proved to be an effective way of identifying the units with the worst performance levels. The available procedures for bias correction and inference, however, proved ineffective for frontier firms. On the other hand, the importance of utilising the new order m and quantil estimators was evidenced. Several observations from our sample were detected as extreme values, which affected the estimated efficiency scores from other observations. The use of the aforementioned estimators, more robust to extreme values and outliers, generated more trustworthy results. Finally, an attempt was made to investigate whether or not capital control, business segment and banks’ size affect their efficiency, besides investigating the sector’s performance behaviour during the target period.
6

Análise dos modelos não paramétricos de avaliação de eficiência e a performance dos bancos comerciais brasileiros

Silva, Tarcio Lopes da January 2006 (has links)
O principal objetivo deste trabalho é descrever e analisar do ponto de vista teórico os principais métodos não paramétricos de avaliação de eficiência e avaliar empiricamente os resultados gerados por esses diferentes métodos aplicados a um mesmo conjunto de dados. Com essa finalidade, realizou-se diversas simulações teóricas para fins comparativos dos tradicionais estimadores DEA e FDH, inclusive de seus procedimentos para inferência e correção de viés, e dos novos estimadores de ordem m e quantil. Empiricamente, utilizamos uma amostra de 184 bancos comerciais brasileiros no período de Junho/1995 à Junho/2004. Os resultados mostraram que as diferentes suposições impostas ao conjunto de produção pelos estimadores DEA e FDH afetam sensivelmente os índices de eficiência de várias firmas. Apesar disso, o uso de mais de um estimador mostrou-se um bom artifício para identificação das unidades com pior desempenho. Os procedimentos disponíveis para correção de viés e inferência, entretanto, mostraram-se deficientes principalmente para as firmas localizadas ao longo da fronteira estimada. Por outro lado, a importância da utilização dos novos estimadores não paramétricos quantil e de ordem m ficou evidente devido a presença de observações consideradas valores extremos, que distorcem os índices de eficiência estimados de outras observações. O uso de tais estimadores, mais robustos a valores extremos e outliers, gerou resultados mais confiáveis. Finalmente, procurou-se investigar se o controle de capital, o segmento de atuação e o porte dos bancos afetam sua eficiência, além de investigar o comportamento da performance do setor durante o período de análise. / The purpose of this piece of work has been to analyse and describe the main non parametric efficiency evaluation methods from a theoretical point of view and to empirically analyse the efficiency scores generated by such methods using the same data sample. For that purpose, theoretical simulations were used in order to compare traditional DEA and FDH estimators, their inference and bias correction procedures included. The new order m and quantil estimators were also used. We empirically used a sample of 184 Brazilian commercial banks spanning the June/1995 to June/2004 period. The results show that several banks’ efficiency scores were noticeably affected by the different assumptions made by the DEA and FDH estimators towards the production set. Nevertheless, the use of more than one estimator proved to be an effective way of identifying the units with the worst performance levels. The available procedures for bias correction and inference, however, proved ineffective for frontier firms. On the other hand, the importance of utilising the new order m and quantil estimators was evidenced. Several observations from our sample were detected as extreme values, which affected the estimated efficiency scores from other observations. The use of the aforementioned estimators, more robust to extreme values and outliers, generated more trustworthy results. Finally, an attempt was made to investigate whether or not capital control, business segment and banks’ size affect their efficiency, besides investigating the sector’s performance behaviour during the target period.
7

[en] GRADUATION METHODS UNDER PARAMETRIC AND NON-PARAMETRIC MODELS FOR SELECT AND ULTIMATE TABLES / [pt] METODOLOGIAS DE CONSTRUÇÃO DE TÁBUAS BIOMÉTRICAS SELETAS E FINAIS A PARTIR DE MODELOS PARAMÉTRICOS E NÃO-PARAMÉTRICOS

FABIO GARRIDO LEAL MARTINS 04 March 2008 (has links)
[pt] O estudo aborda as diversas metodologias de construção de tábuas biométricas: desde as técnicas de graduação tradicionalmente utilizadas para os casos em que há grande quantidade de dados, até um método específico de aplicação para o caso de poucos dados. Inclui uma discussão sobre as formas de construção de tábuas seletas, em particular de sobrevivência de inválidos. A população de servidores públicos estatutários da administração direta do município do Rio de Janeiro é utilizada para a graduação de tábuas de sobrevivência de válidos e de inválidos, enquanto que a dos aposentados urbanos por invalidez do INSS serve de base para a tábua seleta de sobrevivência de inválidos. / [en] This study represents an approach to the main methods of life tables construction. It shows traditional graduation techniques for cases including high exposure data, as well a methodology for few data. Further more, this study generates a discussion about select life tables construction, in particular disability mortality tables. Data set from Rio de Janeiro officials population were used for mortality and disability mortality tables construction. In addition, a select disability mortality table was constructed based on the INSS urban disability retired population.
8

Data Driven Visual Recognition

Aghazadeh, Omid January 2014 (has links)
This thesis is mostly about supervised visual recognition problems. Based on a general definition of categories, the contents are divided into two parts: one which models categories and one which is not category based. We are interested in data driven solutions for both kinds of problems. In the category-free part, we study novelty detection in temporal and spatial domains as a category-free recognition problem. Using data driven models, we demonstrate that based on a few reference exemplars, our methods are able to detect novelties in ego-motions of people, and changes in the static environments surrounding them. In the category level part, we study object recognition. We consider both object category classification and localization, and propose scalable data driven approaches for both problems. A mixture of parametric classifiers, initialized with a sophisticated clustering of the training data, is demonstrated to adapt to the data better than various baselines such as the same model initialized with less subtly designed procedures. A nonparametric large margin classifier is introduced and demonstrated to have a multitude of advantages in comparison to its competitors: better training and testing time costs, the ability to make use of indefinite/invariant and deformable similarity measures, and adaptive complexity are the main features of the proposed model. We also propose a rather realistic model of recognition problems, which quantifies the interplay between representations, classifiers, and recognition performances. Based on data-describing measures which are aggregates of pairwise similarities of the training data, our model characterizes and describes the distributions of training exemplars. The measures are shown to capture many aspects of the difficulty of categorization problems and correlate significantly to the observed recognition performances. Utilizing these measures, the model predicts the performance of particular classifiers on distributions similar to the training data. These predictions, when compared to the test performance of the classifiers on the test sets, are reasonably accurate. We discuss various aspects of visual recognition problems: what is the interplay between representations and classification tasks, how can different models better adapt to the training data, etc. We describe and analyze the aforementioned methods that are designed to tackle different visual recognition problems, but share one common characteristic: being data driven. / <p>QC 20140604</p>
9

Sensor Validation Using Linear Parametric Models, Artificial Neural Networks and CUSUM / Sensorvalidering medelst linjära konfektionsmodeller, artificiella neurala nätverk och CUSUM

Norman, Gustaf January 2015 (has links)
Siemens gas turbines are monitored and controlled by a large number of sensors and actuators. Process information is stored in a database and used for offline calculations and analyses. Before storing the sensor readings, a compression algorithm checks the signal and skips the values that explain no significant change. Compression of 90 % is not unusual. Since data from the database is used for analyses and decisions are made upon results from these analyses it is important to have a system for validating the data in the database. Decisions made on false information can result in large economic losses. When this project was initiated no sensor validation system was available. In this thesis the uncertainties in measurement chains are revealed. Methods for fault detection are investigated and finally the most promising methods are put to the test. Linear relationships between redundant sensors are derived and the residuals form an influence structure allowing the faulty sensor to be isolated. Where redundant sensors are not available, a gas turbine model is utilized to state the input-output relationships so that estimates of the sensor outputs can be formed. Linear parametric models and an ANN (Artificial Neural Network) are developed to produce the estimates. Two techniques for the linear parametric models are evaluated; prediction and simulation. The residuals are also evaluated in two ways; direct evaluation against a threshold and evaluation with the CUSUM (CUmulative SUM) algorithm. The results show that sensor validation using compressed data is feasible. Faults as small as 1% of the measuring range can be detected in many cases.
10

ANNOTATION MECHANISMS TO MANAGE DESIGN KNOWLEDGE IN COMPLEX PARAMETRIC MODELS AND THEIR EFFECTS ON ALTERATION AND REUSABILITY

Dorribo Camba, Jorge 12 January 2015 (has links)
El proyecto de investigación propuesto se enmarca dentro del área de diseño de producto con aplicaciones de modelado sólido CAD/CAM (Computer Aided Design/Computer Aided Manufacturing). Concretamente, se pretende hacer un estudio de las herramientas de anotación asociativas disponibles en las aplicaciones comerciales de modelado CAD con el fin de analizar su uso, viabilidad, eficiencia y efectos en la modificación y reutilización de modelos digitales 3D, así como en la gestión y comunicación del conocimiento técnico vinculado al diseño. La idea principal de esta investigación doctoral es establecer un método para representar y evaluar el conocimiento implícito de los ingenieros de diseño acerca de un modelo digital, así como la integración dinámica de dicho conocimiento en el propio modelo CAD, a través de anotaciones, con el objetivo de poder almacenar y comunicar eficientemente la mayor cantidad de información útil acerca del modelo, y reducir el tiempo y esfuerzo requeridos para su alteración y/o reutilización. / Dorribo Camba, J. (2014). ANNOTATION MECHANISMS TO MANAGE DESIGN KNOWLEDGE IN COMPLEX PARAMETRIC MODELS AND THEIR EFFECTS ON ALTERATION AND REUSABILITY [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/45997 / TESIS

Page generated in 0.0613 seconds