Spelling suggestions: "subject:"coequality control istatistical methods."" "subject:"coequality control bystatistical methods.""
11 |
Process capability analysis using Motorola's six sigma characterization methodologyIsmail, Slim Ben 01 July 2001 (has links)
No description available.
|
12 |
Using economic models for process improvement to evaluate the performance of control chartsPraisont, Chintanai 01 July 2002 (has links)
No description available.
|
13 |
Forecasting corporate performanceHarrington, Robert P. January 1985 (has links)
For the past twenty years, the usefulness of accounting information has been emphasized. In 1966 the American Accounting Association in its State of Basic Accounting Theory asserted that usefulness is the primary purpose of external financial reports. In 1978 the State of Financial Accounting Concepts, No. 1 affirmed the usefulness criterion. "Financial reporting should provide information that is useful to present and potential investors and creditors and other users..."
Information is useful if it facilitates decision making. Moreover, all decisions are future-oriented; they are based on a prognosis of future events. The objective of this research, therefore, is to examine some factors that affect the decision maker's ability to use financial information to make good predictions and thereby good decisions.
There are two major purposes of the study. The first is to gain insight into the amount of increase in prediction accuracy that is expected to be achieved when a model replaces the human decision-maker in the selection of cues. The second major purpose is to examine the information overload phenomenon to provide research evidence to determine the point at which additional information may contaminate prediction accuracy.
The research methodology is based on the lens model developed by Eyon Brunswick in 1952. Multiple linear regression equations are used to capture the participants’ models, and correlation statistics are used to measure prediction accuracy. / Ph. D.
|
14 |
Process parameter optimisation of steel components laser forming using a Taguchi design of experiments approachSobetwa, Siyasanga January 2017 (has links)
A research report submitted to the Faculty of Engineering and the Built Environment,
University of the Witwatersrand, Johannesburg, in partial fulfilment of the requirements for
the degree of Master of Science in Engineering.
Date: September 2017, Johannesburg / The focus in this research investigation is to investigate the Process Parameter
Optimisation in Laser Beam Forming (LBF) process using the 4.4 kW Nd: YAG laser
system – Rofin DY 044 to form 200 x 50 x 3 mm3 mild steel - AISI 1008 samples. The
laser power P, beam diameter B, scan velocity V, number of scans N, and cooling flow
C were the five input parameters of interest in the investigation because of their
influence in the final formed product. Taguchi Design of Experiment (DoE) was used
for the selection and combination of input parameters for LBF process. The
investigation was done experimentally and computationally. Laser Beam Forming
(LBF) input parameters were categorised to three different levels, low (L), medium (M),
and high (H) laser forming (LBF) parameters to evaluate parameters that yield
maximum bending and better surface finish/quality. The conclusion drawn from LBF
process is that samples which are LBFormed using low parameter settings had
unnoticeable bending and good material surface finishing. On the other hand, samples
LBFormed using medium parameters yielded visible bending and non-smooth surface
finishing, while samples processed using high LBF parameters yielded maximum
bending and more surface roughness than the other two process parameters. / MT2018
|
15 |
Determining the most appropiate [sic] sampling interval for a Shewhart X-chartVining, G. Geoffrey January 1986 (has links)
A common problem encountered in practice is determining when it is appropriate to change the sampling interval for control charts. This thesis examines this problem for Shewhart X̅ charts. Duncan's economic model (1956) is used to develop a relationship between the most appropriate sampling interval and the present rate of"disturbances,” where a disturbance is a shift to an out of control state. A procedure is proposed which switches the interval to convenient values whenever a shift in the rate of disturbances is detected. An example using simulation demonstrates the procedure. / M.S.
|
16 |
The Fixed v. Variable Sampling Interval Shewhart X-Bar Control Chart in the Presence of Positively Autocorrelated DataHarvey, Martha M. (Martha Mattern) 05 1900 (has links)
This study uses simulation to examine differences between fixed sampling interval (FSI) and variable sampling interval (VSI) Shewhart X-bar control charts for processes that produce positively autocorrelated data. The influence of sample size (1 and 5), autocorrelation parameter, shift in process mean, and length of time between samples is investigated by comparing average time (ATS) and average number of samples (ANSS) to produce an out of control signal for FSI and VSI Shewhart X-bar charts. These comparisons are conducted in two ways: control chart limits pre-set at ±3σ_x / √n and limits computed from the sampling process. Proper interpretation of the Shewhart X-bar chart requires the assumption that observations are statistically independent; however, process data are often autocorrelated over time. Results of this study indicate that increasing the time between samples decreases the effect of positive autocorrelation between samples. Thus, with sufficient time between samples the assumption of independence is essentially not violated. Samples of size 5 produce a faster signal than samples of size 1 with both the FSI and VSI Shewhart X-bar chart when positive autocorrelation is present. However, samples of size 5 require the same time when the data are independent, indicating that this effect is a result of autocorrelation. This research determined that the VSI Shewhart X-bar chart signals increasingly faster than the corresponding FSI chart as the shift in the process mean increases. If the process is likely to exhibit a large shift in the mean, then the VSI technique is recommended. But the faster signaling time of the VSI chart is undesirable when the process is operating on target. However, if the control limits are estimated from process samples, results show that when the process is in control the ARL for the FSI and the ANSS for the VSI are approximately the same, and exceed the expected value when the limits are fixed.
|
17 |
A Clinical Decision Support System for the Identification of Potential Hospital Readmission PatientsUnknown Date (has links)
Recent federal legislation has incentivized hospitals to focus on quality of patient
care. A primary metric of care quality is patient readmissions. Many methods exist to
statistically identify patients most likely to require hospital readmission. Correct
identification of high-risk patients allows hospitals to intelligently utilize limited resources
in mitigating hospital readmissions. However, these methods have seen little practical
adoption in the clinical setting. This research attempts to identify the many open research
questions that have impeded widespread adoption of predictive hospital readmission
systems.
Current systems often rely on structured data extracted from health records systems.
This data can be expensive and time consuming to extract. Unstructured clinical notes are
agnostic to the underlying records system and would decouple the predictive analytics
system from the underlying records system. However, additional concerns in clinical
natural language processing must be addressed before such a system can be implemented. Current systems often perform poorly using standard statistical measures.
Misclassification cost of patient readmissions has yet to be addressed and there currently
exists a gap between current readmission system evaluation metrics and those most
appropriate in the clinical setting. Additionally, data availability for localized model
creation has yet to be addressed by the research community. Large research hospitals may
have sufficient data to build models, but many others do not. Simply combining data from
many hospitals often results in a model which performs worse than using data from a single
hospital.
Current systems often produce a binary readmission classification. However,
patients are often readmitted for differing reasons than index admission. There exists little
research into predicting primary cause of readmission. Furthermore, co-occurring evidence
discovery of clinical terms with primary diagnosis has seen only simplistic methods
applied.
This research addresses these concerns to increase adoption of predictive hospital
readmission systems. / Includes bibliography. / Dissertation (Ph.D.)--Florida Atlantic University, 2017. / FAU Electronic Theses and Dissertations Collection
|
18 |
Multivariate Quality Control Using Loss-Scaled Principal ComponentsMurphy, Terrence Edward 24 November 2004 (has links)
We consider a principal components based decomposition of the
expected value of the multivariate quadratic loss function, i.e.,
MQL. The principal components are formed by scaling the original
data by the contents of the loss constant matrix, which defines
the economic penalty associated with specific variables being off
their desired target values. We demonstrate the extent to which a
subset of these ``loss-scaled principal components", i.e., LSPC,
accounts for the two components of expected MQL, namely the
trace-covariance term and the off-target vector product. We employ
the LSPC to solve a robust design problem of full and reduced
dimensionality with deterministic models that approximate the true
solution and demonstrate comparable results in less computational
time. We also employ the LSPC to construct a test statistic called
loss-scaled T^2 for multivariate statistical process control.
We show for one case how the proposed test statistic has faster
detection than Hotelling's T^2 of shifts in location for
variables with high weighting in the MQL. In addition we
introduce a principal component based decomposition of Hotelling's
T^2 to diagnose the variables responsible for driving the
location and/or dispersion of a subgroup of multivariate
observations out of statistical control. We demonstrate the
accuracy of this diagnostic technique on a data set from the
literature and show its potential for diagnosing the loss-scaled
T^2 statistic as well.
|
19 |
An analysis of the California State Department of Parks and Recreation's "Quality Management Program"Turney, Celena 01 January 1997 (has links)
No description available.
|
20 |
Development of confidence intervals for process capability assessment in short run manufacturing environment using bootstrap methodologyKnezevic, Zec Gorana 01 October 2003 (has links)
No description available.
|
Page generated in 0.131 seconds