(a) State whether the following statements are True or False. Give reason in support of your answer:
(i) The R\mathrm{R} – chart is suitable when subgroup size is greater than 10 .
Answer:
The statement "The RR – chart is suitable when subgroup size is greater than 10" is generally considered to be False.
Justification:
The RR – chart, or range chart, is used in statistical process control to help evaluate the stability of a process in terms of the variation among the values within a sample (subgroup). However, there are some considerations regarding subgroup size:
Small Subgroup Sizes: The RR – chart is most effective and commonly used when the subgroup size (nn) is small, typically between 2 and 10. This is because the range is a simpler and more efficient calculation when dealing with smaller subgroup sizes.
Large Subgroup Sizes: When subgroup sizes are larger than 10, the RR – chart may not be the most appropriate choice for evaluating within-group variability due to the following reasons:
Sensitivity: The range is less sensitive to the spread of data as the subgroup size increases. It only considers the largest and smallest values and ignores the distribution of all the values in between.
Normality Assumption: The distribution of the range is highly dependent on the underlying distribution of the process data, and this dependency increases with subgroup size. For larger subgroups, if the data is not normally distributed, the control limits calculated for the RR – chart may not be accurate or reliable.
Alternative Charts: The SS – chart (standard deviation chart) is often recommended for larger subgroup sizes because it considers all data points in the subgroup, providing a more accurate and reliable measure of dispersion, especially when n > 10n > 10.
Example:
Consider a subgroup of size 15 with the following values:
The range (RR) would be calculated as the difference between the largest and smallest values: R=13-5=8R = 13 – 5 = 8.
However, this doesn’t give any insight into the variability of all the other values in the subgroup.
If there were outliers or shifts within the subgroup, the RR – chart might not detect them effectively due to its insensitivity to the distribution of internal values.
Conclusion:
While the RR – chart is a valuable tool in statistical process control, its suitability is generally constrained to scenarios with smaller subgroup sizes. For larger subgroup sizes, alternative charts, such as the SS – chart, are typically recommended to provide a more accurate and reliable analysis of process variability.
(ii) In single sampling plan, if we increase acceptance number then the OC curve will be steeper.
Answer:
The statement "In single sampling plan, if we increase acceptance number then the OC (Operating Characteristic) curve will be steeper." is generally considered to be False.
Justification:
In the context of statistical quality control, the Operating Characteristic (OC) curve is a graph that depicts the discriminating power of a sampling plan. It shows the probability of accepting a batch (lot) of items given various levels of quality (fraction nonconforming). The OC curve is influenced by the acceptance number in a single sampling plan.
Acceptance Number: The acceptance number in a single sampling plan is the maximum number of defective items allowed in the sample for the lot to still be accepted.
OC Curve Characteristics:
The x-axis of the OC curve represents the quality level of the lot (often the fraction or percentage nonconforming).
The y-axis represents the probability of accepting the lot at each quality level.
Explanation:
Increasing the Acceptance Number: If we increase the acceptance number in a single sampling plan:
The probability of accepting a lot (even with a higher defect rate) increases because we are allowing more defective items in the sample before rejecting the lot.
Therefore, the OC curve becomes less steep and shifts upwards, indicating a higher probability of accepting lots even as their quality decreases.
Example:
Consider two single sampling plans:
Plan A: Sample size = 50, Acceptance number = 1 (accept the lot if there is 1 or fewer defective item in the sample).
Plan B: Sample size = 50, Acceptance number = 5 (accept the lot if there are 5 or fewer defective items in the sample).
If we plot the OC curves for these two plans:
Plan A will have a steeper OC curve because the probability of accepting a lot decreases more rapidly as the quality level decreases (due to the lower acceptance number).
Plan B will have a less steep OC curve because it is more lenient (due to the higher acceptance number), and the probability of accepting a lot decreases more slowly as the quality level decreases.
Conclusion:
Increasing the acceptance number in a single sampling plan generally makes the OC curve less steep, not steeper, because it increases the probability of accepting lots even with lower quality levels. This is because a higher acceptance number allows for more defective items to be present in the sample without leading to rejection of the lot.
(iii) If the effect of summer and winter is not constant on the sale of AC then we use the additive model of the time series.
Answer:
The statement "If the effect of summer and winter is not constant on the sale of AC then we use the additive model of the time series." is generally considered to be False.
Justification:
Time series models are used to analyze the trend, seasonality, and other components in a data series over time. The two primary models used to describe time series data are:
Additive Model: Assumes that the components of the time series (trend, seasonality, and residual) are added together.
Y_(t)=T_(t)+S_(t)+R_(t)Y_t = T_t + S_t + R_t
Where:
Y_(t)Y_t is the observed data at time tt,
T_(t)T_t is the trend component,
S_(t)S_t is the seasonal component, and
R_(t)R_t is the residual or error component.
The additive model is typically used when the magnitude of the seasonal fluctuations does not depend on the level of the time series (i.e., the fluctuations are roughly constant over time).
Multiplicative Model: Assumes that the components of the time series are multiplied together.
The multiplicative model is typically used when the magnitude of the seasonal fluctuations depends on the level of the time series (i.e., the fluctuations increase or decrease as the data values increase or decrease).
Explanation:
If the effect of summer and winter (seasonality) is not constant on the sale of AC (i.e., the seasonal effect increases or decreases as the overall sales level changes), then the multiplicative model would generally be more appropriate, not the additive model.
Non-Constant Seasonal Effect: If the seasonal effect is not constant and varies with the level of the time series (e.g., the increase in sales during summer is proportionally larger when overall sales levels are higher), this implies a multiplicative relationship between the components.
Example:
Consider two scenarios:
Scenario 1: Every summer, AC sales increase by approximately 1000 units, regardless of the overall sales level.
Scenario 2: Every summer, AC sales double, regardless of the overall sales level.
In Scenario 1, the seasonal effect (increase of 1000 units) is constant, so an additive model might be appropriate. In Scenario 2, the seasonal effect is not constant (it depends on the overall sales level), so a multiplicative model might be more appropriate.
Conclusion:
When the seasonal effect is not constant and depends on the level of the time series, the multiplicative model is typically more appropriate than the additive model. Therefore, the statement is false.
(iv) If a researcher wants to find the relationship between today’s unemployment and that of 5 years ago without considering what happens in between then the partial autocorrelation is the better way in comparison to autocorrelation.
Answer:
The statement "If a researcher wants to find the relationship between today’s unemployment and that of 5 years ago without considering what happens in between then the partial autocorrelation is the better way in comparison to autocorrelation." is generally considered to be True.
Justification:
Autocorrelation: Autocorrelation, also known as serial correlation, measures the linear relationship between lagged values of a time series. It does not account for the correlations at all intervening lags. In other words, it does not control for the potential influence of other time lags.
Partial Autocorrelation: Partial autocorrelation, on the other hand, measures the linear relationship between the observation at time tt and the observation at time t-kt-k, controlling for the observations at all intervening time points. It essentially isolates the correlation at lag kk by removing the effect of correlations at shorter lags.
Explanation:
If a researcher wants to understand the relationship between today’s unemployment rate and that of 5 years ago, without considering the intervening years, the partial autocorrelation function (PACF) would indeed be the more appropriate tool.
The PACF will show the direct effect (correlation) of the unemployment rate 5 years ago on today’s rate, after removing the effects of the unemployment rate in the intervening 4 years. This allows the researcher to understand the isolated relationship between the unemployment rate at these two points in time, without the influence of the years in between.
In contrast, the autocorrelation function (ACF) would show the total effect (correlation) of the unemployment rate 5 years ago on today’s rate, without controlling for the intervening years. This means that the ACF could be capturing indirect effects that are not of interest to the researcher in this scenario.
Example:
Suppose we have a time series data of unemployment rates for several years.
If we calculate the autocorrelation at lag 5, it will give us the correlation between the unemployment rate today and that of 5 years ago, but it will not control for the unemployment rates in the intervening years.
If we calculate the partial autocorrelation at lag 5, it will give us the correlation between the unemployment rate today and that of 5 years ago, controlling for (or removing the effects of) the unemployment rates in the intervening years.
Conclusion:
In scenarios where a researcher is interested in understanding the direct relationship between observations at two points in time, without the influence of observations at intervening time points, the partial autocorrelation is indeed a more suitable method compared to autocorrelation. Therefore, the statement is true.
(v) A system has four components connected in parallel configuration with reliability 0.2,0.4,0.5,0.80.2,0.4,0.5,0.8. To improve the reliability of the system most, we have to replace the component which reliability is 0.2 .
Answer:
The statement "A system has four components connected in parallel configuration with reliability 0.2,0.4,0.5,0.80.2,0.4,0.5,0.8. To improve the reliability of the system most, we have to replace the component which reliability is 0.20.2." is generally considered to be True.
Justification:
Parallel System Reliability: In a parallel configuration, the system will function if at least one of the components functions. The reliability of a parallel system, R_(s)R_s, is calculated as:
where R_(i)R_i is the reliability of the i^(th)i^{th} component and nn is the number of components.
Improving Reliability: To improve the reliability of a parallel system most effectively, it is generally best to improve the component with the lowest reliability because it has the most room for improvement and will contribute the most to increasing the overall system reliability when improved.
Explanation:
Given the reliabilities of the four components as 0.2,0.4,0.5,0.80.2, 0.4, 0.5, 0.8, let’s consider the impact of improving the reliability of the component with reliability 0.20.2 (let’s call it Component 1).
The current reliability of the system, using the formula for parallel reliability, is:
If we improve the reliability of Component 1 (the component with reliability 0.20.2), it will have a larger impact on the overall system reliability than improving a component with higher reliability, because the unreliability (1-R_(1))(1 – R_1) of Component 1 is currently the largest among the components, and reducing it will have a larger impact on reducing the overall unreliability of the system.
Example:
Let’s consider improving Component 1 from 0.20.2 to 0.90.9. The new system reliability would be:
Comparing R_(s)^(‘)R_s’ with the original R_(s)R_s, we would likely see a more significant improvement in system reliability than if we improved one of the components with higher original reliability.
Conclusion:
Improving the component with the lowest reliability (in this case, 0.20.2) will generally have the most impact on improving the overall reliability of a system configured in parallel. Therefore, the statement is true.