Handbook of Reliability, Availability, Maintainability and Safety in Engineering Design - Part 44 pptx

10 170 0
Handbook of Reliability, Availability, Maintainability and Safety in Engineering Design - Part 44 pptx

Đang tải... (xem toàn văn)

Thông tin tài liệu

4.2 Theoretical Overview of Availability and Maintainability in Engineering Design 413 Average output rate =[1×(0.95×0.9×0.9)] +[0.5×(0.95×0.9×0.1)] +[0.5×(0.95×0.1×0.9)] = 0.7695 + 0.04275+ 0.04275 = 0.855 = 85.5% It must be noted that this average output rate is expressed as a percentage of the pos- sible output that can be achieved at maximum design capacity and 100% utilisation. It is thus an expression of the output capability of the plant as a whole, depending on the percentage capacity for each state of the plant, as a result of the states of each individual system, multiplied by its availability. The above example of the simple three-system plant is insightful as an introduc- tion to a plant having several states in which outages are regarded as full outages. If, however, the system outages are such that, over a specific period of time T, the sys- tems could experience full outages as well as partial outages that are limited to 50% of system output, then a table similar to the previous state matrix can be developed. To calculate the expected or average output rate of the plant over an operating period T, the percentage capacity (% of MDC) for each state during the shift pe- riod T is multiplied by the equivalent availability of each system. The partial state matrix given in Table 4.2 presents the possible states of each system over a period of time T. These states are either 100% down (full outage) or 50% down (partial outage). Table 4.2 Double turbine/boiler generating plant partial state matrix Period T Period T In period T In period T %MDC 1 Up Up Up 100% 2 Up Up Down 50% 75% 3 Up Up Down 100% 50% 4 Up Down 50% Up 75% 5 Up Down 100% Up 50% 6 Down 50% Up Up 50% 7 Down 100% Up Up 0% 8 Down 50% Up Down 50% 50% 9 Down 50% Up Down 100% 50% 10 Down 50% Down 50% Up 50% 11 Down 50% Down 100% Up 50% 12 Down 50% Down 50% Down 50% 50% 13 Down 50% Down 50% Down 100% 25% 14 Down 50% Down 50% Down 50% 25% 15 Down 50% Down 100% Down 100% 0% 16 Down 100% Do wn 100% Down 100% 0% 414 4 Availability and Maintainability in Engineering Design If, for simplicity, the likelihood and duration of each state in the above partial state matrix table is considered to be equal, and each state could occur only once in the operating period T, then according to Eq. (4.141) given above, the equivalent availability of each system can be calculated as: EA = ∑ [(T o ) ·n(MDC)] T ·MDC . In this case: T o = operational period for each state T = total operating period or cycle n(MDC)= capacity as % of MDC for each system n = fraction of process output MDC = maximum demand capacity. Thus, the equivalent availability for the boiler in period T (i.e. working down the second column of Table 4.2), and u sing Eq. (4.141), can be calculated as: EA Boiler = (1.5)[5×1 + 0.5+ 0+ 8×0.5+ 0] 24×1 = (1.5)9.5 24 = 0.59375 = 59.375% . Similarly, the equivalent availability for #1 turbine and #2 turbine can also be cal- culated. The equivalent availability for the #1 turbine in period T: (i.e. working down the third column of Table 4.2) EA #1 Turbine = (1.5)[7×1 + 4×0.5+ 5×0] 24×1 = (1.5)9 24 = 0.5625 = 56.25% . The equivalent availability for the #2 turbine in period T: (i.e. working down the fourth column of Table 4.2) EA #2 Turbine = (1.5)[7×1 + 4×0.5+ 5×0] 24×1 = (1.5)9 24 = 0.5625 = 56.25% . 4.3 Analytic Development of Availability and Maintainability in Engineering Design 415 With the system equivalent ava ilabilities for the boiler, #1 turbine and #2 turbine now calculated from all the possible partial states (up, 50% down, or 100% down), what would be the expected or average plant output rate when the equivalent avail- ability for the boiler is 59.375% and the equivalent availability for the turbine gen- erators are 56.25% each? Taking into consideration states with reduced utilisation as a result of partial outages, the expected or average plant output rate is calculated as: Average plant output rate with partial outages = Σ (capacity of plant state at full and partial utilisation of systems that are operational × availability of each integrated system) = [1.0 × (0.59375 × 0.5625 × 0.5625)] + [0.75 × (0.59375 × 0.5625 × 0.5625)] + [0.5 × (0.59375 × 0.5625 × 0.4375)] + [0.75 × (0.59375 × 0.5625 × 0.5625)] + [0.5 × (0.59375 × 0.4375 × 0.5625)] + [0.5 × (0.59375 × 0.5625 × 0.5625)] + [0.5 × (0.59375 × 0.5625 × 0.5625)] + [0.5 × (0.59375 × 0.5625 × 0.4375)] + [0.5 × (0.59375 × 0.5625 × 0.5625)] + [0.5 × (0.59375 × 0.4375 × 0.5625)] + [0.5 × (0.59375 × 0.5625 × 0.5625)] + [0.25 × (0.59375 × 0.5625 × 0.4375)] + [0.25 × (0.59375 × 0.4375 × 0.5625)] = 0.18787 + (2 × 0.14090) + (4 × 0.09393) + (4 ×0.07306) + (2 × 0.04697) = 0.18787 + 0.2818 + 0.37572 + 0.29224 + 0.09394 = 85.5% The expected or average plant output rate, taking into co nsideration states with re- duced utilisation as a result of partial outages, is 85.5%. 4.3 Analytic Development of Availability and Maintainability in Engineering Design Several techniques are identified for availability and maintainability prediction, as- sessment and evaluation, in the conceptual, preliminary and detail design phases respectively. As with the analytic development of reliability an d performance of Sect. 3.3, only certain of the availability and main tainability techniques have been considered for further development. This approach is adopted on the basis of the transformational capabilities of the techniques in developing intelligent computer automated methodology using optimisation algorithms (OA). These optimisation algorithms should ideally be suitable for application in artificial intelligence-based (AIB) modelling, in which development of knowledge-based expert systems within a blackboard model can be applied in determining the integrity of engineering de- sign. Furthermore, the AIB model must be suited to applied concurrent engineering design in an integrated collaborative design environment in which automated con- tinual design reviews may be conducted during the engineering design process by remotely located design groups communicating via the internet. 416 4 Availability and Maintainability in Engineering Design 4.3.1 Analytic Development of Availability and Maintainability Prediction in Conceptual Design A technique selected for further development as a tool for availability and main- tainability prediction in determining the integrity of engineering design during the conceptual design phase is modelling based on measures of system performance. This technique h as already been considered in part for reliability prediction in Sect. 3.3.1, and needs to be expanded to include prediction of reliability as well as inherent availability and maintainability. System performance analysis through the technique of simulation modelling is also considered, specifically for prediction of system characteristics that affect system availability. Furthermore, the technique of robust design is selected for its preferred application in decision-making about engineering design integrity, particularly in considering th e various uncertainties involved in system performance simulation modelling. Monte Carlo simulation is used to propagate these uncertainties in the applica- tion of simulation models in a collaborative engineering design environment. The techniques selected for availability and maintainab ility prediction in the conceptual design phase are thus considered under the following topics: i. System performance measures and limits of capability ii. System performance analysis and simulation modelling iii. Uncertainty in system performance simulation modelling. 4.3.1.1 System Performance Measures and Limits of Capability Referring back to Sect. 3 .3.1, it was pointed out that, instead of using actual per- formance values such as temperatures, pre ssures, etc., it is more meaningful to use the proximity of the actual performance value to the limit of capability of the item of equipment. In engineer ing design review, the proximity of performance to a limit closely relates to a m easure of the item’s safety margin, which could indicate the need for design changes or selecting alternate systems. Non-dimensional numeri- cal values for system performance may be obtained by determining the limits of capability,C max andC min , with respect to the performance parameters for system in- tegrity (i.e. reliability, availability,maintainability and safety). The nominal range of integrity values for w hich the sy stem is designed (i.e. 95 to 98% reliability at 80 to 85% availability) must also be specified. Thus, a set of data points are obtained for each item of equipment with respect to the relevant performance parameters, to be entered into a parameter performance matrix. a) Performance Parameters for System Integrity For predicting system availability, the per formance measures for reliability and maintainability are estimated, and the inherent availability determined from these 4.3 Analytic Development of Availability and Maintainability in Engineering Design 417 measures. As indicated previously, system reliability can be predicted by estimating the mean time between failures (MTBF), and maintainability performance can be predicted by estimating the mean time to repair (MTTR) . Inherent availability, A i , can then be predicted according to: A i = MTBF (MTBF+ MTTR) (4.148) In the case of reliability and maintaina bility, there are no operating limits of capa- bility but, instead, a prediction of engineering design performance relating to MTBF and MTTR. Data points for the parameter performance matrix can be obtained through expert judgement of system reliability by estimating the mean time between failures (MTBF), and of maintainability performance by estimating the mean time to repair (MTTR) of critical failures (Booker et al. 2000). (Refer to Sect. 3.3.3.4 dealing with expert judgement as data.) A reliability data point x ij can be generated from the predicted MTBF(R),amax- imum acceptable MTBF(R max ), and a minimum acceptable MTBF(R min ),where: x ij = (R−R min ) ×10 R max −R min (4.149) Similarly, a maintainability data point x ij can be generated from the predicted MTTR(M), a minimum acceptable MTTR(M min ), and a maximum acceptable MTTR(M max ),where: x ij = (M max −M) ×10 M max −M min (4.150) b) Analysis of the Parameter Profile Matrix The p erformance measures of a system can be described in matrix form in a pa- rameter profile matrix (Thompson et al. 1998). The matrix is compiled containing data points relating to all salient parameters that describe a system’s p erformance. The rows and columns of the matrix can be analysed in order to predict the charac- teristics of the designed system. Figure 3.22 is reproduced below as Fig. 4.13 with a changeto the column heading from process systems to equipment items, as a single system is being considered. Consider one row of the matrix. Each data point x ij refers to a single performance parameter; looking along a row reveals whether the system design is consistently good with respect to this parameter for all the system’s equipment items, or whether there is a variable performance. For a system with a large number of equipment items (in other words, a high level of complexity), a good system design should have a high mean and a low standard deviation of x ij scores for each parameter. These scores are calculated as indicated in Figs. 3.23, 3.24 and 3.25 of Chap. 3, Sect. 3.3.1. Furthermore, a parameter performance index (PPI) that constitutes an 418 4 Availability and Maintainability in Engineering Design Equipment items x 11 x 12 x 13 x 14 x 1i x 21 x 22 x 23 x 24 x 2i Performance x 31 x 32 x 33 x 34 x 3i parameters x 41 x 42 x 43 x 44 x 4i x j1 x j2 x j3 x j4 x ji Fig. 4.13 Parameter profile matrix analysis of the rows of the parameter profile matrix can be calculated(Thompson et al. 1998): PPI = n  n ∑ j= 1 1/c ij  −1 (4.151) where n is the number of design alternatives. The inverse method of calculation of an overall score is advantageous when the range of scores is 0 to 10, as it highlights low scores, whereas a straightforward addition of scores may not reveal a low score if there are high scores in the group. However, the inverse calculation method is less sensitive to error than a multiplica- tion method. The numerical value of PPI lies in the range 0 to 10, no matter how many data points are included in the calculation. Thus, a comparison can be made to judge whether there is acceptable overall performance with respect to all the param- eters, or whether the system design is weak in any respect—particularly concerning the parameters of reliability, inherent availability, and maintainability. A similar calculation to the parameter performance index can be made for each column of the parameter profile matrix, whereby an equipment or device perfor- mance index (DPI) is calculated as (Thompson et al. 1998): DPI = m  m ∑ j= 1 1/c ij  −1 (4.152) where m is the number of performance parameters relating to the equipment item of column j. A comparison of DPIs reveals those equipment items that are contributing less to the overall performance of the system. For an individual equipment item, a good design is a high mean value of the x ij scores with a low standard deviation. This system performance prediction meth od is intended for complex systems compris- ing many sub-systems. Overall system performance can be quite efficiently de- termined, as a wide range of system performance parameters can be included in the PPI and DPI indices. However, in the case of reliability, inherent availability, and maintaina bility, only the two parameter s MTTR an d MTBF are included for prediction, as indicated in Eq. (4.148). From an engineering design integrity point of view, the method collates relevant design integrity data (i.e. reliability, inherent 4.3 Analytic Development of Availability and Maintainability in Engineering Design 419 availability, and maintainability) obtained from expert judgement predictions, and compares these with design performance criteria, requirements and expectations. c) The Design Checklist There are many qualitative factors th at influence r eliability, inherent availability, and maintainability. Some ar e closely related to operability, and ther e are no clear demarcation lines. In order to expand design integrity prediction, a study of the many factors that influence these parameters must be carr ied out. An initial list is first derived in the form of a design checklist. The results of previous research into reliability, availability and maintaina bility problems are carefully considered when devising the checklist questions (McKinney et al. 1989). The checklist is intended for general application, and includes factors affecting design operability. In many cases, there will be questions that do not apply to the design being reviewed, which are then omitted. The questions can be presented to the analyst in the form of a specific knowledge- based expert system within an artificial intelligence-based (AIB) blackboard model for design review during the design process. Results are presented in a format that enables design review teams to collaboratively make reasonable design judgements. This is important, in v iew of the fact that one design team may not carry out the complete design of a particular system as a result of multidisciplinary engineering design requirements, and the design review teams may not all be grouped at one location, prompting the need for collaborative engineering design. This scenario is considered later in greater detail in accounting for various uncertainties involved in system performance simulation modelling for engineering design, utilising the rob ust design technique. Furthermore, knowledge-based expert systems within AIB blackboard models are given in Sect. 3.4, Sect. 4.4, and Sect. 5.4. A typical example of a checklist question set, extending from conceptual to schematic design, is the following: Question set Is pressure release and drainage (including purging and venting) pro- vided? Are purge points considered? If there are no purge points, then this may mean drainage via some or other means that could increase exposure to maintenance per- sonnel requiring the need for protection. i. Purge points not present, requiring some other means 0 ii. Purge points present but accessibility will be poor 1 iii. Purge points present and accessibility will be good 2. A series of questions is posed for each design, and each answer is given a score 0, 1 or 2. The total score is obtained by the summation o f the scores, calculated as a percentage of the total of all the relevant questions. Therefore, questions that do not apply are removed from the analysis. The objective is to obtain overall design integrity ratings for the process and for each device, in order to highlight weak de- sign integrity considerations. A high percentage score indicates good performance 420 4 Availability and Maintainability in Engineering Design where the scores complement the MTTR and MTBF calculations. Where there is a mismatch—for example, a high estimated MTBF but a low reliability score, or a high estimated MTTR but low maintainability score—then fu rther design investi- gation is required. d) Integrity Prediction of Common Items of Equipment The prediction method is intended for those process systems that comprise many well-known items (as the design is still in its concep tual phase). It could be expected that certain item s of equipment may exhibit c ommon maintainability req uirements. In order to save time in data estimates, typical maintenance requirements for com- mon devices are prepared as data sheets. Thus, if a centrifugalpump is selected, then a common data sheet would be avail- able for this item. The data sheet can be reviewed and accepted as it is, or it can be edited to reflect certain particular circumstances, or the checklist can be completed from a blank form to compile a new set of data for that item. In addition to the re- sponses to questions for each individual item, a response to each particular question regarding total systems integration may be considered for all relevant items. For ex- ample, in the case of maintainability, the question might refer to the ability to detect a critical failure condition. If the response to this question is estimated at 60%, then it would suggest that 40% of all items for which the question is relevant would re- main undetected. It is thus possible to review the responses to questions across all the integrated systems, to gain an understanding of the integrity of the conceptual design as a whole. e) Design Reviews of Performance Parameters for System Integrity Design review practices can take many forms. At the lowest level, they consist of an examination of drawings before manufacture begins. More comprehensive de- sign reviews include a review at different phases of the engineering design process: the specification (design requirements), conceptual design, schematic or prelimi- nary design, and detail design. There are particular techniques that the design re- view team implement at these different phases, according to their suitability. The method proposed here is intended for use when very little detail of the equipment is known. In particular, it is intended for use during the conceptual design phase in preparation for the follow-up schematic design phase when systems are synthesised using manufactured equipment. Many engineered installations are designed in this way. Design reviews can be quantitative or qualitative. The advantage of quantita- tive reviews is that they present clearly a justification for a decision that a particular design is either satisfactory or unsatisfactory with respect to essential performance specifications. Therefore, if it is possible, there ar e advantages to a quantitative review as early as possible in the engineering design process. A design review should add value 4.3 Analytic Development of Availability and Maintainability in Engineering Design 421 to each design. Design calculation checks are taken care of by all good, traditional design practices; however, a good design review will be repaid by reduced com- missioning and start-up problems and better ramp-up operations. The design review method proposed here seeks to provide a quantitative evaluation that adds value to engineering design by integrating process performance parameters such as mass flows, pressures and temperatures with reliability, availability and maintainability. Performance data required are the same as those used to specify the equipment. Therefore, there is no requirement for extra data in excess of those that would be available for process systems that comprise many well-known items. The only ad- ditional requirement is a value judgement of acceptable and unacceptable MTBF and MTTR. These data are then compiled into a parameter profile matrix using data points derived from the proximity of a required operating point to the performance limit of the equipment. As indicated previously,the use of the proximity of the nom- inal design performan ce to a limit of equipment capability is similar to the concept of a safety margin. Similarly, the estimated MTBF and MTTR data points reflect the closeness of predicted performance to expectations. Having compiled the ma- trix, analysis can be performed on particular variables for all items of equipment, or on all variables for a particular item of equipment, yielding PPI and DPI indices respectively. On a large engineering design project, data can be generated by several design teams, compiled and analysed using the AIB blackboard model for automated de- sign reviews throughoutthe engineering design process. MTBF and MTTR expecta- tions can be varied in a sensitivity analysis. The computer automated methodology can highlight matrix cells with low scores and pick out performance variables and equipment that show poor performance. Therefore, the data handling and calcula- tion aspects of the design verification d o not impose excessive requirements. The flexibility of the approach, and the meth od of data point derivation are especially useful in process engineering enhancement projects. Inevitably, there are instances when engineered installations are subject to modifications either during construc- tion or even after ramp-up (e.g. there may be advantages in processing at different pressures and/or temperatures), necessitating review of the equipment performance data after design completion. Implications of changes to temperature and pressure requirements can be readily explored, since the parameter profile analysis method will immediately identify when the performance of an item is in close proximity to a limit. Furthermore, engineered installations may be required to process materials that are different to those that they were originally designed to process. As the equip- ment data are already available in the AIB blackboard model, all that is needed is to input new process data for further analysis. The time taken to set up the equipment database during a post-design review will be justified, since the data can be used throughout the life o f the design. Reliability, inherent availability, and maintainabil- ity are included in the parameter pro file matrix evaluation by estimating MTBF and MTTR times, and then calculating the inherent availability. In order to expand on these important variables and to explore the various factors that influence reliability, inheren t availability, and maintainability, checklists are developed that m ay be used 422 4 Availability and Maintainability in Engineering Design in different ways, either on their own or to complement the parameter profile anal- ysis. In a design review, many of the questions may be answered to obtain a view of a system’s integrity, or to obtain an overall view of the integrity of the integrated systems design. Another use for the checklists would be in a process enhancement study whereby an existing engineered installation is audited to identify precisely the improvements required. f) Reliability and Maintainability Checklists The checklists are designed for use by engineering designers in the case of a design review exercise, o r by process engineers in the case of required post-design process enhancements. Although the question sets listed in the AIB blackboard models pre- sented in Sects. 3.4, 4.4 and 5.4 are somewhat different from the example checklists for reliability and maintainability, the relevant pr inciples and intended use remain the same. A segment of an AIB blackboard model Expert System tab page is given below, showing a particular question set. The following question sets indicate the general content of the checklists(Thompson et al. 1998). Reliability checklist Q1. Is the component a single unit, active redundant or stand-by redundant? Q2. Are the demands on the equipment short, medium or long pulses? Q3. Does the duty cycle involve any thermal, mechanical or pressure shocks? Q4. Is the pH of the process fluid high o r low? Q5. Is the component used for high, medium or low flow rates? Q6. Is the component in a physical situation where external corrosive attack can occur from: open weather, other machinery being submerged? Q7. Are solids present in the process fluid? Q8. Are gases present? Q9. Is the fluid of a viscosity that is likely to increase loads on the equipment? . communicating via the internet. 416 4 Availability and Maintainability in Engineering Design 4.3.1 Analytic Development of Availability and Maintainability Prediction in Conceptual Design A technique. engineering design integrity point of view, the method collates relevant design integrity data (i.e. reliability, inherent 4.3 Analytic Development of Availability and Maintainability in Engineering Design. early as possible in the engineering design process. A design review should add value 4.3 Analytic Development of Availability and Maintainability in Engineering Design 421 to each design. Design calculation

Ngày đăng: 02/07/2014, 10:20

Từ khóa liên quan

Tài liệu cùng người dùng

Tài liệu liên quan