210 likes | 367 Views
VTT-STUK assessment method for safety evaluation of safety-critical computer based systems - application in BE-SECBS project. CONTENTS 1. Method description 2. Qualitative assessment 3. Quantitative evaluation 4. Conclusions. 1. METHOD DESCRIPTION. 1. METHOD DESCRIPTION
E N D
VTT-STUK assessment method for safety evaluation of safety-critical computer based systems - application in BE-SECBS project
CONTENTS 1. Method description 2. Qualitative assessment 3. Quantitative evaluation 4. Conclusions
1. METHOD DESCRIPTION Quantitative analysis 1. Development of the map of evidence • identification of pieces of evidence to be included in the model (actually a result of qualitative safety evaluation) • identification of the relationship between the pieces of evidence by engineering judgement • identification of the relationship between the evidence and the failure probability of the system
1. METHOD DESCRIPTION Quantitative analysis 2. Definition the structure of the Bayes network model • definition of the variables (or the nodes) of the model; i.e. the definition of the variables measuring the degree of quality for evidence analysed in the tasks of the qualitative analysis • definition of the measurement of rating scales for each variable (usually a discrete or ordinal scale), by using expert judgement • definition of the probabilistic relationships and dependencies between the variables (e.g. the relationship between coverage of testing and failure probability) by expert judgement
1. METHOD DESCRIPTION Quantitative analysis 3 Quantification of the model • quantification of the variable ratings by expert judgement • quantification of the needed probability distributions by expert judgement (i.e. quantification of the weight of different pieces of evidence) • propagation of uncertainties through the Bayes network model
1. METHOD DESCRIPTION Quantitative analysis 4 Interpretation of the results • sensitivity analyses • importance analyses and analysis of need for further data or information
2 QUALITATIVE ASSESSMENT • the evidence from each life cycle phase was analysed qualitatively on the basis of the material provided by the vendor • additional information was asked during two meetings with the vendor • no formal/computer-based tools were used • the material was compared with the requirements of the STUK regulatory guide YVL-5.5 • the aim of the qualitative analysis was to create “a map of evidence” • evidence from platform development process was analysed separatelly
2 QUALITATIVE ASSESSMENT Analysis of the platform software • Three types of information was introduced to the assessment teams: • Overall documentation of the platform • Reference list of the operational experience • References to the type tests
2 QUALITATIVE ASSESSMENT Analysis of the platform software • Analysis of operating experience • Detailed knowledge of the operational experience of the software was presented to the assessment team. Additional information was obtained during assessment meetings • Operational experience of the platform is stored in the database with the developmental proposals and requests. The vendor was asked to demonstrate the use of database, detailed analyses were not possible
2 QUALITATIVE ASSESSMENT Analysis of the platform software • Analysis of type-testing documentation • the quality oftype-testing documentation was evaluated (partially with the vendor personnel) • Analysis of development tools
2 QUALITATIVE ASSESSMENT The analysis of life-cycle phases analysed • requirement specification • concepts design (in documentation called also as system specification) • detailed design • code generation • SIVAT simulation testing • Code compiling and linking • Testing
2 QUALITATIVE ASSESSMENT Analysis of the equirement specification • critical analysis of documentation • analysis of different blocks used in the block diagram presentation • analysis with respect to IEEE 830 • relevant events, response times, input and output signals,relation to other safety functions, signal identification scheme • analysis of redundancy issues • specification of system states • ambiguity of notation
2 QUALITATIVE ASSESSMENT Analysis of the application software development process • definition of the application development lifecycle • assessment was made against software engineering references • the end product is software and the development propcess is analogous to the software engineering process • software engineering standards offer rigorous references to the assessment purposes (IEC60880) • existence of V&V checklists etc. • existence of quality targets • existence of quality guidelines
2 QUALITATIVE ASSESSMENT Analysis of concept design • dependence of the concept design on the platform: description consequences of the assumptions about the platform properties • documentation of the selection of design solutions • documentation of the test plan Analysis of detailed design • documentation of design solutions • documentation of the verification procedures
2 QUALITATIVE ASSESSMENT Analysis of code implementation and generation • analysis of resource metrics • analysis of simulation testing • analysis of the possibilities of engineers to influence for the end result C-code simulation testing • analysis of documentation practices and test strategy • functional coverage of testing Code compiling and linking • tools and their operational experience
2 QUALITATIVE ASSESSMENT Analysis of testing • Although actual test results does not exist for the benchmark, information was gathered about the ‘typical’ test execution for assessment purposes • test strategy, test acceptance criteria, test coverage analysis
3 QUANTITATIVE EVALUATION • the "Bayes network model" is based on the life-cycle of the system • the model was created as a expert judgement process • the quality characteristics were quantified in expert panels • the quantification is based on the observations from the qualitative analysis
2 QUANTITATIVE EVALUATION • a “good standard” for rating the steps of the life cycle is needed • the reliability estimates were not determined • not enough information • lot of uncertainties => difficult to interprete the evidence => difficult to make probability estimates
4, CONCLUSIONS • the analysis of the (system and) application development was mainly qualitative • no software based tools were used • a map of evidence was created (as a Bayes network) • makes it possible to see the relationships between different pieces of evidence • the quality of the requirement specification is important in the case of automated code generation • quantitative analysis was experimented