1 / 43

Development of Verification and Validation Procedures for Computer Simulation use in Roadside Safety Applications NCHRP

Development of Verification and Validation Procedures for Computer Simulation use in Roadside Safety Applications NCHRP 22-24. VERIFICATION AND VALIDTION METRICS Worcester Polytechnic Institute Battelle Memorial Laboratory Politecnico di Milano. Meeting Agenda.

dinos
Télécharger la présentation

Development of Verification and Validation Procedures for Computer Simulation use in Roadside Safety Applications NCHRP

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Development of Verification and Validation Procedures for Computer Simulation use in Roadside Safety ApplicationsNCHRP 22-24 VERIFICATION AND VALIDTION METRICS Worcester Polytechnic Institute Battelle Memorial Laboratory Politecnicodi Milano

  2. Meeting Agenda 9:00-9:30 Introductions/Instructions (Niessner, Focke) 9:30-10:30 Definitions and V&V Procedures (Ray) 10:30-11:30 ROBUST Project Summary (Anghileri) 11:30-Noon Survey Results Noon-1:00 Lunch 1:00- 2:30 V&V Metrics (Ray) 2:30-4:00 Future Work for Task 8 (Ray)

  3. BACKGROUND • Validation and verification metrics: • Quantify the level of agreement between two solutions. • Calculate the error (i.e., difference) between two solutions. • Specification and use of validation/verification metrics is important because different metrics give different scores for the same time-history pairs. • Need an automated, consistent, repeatable, non-subjective procedure.

  4. PARTS OF A METRIC METRIC PARAMETER ACCEPTANCE CRITERION A V&V metric has two implicit parts. : • The metric parameter is the item that is calculated to compare two results. • It may be: • Domain specific • Shape comparison • It may be: • Deterministic • Probabilistic • The acceptance criterion indicates if the level of agreement between the two results is adequate. • It may be: • Deterministic • Probabilistic (i.e., confidence interval)

  5. Evolution of Metrics Deterministic metric with deterministic acceptance criteria Deterministic metric with probabilistic acceptance criteria Qualitative Comparison Stochastic metric with probabilistic acceptance criteria (from Trucano, et al 2002)

  6. DOMAIN SPECIFIC METRICS A measure of the “reality of interest” in the specific technical domain. • If we are designing an aircraft seat, the head injury criteria (HIC) of the occupant is in our reality of interest. • The HIC is an appropriate domain specific metric for this problem. • If we are designing an aircraft wing, the wing tip deflection is in our reality of interest. • Wing-tip deflection is an appropriate domain specific metric for this problem.

  7. DOMAIN SPECIFIC VALIDATION METRICS • We already have well defined evaluation criteria for roadside safety crash tests: • Report 350 • Report 350 Update • EN 1317 • Which criteria would make good domain specific metrics for validation and verification? • Some criteria are numerical (i.e., H and I) • Some criteria are phenomological (i.e., C, D, F, G, K and L) • We can evaluate a simulation using the same domain specific metrics as we use in crash tests. Comparison Report 350 evaluation criteria matrix for test 3-31 on the Plastic Safety System CrashGard Sand Barrel.

  8. RECOMMENDATION • We should use the same evaluation criteria for comparing the results of full-scale crash tests to simulations that we already use to evaluate full-scale crash tests. • These would be deterministic domain specific metrics. • Validation – for comparing to experiments. • Verification – for comparing to known solutions (e.g., benchmarks) • Discussion? Comparison Report 350 evaluation criteria matrix for test 3-31 on the Plastic Safety System CrashGard Sand Barrel.

  9. ACCEPTANCE CRITERIA • For phenomenological criteria (i.e., A, B, C, D, E, F, G, K and N), the acceptance criteria is binary – YES or NO. • Examples • In a simulation the vehicle penetrates the barrier but in the crash test the vehicle did not penetrate. • Simulation Criterion A: Pass • Crash Test Criterion A: Fail • Valid? NO • In both the simulation and validation crash test, the vehicle rolls over as it is redirected. • Simulation Criterion G: Fail • Crash Test Criterion G: Fail • Valid? YES • For numerical criteria (i.e., H, I, L and M) we could pick a criteria based on (1) engineering judgment or (2) an uncertainty analysis. • Example: Assume our acceptance criterion is that the difference must be less than 20 percent. • Criterion H in the table is satisfied so the comparison is valid. • If the acceptance criterion were 10 percent, criterion H would be invalid. Comparison Report 350 evaluation criteria matrix for test 3-31 on the Plastic Safety System CrashGard Sand Barrel.

  10. DOMAIN SPECIFIC VERIFICATION METRICS • Verification is the comparison of simulation results to a known solution. • In roadside safety we do not really have known solutionsbut… • Simulation must obey some basic laws of mechanics: • Conservation of energy, mass and momentum. • Roadside safety verification metrics: • Total energy must be conserved ± 10. • Energy in each part must balance ± 10. • Hourglass energy in each part must be less than ± 10. • Time history sampling rate must ensure that the time histories yield correct data (i.e., acceleration integrate to correct velocity). • Mass scaling must not affect the finial solution. • These verification checks ensure that the simulation results at least conform to the basic physics represented by the model.

  11. SHAPE COMPARISON METRICS • Frequency domain metrics • Point-wise absolute difference of amplitudes of two signals. • Root-mean-squared (RMS) log spectral difference between two signals. • Time domain metrics • Relative absolute difference of moments of the difference between two signals • Root mean square (RMS) log measure of the difference between two signals • Correlation Coefficient • Geer MPC Metrics • Geer • Sprague-Geer • Knowles-Gear • Russell Metrics • ANOVA Metrics • Ray • Oberkampf and Trucano • Global Evaluation Method (GEM) • Velocity of the Residual Errors • Weighted Integrated Factor (WIFac) • Normalized Integral Square Error (NISE) For dynamic problems it is also useful to compare time histories. How do we quantify the similarity/difference between time history data? How do we compare shapes?

  12. SHAPE COMPARISON METRICS • We can group these into similar metrics: • NARD Metrics • Point-wise absolute difference of amplitudes of two signals. • Root-mean-squared (RMS) log spectral difference between two signals. • Relative absolute difference of moments of the difference between two signals • Root mean square (RMS) log measure of the difference between two signals • Correlation Coefficient • MPC (magnitude-phase-combined) Metrics • Geer • Sprague-Geer • Knowles-Gear • Russell Metrics • ANOVA Metrics • Ray • Oberkampf and Trucano • Velocity of the Residual Errors • Global Evaluation Method (GEM) • Weighted Integrated Factor (WIFac) • Normalized Integral Square Error (NISE) Nard frequency domain metrics very rarely used. Nard time domain metrics seldom used. Several are components of the MPC metrics so using them is redundant. All four MPC metrics are similar. For illustration we will use Sprauge-Geer MPC metrics. Ray and Oberkampf methods almost identical. GEM is part of a proprietary software package so details are difficult to know. Use Oberkampf’s version for illustration.

  13. SHAPE COMPARISON METRICSSprauge-Geers MPC Metrics The magnitude component measures relative differences in magnitude between the computational (i.e., c) and measured (i.e., m) responses. The phase component (i.e., P) measures relative differences in phase or time of arrival between the computational (i.e., c) and measured (i.e., m) responses. The combined component combines the affect of P and M into one parameter. P and M are coordinates on a circle of radius C. 0 ≤ M, P and C ≤ 1.0 Perfect agreement is M=P=C=0.0 No relation is M=P=C=1.0

  14. SHAPE COMPARISON METRICSANOVA Metrics This is the average residual error between the measured (i.e., m) and computed (i.e., c) signals at each time step. This is the standard deviation of the residual errors. If we know the mean and standard deviation and we assume error is random (i.e., normally distributed) we can perform a standard statistical hypothesis test. Calculate the T statistic for confidence level α and N samples. For crash tests, N is always very large -- effectively ∞. This is a very well established statistical technique applied to error estimation. Identical curves would give a mean, standard deviation and T score of zero. All three scores run between zero (best) and one (worst) like the MPC metrics.

  15. SHAPE COMPARISON METRICS • All the time-domain metrics are point-to-point comparisons. It is important that the measured (i.e., m) and calculated (i.e., c) data should: • Be collected at the same time interval so each m and c value matches in time. • This is highly desirable although not totally essential. Values can be interpolated in order to accomplish the same thing but this introduces another source of random error. • Be filtered in the same way to avoid numerical noise from giving in-accurate results • Highly desirable but not totally essential. If the data are filtered differently there may be poor magnitude or error comparisons that are a result of the filtering rather than the shape comparison. • Be shifted so that the impact time is coordinated. • Highly desirable but not totally essential. There are methods to account for a time shift (i.e., Knowles-Gear accounts for this internally).

  16. SHAPE COMPARISON METRIC EVALUATIONS Three studies comparing different shape comparison metrics were found: • Ray (1997) • Plaxico, Ray & Hiranmayee(2000) • Compared: • Geer MPC • Relative Moments • Correlation Coefficient • Log RMS • Log spectral distance • Ray ANOVA • Velocity of Residuals • Selected: • Geer MPC • Ray ANOVA • Velocity of Residuals • Schwer (2004) • Schwer (2006) • Compared: • - Sprauge-Geer MPC • - Russell MPC • Oberkampf ANOVA • RMS • Selected: • - Sprauge-Geer MPC • - Knowles-Gear MPC • Oberkampf ANOVA • Moorecroft (2007) • Compared: • Sprauge-Geers MPC • WiFAC • GEM • NISE • Selected: • -Sprauge-Geers MPC All three recommended: - One of the Geer MPC metrics - One of the ANOVA metrics

  17. SHAPE COMPARISON METRIC EVALUATIONRay et al Ray used the NARD metrics, the ANOVA metrics and the Geer MPC metrics to six “identical” rigid pole crash tests performed by Brown.

  18. SHAPE COMPARISON METRIC EVALUATIONRay (1997) • Based on the ANOVA results Ray suggested: • e < 5 % • σ < 20 % • V < 10 % 0th though 2nd moments less than 20 % 3rd though 5th moments do not seem to be discerning. • RMS > 70 % ? • ρ > 0.90 • ΔAB < 0.35 ? • D ? M, P and C < 10 %

  19. SHAPE COMPARISON METRIC EVALUATIONSchwer • Schwer examined four metrics: • Oberkampf’s version of the ANOVA metrics • Sprauge-Geers MPC metrics • Russell metrics • Root-mean square (RMS) metric • Results should be consistent with Subject Matter Expert (SME) opinions. • Increased use of metrics, but selection rationale is rarely specified.

  20. SHAPE COMPARISON METRIC EVALUATIONSchwer • Schwer examined: • An ideal decaying sinusoidal wave form with: • +20% magnitude error • -20% phase error • +20% phase error • An example experimental time history with five simulations:

  21. SHAPE COMPARISON METRIC EVALUATIONSchwer • Magnitude comparison: • All but the Russell metric are biased toward the measured value (this is good). • Sprauge-Geer and RMS give values similar to the actual error. • Phase comparisons: • All four metrics are symmetric with respect to phase. • RMS and ANOVA are very sensitive to the phase.

  22. SHAPE COMPARISON METRIC EVALUATIONSchwer • The Oberkampf ANOVA metric has some interesting features: • Plotting the residuals shows exactly where in the time history the error is largest. • Top curve: • Initial peak has large error • Tail of the curve is very prone to error • Middle of the curve is pretty good • Bottom curve: • Initial peak has a lot of error. • Initially relaxes to a more reasonable error. • Error gets progressively worse as the event goes on.

  23. SHAPE COMPARISON METRIC EVALUATIONSchwer Schwer also sent five time history comparisons to 11experts and asked them to rate the agreement between zero and unity (unity being the best).

  24. SHAPE COMPARISON METRIC EVALUATIONSchwer • The subject matter expert results agree with both the Sprauge-Geers and Knowles-Gear metrics. • The expert results generally have the same range (note the error bars) and the metrics tend to be positioned in the mid-range of expert opinions. • Both the Sprauge-Geers and Knowles-Gear metrics appear to be good predictors of expert opinion.

  25. Sprague and Geers (S&G) General purpose curve shape metric Implemented in a spreadsheet Weighted Integrated Factor (WIFac) Automotive curve shape metric Implemented in a spreadsheet Global Evaluation Method (GEM) Automotive curve shape plus peak and timing Requires ModEval (stand alone program) Normalized Integral Square Error (NISE) Biomechanics curve shape plus magnitude and phase Implemented in a spreadsheet SHAPE COMPARISON METRIC EVALUATIONMoorecroft Four Curve Shape Metrics Evaluated

  26. SHAPE COMPARISON METRIC EVALUATIONMoorecroft 20% Magnitude Error Only +/- 20% Phase Error Only Head Acceleration Moorecroft examined: Three ideal wave forms: +20% magnitude error -20% phase error +20% phase error And a typical head accel. curve.

  27. SHAPE COMPARISON METRIC EVALUATIONMoorecroft Three Ideal Waveforms and Head Acceleration Sprauge-Geers is a direct measure of the error and works well for phase and magnitude. WIFac is more sensitive to phase errors than magnitude. NISE is more sensitive to phase errors than magnitude. GEM is more sensitive to magnitude than phase.

  28. SHAPE COMPARISON METRIC EVALUATIONMoorecroft Subject Mater Expert Survey • 16 experts (industry, gov’t, academia) submitted evaluations of 39 test/simulation time history curves. • Evaluations consisted of a score (excellent, good, fair, poor, very poor) for magnitude, phase, shape, and overall agreement. • The data represent accel, vel, pos, angle, force, and moment time histories derived from both occupant & structural responses. • Data normalized such that highest peak = 1.

  29. SHAPE COMPARISON METRIC EVALUATIONMoorecroft Example Curve (Pair 19/SME 1)

  30. SHAPE COMPARISON METRIC EVALUATIONMoorecroft Mag. % Error vs. SME Mag. Score

  31. Phasing Average Difference

  32. Metric Avg. vs. SME Shape

  33. SHAPE COMPARISON METRIC EVALUATIONMoorecroft Curve Shape Results • S&G most closely reproduced the reference errors for idealized waveforms. • S&G and GEM performed best in the discrimination evaluation. • S&G, GEM, and NISE were all consistent with the SME evaluations. • Curve shape error matched the error ranges suggested from the magnitude data.

  34. SHAPE COMPARISON METRIC EVALUATIONMoorecroft Moorecroft’s Curve Shape Recommendations • Simple, deterministic metric. • Easy to implement in a spreadsheet. • Limited number of seat tests. • Individual error scores should be consistent. • i.e., 10% is “good” for all features (e.g., M, P and C). • Error metric biased towards the experiment. • Consistent with certification activities. • Appropriate results for idealized curves. • Metric results consistent with SME values. • Sprague & Geers metric meets these specifications and appears to be the best choice for validating numerical seat models.

  35. Recommendation Adopt the Sprauge-Geers MPC metrics and the ANOVA metrics as the shape comparison metrics for both validation and verification. Sprauge-Geers ANOVA

  36. Recommendation Adopt the Sprauge-Geers MPC metrics and the ANOVA metrics as the shape comparison metrics for both validation and verification.

  37. Phenomena Importance Ranking Tables • A PIRT is a table that lists all the phenomena that a model is expected to replicate. • Each phenomena in the PIRT should be validated or verified as appropriate. • There will be a PIRT for: • Each vehicle type in Report 350 (i.e., 820C) • Each type of test (i.e., 3-10)

  38. Roadside Safety Simulation PIRTSExample: 820C Model

  39. Roadside Safety Simulation PIRTSExample: 820C Model We can have more general confidence in models that have a higher level of validation – more specific phenomena have been represented.

  40. Roadside Safety Simulation PIRTSExample: Guardrail We can have more general confidence in models that have a higher level of validation – more specific phenomena have been represented.

  41. Roadside Safety Simulation PIRTSExample: 820C Model We can have more general confidence in models that have a higher level of validation – more specific phenomena have been represented.

  42. Roadside Safety Simulation PIRTSExample: 820C Model We can have more general confidence in models that have a higher level of validation – more specific phenomena have been represented.

More Related