XRF and Appropriate Quality Control u CLUIN Studios

  • Slides: 44
Download presentation
XRF and Appropriate Quality Control u CLU-IN Studios Web Seminar u August 18, 2008

XRF and Appropriate Quality Control u CLU-IN Studios Web Seminar u August 18, 2008 u u Stephen Dyment USEPA Technology Innovation Field Services Division dyment. stephen@epa. gov 1

How To. . . u Ask questions » “? ” button on CLU-IN page

How To. . . u Ask questions » “? ” button on CLU-IN page u Control slides as presentation proceeds » manually advance slides u Review archived sessions » http: //www. clu-in. org/live/archive. cfm u Contact instructors 2

Q&A For Session 4 – DMA 3

Q&A For Session 4 – DMA 3

What Can Go Wrong with an XRF? u Initial or continuing calibration problems u

What Can Go Wrong with an XRF? u Initial or continuing calibration problems u Instrument drift u Window contamination u Interference effects u Matrix effects u Unacceptable detection limits u Matrix heterogeneity effects u Operator errors 4

Within-sample variability can impact data quality more than the analytical method (continued) 5 -5

Within-sample variability can impact data quality more than the analytical method (continued) 5 -5

Within-sample variability can impact data quality more than the analytical method 5 -6

Within-sample variability can impact data quality more than the analytical method 5 -6

95% Confidence Interval (CI) Bar Graphs Sometimes ICP is more precise… Arsenic CI Graph

95% Confidence Interval (CI) Bar Graphs Sometimes ICP is more precise… Arsenic CI Graph = 294 …sometimes XRF is = 203 Lead CI Graph = 174 = 137 Triplicate subsamples of the same (<10 mesh) sample analyzed by both ICP & XRF CI results for As & Pb shown here 5 -7

95% CI Bar Graphs for Another Sample Sometimes XRF is more precise… Arsenic CI

95% CI Bar Graphs for Another Sample Sometimes XRF is more precise… Arsenic CI Graph = 129 …sometimes ICP is = 91 Lead CI Graph = 864 Here the situation is reversed for another sample from the same yard. = 538 5 -8

95% CI Bar Graphs for Particle Size Effects Particle Sizes Ave = 533 Ave

95% CI Bar Graphs for Particle Size Effects Particle Sizes Ave = 533 Ave = 397 Ave = 548 = 438 = 129 all <10 mesh = crushed soil put thru 10 -mesh sieve & analyze all going thru 10 -60 mesh = above then put thru 60 -mesh sieve & what is retained on 60 mesh is analyzed Ave = 375 Ave = 314 Ave = 392 = 129 <60 mesh = that going thru 60 mesh sieve & is analyzed 5 -9

What Can We Do? (continued) 10

What Can We Do? (continued) 10

What Can We Do? u Recognize the advantages and limitations of XRF and laboratory

What Can We Do? u Recognize the advantages and limitations of XRF and laboratory methods u Recognize that uncertainty exists u Perform a demonstration of method applicability study (DMA) u Structure your QA/QC program to adaptively manage uncertainty u Use collaborative data sets- powerful weight of evidence and take advantage of both methods 11

NIST – Recognition of Variability u Certified concentrations based on two or more independent

NIST – Recognition of Variability u Certified concentrations based on two or more independent methods requiring complete sample decomposition or nondestructive analysis u Some of the most homogenous and well characterized material out there u Yet. . . . NIST 2709 Certified Values 12

NIST – Recognition of Digestion Issues See NIST 2709, 2710, 2711 Addendums 13

NIST – Recognition of Digestion Issues See NIST 2709, 2710, 2711 Addendums 13

Your Quality Control Arsenal. . . Weapons of Choice. . . u Energy calibration/standardization

Your Quality Control Arsenal. . . Weapons of Choice. . . u Energy calibration/standardization checks u NIST-traceable standard reference material (SRM), preferably in media similar to what is expected at the site u Blank silica/sand u Well-characterized site samples u Duplicates/replicates u In-situ reference location u Matrix spikes u Examination of spectra 14

Standardization or Energy Calibration u Completed upon instrument start-up or when instrument identifies significant

Standardization or Energy Calibration u Completed upon instrument start-up or when instrument identifies significant drift u X-rays strike stainless steel plate or window shutter (known material) u Instrument ensures that expected energies and responses are seen u Follow manufacturer recommendations (typically several times a day) 15

Initial Calibration Checks u Calibration SRMs and SSCS typically in cups u Perform multiple

Initial Calibration Checks u Calibration SRMs and SSCS typically in cups u Perform multiple (at least 10) repetitions of measuring a cup, removing the cup, and then placing it back for another measurement u Compare observed standard deviation in results with average error reported by instrument u Compare average result with standard’s “known” concentration u Use observed standard deviation for evaluating controls for on-going calibration checks (DMA) 16

Initial Calibration Check Example Known Sample # of Measurements U Reported Moly U Moly

Initial Calibration Check Example Known Sample # of Measurements U Reported Moly U Moly Si. O 2 Blank 1 <LOD 50 ppm U 3 50 NA <LOD 14 150 ppm U 3 150 NA 116 23 50 ppm Moly 3 NA 50 55 42 150 ppm Moly 3 NA 150 <LOD 134 100 ppm U/Moly 6 100 68 112 10 100 NA 230 21 Archived Site Sample 17

Continuing Calibration Checks u At least twice a day (start and end), a higher

Continuing Calibration Checks u At least twice a day (start and end), a higher frequency is recommended u Frequency of checks is a balance between sample throughput and ease of sample collection or repeating analysis u Use a series of blank, SRMs, and SSCS u Based on initial calibration check, how is XRF performing? u Watching for on-going calibration check results that might indicate problems or trends » Typically controls set up based on DMA and initial calibration check work (i. e. , a two SD rule) 18

Control Charting Your Continuing Calibration Checks 19

Control Charting Your Continuing Calibration Checks 19

Continuing Calibration Checks Example of What to Watch for… u u u Two checks

Continuing Calibration Checks Example of What to Watch for… u u u Two checks done each day, start and finish 150 ppm standard, w/ approx. +/- 9 ppm for 120 second measurement Observed standard deviation in calib check data: 18 ppm Average of initial check: 153 ppm Average of ending check: 138 ppm 20

Interference Effects u Spectra too close for detector to accurately resolve u Result: biased

Interference Effects u Spectra too close for detector to accurately resolve u Result: biased estimates for one or more quantified elements u DMA, manufacturer recommendations, scatter plots used to identify conditions when interference effects would be a concern u “Adaptive QC”…selectively send samples for confirmatory laboratory analysis when interference effects are a potential issue 21

Potential Interferences Element ke. V Shell Region (ke. V) Min Possible Interference Max P

Potential Interferences Element ke. V Shell Region (ke. V) Min Possible Interference Max P 2. 02 Ka 1. 85 2. 2 Y, Zr, Nb, S S 2. 31 Ka 2. 2 2. 45 Nb, Mo, Tc, P, Cl Cl 2. 62 Ka 2. 45 2. 82 Tc, Ru, Rh K 3. 31 Ka 3. 12 3. 42 Ag, Cd, In Ca 3. 69 Ka 3. 54 3. 87 Sb, Te Ti 4. 51 Ka 4. 3 4. 7 Sc, Cs, Ba Ba 4. 83 Lb 4. 7 5. 1 V, Ti, La, Pr Cr 5. 41 Ka 5. 2 5. 6 Pm, V Mn 5. 9 Ka 5. 7 6. 1 Eu, Cr Fe 6. 4 Ka 6. 2 6. 6 Gd, Tb, Mn Co 6. 93 Ka 6. 73 7. 13 Er, Ho, Fe Ni 7. 48 Ka 7. 28 7. 68 Yb, Ho Cu 8. 05 Ka 7. 85 8. 25 Hf, Tm Zn 8. 64 Ka 8. 44 8. 84 Re, Lu W 9. 67 Lb 9. 5 9. 9 Au, Ge Hg 9. 99 La 9. 8 10. 2 Re, Ge, As As 10. 54 Ka 10. 3 10. 7 Pb (continued) 5 -22

Potential Interferences Element ke. V Shell Region (ke. V) Min Possible Interference Max Se

Potential Interferences Element ke. V Shell Region (ke. V) Min Possible Interference Max Se 11. 22 Ka 11 11. 4 Po Au 11. 44 Lb 11. 3 11. 675 At Br 11. 92 Ka 11. 7 12. 1 Hg, Fr Pb 12. 61 Lb 12. 4 12. 8 Ac, Kr, Se Bi 13. 02 Lb 12. 85 13. 15 Rb 13. 39 Ka 13. 2 13. 6 U 13. 61 La 13. 5 13. 9 Sr 14. 16 Ka 13. 9 14. 3 Zr 15. 77 Ka 15. 4 16 Mo 17. 48 Ka 17. 1 17. 7 Ag 22. 16 Ka 21. 804 22. 404 Cd 23. 17 Ka 22. 81 23. 14 Sn 25. 27 Ka 24. 894 25. 494 Sb 26. 36 Ka 25. 974 26. 574 Pa, Po, Br Kr Sr, Cf, Ac 23

5 -24 Periodic Table Version 24

5 -24 Periodic Table Version 24

Lead/Arsenic Interference Example (continued) 25

Lead/Arsenic Interference Example (continued) 25

Lead/Arsenic Interference Example Pb = 3, 980 ppm Pb = 3, 790 ppm 26

Lead/Arsenic Interference Example Pb = 3, 980 ppm Pb = 3, 790 ppm 26

Arsenic in the Presence of Lead One Vendor’s Answer Algorithm predicts lead Lα in

Arsenic in the Presence of Lead One Vendor’s Answer Algorithm predicts lead Lα in 10. 5 ke. V spectral region based on the “clean” lead Lβ signal. The lead contribution is subtracted leaving the arsenic Kα. 27

The Skeptical Chemist…. u Difficulty in resolving As concentrations when Pb was greater than

The Skeptical Chemist…. u Difficulty in resolving As concentrations when Pb was greater than 10 X the As u “ 10 Times Rule” empirical rule of thumb » “J” any XRF detected values for arsenic below 1/10 of the lead value u Example » Pb detected at 350 + 38 ppm — As detected at 28 + 6 ppm report as estimated “J” — As detected at 48 + 10 ppm would not require a “J” 28

Monitoring Detection Limits u Detection limits for XRF are not fixed for any particular

Monitoring Detection Limits u Detection limits for XRF are not fixed for any particular element u Measurement time, matrix effects, the presence of elevated contaminants…all have an impact on measurement DL u Important to monitor detection limits for situations where they become unacceptable and alternative analyses are required 29

As an Example…. Chemical Abstract Series Number Innov-X 1 120 sec acquisition (soil standard

As an Example…. Chemical Abstract Series Number Innov-X 1 120 sec acquisition (soil standard – ppm) Innov-X 1 120 sec acquisition (alluvial deposits - ppm) Innov-X 1 120 sec acquisition (elevated soil - ppm) Antimony (Sb) 7440 -36 -0 61 55 232 Arsenic (As) 7440 -38 -0 6 7 29, 200 Barium (Ba) 7440 -39 -3 NA NA NA Cadmium (Cd) 7440 -43 -9 34 30 598 Calcium (Ca) 7440 -70 -2 NA NA NA Chromium (Cr) 7440 -47 -3 89 100 188, 000 Cobalt (Co) 7440 -48 -4 54 121 766 Copper (Cu) 7440 -50 -8 21 17 661 Iron (Fe) 7439 -89 -6 2, 950 22, 300 33, 300 Lead (Pb) 7439 -92 -1 12 8 447, 000 Manganese (Mn) 7439 -96 -5 56 314 1, 960 Mercury (Hg) 7439 -97 -6 10 8 481 Molybdenum (Mo) 7439 -93 -7 11 9 148 Nickel (Ni) 7440 -02 -0 42 31 451 Potassium (K) 7440 -09 -7 NA NA NA Analyte 5 -30

Monitoring Dynamic Range u Periodic, in response to XRF results exhibiting characteristics of concern

Monitoring Dynamic Range u Periodic, in response to XRF results exhibiting characteristics of concern (e. g. , contaminants elevated above calibration range of instrument)…sample sent for confirmatory analysis » Is there evidence that the linear calibration is not holding for high values? » Should the characteristics used to identify samples of concern for dynamic range effects be revisited? 31

Matrix Effects u In-field use of an XRF often precludes thorough sample preparation u

Matrix Effects u In-field use of an XRF often precludes thorough sample preparation u This can be overcome, to some degree, by multiple XRF measurements systematically covering “sample support” surface u What level of heterogeneity is present, and how many measurements are required? u “Reference point” for instrument performance and moisture check with in-situ applications 32

Worried About Impacts From Bags? u We’ve evaluated a variety of bags and found

Worried About Impacts From Bags? u We’ve evaluated a variety of bags and found little impact u Analyze a series of blank, SRM, and SSCS by analyzing replicates or repetitions through the bag u Exceptions include bags with ribs and highly dimpled, damaged, creased bags » Result in elevated DLs, reported errors 33

Examination of Spectra u Spectral response is actually a range in the 100’s of

Examination of Spectra u Spectral response is actually a range in the 100’s of electron volts u Resolution of latest detectors <190 -230 e. V u Older models ~280 -300 e. V u Can use spectra to evaluate high NDs or errors for target metals 34

Controlling Sample Heterogeneity Estimating Measurement Number Requirements u Goal: reducing error due to heterogeneity

Controlling Sample Heterogeneity Estimating Measurement Number Requirements u Goal: reducing error due to heterogeneity to at least 30% and at most comparable to analytical error at action level (<10%) u Ten measurements from a “sample support” at action level to estimate variability u Aggregate error = st. dev. /sqrt(n) where n is the number of samples contributing to aggregate u Typically takes from four to sixteen aggregated measurements to achieve 35

Arsenic Concentration ppm Data Set 1 74 265 38 124 399 58 89 17

Arsenic Concentration ppm Data Set 1 74 265 38 124 399 58 89 17 41 103 203 94 78 82 22 117 155 Std Dev 29. 59 Data Set 2 Ag Error 9. 36 Std Dev 127. 01 Ag Error 40. 16 5 -36

Assessing Instrument Precision Duplicates and Replicates 37

Assessing Instrument Precision Duplicates and Replicates 37

Field Based Action Levels Part of Your QC Program? u As part of your

Field Based Action Levels Part of Your QC Program? u As part of your QC program you will likely choose collaborative samples » Helpful to ensure the quality of XRF decisions, monitor predictive relationships u Some samples are obvious » Spectral interferences, problematic matrices, samples outside instrument linear range, etc. u Maximize the value of these samples » Focus around action levels “too close to call” » Some high, some low values » Watch decision error rates 38

Developing Predictive Relationships u. Non-parametric techniques allow focus on the decision: What is the

Developing Predictive Relationships u. Non-parametric techniques allow focus on the decision: What is the probability of being above or below regulatory requirements? u. No significant statistical assumptions being made (e. g. , normality). u. Results are not affected by outliers and/or non-detects. Proposed Field-Action Level action decisions are yes/no decisions. Non-Definitive Technique u. Most remedial “False” Pos Decision Error “True” Neg Decision “True” Pos Decision The ideal = the perfect parametric regression line “False” Neg Decision Error Reality = non-parametric: count how many occur in each category Regulatory Action Level “Definitive” Technique 5 -39

Field Based Action Levels 59 Total pairs 3 False Positive Errors=7. 7% True Positive

Field Based Action Levels 59 Total pairs 3 False Positive Errors=7. 7% True Positive 19 Pairs 1 False Negative Error= 5% True Negative 36 Pairs (continued) 5 -40

Field Based Action Levels 59 Total pairs 10 False Positive Errors= 26% True Negative

Field Based Action Levels 59 Total pairs 10 False Positive Errors= 26% True Negative 29 Pairs True Positive 20 Pairs 0 False Negative Error= 0% (continued) 5 -41

Field Based Action Levels 3 Way Decision Structure With Region of Uncertainty 59 Total

Field Based Action Levels 3 Way Decision Structure With Region of Uncertainty 59 Total pairs 3 False Positive Errors=7. 7% True Positive 19 Pairs 11 Samples for ICP True Negative 26 Pairs 0 False Negative Error= 0% 5 -42

Questions? 43

Questions? 43

Thank You After viewing the links to additional resources, please complete our online feedback

Thank You After viewing the links to additional resources, please complete our online feedback form. Thank You Links to Additional Resources Feedback Form 5 -44