Advisory Committee on Nuclear Waste 133rd Meeting, March 19, 2002
Official Transcript of Proceedings
NUCLEAR REGULATORY COMMISSION
Title: Advisory Committee on Nuclear Waste
133rd Meeting
Docket Number: (not applicable)
Location: Rockville, Maryland
Date: Tuesday, March 19, 2002
Work Order No.: NRC-283 Pages 1-116
NEAL R. GROSS AND CO., INC.
Court Reporters and Transcribers
1323 Rhode Island Avenue, N.W.
Washington, D.C. 20005
(202) 234-4433. UNITED STATES OF AMERICA
NUCLEAR REGULATORY COMMISSION
+ + + + +
ADVISORY COMMITTEE ON NUCLEAR WASTE (ACNW)
133RD MEETING
+ + + + +
TUESDAY,
MARCH 19, 2002
+ + + + +
ROCKVILLE, MARYLAND
+ + + + +
The meeting commenced at 10:00 a.m. in
Room T2B3, Two White Flint North, Rockville,
Maryland, George M. Hornberger, Chairman, presiding.
COMMITTEE MEMBERS PRESENT:
GEORGE M. HORNBERGER, Chairman
RAYMOND G. WYMER, Vice Chair
B. JOHN GARRICK, Member
MILTON N. LEVENSON, Member
STAFF PRESENT:
JOHN T. LARKINS, Executive Director, ACRS-ACNW
SHER BADAHUR, Association Director, ACRS-ACNW
HOWARD J. LARSON, Special Assistant, ACRS-ACNW
LYNN DEERING, ACNW Staff
LATIF HAMDAN, ACNW Staff
MICHAEL LEE, ACNW Staff
RICHARD K. MAJOR, ACNW Staff
WILLIAM HINZE, ACNW Staff
CAROL A. HARRIS, ACRW/ACNW Staff
RICHARD P. SANVIO, ACRS/ACNW Staff
Also Present:
CAROL HANLON, DOE
PETER SWIFT, Bechtel SAIC
WILLIAM BOYLE, DOE
. I-N-D-E-X
Opening Statement. . . . . . . . . . . . . . . . . 3
ACNW Planning and Procedures
Update on DOE Performance Assessment
Program (BJG/MPL). . . . . . . . . . . . . . 6
Carol Hanlon . . . . . . . . . . . . . . . . 6
Peter Swift. . . . . . . . . . . . . . . . .14
Bill Boyle . . . . . . . . . . . . . . . . .46
Adjourn. . . . . . . . . . . . . . . . . . . . . 116
. P-R-O-C-E-E-D-I-N-G-S
(10:05 a.m.)
CHAIRMAN HORNBERGER: The meeting will
come to order. This is the first day of the 133rd
meeting of the Advisory Committee on Nuclear Waste.
My name is George Hornberger, Chairman of the ACNW.
Other Members of the Committee present are Raymond
Wyner, Vice Chairman, John Garrick and Milton
Levenson. And also present we have a consultant with
us today, Bill Hinze.
During today's meeting, following the
planning and procedures session the Committee will (1)
hear an update from DOE on its performance assessment
program; (2) finalize the annual research report to
the Commission, and (3) discuss preparations for
tomorrow's meeting with the Commissioner.
John Larkins is the designated federal
official for today's initial session.
This meeting is being conducted in
accordance with the provisions of the Federal Advisory
Committee Act. We received no requests for time to
make oral statements from members of the public
regarding today's session. Should anyone wish to
address the Committee, please make your wishes known
to one of the Committee staff. We have received one
written comment from Mr. Mel Silberberg, on the
research program. His letter will be inserted into
the record at this meeting.
It is requested that speakers use one of
the microphones, identify themselves and speak with
sufficient clarity and volume so that they can be
readily heard.
Before proceeding, I would like to cover
some brief items of current interest. Items of
interest, (1) Dr. Victor Ransom has been appointed as
the eleventh Member of the ACRS. He is a Professor
Emeritus of Nuclear Engineering, Purdue University.
Prior to this, he was a Scientific and Engineering
Fellow at the Idaho National Engineering and
Environmental Laboratory. Mr. Timothy Cobetz and Mr.
Robert Elliott have been selected the ACRS/ACNW
Technical Staffs. Rob, who returns to the ACRS staff
having previously served on a rotational assignment
comes from NRR and will replace Noel Dudley on the
ACRS staff. Tim, who joins the Staff from the Spent
Fuel Project Office, will assist both Committees as
the work load dictates.
Dr. Margaret Chu has been approved by the
Senate as Director, Office of Civilian Radioactive
Waste Management. She comes to DOE from Sandia
National Laboratories where she has been in charge of
the Nuclear Waste Management Program. Prior to that
she was Deputy Manager for WIP.
The attached, at least attached in our
book here, February 13, 2002, a paper by Commissioner
Dicus, "The Future of Environmental Protection, a U.S.
Regulator's Perspective" provides a most interesting
perspective on this topic and I'm sure that anyone who
wants it can get a copy of this document.
Any other items? Okay, good. We are
going to move to our first topic which is an update on
DOE performance assessment and John Garrick will chair
this section of the meeting.
MEMBER GARRICK: I'm going to waive any
opening remarks for the benefit of having the time to
ask questions and what have you and I think we have
three people that we're going to hear from: Carol
Hanlon, Peter Swift and Bill Boyle. And I would ask
each of them to give us a quick statement of their
assignment or their role for the benefit of the record
and the Committee and those in attendance.
So Carol?
MS. HANLON: Thanks, Dr. Garrick. Is this
on? Can you hear me? Good morning. I am Carol
Hanlon with the Department of Energy. I'd like to
introduce to you my colleagues, Dr. William Boyle and
Dr. Peter Swift and ask them to join us up here. They
will be giving the main presentations.
Peter is with the Sandia National Labs,
Performance Assessment, and he has had a very main
role in helping us with our performance assessment
activities as well as the prioritization effort going
forward.
Dr. Boyle, as you know, is a Technical
Advisor, with Yucca Mountain and has strong
underground geotechnical expertise.
So the gentlemen will be making the
presentations.
You know that the Committee has been
carefully following our process and are particularly
concerned both with the technical aspects as well as
the performance assessment. We've briefed you many
times and especially last year on several of these
topics, including the Supplemental Science and
Performance Analysis Document, the Preliminary Site
Suitability Sites and Engineering Report and I know
you've been at many of the key technical issue
technical exchanges. So you're very familiar with
these issues.
We're also familiar with and we have
carefully considered the letters that you provided,
especially the letter on performance assessment and
we're hoping that you will see some of your
recommendations included in our path forward.
So I'm pleased to be able to speak with
you today and give you an update on some of the
information that has come out, some of the reports
that have come out since last summer.
I've introduced Dr. Boyle and Dr. Swift
and if I may just briefly cover some of the
information as an introduction.
This is our snapshot on our home page
which is available at www.yuccamountain --
www.ymp.com and it pretty nicely captures the major
efforts, the major accomplishments we have had during
the last year or so, the release of the Yucca Mountain
Site Suitability Evaluation, Rev. 1 of the Science and
Engineering Report, the SR Comment Summary Document,
Supplemental Comment Summary Document, those
responding to and summarizing comments that we
received during our comments period; the final
environmental impact statement and some other
information as well as the state and county impact
reports.
CHAIRMAN HORNBERGER: In the spirit of
engelbrecht, you did say ynp.com and I didn't know
that DOE had become a dot com.
(Laughter.)
MS. HANLON: Thank you very much for
helping me. Did I say dot com? Thank you.
Everyone will correct me. YMP.gov. And
I will never use an acronym again.
(Laughter.)
So the presentations that follow address
these technical updates and comments on preliminary
site suitability evaluation. There are two types of
them. One that evaluates the evaluation, the impacts
of the final regulatory standards including the
Environmental Protection Agency Standard 40 CFR Part
197 as well as Nuclear Regulatory Commission's 10 CFR
Part 9, excuse me, 63.
In addition, the technical updates
consider the evaluations of additional information
which was available since release of the supplemental
science report and analysis, the science and
engineering report and the preliminary site
suitability evaluation report, that information that
was continuing to be collected and analyzed over the
summer.
Another topic that we will discuss is the
treatment of uncertainty in the total system
performance assessment for the license application,
both the uncertainty analysis and strategy and
discussion of treatment forward of uncertainty and
finally, the path forward for the Yucca Mountain
performance assessment focusing on uncertainty that
matters and risk-informed prioritization for
performance assessment.
And you have in your book and in the
presentation again these major developments, on-going
technical exchanges with the staff during the year and
we had another technical exchange last week in San
Antonio; the release in May of the Science and
Engineering Report which was based on the total system
performance assessment in July; in August, releasing
supplemental science and performance analyses as well
as a preliminary site suitability evaluation; and
including the updates later to total system
performance assessment, staff recommendation and the
technical basis which Peter will say something about.
CHAIRMAN HORNBERGER: Carol, what was the
technical exchange last week?
MS. HANLON: It was on -- what was the
title again?
DR. BOYLE: Laboratory design.
MS. HANLON: And the final regulatory
standards not in July, but in June, 40 CFR 197 which
was finalized in November, 63 -- 10 CFR Part 63 was
finalized and also in November, the Department's 10
CFR 963 was released.
In December 2002, we had -- I think that's
an error -- 2001, additional information documented
was presented in four Letter of Reports which we'll
discuss with you today and in February, the site
recommendation went forward from the President. So
we're in the process of realigning our science and
performance assessment activities within BSC and
moving forward with a consistent direction on
treatment of uncertainty as well as focusing on the
risk-informed performance-based approach.
So with, unless you have any questions on
that brief introduction, I'd like to turn the
microphone over to Dr. Swift.
CHAIRMAN HORNBERGER: Just a quick one,
Carol. What's BSC?
MS. HANLON: Bechtel.
DR. SWIFT: BSC is Bechtel SAIC Company.
It's the management operating contractor and this
first presentation is the four letter reports that
Carol mentioned. I'll go through them, fairly
quickly, but just summarize what new information there
is relevant to performance assessment since the major
documents of last summer.
I should credit many other people, Jerry
McNish, the manager of the Total System Performance
Assessment Department, in particular; and Mike DeLugo,
who was the lead on one of the four letter reports,
the largest, that Update Impact letter report.
And just to clarify, there was one mention
made there on Carol's side on realigning science and
performance assessment activities within Bechtel SAIC
and what has been done is that the Post-closure
Science Programs have been brought together with
performance assessment into a single organization
called the Performance Assessment Project. Bob
Andrews is the manager of that.
And the performance assessment
calculations, the TSPA, Total Systems Performance
Assessment, is one department within that larger
Performance Assessment Project. In fact, there are
several subprojects. TSPA now actually reports to me
in this group called Performance Assessment Strategy
and Scope.
The science programs we're familiar with
for years also now report directly to Bob Andrews
within Performance Assessment.
A couple of overview slides here, just to
go through quickly. What we have here first, there's
a body of information that is the Total System
Performance Assessment for the Site Recommendation,
TSPASR, documentation and with that I'm including the
Supplemental Science Performance Analyses from last
summer, last spring and summer.
This, the SSPA and the other documents
that are associated with that, I believe have already
been presented to this group, so what I'm focusing on
are things that follow that, that's this page and the
next one in the handout. A Letter Report in
September, completed in September, looking at the
impacts of the final EPA rule and also supporting the
final environmental impact statement and then a Letter
Report in December on the impacts of the NRC's final
rule which was, we felt, there are enough things in
that to run traditional analyses.
And then this technical update impact
Letter Report, known by its acronym as the TUILR.
These, so you can -- a graphic showing you what the
documentation is, two pages of this. First page is
performance assessment documents, going all the way
back to September of 2000, a document called the TSPA,
SR-Rev 0 and it's updated. ICN stands for Interim
Change Notice. That's basically the revision.
Updated in December, that's the version which people
are most familiar with. That supported the site
recommendation and the upper-tier documents that were
released that spring and summer, but it was also
updated in the spring, the supplementary analyses were
published in July in something called the
Supplementary Science Performance Analyses Volume 2,
SSPA Volume 2.
Then September and December, new results
that you probably have not seen yet. The Part 197
update and the Part 63 update.
MEMBER GARRICK: Peter, when you get
around to doing the TSPA-LA, will it integrate all of
these documents into the TSPA-LA?
DR. SWIFT: The TSPA-LA will be stand
alone in the sense that it will be a complete
documentation of its own set of analyses. It will
probably most closely resemble the models used in
these ones, but does that answer your question?
We don't have to keep sending you back to
a lower tier or older documents.
MEMBER GARRICK: Okay, thank you.
DR. SWIFT: This talk is about TSPA, but
it's worth keeping tracking of the non-TSPA documents
also, the upper tier documents of the science
documents.
I've lumped them both together on this
side. Go back to 2000, you have the Process Model
Reports and the Analysis Model Reports prior to the
scientific basis or TSPA-SR. They fed into an upper
tier DOE document released last May, the Yucca
Mountain Science and Engineering Report. These were
contract reports. This is a DOC document. This is a
primary technical basis for site recommendation, a
thing called the Science and Engineering Report,
published in early May 2001.
The scientific basis was updated again in
the spring of 2001 in this Supplementary Science and
Performance Analyses by one which was a scientific
basis. This document, published in July as a DOE
document, I believe, has new science that was not in
this one. And also in Volume 2 it has new TSPA
analysis.
Together, these two supported the
preliminary site suitability evaluation. This is the
document that actually makes the site recommendation
case. That was a DOE document published in August.
Thecover date is July,but wasn'treleased until August.
And this is all material you've seen or
was available. This is the new part over here, the
November 2001, Technical Update Impact Letter Report.
(Slide change.)
DR. SWIFT: Now the Letter Report on the
Final EPA Rule and it's worth actually noting the
footnote. If you try to do a search in any records,
data base, looking for that document, you'll discover
that the title of it says it's input to the final
environmental impact statement. That's correct.
Informally, we think of it as the update report on the
EPA rule and it was originally planned prior to the
completion of the EPA rule. It was originally planned
as an EIS update.
So the TSPA was modified to meet
specifications in Part 197. We went from the average
member of the critical group to the reasonably
maximally exposed individual. We went from 20
kilometers to 18 kilometers, both for groundwater
release and for the volcanic disruption scenario and
ashfall. And the EPA rules specified 3000 acre/feet
per year for groundwater protection. So we ran those.
(Slide change.)
DR. SWIFT: We also, these were the ones
that were aspects of the analysis that was planned for
the EIS, looked at both the base case waste policy act
inventory and a possible expanded inventory. And that
was the main point of the EIS.
We also ran some updated igneous activity
scenarios. We reran human intrusion which we had not
run since December of 2000 and we looked at two
different times for human intrusion.
(Slide change.)
DR. SWIFT: So far those changes were all
driven by regulation or assumption. We also did make
changes in the model itself since the model used in
the spring of 2001. I listed the most important one
here first. Waste-package corrosion calculations for
the results of that show -- used a general corrosion
model that was independent of temperature.
In the SSPA, the supplementary results
from next spring, we had used a temperature
independent corrosion model which basically showed
corrosion slowing at lower temperatures. We felt
there was insufficient technical basis to support that
for the site recommendation. You know it was already
published in the SSPA, so we took it back out and that
one change there counts for most of what you're going
to see in these slides.
We found an error in our in-drift
thermal-hydrology work which we omitted heat transfer.
It made a whole lot of difference. So we put it back
in and got it right.
We had omitted portal transport from the
portal due to intrusion. We corrected that. We had
an updated version of a waste package degradation
model. And we modified the inventory slightly at the
request of the Naval programs to treat their fuel as
part of the commercial inventory, whether it's a DOE
inventory. It's a small fraction anyway and would
make no difference.
(Slide change.)
DR. SWIFT: Results. These are mean
annual doses and millirems per year. I'm not showing
the complete panel of the doses that generated that,
but these are means drawn from 300 realizations. The
black curve here is the mean from TSPA-SR in December
of 2000. The red here is a single curve shown from
SSPA June-July of 2001. And this happens to be for
the high temperature operating mode that we looked.
This was only the high temperature mode here. In
SSPA, we looked at high and low. And then here, blue
and green, you can hardly tell the difference between
them, this new modified model run for both high and
low temperature for the updated model.
MEMBER GARRICK: Has the red curve not
reached its peak yet?
DR. SWIFT: That is correct, the red curve
-- unless that is its peak. We don't know that. But
in the actual highest point on the curve is here.
That's due a climate spike. By inference, we believe
that -- we can't rule out the possibility it might
have achieved a higher peak if it ran longer, but
there actually is a peak in there.
Taking out the temperature-dependent
corrosion, basically moves the time of large scale
package failure from here to 740,000 years and what
that has done is basically it leaves the -- in the red
curve corrosion rates slowed as temperature dropped
later and the green and blue curves, they do not.
They stay at a higher corrosion rate throughout.
CHAIRMAN HORNBERGER: What's the change to
explain the differences at early times?
DR. SWIFT: It'll come to me in a minute.
CHAIRMAN HORNBERGER: Is it an assumption
on juvenile failures or is it igneous activity or what
is it?
DR. SWIFT: No, it's juvenile failure.
For the SR, we had input from our waste package
engineers, but they saw no credible mechanism for
juvenile failures, so we had none. This is the
earliest general corrosion failure showing up here on
the black curve.
For the updates for both SSPA and the more
recent work early last fall, we have the first general
corrosion failures later. They're out in here. But
we do now have a model for juvenile failures, early
failures, due to improper heat treatment of lid wells.
The number of failures, in about a quarter of our
realizations, we had one or two packages out of 11,000
failing. So it's a very small failure rate, but it
produces a non-zero dose. It gives you small numbers.
This is a non-zero dose out to there that is largely
driven by igneous iodine and Carbon 14 in groundwater
transport.
MEMBER GARRICK: Are you going to later
get into a little more detail about impact of the
changes in this -- in the model in relation to the
difference in the assumptions between the TSPA-SR and
these results? I'm thinking of things like if you've
introduced this corrosion model now, has that brought
seepage back into the picture as an important
phenomena because in the TSPA-SR it was not an
important phenomena.
DR. SWIFT: It's still not particularly
important here. It matters for transport away from a
package, but the corrosion model is still independent
of water saturation. As long as you have humidity,
you have corrosion.
MEMBER GARRICK: And you still have the
same model inside the waste package of the saturated
water environment, those kinds of things?
DR. SWIFT: Uh-huh. Yes. The end package
transport model, I think is when --
MEMBER GARRICK: So it's still diffusive
transport that's the main?
DR. SWIFT: Yes. One significant
difference between and this applies for both the red
and blue-green here. A significant difference between
these two curves and this one is that the -- in an
attempt to put a little more realism in that diffusive
transport pathway of the package, we now split the
transporting waste that are transported by diffusive
properties when they reach the drift wall, the rock.
We put the diffusive transport fraction into the
matrix of the rock and we put the effective, if there
is effective, transport and that would that synchrony.
We put that fraction and it fractures.
Previously, we put it all into the fractures, this
curve, put all the waste and the fractures and that
didn't seem realistic, simply based on the surface
area available for diffusive transport, most of it is
going to go into the largest part of the surface area
which is matrix.
So that's the only change that I think
comes to mind for me anyway, between -- for the in-
drift transport model, between this, these curves and
that one. Probably more realistic with the splitting
of the diffusive.
Ask questions as I go. The time and the
fact that I'm only one person, I wasn't planning to go
out for a lot of detail in this stuff, so go ahead and
ask question.
VICE CHAIRMAN WYMER: I have a question.
How important was the microbiological corrosion? Was
it important at all?
DR. SWIFT: Bill, do you want to field
that one?
DR. BOYLE: I'm sorry, I don't have the
answer.
VICE CHAIRMAN WYMER: Is it a minor
player?
DR. SWIFT: No, I don't think it's a
player at all.
VICE CHAIRMAN WYMER: The other question
is what is meant by aging multipliers of inside out
corrosion?
DR. SWIFT: Aging multipliers for inside
out corrosion. That is pretty cryptic. The model
does not have an explosive treatment of the behavior
of alloy 22 as it ages. Instead, we apply a
multiplier to the corrosion rate to account for aging,
changes in the alloy aging. I don't know what the
update was, but someone felt, I suspect that in the
SRR model we had an aging multiplier only on outside
in corrosion.
Somebody pointed out we should have it on
the inside out corrosion also. But it's an uncertain
parameter. It's a parameter that has a range on it to
account for our uncertainty in the effects of aging
and corrosion.
VICE CHAIRMAN WYMER: So the multiplier
then --
DR. SWIFT: Accelerates the rate.
VICE CHAIRMAN WYMER: Accelerates the
corrosion, in some arbitrarily decided way?
DR. SWIFT: Uh-huh. In some -- I hope
it's more than arbitrary, but it's not physics based.
MEMBER LEVENSON: Is the uncertainty ever
symmetrical? It's always in a more dangerous
direction?
DR. SWIFT: We'll come to that in the next
set of talks. For these analyses, I believe it is
asymmetrical in many cases. I would like to see more
symmetry.
MEMBER GARRICK: Generally, more of a log
normal than a uniform?
DR. SWIFT: I know where you're headed
with the question. Keep asking it.
(Slide change.)
DR. SWIFT: The igneous activity results.
I think that's the next -- yeah. This same
presentation or say similar presentation was given by
Jerry McNish to the Review Board in January and this
figure drew quite a lot of attention from the Board
who were displeased with the lack of prominence given
to the word "probability weighted" here. I want to
make clear of that right now. These are probability
weighted mean annual doses. This is consistent with
what's in Part 63. This is not being included with
obvious dose you'd expect to see, but it is -- I'll go
through that in a couple of slides here, what it
really is.
This is the regulatory dose of volcanic
activity. The black curve here is what was shown in
TSPA-SR in December of 2000 and the blue and red and
curves here were updated in September and these, by
the way are essentially identical. I was also updated
in the SSPA in June and July.
The red and blue were a perfect overlay
here for high temperature and low temperature
operating modes. The volcanos are pretty insensitive
to temperature of the depository.
Changes here, recent updates, since SSPA,
specifically for this analysis, we move the location
from 20 kilometers to 18 kilometers and we updated the
biosphere dose conversion factors. We also made all
the changes I just talked about in the nominal model.
That's the other feature that's here.
There are a series of other changes not
described here which were updated as part of SSPA in
the spring. They are what account for this vector of
25 increase from here to here, interruptive dose and
the decrease over here at later times.
The smooth curve to here or all the way
out to here is driven by the volcanic ashfall dose and
the irregular curve here and here is from the
groundwater release from damaged packages and at some
point in the future the basic weight, probability
weighted dose from the groundwater pathway from
packages damaged by igneous activity will cross over
and exceed the corruptive ones. If you put edging for
the black line, you adjust the eruptive half, we have
a curve that kept on going out like that where it's a
groundwater curve, goes like that. So at the
crossover point, you see that the curve changed from
being smooth to being irregular.
The sharpness here is due to a long term
climate change in the model, it's spiking here. These
are glacial climates.
The other major changes here between
basically we -- at suggestions from the center and the
NRC staff, we looked at a different wind speed data
set which led to an increase of about, a factor of 2.5
from here to here. We updated our biosphere dose
conversation factors --
MEMBER GARRICK: Does that mean you will
look more at a wind row than a --
DR. SWIFT: No, the wind direction is
still assumed to be fixed towards the location of the
REMI for these. So that would have to be a factor of
4 or 5.
MEMBER GARRICK: Yes.
DR. SWIFT: It's not a huge player, but
yeah --
MEMBER GARRICK: It is quite a huge
player.
DR. SWIFT: The 4 or 5 add up.
MEMBER GARRICK: Yes.
DR. SWIFT: We looked at wind from a
higher altitude. They had pointed out that we had a
data set that went to higher altitude than we could
have used and we used that and that was part of the
difference here.
As a matter of fact, we unrealistically
used only the highest altitude, the 300 millibar data
only went into that, whereas for this one, we used a
somewhat lower altitude data set against the full
column of wind speeds and got the elevation up.
This also has an increase in the number of
packages involved in the eruption, due to a
recalculation of how we did that. Has increased dose
conversion factors due to reconsideration of the nasal
ingestion pathway. That's the larger particles lodged
in the nose. We ended up putting in the long --
MEMBER GARRICK: But you continued to use
the assumption that all the waste packages were
degraded that were in the intersect?
DR. SWIFT: Yes. All packages in the --
we were conceptualizing the volcano as a conduit, a
cylinder that rises up through it. It's also got an
intrusive dike, a tabular body that may cross many
drifts, but the portion that erupts, we're assuming is
a cylinder with a mean diameter of about 50 meters of
- or medium diameter.
Yes, all packages in that cutout by the
cylinder are assumed to be fully destroyed. The
phrase is damaged sufficiently to provide no further
protection. And the waste within them is produced to
the grain size of the particles which is by-products.
MEMBER LEVENSON: Is there any
significant, for this type analysis, is there any
significant difference in the footprint of the high
level versus the low level?
DR. SWIFT: The high-temperature/low-
temperature?
MEMBER LEVENSON: Yes.
DR. SWIFT: It's a simple scaling. It
affects the probability that the event will hit it at
all. And if you need to have a larger footprint for
a lower temperature operating mode, then the
probability of the event scales -- it's not precisely
linear because -- it's close enough. If you double
the footprint, you're going to double the probability.
MEMBER LEVENSON: I would have expected to
see a little difference between the high temperature
--
DR. SWIFT: Oh, thank you. Thank you. We
didn't do it. We said that and that's a caveat that
is in the text and I should say that. We simply used
the one footprint for this and in text we discuss how
to use the weighting factor if you want to.
It's not clear that we will have different
footprints. One of the options was for low
temperature, was to use the same footprint and a
longer and more rapid ventilation period.
So we weren't quite sure what to do with
that and it was going to be a nuisance to --
MEMBER GARRICK: How about the erosion
time of the 1000-year erosion time for the 15
centimeter layer? You're still using that?
DR. SWIFT: No. Let me explain what we
actually did here. This is -- this is a good slide to
do it with.
This is the -- what we call the
conditional dose, the dose that you would get -- this
is a figure that we probably should have showed them
ERB in January, but didn't.
If an event happened at 100 years and
these were calculated by the way it would be SR
modeled, black being the curve on the previous page.
If an event were to happen at 100 years, a person
living after that would receive a dose as shown here.
So a person alive at 2000 years might receive a dose
somewhere in this bandwidth here, the mean being red,
so that would be 95th shown, both shown in black.
Clearly, the dose would be worse if you
were alive at the year of the volcano. There's no
probability weighting shown here. The uncertainty
between the lowest and the highest curves reflects
uncertainty basically in the inputs to our ASHPLUME or
transport model, things like windspeed and the conduit
diameter. That's basically how many packages are
effective. And also in our biosphere conversion
factors.
The slope of the curve, how fast it drops
off through time is a factor of two things. One is
there's radioactive decay and the other is how quickly
that contaminated ash layer erodes away. And -- all
right, that's the top figure.
The bottom figure down here is just mean
curves. The red here. Now it's just the mean curve
shown for condition events at different times at 100,
500, 1,000, 5,000 years.
This, if you were to draw a curve,
connects the dots through the tops of them, that's the
radioactive decay curve. So these curves then are our
soil removal factor. That's the rate at which soil
will be contaminated and ash layers being eroded away.
However, our treatment is not quite as
simple, John, as the way you describe it. What we're
doing is we are assuming that the top layer of top
soil erodes at a rate of 6 to 8 millimeters per year.
I believe that's right. However, we're assuming that
soil is plowed annually, so it's constantly being
remixed to 15 centimeters, so that any way -- how
thick the ash layer is, the radionuclides get mixed in
to a 15 millimeter soil layer every year and the top
of it gets skimmed off every year. So it's an
exponential decay in our soil removal rather than a
simple decay. There's always some still left there.
And so if we weren't mixing, we would take
off that 15 centimeters in several hundreds of years.
It would be relatively rapid. We are mixing, so that
we're always creating a -- we've always moving
radionuclides deeper down in that soil layer with each
year's plowing and erosion.
Clearly, we are fairly sensitive to the
way we treat erosion. If we had zero erosion, if the
soil layer stayed there forever, this curve would look
like the connecting the dots from the top of there,
simply go out like that. Would be a radioactive decay
curve.
CHAIRMAN HORNBERGER: This is a pretty
critical part of the model for the first few thousand
years?
DR. SWIFT: Yes, it is.
MEMBER GARRICK: And it struck me as an
extremely conservative assumption.
DR. SWIFT: The assumptions that go into
that have to do with whether you think we're dealing
with agricultural land or stable desert soil. If
we're dealing with agricultural land that really is
being plowed every year, this may not be that
unrealistic. We have a fairly high, compared to what,
for example, the NRC staff has recommended, we have a
relatively rate at which stuff blows off, but because
we're plowing and mixing, that is consistent with what
you'd expect to see on crop land.
On the other hand, if we didn't have this
plowing and mixing going on, we had stable desert
soil, we shouldn't have such air mass loading or such
rapid erosion. We have pretty high air mass loading
in our BACS. It's dusty air people are breathing in,
consistent with agricultural land. It's blowing
around.
It's what we did anyway. I just want to
show one other slide here.
(Slide change.)
DR. SWIFT: This is how you get from those
conditional doses to dose mean doses because this is
something that it's not intuitive and this is just a
question of probability space rather than a real
phenomenon. This is what the role asks for and I
believe it makes sense.
Think of these as mean doses from the
previous curve, the mean conventional dose. If an
eruption happened, Volcano 1 happened in Time 1 and
you dropped the time axis here, this is dose/time, a
person alive in the future could get in the Year T-1,
they would get that dose. If they were in Year T-5,
but the eruption was in Year T-1, they would get this
dose here off that curve there.
If, on the other hand, they were in Year
T-5 and an eruption happened in Year T-6 out here,
they'd get a zero dose. The eruption hasn't happened
yet.
Now put it into probability weighting, the
probability that a person living out here in the Year
T-5 could get a dose from an eruption that already
happened back in Year 1, Year 2, 3, 4 and so on, well,
the probabilities of those were all the same, similar
to the process that has a time constant probability
and so the probability weighted mean dose we'd get out
here is simply the sum of the probability of all the
events in the time interval of interest, 0 to 1,000
years times the doses associated with each one of
those events at the time you're interested in.
So at Time 5, this person living here
could be getting a dose from this event, from this
event, this event, or that event. That one is a zero.
And each one of them has equal probability and you
multiply them and sum them up. And what you get when
you do this, this is actually what we do, but a little
thought experiment suggests that at early times,
although the consequences are highest, the probability
that the event happens in that year or has already
happened, the probability is low. As you go out in
time, the probability accumulates.
So the probability if you're living out
here at the Year 10,000, the probability that the
event has already happened is 10,000 times the
probability in the first year. So in this sum here,
the doses go down at later times, but the
probabilities accumulate and you'd expect to see a
mean curve that starts out low and climbs to some
intermediate peak and then falls off again as doses
decay from radioactive decay.
And that actually is what -- that's what
the blue curve is here or the black one. The peak is
around 3000 years and after that radioactive decay
takes over and starts to drop off.
So that's -- the point of that explanation
is just to say, show how we got from things that look
like this to the probability of weighted sum that the
regulation asks for.
I've got to speed up here.
(Slide change.)
DR. SWIFT: Human intrusion scenarios.
This is the forced assumption that a driller drills
through the waste package. Part 197 says one waste
package, made a pathway to the borehole pathway to the
saturated zone and assume it occurs at a time when the
waste package is degraded enough that the grower would
not recognize it. And this then is -- we picked
30,000 years. And this is an intrusion of 30,000
years as our annual doses, a full set of 300 of them
with a mean shown.
And these are -- we also reran it for the
proposed NRC rules, was prior to finalization of 197.
We used the 100-year time. What you see here, for
example, from 100 out to the first arrivals coming in,
this is basically your minimum saturated zone
transport time. The spread of arrival is out from
time after that show the spread and saturated zone of
transport. So some realizations showed first arrivals
here. Some didn't have them arriving too well out
there.
(Slide change.)
DR. SWIFT: The December Report looked at
the impact of Part -- final Part 63. The main
difference here was the rule now requires us to use
3000 acre/feet per year for individual protection
which is something the EPA has not clarified in their
rule. So now we were using a sample value previously.
We also in this report, if you get a hold of the
report and read it, will discover we ran a couple of
cases that are now moot following clarification of the
word "unlikely" and the new proposed rule.
We went ahead and ran a case with an
igneous intrusion eruption for the groundwater
protection standard and also for the human intrusion.
And with a clarification of the rule in those cases
are moot.
What happened with the 3,000 acre/feet.
The result was to scale those by approximately two
thirds.
CHAIRMAN HORNBERGER: Two thirds or three
halves?
DR. SWIFT: Two thirds. We're diluting
them.
CHAIRMAN HORNBERGER: Oh, that's dilution.
DR. SWIFT: This was a sample value.
There's the range given, with a mean of about 3,000
acre/feet in our -- this is what we found from our
survey of these in the region. Well, this just pushed
us to the upper portion of the range in the rule and
produces a little more dilution.
These -- the numbers shown here are the
nominal performance only. These are the numbers of
the doses due to the juvenile failures of nominal
performance. We took the volcano out to show that.
CHAIRMAN HORNBERGER: It's two-thirds
because you put everything into the volume.
DR. SWIFT: The larger volume.
CHAIRMAN HORNBERGER: Rather than have a
concentration?
DR. SWIFT: Yes.
(Slide change.)
DR. SWIFT: The Technical Update Impact
Letter Report, this is the fourth of them. I said a
letter report. This is -- this report is underlying
science rather than TSPA. I had a much smaller role
in this report, but Mike DeLugo is the person who did
most of the coordinating of it.
The point here was documenting additional
information since completion of the underlying science
for the Science and Engineering Report and the Yucca
Mountain Preliminary Site Suitability Evaluation. So
this updates in science since roughly the spring of
2001. In some cases it goes back a little further
than that. But this was a new work that was going on
last summer and early fall in experimental programs.
And then the impacts of this work were evaluated on
TSPA and preclosure, basically to make sure there
weren't any -- wasn't any new information that would
necessitate a re-evaluation of the said
recommendation.
It's a thick report. It's almost 400
pages long. It includes 11 White Papers in each of
the topical areas where the technical staff was sent
back to just document what is their new information.
Then we had a rapid series of workshops where we
looked at the impacts. To do this, we got the
technical experts who wrote those White Papers
together with TSPA analysts and we had our workshop
setting. We went through each topic and on the spot
estimated elicited impacts on total systems.
A couple of pages here of examples of the
sorts of things. This isn't a very complete list.
Just the sorts of information that was available:
fracture data, seepage data. These were things that
were written up and then people were asked how would
this affect the input for models and the modelers were
asked would this have an effect. And so on, the high
profile one here, the discovery of high concentrations
of chloride in seepage waters at low temperatures.
VICE CHAIRMAN WYMER: In your technical
update, you didn't include a couple of processes.
DR. SWIFT: To the extent that they're
captured in the on-going work related to the --
usually the air field environment and the engineered
barrier system, yes, we did. It was structured around
the existing science programs. We didn't force a
White Paper on a couple of processes.
You want to deal with that?
DR. BOYLE: That fluoride example is
actually -- falls within the realm of a couple of
processes. As it turns out the flourine came from the
materials that were introduced by jackets, but it was
postulated that it could have been coming from the
fluoride. That's the small amount of fluoride that's
present in the rocks.
VICE CHAIRMAN WYMER: So to that extent
you put it in.
DR. SWIFT: Yes. There were some portions
of this were entirely preclosure, for example, updated
data on aircraft activity. We have a new survey of
aircraft traffic in the area, know that that would
change the risk of accidental airplane crash, which
seems moot now.
(Slide change.)
DR. SWIFT: So what were the results of
this and if you've got a copy of the technical update,
the TUILR, I recommend you go straight to the very
back end of it on pages 350 on where there's an
appendix that discusses impacts on post closure
assessment and there are a series of chapters in that
appendix, a weight for each of the various measures of
the components. And here's the conclusion. All
impacts of all the new work are insignificant, except
for these two.
First, the Transport Team believes that
they're now able to show a reduction of nominal dose,
igneous dose, but the nominal performance. They thing
it may show a more delayed retardation in unsaturated
zone perhaps lowering the 10,000 year dose up to one
order of magnitude. This is lowering, but it's also
just pushing it up further in time. It's slowing the
transport.
VICE CHAIRMAN WYMER: What's that due to?
DR. SWIFT: Excuse me?
VICE CHAIRMAN WYMER: What slows it down?
DR. SWIFT: It's a change in the way
they've been treating diffusion near the matrix and
this is not my field. It's out of the realistic case
AMR and so-called realistic case AMR and that it was
a change -- actually, I believe a numerical treatment
in the model of the -- I believe it added more cells
in the matrix, so you -- instead of having the matrix
represented by a single cell and diffusion occurring
all the way to the center at once and way back out
again in a single step, I believe in the numerical
model there, so it takes longer for the diffusion to
get in and out. And it's the back out part where
we're seeing the benefit.
The model probably will not be permitted
in the TSPA because it's numerical intensive. But
anyway, and also we weren't too excited about a
possible one order of magnitude reduction in a dose
that's already 10-5. But it's worth knowing anyway.
It may be there.
Possible increases in the eruptive dose.
Basically, we looked at the impacts of the Center's
model and the conclusion of our staff was that no more
than one order of magnitude increase there. That's
based simply on looking at the total number of
packages that might be involved. The largest
difference is between -- from a performance point of
view, the largest difference is between the Center's
model and the one we're using as the Center's proposed
mechanism has more packages.
MR. HINZE: Peter, if I may, Bill Hinze.
You have eruptive there. What about intrusive? Is
that considered in that?
DR. SWIFT: Yes. The effective is only on
the eruptive side here. That's deliberate to say
eruptive in the dose there.
The Center's model that basically we're
concerned about here, we worry about, is the one that
calls for what I call a dog-leg eruption where magma
rises, hits a drift, flows down a drift and goes back
up again. Ours goes straight through and they
proposed, and we can't rule out the possibility it
would go a dog-leg path and sweep the entire drift and
of course, the eruption. And if so, more packages
will be involved.
MR. HINZE: Are you considering it at all
in terms of the intrusive? Is that going to be coming
along?
DR. SWIFT: We are reevaluating our model
for the intrusive effects. We don't see a big impact
there on dose. We believe our model needs more work
to be ready for the LA, but we don't think that's
going to change much.
MR. HINZE: Are you eliminating the zones
that you had in terms of disruption of the canisters?
DR. SWIFT: That may be modified for LA.
It may not. We're working on that right now. Some
version of that is likely to stay at the Zone 1 of
extreme damage and Zone 2, lesser damage, but in fact,
there are igneous geniuses are working on that
question right now.
MR. HINZE: Thank you.
MEMBER GARRICK: If the new data that's
being talked about now on igneous activity results in
an increase in the likelihood term, what is that going
to do to your results?
DR. SWIFT: In terms of increasing the
probability of a volcanic event or an eruption, those
are separate probabilities.
MEMBER GARRICK: Yes.
DR. SWIFT: Increasing the probability of
either of those at the site is pretty much a direct
scaler on the probability, the way it goes.
MEMBER GARRICK: Right.
DR. SWIFT: There is a question about the
air magnetic data and whether or not that will change
the probability.
MEMBER GARRICK: So if the cases increases
that a 10-7 number is not even justified on the basis
of the supporting evidence and it may be more like
10-6, if that happens it's going to be pretty much a
linear effect?
DR. SWIFT: Yes. We don't think that's
going to happen. Our of our impact assessment, we
don't think that probability is changing much.
And I think that does it. No, I've got
the summary slide.
(Slide change.)
DR. SWIFT: Just for completeness, to note
that we did look at impacts on pre-closure performance
of new data. Also, we didn't see anything there.
I think that's simply a summary side.
Let's me sit down.
The analyses that I've just summarized
here, basically provide confidence in the adequacy,
appropriateness of the SR.
MEMBER GARRICK: I know we're running a
little behind and I want to give the Committee a
chance to ask questions, but you'll be hanging around,
will you not?
DR. SWIFT: Actually, I should have said
that right off. No, I have a 2:55 flight to catch.
MEMBER GARRICK: Oh, I see. Well, then
let's give the Committee the benefit of your presence
and see if there are any questions.
DR. SWIFT: Bill and I are your speakers
until 12:30.
MEMBER GARRICK: And you have to leave --
yes. Okay.
Ray, go ahead.
VICE CHAIRMAN WYMER: In your backup slide
24, you indicate that Carbon-14 is rated Class C.
DR. SWIFT: Yes.
VICE CHAIRMAN WYMER: Why did you put that
in there?
DR. SWIFT: I don't actually know where
that came from, Class C. So I guess I can't answer
your question. I realize it was likely to come. I
don't know.
VICE CHAIRMAN WYMER: That's the first
time I really heard it talked about the Carbon-14
being rated Class C.
DR. SWIFT: I can say that we do not have
a realistic model for groundwater transport. This is
based on the assumption that Carbon-14, carbon, in
general, is a nonreactive species for groundwater
transport, our groundwater chemists just don't like
that.
(Laughter.)
DR. SWIFT: So basically that's an upper
bound on Carbon-14.
VICE CHAIRMAN WYMER: You don't know where
that came from?
DR. SWIFT: No, I don't.
MR. BOYLE: Good morning. Thank you for
this opportunity. Peter had talked about some
updates, and these next two talks -- the first by me
and then I'll be followed by Peter -- are going to
deal with uncertainty analyses and what we're doing
with uncertainties.
For those of you that were present at the
NWTRB meeting at the end of January in Pahrump, I made
this presentation there, and I'm pretty much going to
make the same presentation. And I think Peter will as
well, for the most part.
This report -- it's available at our
website, if you haven't seen it already. And it
represents the work of others, in particular the two
people whose signatures are on the report -- Kevin
Coppersmith of Coppersmith Consulting, and Jerry
McNish of BSC.
And Chapter 2 of the report was prepared
by Jerry and the various process model leads. Chapter
3 was prepared by Kevin, and with input from Peter and
Bob Andrews, comments from them. And Chapter 4 was
prepared by Karen Jenny and Tim Nieman of GeoMatrix.
Now, the overview of the next two talks --
the first is by me on the report itself, "Uncertainty
Analyses and Strategy Report," and Peter is going to
talk about how to -- the implementation of a
consistent treatment of uncertainty in the TSPA, total
system performance assessment, for license
application.
This is the title of Section 1 of the
report. It's "Introduction," and the three main goals
of the report are listed on page 2 of the report.
I've distilled them here in these three bullets. This
is what is done in Section 2 of the report. Summarize
and discuss what we at the project have done to
evaluate, clarify, and improve the representation of
uncertainty in the total system performance
assessment. That's Section 2, and it also gets at
comments made by other groups.
Based on this discussion, Section 3
develops a strategy for how to handle uncertainties,
and it also proposes some improvements for the future.
And then, Section 4 deals specifically with how to
communicate uncertainties to various groups,
decisionmakers, technical people, and also proposes
some improvements for the future.
The next I think it's six or so -- I think
it's up through page 9 -- pages 4 through 9 of the
package you have are a table. And it's related to
something that's in Section 2 of the report. Here is
the title of Section 2 of the report, "Evaluation of
Uncertainty Treatment in the TSPA and the Significance
of Uncertainties."
On pages 30 and 31 of the report, there is
Table 2.2 and it's called "Key Remaining
Uncertainties," and it deals in the table with these
first four columns. And in the report there is in
that table in the report, this fifth column isn't
there. The information that's in the fifth column I'm
showing you here is in the report, but it's in the
text of the report. But we had a request from people
at headquarters to distill down those paragraphs and
pages in the report and create this fifth column.
As I did at the NWTRB meeting in Pahrump
-- we'll be here all day if we go through each and
every item in this table. The main point that I want
to get across with respect to Section 2 is the various
technical investigators were asked to summarize the
state of uncertainties. What I asked them to do is I
asked them, how can you sleep at night knowing that
there was a potential at that time that a decision was
going to be made? How can you sleep at night with the
remaining uncertainties? And that's what this table
and those parts of the report tried to capture.
We got back two very common answers of why
these people were able to sleep at night. One is the
uncertainties really didn't matter. They looked at a
broad range, and for some of the items it didn't
really affect the dose at 18 or 20 kilometers.
MEMBER GARRICK: But, Bill, isn't that
dependent upon the model?
MR. BOYLE: Sure.
MEMBER GARRICK: Because in the VA, for
example, seepage was a very important phenomena.
MR. BOYLE: Right.
MEMBER GARRICK: And so you changed your
corrosion model, so that in the site recommendation
report it's not an important phenomena.
MR. BOYLE: Right.
MEMBER GARRICK: And I think it's those
kinds of connections that are very important.
MR. BOYLE: Right. And that's the point
I would say when they -- when I say that there wasn't
-- it really didn't have an effect or it wasn't
important, it is with respect to the insights that
were being gained by an implementation of either the
TSPA itself or some subsystem.
But, you know, the answer is both.
Sometimes it was -- when carried all the way through
to the end of the TSPA calculation, it showed that it
didn't matter, which then just raises the question,
what if the underlying models really aren't right?
But those were the answers I got back from
the PIs. One is it really didn't matter, it seemed,
over a range of uncertainty. But the second answer
that came back quite frequently, and is represented in
the far right column, in various words is, "Well, I
was conservative." You know, I took a bound, like the
one that deals with the rock -- acknowledge that the
analyses were very conservative.
Whether that's a palatable approach in the
end, that is what was used at this point, and that's
the answer that was given.
So with that, we can't possibly spend a
lot of time on all these technical items. In January
in Pahrump, I jumped up to slide 10. I'm going to do
it here today again. And it's -- we're jumping to a
new section of the report, and this was a very
important section of the report, Section 3, and that's
the title of it up there, "Strategy for the Future
Treatment of Uncertainties."
And Section 3.1 of the report has a
compilation of words from the regulation. It quotes
from the EPA's regulation on how uncertainties should
be treated. It has quotes from what at that point was
-- probably started with the draft of 63, and then we
may have stayed with the draft or perhaps we got the
final comments from 63. I think we did get the final
comments from 63, but also comments from this
committee, the Nuclear Waste Technical Review Board,
the NEA/IAEA peer review group for the TSPA, and also
the peer review group we had for the TSPA-VA.
So we synthesized all of those -- you
know, provided the quotes and synthesized those
comments in Section 3.1. And then, in Section 3.2,
came up with a strategy for the future. And on these
next two slides, slide 10 here and 11, there were
eight recommended things to do. And these are the
quotes from those eight things. The first four are
shown here.
And if you read the report, each of the
eight recommendations starts off with a section in
bold, and that's what's reproduced here. And so they
are develop a total system performance assessment that
meets the intent of reasonable expectation. That's
defined in the EPA rule and also the NRC's word-for-
work exactly the same.
Quantify uncertainties in inputs to
performance assessment. Identify processes that
encourage the quantification of uncertainties and gain
concurrence on approaches with the Nuclear Regulatory
Commission. And provide the technical basis for all
uncertainty treatment.
Also, the fifth recommendation was to
address conceptual model uncertainty. Develop a
consistent set of definitions and methods for bounds
and conservative estimates. Develop and communicate
information that can be used by decisionmakers. And
this is dealt with more explicitly in Section 4 in the
next few slides. And also, develop detailed guidance
and provide for its implementation.
After the report came out, the DOE sent a
technical direction letter over from our contracting
officer over to Bechtel SAIC and told them to develop
this detailed guidance based upon a strategy, either
this strategy or one similar to it, and incorporate
that strategy into the planning exercises they were
doing to get us out to license application. And
that's what Peter is going to talk about in the next
talk.
At the meeting in Pahrump of the NWTRB,
detailed implementation was being developed -- a
document. I have a copy of it here somewhere. It was
being developed at that time, but now it actually has
been developed. And Peter will talk about that.
Now, Chapter 4 -- or Section 4 of the
uncertainty analyses and strategy report -- that's the
title of it -- "Communication of Uncertainties." This
exact figure is not actually in the report. There's
a very similar figure in the uncertainty report. I
think it's -- I wrote it down. It's Figure 2-13 on
page F-18 that's very similar to this. But this is
the slide I showed in January.
And what I wanted to get across -- for
those of you that -- you saw Peter's slide this
morning, slide 9, that showed the black, the blue, the
green, and the red curves. Carol in September showed
a similar such figure when she was making a
presentation for somebody else at a Nuclear Waste
Technical Review Board meeting. Tim Sullivan was
sick, and so Carol made that presentation with a very
similar figure.
And there were comments from Dr. Knoppman,
a member of the NWTRB, on the fact that that figure
doesn't show any uncertainty. It just shows means.
And so we took that comment to heart.
And if you go back and you look at the
preliminary site suitability evaluation document that
was out last summer, that also had figures of that
type which didn't show any uncertainty, where now if
you go and look at the final site suitability
evaluation documents you'll see this figure and some
of the other figures that I'm going to show in this
talk.
At the time of the talk in Pahrump in
January, the site suitability evaluation documents
weren't final yet, so I couldn't reference them. But
I was pretty sure that this figure might end up in it.
This figure is also -- Peter showed, I believe it was
slide 10, this morning, the one that he had labeled as
the probability weighted dose axis.
There was -- you could have read about the
controversy about that figure even in the general
press. It made the Las Vegas Review-Journal, and The
Sun, and also some of the energy-related documents.
The Nuclear Waste Technical Review Board,
even in their most recent letter to DOE, had concerns
essentially that generated from this figure and the
one that Jerry McNish had shown in the presentation
before, in that -- and I'm reproducing it here exactly
how it was shown in January to show -- it's
interesting that it comes up in a talk about
communication of uncertainties.
The concern is is that it's just labeled
as total annual dose, with no recognition that it's
probability weighted. And there were some concerns
perhaps that things were not being communicated quite
clearly. But as I said at the meeting in Pahrump, if
you go to the uncertainty analyses report, you'll see
an explanation down here that does describe it as
probability weighted.
Or if you go to the SSE, the site
suitability evaluation document, you'll see a big
paragraph that explains the fact that it's probability
weighted. But for a PowerPoint presentation, in order
to have a nice, big figure, that was stripped out.
So, you know, there were no ill intentions, but it
just shows that in communicating sometimes there can
be unintended consequences.
Now, all of these charts -- these next few
charts deal, as Peter has already described these
charts -- these have to do -- when it says "total," it
takes the disruptive igneous event doses and adds them
to the nominal. In a sense, they just look -- because
of the magnitude of the igneous doses, they just look
like the igneous doses.
I would much rather show the nominal
results. But by the time I get a few slides in you'll
see that in order to make meaningful graphs of some of
these results we have to go with something like the
igneous results, not the nominal results, because the
nominal results produce too many zero doses and they
don't make very meaningful graphs.
CHAIRMAN HORNBERGER: So I take it you've
solved this problem, and you now know how to
communicate --
MR. BOYLE: Yes.
CHAIRMAN HORNBERGER: -- clearly what a
convolution integral is to the lay public.
MR. BOYLE: We try.
(Laughter.)
I will say the January meeting ended at
midday. I drove right back and met with Ken DeLugo
for -- he was working on the site suitability
evaluation documents. We went through every figure in
there to make sure that the Y-axis was correctly
labeled and that we had the big paragraph explaining
it.
So we did take the Board's comment to
heart. We did not want to be misrepresenting anything
to anyone.
MEMBER GARRICK: Did the Board want you to
continue to show all the realizations, given that --
MR. BOYLE: I'll get to that. But wait
until you get to the next slide. One of the
recommendations in this Section 4 of the report is,
with respect to communicating uncertainties, there are
different audiences. Some people are much more
comfortable with a lot of detail. And decisionmakers,
or those that don't have a background in mathematics,
or in TSPA in particular, perhaps need less.
This is full-blown. But even in the
preliminary site suitability evaluation, we never
showed any such thing, which led to the comment about
Carol's presentation in September. So we did want to
show -- this shows all -- this shows probably a
maximum amount in terms of what you would want to show
in the results.
But since some people have difficulty with
the horsetail diagrams, one of the recommendations of
Section 4 is we'll thin it out some, if you will, you
know, clear it up. So these are essentially the same
results, but it's just shaded in between the 5th and
95th percentile, still showing a mean.
To remove some of the distractions of all
of the horsetails, try and get it across simpler, that
-- if you will, that this is an air band, if you want
to think of it that way, and it was shaded in to show
the possible range of results between the 5th and 95th
results. And this slide also is now labeled
probability.
CHAIRMAN HORNBERGER: Why the mean and not
the median?
MR. BOYLE: Why the mean and not the
median? Because it's the regulatory measure. That's
-- just make it simpler. You know, my wife
understands the difference between the mean and
median. She had to take a course. But many people do
not, so --
MS. HANLON: Bill, before you go on, I
just want to mention, since Bill is -- has talked
about the fact that Dr. Knoppman, as well as Dr.
Cohen, mentioned several times that they were unhappy
with the level of treatment of uncertainty, in the
final site suitability evaluation we did spend a great
deal of time, both in the executive summary as well as
in Chapter 4, going into a discussion of uncertainty
and putting more treatment in with what Bill is
talking about.
MR. BOYLE: Yes. And we may have added
the first -- the two figures, the full horsetail
diagram, which was a change from the preliminary site
suitability evaluation. I believe we added this one,
and I believe we added this one. And this is the
figure that gets across why I'm showing the igneous --
the combined total doses rather than the nominal
results.
This represents a cumulative distribution
function and a relative occurrence of PDF, if you
will, of the 5,000 realizations for the igneous doses.
And we get a nice, smooth cumulative distribution
function based on those 5,000 realizations. It goes
all the way from zero to one.
Whereas, in the nominal, within the 10,000
years, which is what the site suitability evaluation
dealt with, some 70 or 80 percent of all the
realizations for the nominal case are actually zero.
And it makes -- you end up with a funny-looking
cumulative distribution function, which I didn't want
to have to go into all that explanation, so we chose
a data set that gave a nice, smooth one.
And this figure is in the site suitability
evaluation, and what it represents is at the time of
the peak in this plot, at 312 years, right here, we
looked at all 5,000 realizations and plotted them up
as a cumulative distribution function and as a
relative occurrence of probability density function,
if you will.
And it can be seen just at first glance
because of the log scale, but it's a first cut. And
so they look approximately normal, so it's a log
normal result that's a first cut.
Then, my last slide, I ended with a quote
from Charles Darwin. I thought it was appropriate
relative to TSPA and uncertainties first, and so I
don't -- I haven't read this book by Darwin, but I got
the quote out of a book of quotes. And I don't know
in which -- what context he made this.
But it's interesting that it's by Darwin
and that our TSPA has been evolving, not by natural
selection, but we hope by survival of the fittest --
you know, the better models for surviving.
Also, TSPA -- it relates to TSPA in that
we are looking at the future in a TSPA, and we also
must make judgments based with conflicting and vague
probabilities. And with that, I turned it over to
Peter with one last explanation.
I think as perhaps this committee knows
full well, that there apparently are perhaps two types
of analysts, those that are very comfortable with
bounding, conservative approximations, and others that
want a fuller representation of the uncertainties
involved.
And I had -- after I put these slides
together I attended a National Academy of Sciences
meeting, Committee on Geological and Geotechnical
Engineering, where that discussion came up of the
frustrations when the two groups collide.
And it had nothing to do with Yucca
Mountain, but it put it in perspective for me that
we're not the only project that deals with this choice
of, do we just bound it and get on with it, and remove
some of the information, or should we deal with the
uncertainties more fully?
And I said at the January meeting in
Pahrump that the two different approaches, when viewed
in the extreme by the proponents of the other
approach, can be viewed as an unyielding rock, if you
will, one that doesn't yield any sort of information,
whereas the other can be viewed as this big whirlpool
that sucks in all available time and money.
And with that image of a rock and a
whirlpool, between which a path has to be charted,
brought to mind Odysseus sailing between Scylla and
Charybdis. And for us I said, "Peter is our Odysseus,
who is going to tell us how he was to chart a course
and the detailed implementation of how we were to
treat the uncertainties."
And at that point, the guidelines that I
showed you were in the process of being prepared, have
now been prepared. I think they provide a proper
course on how to deal with uncertainties. I'd like to
think that this committee would feel the same way, but
there's always a little caution in that, you know, the
answers in the implementation, you know, that the
guidelines are not that prescriptive in terms of
everybody would follow them exactly the same way.
So time will tell, but I'm heartened by
the approach that Peter and his staff have developed.
And I think he will tell you about it now.
MEMBER GARRICK: Just to telegraph
something that may be for the benefit of Peter is that
the problem is not whether the situation lends itself
to a bounding analysis or a probabilistic analysis.
The problem is that when you do a bounding
type analysis and you try to embed it in a
probabilistic analysis with language that's very
confusing, an example of which is to say, "Well, I
don't know what the solubility is, and I don't want to
put a distribution on it. So I'm going to assume that
this is what it is, and it's an upper bound." And
then you later say that there's no uncertainty
associated with the solubility because you assumed a
point value and as an upper bound.
Now, that's where you throw the system
into total turmoil, and that particular flaw is very
evident in the TSPA-SR. It's one thing to use
bounding analysis in a screening capacity, and what
have you, but it's another thing to use bounding
analysis on something about -- something that's very
uncertain, and then, in the wrap up say that there is
no uncertainty associated with it because you bounded
it.
And that's the same as ignoring the
uncertainty, and that's something that we have real
concern with.
MR. BOYLE: Right. And I think on that
same issue the Nuclear Waste Technical Review Board,
they used different words, but it's the same issue of
how --
MEMBER GARRICK: More elegantly, I'm sure.
(Laughter.)
MR. BOYLE: We ended up, particularly in
the TSPA-SR, they commented on it in a letter of
March 20th, 2000. We have this mix of where we've
incorporated uncertainties for some parts, did not for
other parts, and we've got this mix. The guidelines
that Peter is going to talk about I think will try --
will end up in a better situation.
Hopefully, at the end of the
implementation of those guidelines we won't have this
unknown mix of uncertainties. We may still have some,
you know, approximations and bounds in it, but
hopefully we'll have a better handle on it. And
that's what those eight bullets were supposed to get
at, and then Peter was to implement it.
MEMBER GARRICK: Okay.
MR. HINZE: John, can I ask a -- Bill, can
I get to your fifth column, a detail on your fifth
column, which is kind of an ominous title. On page 9,
you have, "New analyses may lead to reduction of the
probability of explosive, eruptive phenomena." What
analyses are these? Could you explain that a bit to
us?
MR. BOYLE: You know, I would have to --
I didn't --
MR. SWIFT: The question is -- it goes to
the type of volcanic eruption. Some volcanic
eruptions involve violent eruption and ash pushed
quite a long way into the atmosphere. And those are
the ones we're worried about. They're called violent
strombolean eruptions.
They're relatively rare in the geologic
record from Yucca Mountain, but not -- they're there.
But they're not the most common type, which are normal
strombolean eruptions, which produce a cinder cone
directly around the point of eruption and do not
produce ash blankets over a large area.
The question is: what fraction of our
eruptions are actually violent? And when does the
violent phase occur? Is it early in the eruption or
late in the eruption? If it's early in the eruption,
then that's the time we worry about. If it's late in
the eruption, the waste may already have been ejected
into a cinder cone close to the conduit rather than
being pushed out 20 kilometers.
For the SR and for all of the work you've
seen, we took the copout path of bounding it with the
assumption that our eruptions were, indeed, violent --
the strombolean ones. And so if we can justify a
basis for saying that some -- only, say, 10 percent,
20 percent, whatever -- we can justify a value, we'll
try to use that and produce our eruptive probability
that way, our probability of violent eruption.
MR. HINZE: Thanks.
MEMBER GARRICK: Any other questions for
Dr. Boyle before he sits down? Okay. Thank you.
MR. SWIFT: I wasn't completely prepared
for it in Pahrump when Bill introduced me as Odysseus.
I wasn't prepared for that. But it did occur to me
that at least one point was relevant, that Odysseus
had been on the road far too long and -- 22 years, was
it? And whether that was me or the project, I wasn't
quite sure. Also, it didn't have a happy ending
either.
(Laughter.)
So Odysseus is not the analog here. It's
Scylla and Charybdis that we're worried about.
You've got to start thinking about the
treatment of uncertainty with this question of
conservatism versus realism. And these are just some
simple observations here that -- many reviewers of our
TSPA have criticized a lack of realism. There's a
list of them, and this group is right there.
Obviously, there's a common theme. People
are looking for something we're not providing.
The second bullet here is my own
observation that I believe in general these reviewers,
when they review the TSPA and find a lack of realism,
they are in many cases not distinguishing between the
TSPA and the underlying process models. For them, the
TSPA is a window into the process models.
So if our process modelers make the
assumption that they will bound a solubility limit
within a range of uncertainty, we carry that forward
into the TSPA. And, yes, it's a lack of realism.
It's actually, I believe, a lack of realism in the
underlying process models.
This is appropriate. I think a good TSPA
should be a window into the underlying science. It
should be the first place you go to look to see how
well we understood something. But there are
differences, and it's worth keeping those in mind.
There are some places where we may have a more
realistic treatment at the underlying level, and for
good reasons have chosen to simplify it in the TSPA.
All of the reviewers' comments and
expectations with respect to realism -- there's a good
-- excellent summary of them in the Coppersmith and
McNish report that Bill just mentioned, Section 3.1.
But to me, I'm focused on what's in the rule, what has
the NRC asked for in the rule.
And this is the two clauses out of the
definition of "reasonable expectation" that basically
for me sum up the issue pretty well. And I think,
fortunately for the reviewers listed on the previous
slide, these two bullets actually do put the key
thoughts directly into the rule.
Characteristics of reasonable expectation
include -- do not exclude important parameters simply
because they are difficult to precisely quantify. And
this one focused on the full range of defensible and
reasonable parameter distributions, rather than only
upon extreme physical situations and parameter values.
These are the words out of the rule.
I actually take some heart in the site
softness of language here. It's not fully
prescriptive. It doesn't say focus exclusively on the
full range or only use a full range. Rather, it
suggests to me that we're looking for some common
sense here, but, clearly, the goal was -- the goal is
a full treatment of uncertainty.
So what's in our -- the guidance that we
came up with for the project? What we're looking for
is some version of a realistic analysis rather than a
bounding one. But what's admitted right up front,
some conservatisms will remain. Our job is to be
clear about where they are, what the basis is for
them, and what their impact is. There are cases where
the applicant, I believe, is going to end up being
conservative and explaining why and what -- how it
matters.
Focus on a realistic treatment of
uncertainty. That's not the same as a full
understanding of realistic performance. This is a
sticking point within the project. Realistic
treatment of uncertainty sometimes gets equated with
a full deterministic understanding of reality. And
the first here is achievable -- realistic treatment of
uncertainty. The full understanding of realistic
performance is not achievable. That would be the --
that would require 10,000 years yet.
So the bullets that go along with that for
me -- simplified models are okay in the TSPA. Broad
uncertainties are okay, if they're justified and
explained. This is important. Scientists generally
think of their job to be to reduce uncertainty. We
need a shift in mind-set here. Our job for TSPA is
not to reduce uncertainty; it's to make sure we've
adequately included it.
So broaden the range of uncertainty rather
than -- based on present knowledge, if you weren't
confident with the uncertainty bounds you've put in,
make them broader. If you weren't confident in them,
that meant they weren't broad enough. And then see if
they matter.
MEMBER GARRICK: I'm pleased to see that
there. That's a very important issue.
MR. SWIFT: These are just words so far.
We still have to implement these. But that thought --
the shifting of a scientist's mind away from 20 years
of experimental work driven to reduce uncertainty to
the simple statement "give me a broad uncertainty
amount," that's a difficult shift.
Scientists and PA analysts need to work
together to incorporate uncertainty in the TSPA. I'll
have more to say on that. But it -- it can't be done
by either the process scientist or the PA analyst
independently -- and focus on a clear explanation of
what we did, mathematical/conceptual descriptions. If
we're talking about parameter uncertainty, you'll
actually be able to see the equations in which the
parameter was implemented and the traceability.
That's something to strive for.
This thing called the guidelines document.
This is -- Bill described it as having been required
contractually by the DOE in a direction letter in
December. It was delivered on March 1st. It's a
rather dull document. I apologize. Guidelines for
developing a document and alternative conceptual
models, model abstractions, and parameter uncertainty
in TSPA.
It's, I say, dull because we don't want to
call it a procedure. We are not -- it's not a quality
assurance procedure in that sense, but it reads like
a procedure. I wish I knew how to fix that.
It describes the -- it meets the
requirements of the technical direction letter by
implementing the strategy outlined in the report Bill
described. It also addresses some NRC KTI agreements,
and the last page of this handout is the text of those
agreements.
The important thing here is that it uses
a team approach for both models, the alternative
conceptual models and the abstractions. And for the
parameters we set up a three-cornered team -- a
triangular team, with a lead for models, the
abstraction models. I use the same person as the
lead, he or she, same person does the lead for the
alternative conceptual model work. And a parameter
lead, and then a subject matter expert and a TSPA
analyst.
So think of it for parameters, where there
is one parameter team lead, but for each uncertain
parameter in the TSPA there will be a subject matter
expert and a TSPA analyst who -- the three of them
jointly have to agree on the distribution for that
parameter and actually sign off on it. Likewise, the
models.
The model abstractions -- the goal of
abstraction is to capture the important processes, the
processes that are important to system interactions,
and to make sure that the abstraction allows an
appropriate representation of uncertainty.
This is important. The abstraction is
going to use simplified parameters, often lumped
parameters, to capture quite a lot of things. They
have to built with an eye towards, can we actually
assign uncertainty -- representational uncertainty to
those parameters in a meaningful way?
The sections get developed by the subject
matter experts. These are for the scientists --
reviewed by the process model analysts. They're
developed in the scientists' reports. These are the
AMRs, the analysis and model reports.
There is no prescription on how to
actually do an abstraction, recognizing that they can
be everything from -- well, not listed -- you could
just put the full numerical model into the PA. You
could simplify it, simple functions, response
services, parameters.
The implementation in the TSPA gets back-
reviewed by a subject matter expert, and that
implementation gets documented in the TSPA's report.
For alternative conceptual models, there's
a little simple step-through process here that we're
asking our model developers to walk through. For each
process of interest in product alternatives, if any,
with consistent available information, there's no
requirement here to go out and make up alternatives.
In fact, there aren't any that are consistent
available information.
If only one conceptual model is consistent
with all the information, that's good. That means
you've -- you don't have a valid viable alternative to
conceptual models. Instead, you have things that can
be screened out. And you document that at that point.
That basically is part of our FEP screening process.
Things like, for example, seismic rises in the water
table that might flood the repository is not an
alternative conceptual model because it is not
consistent with available information. We believe
that can be ruled out.
If you have multiple viable alternative
conceptual models, evaluate their impacts on subsystem
and component performance. That's the process model
or the specialist in that area. If there are
alternatives, if the alternatives result in the same
subsystem performance, i.e. the same information that
you delivered to the system model, then, again,
alternative conceptual model uncertainty is not a
significant source of uncertainty in the total
analysis. Doesn't matter which alternative we use,
we're getting the same result out of it.
If two or more show different subsystem
performance, develop abstractions for both and deliver
them to TSPA. That takes you back into the
abstraction process. Basically, have them reviewed by
TSPA and implemented.
Here's a point. If the abstractions for
the alternatives are not straightforward, this is a
place where I think you're going to see some
conservative choices come in. I don't really have an
example in my head, but some -- let's suppose somebody
proposes an alternative conceptual model which would
show improved performance but is going to be a heck of
a chore to abstract it into the TSPA. Perhaps the
example I gave earlier of metrics diffusion in the
unsaturated zone might be one.
This is a place where I think the project
will probably take the cost effective approach and
explain why they're being conservative.
TSPA evaluates --
CHAIRMAN HORNBERGER: Peter?
MR. SWIFT: Yes.
CHAIRMAN HORNBERGER: On that point, it
strikes me that what you're -- if what you're saying
is that you have alternative conceptual models that
are consistent with information, then you don't have
a clear way to choose one over the other.
MR. SWIFT: Right. Yes.
CHAIRMAN HORNBERGER: And it strikes me
that all you're saying is that, fine, if they give
different performance we will use the one that shows
the worst performance.
MR. SWIFT: Yes. Well --
CHAIRMAN HORNBERGER: Is that right?
MR. SWIFT: That will do. The question
is, at what level do they show the worst performance?
If they're showing different performance at the
subsystem level, that isn't -- doesn't for sure mean
they're going to show different performance at the
system level. But, yes, other than that I -- same as
what you just said. But the idea is to actually
direct people who document this process of thinking.
MEMBER GARRICK: It seems there's kind of
a corollary rule here that would apply, too, and that
is that if you have multiple conceptual models -- and
let's say that those models provide the same results
-- then you ought to use the simplest model as the
basis. This is the Copenhagen rule for the great
physicist.
CHAIRMAN HORNBERGER: It actually precedes
Copenhagen, because it's William of Ockham in I
believe it was 1674.
(Laughter.)
MEMBER GARRICK: Well, Niels Bohr picked
it up and --
(Laughter.)
-- made the point very elegantly that if
you have multiple theories, and they give you the same
results, we're going to, by damn, take the simplest
one.
MR. SWIFT: The problem is when they don't
give you the same results.
MEMBER GARRICK: Yes, I understand. But
there is that issue that there is a tendency sometimes
for modelers to want to impress you with the
complexity rather than impress you with the
simplicity.
MR. SWIFT: If the two models give the
same subsystem result, my conclusion, alternative
conceptual model uncertainty is not significant. The
under bullet not stated there is that the subject
matter expert then has to document that as to, yes, I
have these multiple alternatives. They all give me
the same result. Therefore, I'm only going to deliver
the simplest one, or the one of their choosing,
forward in the TSPA.
MEMBER GARRICK: Yes.
MR. SWIFT: And that actually is in the
guidance document. That step is there.
VICE CHAIRMAN WYMER: Now, this is turning
out to be a lot of extra work, but -- and looking at
various parameters. But what you haven't said is how
the parameters should be looked at.
MR. SWIFT: Let me get to that on the next
slide when I talk about parameters. Let's imagine
here that the alternative conceptual models are
implemented in TSPA and you actually run a full TSPA
or a subset of TSPA with the different alternatives in
it.
And if the options -- the impacts are
significant, then the options are -- there are
basically two options. One is you can carry the
multiple alternatives all the way through to the
regulatory dose, but then you have to weight the
alternatives, and then you have to be able to defend
those weightings in some way.
So that may not be the -- the first simple
thing, you always give them equal weight, but if they
-- and see if it makes a difference. If they don't
make a difference, then you learn something. If you
can't defend weights, then at that point, again, you
default to the more conservative one if you've gone
through this.
Parameters --
VICE CHAIRMAN WYMER: I'd say selecting a
parameter value is not the same as selecting --
MR. SWIFT: Yes, I'm aware of that. The
assumption here that you just caught me on is we
already know which parameters matter. And the
identification of we're actually -- this is the step
in the process that we're in right now. It's
identifying the parameters we want to treat as
uncertainty parameters in the TSPA.
And we're doing that by -- the TSPA team
is providing a list of the parameters that were
treated as uncertain parameters in previous analyses
back to the subject matter experts in each of their
areas for review and updating.
There are parameters that have been
treated as uncertain parameters in past analyses that
actually aren't doing very much in the analysis. The
analysis would be insensitive to the uncertainty in
them. If that uncertainty was appropriate, you know,
justifiable, defensible, and still was doing nothing,
then that parameter might be a candidate for one to be
switched to a fixed value.
If, on the other hand, the subject matter
expert looks at that list of uncertain parameters and
says, "Whoa. Here's the one that really captures the
process. Better put a distribution on that and get it
in there," that will happen. But the real answer to
your question is that this is -- it's human judgment,
and this is why an iterative analysis better, because
people learn things through time as to which sources
of uncertainty matter at the system level.
We've learned a lot in 10 years of TSPA
and interacting with the process model teams. I
actually do think that we have the right uncertain
parameters, probably more of them than we need, and
there isn't a unique test to make sure you've gotten
them all. That's, from a judgment, an iteration and
review.
But once you've got the list of uncertain
parameters identified, categorized, they get mapped
back to the subject matter experts for documentation
in their AMRs. And the full range of defensible and
reasonable distributions gets documented by the
subject matter experts in their AMRs in that
triangular model with the team lead and an analyst.
There are two things yet to consider in
building uncertainty distribution. First is the
available data. But, second, and this is the part
that typically gets missed, you have to think of how
the parameter is used in the model. Model scaling
issues, what's the cell size in the model. These are
numerical models, and it makes little sense to use
porosity data collected at a sidewall core this big,
use that exactly as is in a model where you might have
cell blocks hundreds of meters on the side. The
distribution is something different.
So think of spatial variability, which is
the example I was offering there, because it affects
the scaling of the parameter, how it's used in the
model. This is the point where you want the modeler,
the person who actually knows what the parameter is
doing in the equation in the model, working with the
subject matter expert most familiar with the data,
working with the team lead who is -- will have -- the
statistician who is supporting them in how to apply
the -- how to build a defensible distribution from the
available data.
And it's typically not a matter of fitting
it with a normal or log normal or some specified
model, because nature doesn't work in statistics like
that. What you want is -- what we want is a
distribution function that doesn't add any new
information, honors information we have, doesn't
create new knowledge.
So the simplest example of such would be
a peacefulized linear distribution. If you actually
thought data itself was appropriate to be used in the
model as was -- as is -- a peacefulized linear fit for
the data is better than trying to force fit a normal
or log normal distribution.
But the distribution is -- ultimately,
it's a subjective decision, and you want it made by
the right experts -- the scientist, the PA modeler,
and a statistician with experience in doing that. And
then you want it documented, and so we'll do that, and
then implement things through a controlled database.
MR. BOYLE: And also, I think in the way
this system is set up, it's the consideration of
alternative models frequently gets at which parameters
are under consideration. For example, if your model
is that the rock is elastic, well, Young's module and
Poisson's ratio is sufficient. If, on the other hand,
you assume that it is viscal plastic, well, then, that
generates a whole new set of parameters for which you
then need values.
MR. SWIFT: The last slide here, this one
actually got edited a little bit in the final review,
and you'll get a kick out of what came out of it here.
This bullet used to say that regulators and reviewers
are not asking for the impossible, and someone felt
that was a little too negative.
(Laughter.)
But I actually believe it. And I'm
getting back to the idea there that if we, the DOE,
were to misinterpret what you're asking for as -- I
hope I'm going to get a head nod here. If we were to
misinterpret that as asking for a full, realistic,
deterministic solution to the future, that is
impossible. We're not going to do it.
But we can commit to a realistic treatment
of uncertainty. Can we actually achieve it? It will
be some version of it, but there will --
pragmatically, there will be conservatisms here and
there. And it's our job to explain what we did.
And this is the last point -- there's no
unique solutions. A lot of credibility comes from how
well we can explain it.
And that's it for that presentation.
MEMBER GARRICK: Before we go to the next,
any comments from members of the committee? Milt?
MEMBER LEVENSON: Well, I'd like to make
one comment. It's sort of a follow-on to the comment
John made earlier, and that is that it's not of major
importance that we reduce uncertainty. The importance
of uncertainty is only to make sure that the true
extent of some risk is not obscure, and that, as in
medical work, false positives are equally to be
treated as false negatives. And the whole objective
is to make sure that you know enough so that you can
evaluate the risk. Now, reducing uncertainty per se
is certainly not an objective of mine.
MR. SWIFT: Others see things differently.
I agree with you completely, but the -- on the
alternative conceptual model side, for example, a
question I got from the ATRB was, well, where is the
step where you go out and design an experimental
program to go back and test those models and throw one
or the other of them out?
And that isn't where I actually was
thinking. I was thinking we're going to make a
decision based on information that we have now. We're
going to decide if the uncertainty matters. And it's
a different way of thinking of things.
MEMBER GARRICK: Ray?
VICE CHAIRMAN WYMER: Well, actually,
you're not going to make a decision based on
information -- based on what you have between now and
the time of your license application. No, I've raised
my questions already.
MEMBER GARRICK: George?
CHAIRMAN HORNBERGER: Again, just a
comment following on what you've just said, Peter. I
think in part, as I read some of the TRB comments, it
would be that the question is whether or not there is
an adequate scientific base to support the models that
you have.
MEMBER GARRICK: Yes. And I like the --
your remarks about -- that the distributions or the
uncertainties need to be driven by the information or
the data. Too often we've seen people spending a
great deal of time and effort and exercise on trying
to choose a distribution that will work for them in
their model.
And there's enough analytical tools
available now that there's no reason for doing that.
We ought to be able to forget about whether it's log
normal or beta or gamma or whatever, and let the
information, however it comes out as a distribution,
be the basis of the model.
And even if it's a histogram, because
there are tools now that very effectively convolute
discrete probability distributions. And that is even
a more -- often a more accurate representation of
what's taking place, and very often much easier to
follow.
So these are encouraging signals that
you're giving in the context of guidance. I think
it's very important. And I think it also addresses
this whole issue of the confusion that sometimes
exists between good science and adequate science with
respect to solving a problem, and we've talked about
that a lot in this committee.
We want good science, but we don't think
it's necessary to reduce an uncertainty between, say,
10-12 and 10-7, even though it's five orders of
magnitude, if the risk is of the order of 10-3. So
that's adequate, even with that wide amount of
uncertainty. So these are steps that are very
encouraging to us.
Any questions from the staff? Yes.
MR. HAMDAN: Uncertainty means different
things to different people. Milt and I and perhaps
everybody in this room understands uncertainty as has
been described, and Milt described it very well. And
on slide 11 from your first presentation --
MR. SWIFT: In my first presentation?
MR. HAMDAN: Yes. Which igneous activity
-- you showed us the slide about the effect of igneous
activity -- if you want to -- this was what obscured
uncertainty. This is very clear. It goes to the
point and evaluates the effects of igneous activity of
a property of one. So this has nothing to do with
uncertainty.
But to the public and people who are on
the street, this is the uncertainty. You are saying
in this slide to them that if -- it were -- this is
the point that you are going to give to them, not to
us in this room. We evaluate the risk and then we
make a recommendation based on risk.
But to the people on the street, this is
uncertainty. And I think this needs to be looked at
and responded to and articulated to the public. So
this is the comment that I make.
I have another question for Peter, and
that is on your slide on the parameter uncertainty.
You probably will need to pull that out. The approach
is fine, and you have articulated it very well. The
real questions come with the -- when you want to
assign probability distribution to a certain
parameter, that's where the rubber meets the road.
There sometimes you don't have enough data
to select a distribution, and that's where the problem
lies. There are a lot of parameters with a
distribution that have validity bases, and I wonder if
you could extend your answers from that -- for
conceptual models to do alternative distributions for
these parameters, and satisfy yourself that, really,
it does not make a difference.
And that will probably be needed because
there is simply a lot of parameters with uncertainty
for which we do not know what the right distribution
is.
MR. SWIFT: My own experience in analyses
like this is that for parameters to which the results
are sensitive, the form of the distribution is less
important than the range. What you're worried about
are the impacts of the tails.
And so the difference between a log
uniform distribution and a log normal distribution
will not be that great. The difference between a log
uniform and uniform distribution may be very
important, but I think picking distributions that span
a broad enough range of uncertainty that the range
itself is defensible -- just what the scientists
believe is a broad enough range -- will take quite a
lot of the concern off the shape of the distribution,
the actual form or function used to fit it.
MEMBER GARRICK: I see this next
presentation is -- yes?
MR. HINZE: Well, I guess wanted to
suggest that the most important, the most critical
phase of this whole thing is making the decision on
when you have done sufficient sites that you have a
model that you can do some calculations with. It's
very hard to put uncertainties on that.
It reminds me a bit of the ore deposit at
Roxby Downs in south central Australia, which is the
most important mineral discovery since the Second
World War. It was found as a result of a very
incorrect model. The answer was just beautiful. It's
correct, but it was a totally incorrect model.
Now, if you use that incorrect model,
which some people have tried to use, in other parts of
the world to find a similar ore deposit, you're just
not going to get there. It seems to me that you can
put parameters -- the uncertainty around these
parameters, but it's the question of when you've done
sufficient science that you understand the process
well enough so that you can make a judgment, and that
judgment will have uncertainties.
MEMBER GARRICK: That's correct. Okay.
I see this next presentation is a big one.
MR. SWIFT: It's not as big as it looks.
I'll give you the fast version.
MEMBER GARRICK: We would like to hold, as
best we can, to our 12:30 recess.
MR. SWIFT: Me, too.
So I will certainly not go through all of
these. I will give --
MEMBER GARRICK: I have an advantage. I
can ask all the questions and extend the time, and
then blame you --
(Laughter.)
-- blame you for overrunning.
MR. SWIFT: Can we have the lights here?
So we can see the screen better. Good. Thank you.
Since December, the project has gone
through a replanning exercise. You'll recall that in
-- starting in the summer of 2001, the project went
through a planning exercise for a multi-year plan
prepared by the M&O contractor, BSC, and then to be
approved by the DOE, for the work to be done to
support a license application.
And the plan, which was submitted in
September, produced a large body of work that was --
scientific work that went out to about -- an
application in about 2006. The dates were somewhat
flexible. But the DOE came back to BSC and said,
"Perhaps you want to replan."
There was not a prescribed date. However,
we felt it was prudent to replan with the idea of 2004
in mind to see if we could, in fact, identify a scope
of work that would allow us to produce a docketable
license application in 2004. Of course, docketable is
the NRC's decision, not ours, but we wouldn't submit
an application unless we felt there was a reasonably
good likelihood that it would be docketed.
So with that in mind, we set out to
prioritize work in the performance assessment and
science activities, which we realigned the project at
the same time, so that the science became part of the
performance assessment project, and focused primarily
on the work that was necessary for license
application, identify and select an overall scope of
work to balance the project management risks. This is
not human dose risk. This is the management risk.
What's our risk of success or failure? And document
it.
And this, then, would be the basis for the
replan that was delivered to DOE March 1st. Back in
December, we were planning ahead. In fact, we did --
BSC did deliver a new multi-year plan to DOE on
March 1st, and that is in DOE review.
And that has not been released yet. Is
that correct? The plan B of -- anyway, it's a big fat
thing, work plans for the outyears, starting
immediately but out to 2004.
And inform these decisions with input from
the TSPA analyses, technical staff working with the
science program staff, line management, project
management -- this will be the senior management team
-- and then project planning. These are the people
who ultimately -- they're the ones that have to
prepare a multi-year plan.
In theory, it's -- no, more than theory,
it is a resource-loaded schedule where you can
actually point to a schedule and see what things cost,
how long they take, and what they do. And that's the
-- so the process we wanted to go through here, just
for -- this is -- remember, we're only prioritizing
work within performance assessment and science.
This does not include design activities.
This does not include licensing activities, quality
assurance activities, the various support activities.
It's only a portion of BSC's budget that was developed
in such studies.
The PA team identified attributes --
basically, a short story here -- we're headed for a
multi-attribute utility analysis. And we've done it.
That's where this clarification is headed.
The PA team defined the attributes at
which the work scope was evaluated. The department
managers of the science departments defined work
scopes to be considered for each model, component.
Think of the unsaturated zone or the saturated zone,
and so on. For each model component, they defined
alternative work scopes they wanted to have
considered. For each of those they should have
estimates of cost and time.
And the department managers and the TSPA
modelers provided initial estimates of the impact of
the proposed work on the attributes. Basically, an
attributes questionnaire. We scored each work scope
description.
There were 25 model components, and each
one had about -- almost each one had three work
scopes, so about 75 different work scope descriptions
were scored against these attributes. And that was
actually done in a workshop in January where we had
the key players all together in a room for three days
and went through scoring the -- first of all, we wrote
final work scope descriptions, we scored them against
the attributes, and I'll go through how we -- what
that means here in a minute.
And then we ran them through the utility
analysis tool that I'll describe in a minute, produced
an initial prioritization, in mid-January had a
management review of it, and provided input to our
budget team at the end of January. And that, in turn,
has gone on to DOE now.
This is a conceptual figure that -- it's
important because this came out of a BSC management
meeting in November -- the idea that, since most of us
think best in only three dimensions, let's find --
think of it in three dimensions. What are the things
that matter to us in making decisions about what
science activities we do on the project?
And we came up with three axes -- a
quantitative performance axis. What is our calculated
total annual dose? And what work are we doing that
moves it up or down? This is, you know, basically,
are we in compliance with 63/113?
Regulatory defensibility and acceptability
-- in a regulatory framework, can we defend the models
and data used to calculate that dose? Have we met the
qualitative requirements of Part 63? This axis is
Part 63 and 197, as implemented through 63.
So things like multiple barrier
requirements to qualitative, descriptive requirement,
but that would live on this so-called X-axis.
Satisfying KTI agreements -- the NRC has
given us a list of what needs to be done to defend the
models. We, in some form or another, need to address
those agreements and produce defensible models. Then
-- and we need, say, for example, quality assurance
requirements, our so-called X-axis attribute.
Then there are the Z -- what we call the
Z-axis out this way. This was a -- Y was up in this
coordinate system when it first appeared on our white
board. So the Z-axis here, qualitative acceptability,
internal and external defensibility, these are issues
that we know we care about them, yet you can't trace
them to anything that's in the rule.
So some of these are -- some of them are
actually quantitative as well as qualitative. But
qualitative things -- defensibility of models, beyond
what's needed for a regulatory framework, the question
of, can we convince people we actually understand the
system well?
Many of the Technical Review Board's
concerns are on this axis, not all of them. I'll
argue that some of the NRC staff and center's concerns
may be on this axis. They're valid. This is not to
say the Z-axis is not important, but there are
technical issues that don't tie directly to the rule.
For a quantitative one, an easy one to
think of is peak dose. There's no regulatory limit
applied to peak dose occurring several hundreds of
thousands of years out. At the peak dose is a
quantitative Z-axis attribute.
So it's not three-dimensional space, and
those axes aren't orthogonal. And they're certainly
not mutually exclusive. We decided to define it as a
16-dimensional space for the purpose of the utility
analysis, and nobody can think in that, but the
spreadsheet does.
There are 16 attributes here that can be
coarsely lumped against those three axes, but, in
fact, for the utility analysis we scored things on
each one of these attributes without considering those
at X-, Y-, Z-axis. That's just there as a
communication tool for our own management team.
For each work scope, we went to the
technical staff and said, "Will your work, if you do
your Level 1, 2, or 3 scope" -- and I'll explain what
those are in a second here -- "will that change
10,000-year mean annual dose?" Which, by the way,
that is driven entirely by the volcano. That's the
10,000-year total. That's the Part 113 dose.
Will it change groundwater concentrations
or human intrusion? And that's it for quantitative
performance.
Regulatory defensibility -- have we
captured all credible FEPs? Have we excluded the ones
that can be excluded according to the criteria? Are
we meeting our requirements to describe -- identify
and describe multiple barriers? And do they link to
specific KTI agreements?
The so-called Z-axis sorts of attributes
impact on conference of internal reviewers, impact on
conference of external reviewers, and some
quantitative ones have come out of the TSPA
calculations. Change in time to 15 millirem. Change
in uncertainty. This would be the distribution spread
from the 9th and 5th, for example. There's no
regulatory driver for that. It's the mean we're
regulating on, but we do care about that spread in the
uncertainty and system outputs.
We looked at a forced early failure case,
peak dose, and this -- I should say associated with
conditional igneous intrusion. We had a question
there about, will your work affect our conditional
igneous dose? Will your work affect a representation
of uncertainty at the parameter level? And our
ability to defend the conceptual models.
The actual questions themselves are shown
in the handout. I have time to show those a little
bit. This goes pretty quick here.
For each model component -- there are 25
of them I'll show up here in a minute -- department
managers define three levels of work. The expectation
was that this would be an increase in cost and/or
time. Level 1 would be the quickest and cheapest.
Level 3 would be the longest and most expensive.
Level 1 -- what work would be required to
complete quality assurance issues and to validate the
existing models? That's not focusing on not
developing new models, but meeting our own internal
validation requirements for the models that we used in
the analyses I showed an hour ago -- the most recent
set of PA models, which are not -- in our own
terminology, they are not qualified and validated yet.
Level 2 scope -- take a so-called risk-
informed approach to going beyond Level 1. Risk-
informed in this sense means to us look at the impact
of the work before you decide if you're going to do
it, and, in particular, this might involve taking PA-
based -- system-level, performance-based approaches to
resolving KTI agreements rather than the literal full
scope of work that was anticipated when we agreed to
the agreement.
In other words, if we can show it doesn't
matter, is that a sufficient way to address a
question?
Level 3 -- and these are both optional.
We went to the work package managers and said, "If you
can close everything at Level 1, don't bother to
define any higher levels for us." Level 3 was
essentially the same as that plan A work scope that
got us to 2006. And with respect to KTI agreements,
it was the full and literal completion of all
activities proposed.
Managers were -- these are the science
managers, department managers, provided input on how
well each proposed work scope meets the defined set of
attribute scores with respect to the defining set of
attributes -- better way of saying that.
And the -- I'm going to skip a couple of
slides here and just go to slide 10, because the --
the same word that's on the intervening slide, where
you've got this helpful equation here. Ignore the
figures. They're not all that helpful.
But for each one of these 16 attributes,
we asked questions of the technical staff, how likely
is your work at this level -- scope of work -- how
likely is it to, for example, increase confidence in
your treatment of parameter uncertainty? How likely
is it to result in an increase in total dose? And the
technical staff provided answers that ranged from very
unlikely to very likely.
We then asked a management team what types
of -- what value they assigned to different types of
answers. That's this V thing here. Actually, it's a
relatively small player in the utility analysis. But
since it's up there -- if someone said their work was
likely to -- or, let's say, it was likely to result in
a change in dose, we then asked the TSPA modelers,
"How big a change might that produce?" These are all
subjective answers, but at least we're asking the
right experts.
And the TSPA modeler might say, "Oh, it
could increase by a factor of greater than 10, or
could have a small change of less than a factor of
10." We wanted to apply -- but this now is a
management decision, what value do you apply to the
different answers.
In this hypothetical example here, we gave
that a weight of one and a weight of zero for a
neutral effect, a small change in dose, and a weight
of .15 for an increase in dose. But that's how the
impact value function was used.
Then this weighting -- this is a
subjective management decision. How important did
management think that question was? And these were
elicited from the management team in the project.
If I go back to that three-dimensional --
you know, this figure -- if our technical staff were
all-knowing, for any scope of work they actually
could, in theory, define a vector in this N-
dimensional space that defined where their work would
put us. Would their work, you know, greatly increase
qualitative defensibility? Would it greatly increase
regulatory defensibility? And so on.
That would be the first term of that
three-term sum that went into the utility analysis.
The other two are management questions of, where does
the project want to be in that -- in this
N-dimensional space? If the project did not care
about qualitative acceptability, external reviewer
type issues, we could truncate the Z-axis and live
entirely in the X/Y plane. It appears to be the bare
minimum needed for licensing under Part 63.
But the project is not willing to accept
that risk, and, you know, the TRB certainly
understands that point. I met with them -- a subgroup
of them, just a couple of days ago, and we talked
about this. Obviously, we're not going to live only
on the X/Y plane.
So where are our decisions headed? And
that's the point of the -- listing the management
weight. And for each of the 16 attributes, for each
of the work scopes, you can define -- we define a
likelihood that the answer will be what we -- what --
that the likelihood of a specified answer will occur.
That comes from the technical staff, the scientists.
And these values and weights comes from managers.
And then, for each work scope you create
a spreadsheet and sum them up. And the -- you get a
utility. It's a dimensionless number. It's
associated with each work scope, and it's a
quantitative -- and fully comparable from one work
scope to the next -- measure of the utility of doing
that piece of work, utility defined in respect to the
questions we asked, what are the attributes, and the
weights management put on those attributes.
Now, caveats that come out of this --
first of all, like any decision analysis exercise, the
results of the model here means the utility model,
utility analysis. It's a decision-aiding tool. The
project had no intention of using it as a direct
decisionmaker. Other -- that's the most important
caveat. It's to inform managers who still have to
make subjective human judgments.
The cost assumptions that went into it
were not always consistent. Results -- we
deliberately did not constrain them by schedule. We
didn't force people to say, "You've got to make
everything end by 2004." There were some differences
among department perspectives and the impacts of their
work, despite the workshop discussions. Surprisingly
few, actually. That workshop was as close to a
consensus as I've seen when people understood what it
was we were doing.
It doesn't include all work scopes. It's
just a -- for those who are looking for the piece of
the management budget or design or testing the
interface with design or the TSPA calculations
themselves. We excluded those from the exercise.
Some questions didn't capture what we were
after. We wrote some bum questions. That happens.
Utility rankings -- we presented utility-
only rankings, and they're in the packet here, and
also utility cost ratio rankings. Utility-only
rankings ignore cost. Utility cost ratios are better.
They're clearly what the tool was designed for.
You're doing a -- we are doing a utility analysis
because cost does matter. We don't have an unlimited
budget.
But when you do a utility cost ratio, you
discover pretty quickly that not all work costs the
same. And very large work packages may perform more
poorly in the evaluation, simply because they've got
a big denominator -- cost -- where, obviously, if
you're a utility people defined their, aggregated it
coarsely, and produced expensive packages. You get a
big denominator in that fraction.
The examples here in the packet are
weights, management decision weights from two people
-- Bob Andrews and myself. In fact, in the back, in
the backup slides, there are different rankings
provided with weights listed from other groups of
people. And in the report that went with this, the
people are identified by name. They were the BSC
management team. We also listed some DOE managers but
did not include their results. That would have been
inappropriate.
What we discovered -- you can see when you
get to the back -- is that actually the management
weights, even though we had some -- quite a broad
range of management types in the exercise, we were not
as sensitive to the management weights as you might
think.
And this last caveat here, that the --
deliberately, we didn't want to emphasize -- didn't
want to focus only on things that show a positive
benefit or a negative impact either way. Both of them
are important. Obviously, we need to know if our work
is going to show poor performance. That's -- we must
know that. But we also want to value work that shows
improved performance.
MEMBER GARRICK: I think we're going to
have to wrap up in about five minutes.
MR. SWIFT: Okay. I'm there.
Just an example here, a couple of
examples. Thank you, Bill. I apologize for this.
These were three different levels of work
defined for engineered barrier system flow and
transport. The way to read this figure -- the bars
here are the amount of utility associated with that
activity for each one of these 16 attributes over
here.
And the first thing that -- we have the
big blue band here -- resolution and closure of KTI
issues. So the experts in the engineered barrier
system department felt that if they did more work, the
Level 2 work, they had a better likelihood of closing
their KTI agreements. And the band is thicker here
than it is here. But they were still -- I believe
this was a likely answer, and that was very likely.
Interestingly, although these two work
scopes are very different, they show an absolute --
this one costs $2 million more, takes two years
longer, and in the space of these questions that we
asked gives you the same answer. So that's an example
where the management decision was pretty much a no-
brainer. We looked at the one that gives you the same
answer more cheaply.
CHAIRMAN HORNBERGER: Peter, you'll have
to forgive Raymond and I. We have another meeting we
have to go to.
VICE CHAIRMAN WYMER: I apologize.
MEMBER GARRICK: That's all right. We'll
carry on here for a while.
MR. SWIFT: The work scopes are sorted by
utility at the Level 2. This was the so-called risk-
informed work scope, the intermediate work scope for
each of these areas. Biosphere scored highest. This
is simply because -- a number of reasons, but
biosphere has a high likelihood of affecting the total
10,000-year dose.
And in the management weighting used to
generate this figure, Bob Andrews and I both felt that
anything that was not moved -- the 10,000-year total
dose was the most important thing. We gave that a
very high weighting, and the biosphere igneous
activity was --
MEMBER GARRICK: Peter, let me comment on
that. Of all the things on here, the biosphere is
probably the most prescriptive. And so -- and one of
the things that prescription does is very often
eliminate a lot of decisionmaking and analysis,
because this -- the biosphere -- the regulations on
the biosphere are pretty binding in terms of how much
flexibility you have in analysis and investigation.
I'm surprised that that would end up on top.
MR. SWIFT: Well, this is a -- let me go
to the next -- a different slide. This one is in the
backups, and it's -- I'll come back to answering that
question.
MEMBER GARRICK: Yes. Okay.
MR. SWIFT: This is incremental utility.
This is --
MEMBER GARRICK: Yes, I understand that.
Yes.
MR. SWIFT: -- how much more you're
getting when you -- how much more utility you get when
you go from Level 1 to Level 2. And biosphere has
dropped well down the list here.
MEMBER GARRICK: Yes.
MR. SWIFT: The difference there is that
our biosphere team felt that even at their lowest
level of work they were likely to show an increase in
the BDCFs. And that's what drove that --
MEMBER GARRICK: I see. Okay.
MR. SWIFT: And I actually -- this is all
subjective. This is judging by work that is not yet
done, and I tried to argue with Tony Smith about this
one in particular. I don't think shifts are going up
as much as he thought they were, but he's the
technical department lead on that. And that's where
his work fell out.
That was the -- did I just put up the
incremental utility plot? I did. Okay. I'm looking
for my summaries here.
MEMBER GARRICK: Are you going to tell us
how this has affected the outcome of the decisions?
MR. SWIFT: Yes, if I can find the slide.
Yes, that was my mistake. They're over here now.
It's actually 22 and 23.
What did we do with this? We brought the
spreadsheet as an electronic tool, and the types of
rankings you see. We brought them to a BSC management
workshop where we had the senior project manager of
the BSC, not the corporate management, but Nancy
Williams, who is the project manager, and her staff --
I could name who they are. She had a -- has an
oversight -- project oversight review board. No, it's
just called project oversight board, POB, that is
people you're familiar with.
It would be Jack Bailey, John Beckman,
Gene Yonker, Nancy Williams herself chairing it.
Representatives from the national laboratories would
be Roland Carson, Joe Farmer, Andrew Worrell, Sal
Peterman from USGS, Tom Cotton, who is on it. These
people met for a fairly intensive three-day meeting to
go through the results of this spreadsheet, quite a
lot of detail.
And then we put the final rankings up on
the screen -- not final, we put the various versions
of the rankings up on the screen. They've been up off
and on for a long time, with the cost utility ratios
displayed and total cost. And we started drawing
lines in the budget where -- what can we afford? And
that would be an example that would be -- slide 20,
for example.
These are -- we do Level 1 scope for each
of them. This is what it costs in FY '02, and then we
start adding in more work at the -- from Levels 2 or
3 at the sort of -- the highest increments of utility
cost ratios first. So the most bang for the buck
principle here.
What did we discover when we started
drawing the line in the budget? That there was not as
much money available as we had happened, and we were
able to come to the conclusion fairly quickly that, in
fact, the Level 1 scope was where we were looking.
The emphasis was going to be, based on the money
available, on validating the models that were already
available. These would be the models we showed
earlier today.
However, we could not afford to move up to
Level 2 and 3 across the board, so we started taking
those work packages apart item by item. And the
management team actually, in real time, went through
all of the work package descriptions and brought
things forward from Level 2 to Level 3 into the budget
on subject human judgment bases.
A primary selection criteria of moving
work forward was to avoid canceling any tests that
were ongoing. One of the lessons learned from other
projects is: don't cancel a test if you've already
paid for startup costs. Go ahead and collect the
data.
So this is just -- this is not all PSF
tests. Some have PSF tests -- some tests are brought
forward, and other examples of testing activities were
brought forward.
And activities that were needed had to be
accelerated to support documentation activities that
could be done. Activities that were at the Level 1
scope of work but were planned to be done too late to
support license application in '04 were accelerated,
and that required bringing extra money forward.
Basically, this was a money management
tool that they exercised about how you use your money
wisely to manage your work.
And then this exercise with project
management took place in late January, and early
February we spent detailing work package descriptions
that were then delivered to DOE on March 1st.
And I apologize for running over. I have
a summary slide here. This was a decision -- the
multi-attribute utility analysis was a decision-aiding
tool, not a decisionmaking tool. We keep saying that
because that last slide included a lot of human
judgment.
You want both the technical and the
management input. The management weights are
important. That's where we decide what it is -- where
we want to be in that X/Y/Z space.
Consideration given to regulatory
requirements, technical defensibility, and money.
And, yes, we will have to reevaluate it as new
information becomes available.
MEMBER GARRICK: Okay. Thank you.
Milt, do you have any comments?
MEMBER LEVENSON: No.
MEMBER GARRICK: Appreciate the
presentations. I think one of the real problems when
you get into this business of trying to come up with
utility functions that contain preference functions is
dealing with the different groups as you have, and
addressing the biases that might exist in those
groups, because all of us think that what we're doing
is the most important. So there has to be some sort
of normalization process.
But you said that this seemed to be --
there seemed to be a lot of harmony in this case. I'm
surprised at that.
MR. SWIFT: There was. The trick I think
was to focus people on giving fair questions -- fair
answers to questions as we asked them. When -- if you
simply say, "Is my work important?" ask somebody to
answer that question, the answer would always be yes.
But if you say, "Will this specific piece of work
change a dose result?" or "Will this close KTI
agreements? And, if so, please name them and explain
how you're going to close them." At that level,
people were quite objective, and they were willing to
say, "No. Actually, this doesn't do anything to
dose."
We had a broad enough set of questions --
MEMBER GARRICK: Were they thinking in the
context of uncertainty when they answered the
question? Because --
MR. SWIFT: Not as much as I had hoped
they would be.
MEMBER GARRICK: Because the risk is
really the uncertainty. And so if what they're saying
is that it doesn't affect the central tendency
parameter, that's one thing. But if it does affect
the tails of the distribution, it could be very
significant.
MR. SWIFT: With respect to the dose
calculations, yes, they were -- they were thinking of
that sort of thing. But they were all thrown for a
curve -- thrown a curve right away by the realization
that 10,000-year total dose is really a question of
igneous activity.
And as soon as people realize that, you
know, if you want to score on the Y-axis, the
quantitative axis in that, you've got to have a
volcano.
MEMBER GARRICK: Yes.
MR. SWIFT: A whole lot of people came
into the room thinking they were going to say,
absolutely, I've got some tail up there that's going
to drive dose. Actually, no, you don't. You may
score on time, 215 millirem. You may score on peak
dose. You may score on the conditional early failure
scenario. But those all get different weights.
And because we had a broad range of
questions, I think every technical staff person was
able to feel like, yes, there is a question that
captures my -- their personal issues. But then, they
didn't know how management was going to weight those
questions, and so they were -- the technical level of
agreement was surprisingly high. People always found
a question they could say, "Yes, that's the one I'm
aiming at."
MEMBER GARRICK: Any questions from the
staff?
One thing I would say is that in the TSPA-
SR you had a couple of appendices that did a very nice
job of delineating the key assumptions, and then going
out on a limb a little bit and indicating what the
impact of these assumptions might be.
I don't know if it's the way you're doing
it, but I think that it would be very helpful with
respect to traceability and transparency for this to
be kind of a reference point and subsequent versions
be measured against this reference point. I think it
would make it very clear what -- which assumptions
have changed and what impact they've had, and which
assumptions are being driven by the decision to go to
an entirely different corrosion model, for example.
By just changing the model, you can end up
with a different set of importance rankings for
contributors. So I found that what you did in the SR-
TSPA very valuable in boiling down just exactly what
the team thinks is important from an assumption set
standpoint.
And I hope that something like that is
carried forward. I'm not saying you should do it,
because we don't advise you; we advise the Commission.
But that was -- I'm just observing that that was an
example of a transparency tool or a traceability tool
that was very helpful. And your presentations were
very helpful, and we thank you very much.
And with that, unless Carol has --
MS. HANLON: Dr. Garrick, Bill has brought
copies of his uncertainty analysis and strategy, as
well as Peter's guidelines. So we're going to leave
these here for you. If you need additional copies,
let us know.
And I'd just like to thank Bill and Peter
again for working around very difficult schedules,
including technical exchanges and Peter's out of the
country trek to be here today.
MEMBER GARRICK: We know they are very
busy men, and we know you're a very busy lady. Thank
you very much.
(Whereupon, at 12:48 p.m., the
proceedings in the foregoing matter went
off the record.)
Page Last Reviewed/Updated Monday, October 02, 2017