ACRS/ACNW Joint Subcommittee Meeting, January 13, 2000
UNITED STATES OF AMERICA
NUCLEAR REGULATORY COMMISSION
ADVISORY COMMITTEE ON REACTOR SAFEGUARDS
***
MEETING: ACRS/ACNW JOINT SUBCOMMITTEE
USNRC, ACRS/ACNW
11545 Rockville Pike, Room T-2B3
Rockville, Maryland
Thursday, January 13, 2000
The subcommittee met pursuant to notice, at 8:30 a.m.
MEMBERS PRESENT:
THOMAS KRESS, ACRS, Co-chairman
JOHN GARRICK, ACNW, Co-chairman
GEORGE APOSTOLAKIS, ACRS, Member
RAYMOND WYMER, ACNW, Member
. P R O C E E D I N G S
[8:30 a.m.]
MR. KRESS: Could we please come to order?
This is a meeting of the Joint Subcommittee of the Advisory
Committee on Reactor Safeguards and the Advisory Committee on Nuclear
Waste.
I am Thomas Kress. I'm co-chairing this joint subcommittee,
and on my right is Dr. John Garrick, who is the other co-chair of the
joint subcommittee.
I guess I'll be mostly in charge of this particular meeting.
Other joint subcommittee members in attendance are Dr.
George Apostolakis of the ACRS, Dr. Ray Wymer of the ACNW, and also
present is Dr. Milt Levenson, who is a consultant to the ACNW.
The purpose of this meeting is for the joint subcommittee to
discuss the defense-in-depth philosophy in the regulatory process,
including its role in the licensing of a high-level waste repository,
its role in revising the regulatory structure for nuclear reactors, and
how the two applications should be related to each other.
The discussion will also include the role of
defense-in-depth in the regulation of nuclear materials applications and
other related matters.
The subcommittee will gather information, analyze relevant
issues and facts, and formulate proposed positions and actions, as
appropriate, for deliberation by the full committees.
Michael Markley is the designated Federal official for the
initial portion of this meeting.
The rules for participation in today's meeting have been
announced as part of the notice of this meeting previously published in
the Federal Register on December 21, 1999.
A transcript of the meeting is being kept, so it's requested
that speakers identify themselves, speak clearly and plainly and into
the microphone, so that the transcripter can get you on tape.
This promises to be a very exciting meeting to me. We have
some very distinguished people here.
We have the staff, who's willing to come and share some of
their views with us, and we have three invited experts with us this
morning, all of them former office directors of the Nuclear Regulatory
Commission and now highly-regarded consultants.
Our three invited experts are Bob Bernero, Bob Budnitz, and
Tom Murley.
I have some introductory comments that talk about these
people. I guess I'll just read them.
Mr. Bernero spent 13 years in naval and space nuclear work
at GE and then served for 23 years, from 1972 to 1995, with the AEC and
NRC regulatory staff.
After five years in reactor and fuel cycle licensing, Bob
began work in regulatory development, including decommissioning
standards and spent fuel licensing.
After investigating the TMI accident, Bob formed the
Division of Risk Analysis in the Office of Research, served later in NRR
licensing divisions, and then went back to NMSS until he retired as
director in 1995.
Dr. Budnitz worked at the University of California Lawrence
Berkeley laboratory from '67 to '78 and held the position of associate
director and head of the Energy and Environmental Division.
In 1978, the joined the Nuclear Regulatory Commission as
Deputy Director of the Office of Research and was appointed Director of
that office in '79.
In 1980, Bob left the NRC to found the Future Resources
Associates, a consulting firm working mostly in risk analysis.
His current consulting activities include PRA, emphasizing
external hazards, upgrading the safety of older reactors, and using risk
in safety regulation, including performance analysis of waste disposal
systems.
Dr. Murley was the Director of NRC's Office of Nuclear
Reactor Regulation from 1987 to 1994. Prior to that, he was the
Regional Administrator of NRC's Region I office, beginning in 1983.
Dr. Murley retired from NRC in 1994 after 25 meritorious
years of service. He is presently a consultant on nuclear management
and safety matters in the U.S. and foreign countries.
In addition to all this brain power and good thoughts,
you're going to be treated early on this morning with some thoughts on
this subject from me and Dr. Garrick and from Dr. Apostolakis, and by
virtue of this awesome power I have as chairing this committee, I've
decided I'll go first and get things started and then turn it over to
John for his comments and then let George run the sprint lap and make up
for all the time we've overrun.
So, I do have view-graphs, so I'm going to do this and move
up to the front.
I am going to give you some thoughts I have on this subject,
to put it in somewhat of perspective. These thoughts are my own, by the
way, and may or may not represent any of the views of the ACRS or the
ACNW. For that matter, I don't even know what the ACRS views are on
this topic, or even if they have any.
So, they are my own.
That disclaimer said, I do have a couple of concerns that I
hope we can at least address in this meeting.
The first concern I have is there are a number of
definitions to defense-in-depth that vary slightly from one to the other
that I've seen.
Most of these definitions have a component of
defense-in-depth is there to compensate for uncertainties in our risk
numbers.
Well, I think we can all agree on that, but the problem I
have with that is I can't use that. That's not enough. That's not a
definition. It's a sort of a description, and I have no way to
implement that in that regulations or to use it when I design some sort
of system to deal with the risk.
So, that's the first problem. I don't know how to design to
that, and we need a better definition.
The second problem is what definitions I have seen don't
lend themselves in any way that I can tell, except in an arbitrary
sense, of determining necessary and sufficiency conditions on
defense-in-depth.
We've had a number of instances where there's been arbitrary
appeals to defense-in-depth to disallow some change or some regulation,
and if we're going to reap the benefits of risk-based or risk-informed
regulation, we have to have a way to put rational limits.
We have to know what defense-in-depth is, we have to be able
to identify it, and we have to be able to say how much of it is enough,
and I hope -- I don't think we'll resolve those two things at this
meeting, but I hope we at least make some headway in addressing it.
MR. APOSTOLAKIS: Tom?
MR. KRESS: Yes, sir.
MR. APOSTOLAKIS: I think language is extremely important
here. So, I would change a little bit something you said earlier.
You said "arbitrary appeals to defense-in-depth." The
appeals do not have to be arbitrary, because defense-in-depth itself is
arbitrary.
MR. KRESS: Yes. Good point, George, and I agree with that.
As a way to approach the subject matter, I hope today we can
-- if you notice, in my title, I had the word "design" defense-in-depth.
I hope we can focus on that, as opposed to operational.
I don't want us to get sidetracked into things like
inspection, procedures, quality assurance, management, and even
emergency response.
While those things are considered components of
defense-in-depth, I think if we're going to address a true definition of
defense-in-depth that has ways to put limits on designing facilities to
deal with risk, we ought to focus on design aspects, and in addition to
that, we have a tendency to lapse into barriers and nuclear reactor
defense-in-depth as it's traditionally been covered or been looked at,
and I think we need to generalize the concept, generalize it in the
sense that it applies to any hazardous activity, and in order to do
that, I've put together what I call four design defense-in-depth
principles that I think are general and would apply to just any
hazardous activity.
The first one is do what you can to prevent accidents from
starting in the first place. That's, I call, initiation or paying
attention to initiating events.
Second is do what you can to stop accidents at very early
stages before they progress to unacceptable consequences. I call that
one intervention.
The third is do what you can to provide for mitigating the
release of the hazard vector. The hazard vector in nuclear power
reactors are the fission products, but it could be toxic gases or fire
and smoke or heat or whatever the hazard is you're dealing with. I call
that one mitigation.
And fourth, provide sufficient instrumentation to diagnose
the type and progress of any accident. Call that, of course, diagnosis.
And I've categories these, the first two, as prevention and,
with some overlap, the second and third one as mitigation and the fourth
one as belonging in both categories.
So, I've categorized defense-in-depth principles in terms of
prevention and mitigation.
Now, with those as sort of principles of defense-in-depth, I
think one could arrive at a definition of defense-in-depth, and I think
we may hear several of those today.
I have one that I prefer, so I'm going to propose it right
now, based on these kind of principles.
A generalized risk-related definition of defense-in-depth
could be -- and I'll just read it -- design defense-in-depth as a
strategy of providing design features to achieve acceptable risk, in
view of the uncertainties, by the appropriate allocation of the risk
reduction to both prevention and mitigation.
I like this definition for a number of reasons.
One, it, I think, captures the essence of what we
traditionally think of as defense-in-depth, and number two, it is linked
explicitly to risk analysis and risk concepts, and number three, I think
it lends itself to being able to provide limits to defense-in-depth, and
you may ask how can I work from this definition to arrive at limits?
Well, the key words are "appropriate allocation."
In order to arrive at limits on defense-in-depth with a
definition like this, first off, you do have to have risk and acceptance
criteria for the activity you're dealing with.
These are things like, in nuclear reactors, early death,
latent fatalities, land interdiction, could be frequency of fission
product release or could even be LERF as a surrogate for all of those,
but you have to have an overall risk acceptance criteria, and not only
that, you have to express these risk acceptance criteria in terms of the
uncertainty.
If we're going to deal with uncertainty by defense-in-depth,
we have to have some quantification of what that uncertainty consists
of.
Now, you may hear that there are two kinds of uncertainties,
those that you can quantify and those that you can't.
I maintain that if we're actually going to put limits on
defense-in-depth, you cannot have un-quantified uncertainties; you have
to quantify the whole thing.
What we normally call quantifiable uncertainties can come
right out of the PRA.
What we normally call un-quantifiable uncertainties, I
think, would have to have some estimate of what those are, and we'll
probably have to get that from expert opinion, for this
activity-specific and maybe even facility-specific activity, and the
acceptance criteria that I'm talking about in terms of uncertainties
have to include both of these.
Now, once you have that risk acceptance criteria, the next
question is you have to allocate it among those four areas of prevention
and mitigation, because that's what defense-in-depth basically is. It's
an allocation of risk. And how do you do that allocation?
Well, there's no differential equation or no technical basis
for doing it. Allocation is a matter of policy, and we have to have a
policy statement of some kind that says how much we value prevention
over mitigation.
Now, that's policy, and I can't say how to do that, but we
could provide guidance.
For example, such an allocation or such a value judgement
could depend on the level of the inherent hazard. The more hazardous an
activity, the more we probably should value prevention.
It could depend on how big the uncertainties are. The more
uncertainty you have, you probably want to put equal balance on things.
It could depend on how much of this uncertainty is
un-quantifiable, as opposed to how much is quantifiable.
You may want to minimize the uncertainty. That would be a
classic optimization problem.
You might have noticed in my title I had "beating a dead
horse with a red herring." The dead horse is defense-in-depth as we
traditionally think of it. This minimization is what I threw in as a
red herring, just to confuse the issue.
It also -- some allocation rationally could be based on
what's called the loss function and decision theory. That's how one
normally allocates things. You ask yourself am I willing to suffer this
loss if I don't prevent? What are the consequences of that? And you
can work from that towards a probability that you want to accept for
that occurring.
With that as my introductory thoughts on the subject, I
guess I'll either ask if there are any questions or turn it over to John
Garrick for his thoughts.
I guess I confused everyone.
MR. BERNERO: Bob Bernero.
Are we going to reserve dialogue for the general discussion
period rather than take one paper at a time?
MR. KRESS: It probably would be a good idea to do it that
way. I think I prefer it that way.
MR. GARRICK: I think we're already in trouble
schedule-wise.
MR. BERNERO: So, I won't slap my forehead now.
MR. BUDNITZ: Bob Budnitz from Berkeley, California.
I have one very specific but, I think, important comment.
If you put a dangerous reactor 100 miles from the nearest
off-site person, then you have kept, as best I can tell from the
technology and what I understand it -- you've kept off-site fatalities
to zero, and that's a piece of defense-in-depth called siting and
mitigation, protective actions.
By the way, if you could do protective actions perfectly,
it's another piece, and you don't have that here. You only had the
piece about keeping the source term -- understanding it or keeping it
low.
MR. KRESS: Bob, I agree.
MR. BUDNITZ: I think that's a crucial leg of this.
MR. KRESS: Yes, I agree with you that that is crucial
defense-in-depth.
My reason for not discussing it, or even excluding it, was
there are lots of reactors out there that don't have that
characteristic, and we're talking about revising the regulations, and
we're talking about a lot of the NMSS activities in hospitals and
dispersed areas.
So, I was trying to say what would it be in terms of design?
I agree with you that that is a good defense-in-depth.
MR. BUDNITZ: But more to the point, if I have two identical
facilities that might be NMSS hospital licensees and one of them is in
the middle of nowhere and the other one's in the middle of New York
City, you might require different engineering at the facility, depending
on the site.
MR. KRESS: Probably not.
MR. BUDNITZ: You might.
MR. KRESS: Probably not.
MR. BUDNITZ: In principle, you could achieve the same
protection with different mixes of your allocation, but you don't even
know about that unless you put that allocation criterion on your slide,
which it wasn't.
So, I'm calling people's attention to the notion that you
have to consider that, I think, as a piece of this overall allocation
mix.
MR. KRESS: Yes. I don't know what all the criteria are for
allocation, I just know that we needed some, and those are good
comments.
John, you're up.
MR. GARRICK: I'm a little sorry I prepared anything,
because I would probably be more constructive if I took what Tom said
point by point and commented on it, but what I would like to do is come
before you not as a co-chairman of this meeting but as a plain vanilla
risk person and approach the problem from the point of view that, if I
had a license to do so, how would I address this question of
defense-in-depth, and again, as Tom said, I'm not speaking for ACNW or
ACRS, but I am trying to look at this as a issue that it's time that the
fuzziness of the issue was removed somewhat and that, in keeping with
the transition to a risk-informed way of thinking, it's time to think
about quantification of defense-in-depth as a way of taking the mystery
out.
So, I looked at this from the standpoint of what might be a
conceptual framework for quantifying defense-in-depth, and I recognize
the various interpretations of what constitutes defense-in-depth from
the three fundamental lines of defense that have been articulated in the
material that we have received -- the plant, the safety systems, and the
consequence-limited systems -- as being somewhat of a classical display
of the three most talked about lines of defense, but even that can be
challenged, because there's the whole soft infrastructure of quality
control, of review, of assessment, of audit that people would argue very
strongly are and should be a part of defense-in-depth.
But the position I'm going to take is what we need to do is
pick a piece of it and start looking at it in terms of how we might
quantify it.
So, the piece that I have picked is to look at a reactor
example, have a license to do that as a risk assessor, and a waste
example, and one of the things, I think, that would help this process a
lot would just be to organize the way in which we talk about it and the
way in which we present it, and one of my favorite presentation formats
is a matrix format, a two-dimensional array, and if we have more than
two variables, I have a tendency to fix those variables in some fashion
and reduce it to a manageable presentation.
So, what I have chosen to do, to illustrate, at least
conceptually, what I'm talking about, is to look at protective systems,
protection systems, again admitting that defense-in-depth is more than
protection systems, but to take a very top-down perspective of it, and
having just spent three days on a safety committee at a boiling water
reactor in a very upbeat situation where it's a plant that had its best
all-time performance year, broke all kinds of records in terms of
capacity factors and availability, had the longest run of any plant, any
boiler in history between outages, received an INPO-1 certification, and
it's kind of exciting, and when I'm at the PWR, maybe I'll do the PWR
example.
But what I'd like to do is to suggest that, if we laid out
the information about a reactor in some fashion similar to this, in a
top-down fashion, this is at the very functional level, and say that the
safety functions are basically those -- reactivity control, inventory
control, by which we mean coolant inventory control, heat removal -- as
we all know, the panic in Three Mile Island in the first two or three
days after the accident was a search for heat sinks, and radio-nuclide
containment, and then, in the vertical direction, we talk about classes
of initiating events, and I won't even claim that this is complete, but
the idea is to make it as complete as possible, and generally, we can
divide that into these three classes -- loss of coolant, transients, and
external events, and generally, we can create information that would
allow us to construct probability curves associated with those kinds of
events, and I think we could also argue that, in most large-scope PRAs,
we could aggregate the information in this form.
So, each of these kind of represent a group of scenarios,
and this is the end state core damage frequency for the group of
scenarios that are initiated by loss-of-coolant events, and then the
question -- and then, of course, if we do this carefully and we
probabilistically sum these end states, that constitutes our core damage
frequency, our total core damage frequency.
Now, the question is what do we put into these grid boxes,
and that's what I'd like to talk about a little bit, and I also would
like to reduce this from the very functional level down to a more
hardware level to give it more physical meaning.
Well, there's any one of a number of things and combinations
of things we could put in those grid boxes, but here's some suggestions.
Certainly, in each function, we could put the function
unavailability in terms of the frequency per demand for that class of
initiating events, and also, we could put something like this.
We could put what the core damage frequency would be at the
end state of that particular class of initiating events, given that that
function or system was unavailable. That's material that we can all
extract from a full-scope risk assessment, with some debate, of course,
but the most important entry might be this one. It's the total core
damage frequency with and without the safety function.
This particular core damage frequency is a result of the
convolution of all of the scenarios, and this is the same thing but
without the safety function, and at least if we did that for each of
these grid boxes, we would begin to see what the perspective was of the
contribution of the various safety functions.
Now, if we look at that at a slightly more detailed level
for something like a BWR -- and every time I look at this, I want to
re-tune the labels, and I'm not going to apologize for the small of the
print, you've got copies, but the safety functions can be reduced
basically into vessel-level make-ups systems and a reactor coolant
system, and the one thing you have to remember, that to a risk analyst,
we don't think in terms of safety-related and non-safety-related
systems.
Every system has to prove that it's non-safety-related.
So, I'm not adopting the classical NRC language here, but I
am adopting the classical risk language as to what these systems are
labeled and look like.
So, we have turned up the microscope on one grid box of that
functional diagram, and that's the grid box "inventory control," this
one.
So, the figure I just showed you is just a blow-up of this
one versus this class of initiating event, and we've decomposed that
into eight safety systems and six categories of initiating events.
These are still categories of initiating events, and so,
when we talk about these entries, we're talking again about the total
core damage frequency being the probabilistic sum of the end states of
all these different categories, and then the curve that we want to
compare that with -- this should be a double curve -- is the curve that
results -- that comes about as a result of making the system of interest
unavailable, recalculating this end state, and adding that recalculated
end state to the rest of these and comparing that with this, gives us an
in-context perspective of what that system is providing us with respect
to the bottom line, and that seems to me one of the things we want to
do.
Now, how do we do this with respect to nuclear waste?
Quite a different problem, because here passive systems
dominate the analysis, not only passive systems but geologic natural
setting are a major part of the analysis, and again, you can think of it
functionally, and I apologize to the performance assessment people for
choosing my own labels here, but I see the performance assessment
problem at the protective barrier functional level as basically these
three things -- water location and flow control, waste package
containment, and source term creation, mobilization, and transport --
and in a sense, you might look at this as the base case, and I have also
put down here geo-technical events to account for earthquakes, igneous
activity, and anything else of that type that you'd care to include, and
in principle, given the way we've set this up, the performance
parameter, in principle you could add these probabilistically.
Now, the way I've eliminated the time dependence of the dose
is to choose the time at which the annual release is the maximum into
the biosphere, and that allows me to keep it in a two-dimensional space,
and so, what this is is the peak annual release to the biosphere in
curies, and of course, this is just an expression of the uncertainty
about that, hopefully reflecting both information uncertainty and
modeling uncertainty.
Now, this time, however, what we want to do is, if we remove
this function, what does this curve become, and compare that risk curve,
which would be the one on the right, with what it is if you had the
function. In other words, this curve, the one on the left there, and
this one is the same, with all systems performing their intended
function. So, what that is is here.
This would be the measure of the performance with and
without the function or the system.
Now, how would we decompose that one, just to, again, reduce
it into more physical descriptive terms? This is how it might be
decomposed.
As far as water flow and spatial control systems, you could
imagine these kinds of systems, systems that would somehow impact the
way in which the water from the rainfall is drained from the site, and
I've distinguished between water diversion systems that are brought
about by doing some engineering of the geology versus bringing
engineering systems into the near field, and as far as waste packaging
-- and I'll let you argue as to whether things like drip shields would
be here or here. I would put them here.
Waste package containment -- I'm talking primarily about the
performance of the waste package, and usually there we think in terms of
the waste package corrosion resistance capability, fuel cladding, and
what have you.
Now, as far as the creation of a source term is concerned,
some of the things that are involved here are whether or not we have a
back-fill for purposes of enhancing geo-chemical conditions, also how
much credit we give to things like solubility, retardation, dilution,
and so forth.
So, again, it's a retaining of this structure such that you
have components where you can get some visibility into the contribution
to the overall performance of the taking away from, modifying, or
changing or adding any particular system/subsystem at any particular
location.
So, I wanted to do this because I think that the hope here
is that we take advantage of what we've learned in the risk field.
I think most of the kind of calculations that we're talking
about here have been done.
We can debate about the quality of them, we can debate about
whether they contain the right kind of uncertainties, but that's okay.
Once we get it in this kind of form, and given that those
kind of issues apply to all the boxes, there is great value in the
comparisons, it seems to me.
So, I wanted to just throw this out as an opening salvo, and
as I say, we're in trouble on time, and the chairman, and particularly
me, have contributed to that, and we'll take questions but probably
later.
George?
MR. KRESS: The reason we're in trouble on time is this is a
particularly long-winded group.
MR. APOSTOLAKIS: I was asked to do two things: one, to
present some thoughts that Dana Powers had, the chairman of the ACRS,
and he couldn't be here to present them, and since I happen to disagree
with him on a lot of things, the committee felt that I was the best guy
to present his ideas, and then, I will present some of my own thoughts.
So, we start with Dana.
He gives us first -- and you have the write-up in front of
you, plus the view-graphs -- a sort of historical background on
defense-in-depth.
This is a concept that has evolved over the years, from the
early days when people realized that there were -- there was a
possibility of catastrophic accidents from reactors, the uncertainties
were very large regarding the likelihood of occurrence, so people
devised this idea of multiple defenses.
It turns out, though, that this safety strategy that's
called defense-in-depth may impose unnecessary burden now on the
licensees.
Everybody says that it has served the reactor safety
community well. I have some doubts about it, but I will go along with
it.
Oh, I'm sorry, I'm presenting Dana's.
Even within the reactor safety community, thoughts have
turned to limiting defense-in-depth.
Now, you probably have seen that paper that several of us
wrote and presented at the PSA conference last August where we
identified two schools of thought.
One is the structural school of thought, defense-in-depth,
and Dana is the primary advocate of that, I believe, which says that,
essentially, defense-in-depth is an idea that is embedded in the
regulations, this idea of multiple defenses.
The rationalist school -- Tom and I happen to push that a
little bit -- advocates that defense-in-depth -- that now that we can
quantify uncertainties, we can use defense-in-depth in a more limited
way for those uncertainties that have not been quantified.
Dana offers a couple of thoughts here, says that the
structuralist approach may be difficult to extend in other areas -- he
has in mind NMSS activities, other than reactor, in other words --
whereas the rationalist approach could be extended to other areas, but
then, since you are relying so much on what can be quantified and what
cannot be quantified, you really have to have the analytical capability
which perhaps does not exist in other areas.
Now, a favorite question that Dana raises is what if you're
wrong? That's why I use defense-in-depth. What if my analysis is
wrong? Okay?
So, he says that it may be a little paradoxical to use
analysis to specify where defense-in-depth is applied when, in fact,
defense-in-depth is used to protect you against the possibility that
your analysis wrong.
So, that's an interesting thought there.
So, again, some of the historical reasons for the
development of defense-in-depth here -- again, always according to Dana
-- at that time there was little experience in the operation of nuclear
power plants, there were no industrial standards for the safe operation
of nuclear reactors, there was confidence that accidents were unlikely
but great uncertainties in the consequences given that they would occur,
that they occurred, potentially consequential accidents would be
difficult to interdict once underway, and finally, if an accident
happened at one facility, it would affect the operation of other
facilities, as well.
So, Dana's conclusions are that, for the four classes of
NMSS activities, which are disposal of high-level waste, engineered
casks for transport of nuclear materials, sealed and unsealed sources --
I don't remember the third one, some sort of waste -- Dana believes that
the consequences for these classes of material licensees can be easily
bounded.
In many cases, there is a wealth of operational experience.
I'm glad he said that, because I want to use it later.
The timing is different. Severe accidents are potentially
-- have large consequences develop slowly, so there is the possibility
to interdict, unlike with reactors. Phenomenological uncertainties are
modest, and the technical basis for rationally limiting defense-in-depth
is not well developed.
So, his main position is that he is against the imposition
of a defense-in-depth philosophy on material licensees, which I guess
includes high-level waste repositories.
Now I will present you my thoughts.
The fundamental question is why do we bother? Why are we
having this meeting? What is it that has changed over the years that
has made us have meetings like this, publish papers, and think about
defense-in-depth and its role in reactor regulation?
I believe most of us would agree that the thing that has
changed is that the uncertainties that forced the pioneers to come up
with defense-in-depth now -- a class of those uncertainties can be
quantified, whereas in those days they could not quantify them.
They knew that the frequencies of these accidents were very
uncertain, the consequences could be very high, but the uncertainty was
not quantifiable at the time.
In the last 25 years, starting with the pioneer in reactor
safety study, of course, we started quantifying a good part of these
uncertainties, and again, people with some experience in the field know
that there is also a class of uncertainties that perhaps we cannot
quantify at this time, un-quantified uncertainties.
The potential conflict, then, is between someone who takes
defense-in-depth as a principle and someone who tries to use the
rationalist approach and use defense-in-depth or its tools as standard
engineering tools used within engineering calculations that include risk
assessments and quantification of uncertainty.
So, what I propose is that we avoid the word "principle" and
simply say limit defense-in-depth and say defense-in-depth is a safety
philosophy that requires that a set of provisions be taken to manage
un-quantified -- not un-quantifiable -- un-quantified uncertainty
associated with the performance of engineered systems.
I believe this is consistent with Tom's presentation.
So, I'm carefully avoiding the word "principle." I'm using
the word "safety philosophy." In this, of course, "un-quantified
uncertainty" are the key words here.
Now, some observations.
Many times, people use the words "defense-in-depth" to mean
multiple barriers.
Now, by "barriers," by the way -- the word "barrier" is very
general here. It includes siting, it includes everything, not just
physical barriers like the primary system coolant boundary, and I want
to make that distinction. They are not identical concepts.
Even within the quantified uncertainties, where I'm going to
be using, you know, risk to decide how much I need, I will use multiple
barriers, otherwise I will never be able to go down to 10 to the minus 4
and 5 per year, but this is not using defense-in-depth; this is using
standard engineering tools.
So, let's start by saying that these two things are not the
same concepts.
Now, where does this un-quantified uncertainty come from?
It's primarily from models. We know that our models are inadequate in
many instances, or we know that some of the things that may be important
we cannot even quantify, we haven't tried. Okay?
So, experienced analysts and practitioners do have an idea
how good these analyses are.
Now, if we focus on these un-quantified uncertainties, then
we have to debate them, and then we will all understand better why these
uncertainties are not quantified.
We may be able to define new activities, research activities
or other kinds of activities, experiments, perhaps, to quantify part of
these uncertainties. So, it's not that I'm ignoring them. I think I'm
placing extra attention on these un-quantified uncertainties.
But the crucial question, as I said earlier, is under what
conditions, if any, is defense-in-depth in principle? I don't think
there are any conditions. It should never be called a principle.
It's a safety philosophy, as I gave in the definition, where
the uncertainty is un-quantified, and the words should not appear at all
within a PRA.
When the uncertainties are quantified, drop
defense-in-depth. You just use the tools to manage your risk and
achieve the uncertainty levels that Dr. Kress talked about.
Now, Dana read this and said, well, I am much more
comfortable with defense-in-depth as a means to address the question of
what if we are wrong in our analysis. This is his favorite question:
What if you're wrong?
You can argue that this is just a kind of uncertainty, as,
indeed, I am arguing, but I think that argument trivializes the problem
or implies that we know more than we do.
Well, instead of defending my position, I will attack his.
This is exactly what's wrong with calling it a principle.
You are telling me, no matter what you do, what if you are wrong? So, I
will impose on you defense-in-depth.
Well, I might as well give up. Why did we even try to
develop PRAs? We spent all these resources the last quarter-century.
Why? What if I'm wrong?
I will have to live with defense-in-depth forever, and
that's exactly what the word "principle" does to you. If you call it a
principle, you can't get out of it. It's impervious to analysis.
And in fact, I'm glad that he said, in his presentation --
it's really kind of unfair that he's not here, but on the other hand,
there is a certain pleasure in this.
Why is this a reason to argue against the imposition of
defense-in-depth on material licenses? Why? Because there are no
un-quantified uncertainties. That's why.
Thank you very much.
MR. MURLEY: My name is Tom Murley.
George, I very much, I guess, like your analysis her, but
are you suggesting that one should not make it a principle and,
therefore, if you are confident enough, you could use PRA to justify
removing a barrier like containment, let's say? Would you push it that
far?
MR. APOSTOLAKIS: First of all, I would not use just PRA; I
would use my total knowledge. Yes, I would. Yes. There is nothing
sacred about the containment. But you better come back with some real
good physics to convince me that the uncertainties are not large.
MR. KRESS: George, if you adopted my principle of
allocation, you might say that allocating risk reduction to CDF and to
containment is a matter of policy, and then you would set values for
that allocation, and you would have a containment, even though you could
throw it away and still achieve your risk acceptance, you would still
have containment, because it's a policy in allocation.
MR. APOSTOLAKIS: On the other hand, you might say that the
policy applies to a certain type of reactors -- LWRs, for example.
If somebody comes up with a new design that is fundamentally
different and can make a convincing case that I don't need the
containment, I don't see why I should.
What's next?
MR. KRESS: I guess now we turn to the rest of the agenda,
if I can find it.
That covers the preliminary presentations by the committee
members, and the second part of the agenda -- and we're only about 25
minutes behind, which is not too bad at all -- is presentations from our
invited experts, and we have first on the agenda Dr. Budnitz.
MR. BUDNITZ: Twenty years ago this week, I appeared before
the ACRS downtown, and Bob Bernero reminded me that I sat up on this
side of the table, with my jacket off, tie off, shoes off, and talked
this way, but I won't do that today, because Chet Siess isn't here. May
he continue to prosper. Those were the informal days.
I also want to point out that the reason I'm first is I'm
the youngest of the three, and another reason why I'm first is because I
was the Director of Research for a very brief micro-second 20 years ago,
and these two guys were two of my division directors.
I'm going to confine my remarks to Yucca Mountain and Part
63, but before I do that, I want to start with a bit of philosophy,
because I want to be sure you understand that I think the argument about
whether it's a principle or a criterion is moot, because it depends on
how it's used, and it's only how it's used that matters.
Let me try to make the point directly.
In Exodus, there are 10 commandments, and the two that, by
the way, are observed almost universally in all societies everywhere are
don't steal and don't murder. Don't steal and don't murder.
Are they requirements? Are they laws? Are they what?
I can tell you that, in the United States, in the year 2000,
we are still arguing about the definitions which goes to the
implementation. What really matters is the implementation of those
things.
For example, we're still arguing today about whether
abortion is murder in this country. So, it's not simple just to say
don't murder.
Second, can I steal from my community property from my wife
in California? It turns out that's ambiguous. There's no real answer
to that in California law.
So, things as simple as don't steal and don't murder, which
are principles which all societies follow -- never minding they're in
the Bible, all societies follow them -- can't be implemented without
implementing rules, and it's the rules that govern our behavior, our
enforcement, our regulations, and not what you call it, whether it's a
principle or a biblical commandment or what.
Same thing is true here, and when you come to see what I'm
going to say about Yucca Mountain, you'll see it directly.
George, I don't know what to call it, but one thing for sure
is that, whatever you call it, at Yucca Mountain or for reactors or for
material licensees, it's -- what matters is how the rules and
regulations of Part 50 or Part 60 or Part 63 or whatever, or any of the
regulations, and all the stuff that goes with them, how it's used in
practice, and that's the real point.
In a way, you can imagine that they're high-level criteria
or high-level requirements which, if you meet this stuff, you meet it,
but you can't meet it by itself. You don't know how to meet it by
itself. You've got to meet this stuff that's down below, and then, by
definition, you meet it.
But using it as a principle, then, or a philosophy or
whatever, is because it provides a intellectual framework or a way of
thinking about how this stuff works or how you got to it, and you can
argue about it, you can argue about the details in light of those
principles which you think about, but you have to keep that in mind.
You can't enforce defense-in-depth anymore than you can
enforce what the Atomic Energy Act in 1954 ordered the AEC or NRC not to
do, which is to ensure adequate protection, but you can't go to any
licensee anywhere and say, sorry, you don't meet adequate protection.
What you say is you don't meet part something-something of Part 50.
That's what you don't meet.
By the way, that got translated later into no undue risk,
and it took the Commission 30 years to decide what undue -- you know, as
Hal Lewis on this committee used to say, you really want them to tell us
how much risk is due. That's the safety goal.
The safety goal finally told us, for reactors, what undue
risk meant, even though undue risk had been used for 30 years before the
safety goal was adopted.
You couldn't regulate on undue risk. You can't regulate on
adequate protection. What you can regulate to is some rule somewhere or
what an inspector is told to look for or what can be enforced, and
that's what I'm going to talk about for Yucca Mountain.
So, now I'm going to talk about the dilemma, and this is
quoting straight from the supplementary information for Part 63 that
came out within the last year, where it says in plain English, or
reading the plain English -- and then we're going to come to, you know,
where the rubber hits the road -- the Commission does not intend to
specify the numerical goals for the performance of individual barriers.
By the way, this is a draft; it still hasn't been finalized.
But were this adopted, it tells us the Commission does not intend to
regulate specific numerical goals for barriers.
But -- and here's the big "but" -- in implementing this
approach -- the defense-in-depth was in the previous sentence, so that
insert is, in fact, completely -- I'm not fooling you -- the Commission
proposing to incorporate flexibility into its regulations by requiring
DOE to demonstrate the repository comprises multiple barriers but does
not prescribe which barriers are important or describe their capability.
Don't steal -- but without telling you what stealing means.
I'm just reading the page. Okay? You can't implement don't murder or
don't steal without the details. You can't, because there are
ambiguities about what it means.
MR. GARRICK: Disagree.
MR. BUDNITZ: Okay.
So, what it says here is kind of odd. Propose to
incorporate flexibility by requiring barriers, not going to prescribe.
Well, of course, they go further. So, it's not quite that bad.
This is just the next, you know, eight lines down.
The proposed requirements will provide for a system of
multiple barriers to ensure defense-in-depth and increase confidence.
Probably what you meant was so that you could increase
confidence, but I'm just reading it, what it says. I mean I'll give you
the benefit of the doubt on how you read it. Increase confidence so
that the objective will be achieved. Okay?
I just have to read it that way.
Now, here's the dilemma.
NRC, NMSS, Part 63, Yucca Mountain -- be sure you understand
the context.
Will NRC use this as a decision criteria? Which is really,
more directly, can DOE's license application flunk based on insufficient
defense-in-depth even if it would otherwise pass?
That's where the rubber hits the road, and then you've got
to get into some details about that, but that's the question, and it's
apparently yes. Of course, the rules aren't finalized yet, Part 63 is
still draft, and EPA has to come in and it has to get changed, but
apparently, yes. I've been reading testimony and talks and various
positions of the staff, and apparently, yes.
Now, if so, how? How will the decision be framed and made?
That's where we need to talk.
Observation -- and this is a crucial observation of mine:
The decision criteria, whatever they will be, need to be clear, they
need to be fair, and they need to be technically logical.
MR. KRESS: In other words, the Commission needs to revisit
this statement that they do not intend to specify numerical goals for
the performance of individual barriers.
MR. BUDNITZ: I'm going to argue that what's there is
ambiguous.
MR. KRESS: Yes.
MR. BUDNITZ: And what piece of it they revisit, I'm not
sure, but what's there is ambiguous, and I know that the staff agrees,
because I've heard the staff say this in public, that more is needed,
and there are even some tentative positions, and I'm thrilled that
that's true.
MR. KRESS: I think somebody needs to specify what those
goals for individual barriers are.
MR. BUDNITZ: Fair enough.
Now, I'm going to switch the order of my slides if you've
got them in front of you, because I'm going to make an observation.
I sent a letter to the docket on June 25, and I also sent a
letter to John Garrick, chairman of the ACNW, but this quote is from
both of them. This is from a letter that I wrote six months ago, seven
months ago. I'll read it to you, but you can read it, too.
When I apply these ideas to Yucca Mountain, I stumble
principally because the notion of so-called independent barriers, one of
which can fail without compromising the overall system, which notion has
been so useful conceptually for achieving and demonstrating power
reactor safety seems not to apply to Yucca Mountain, and everybody that
deals with Yucca Mountain understands this.
As I understand the design concept, one cannot assume total
failure of any of the so-called barriers without seriously compromising
overall performance, and that's not necessarily true, by the way, for a
power reactor.
I can show you power reactors operating in the world for
which, if you didn't have a containment, you could meet all the goals,
safety goals and everything.
MR. APOSTOLAKIS: I'm confused by that.
MR. GARRICK: Just one question. Where in a power reactor
does it say how much liquid control has to contribute to the risk?
MR. BUDNITZ: I understand that. I exactly understand, but
the idea here is -- without arguing about what works for reactors, the
idea here is that, for sure, you can't totally remove -- by the way, the
staff agrees with this -- you can't totally remove barrier number four
or barrier number one and still show it at Yucca Mountain, because it
doesn't work that way.
It's not the same as the fact that, at many power reactors,
you can totally remove the containment and you can still meet all
operating NRC goals, except the goal that says you've got to have a
containment, but you know, the overall safety goals and all that stuff
-- you can meet it.
MR. APOSTOLAKIS: I'm confused by that. This is a question
of clarification.
MR. BUDNITZ: Yes.
MR. APOSTOLAKIS: Are you talking about a particular
technology?
MR. BUDNITZ: I'm talking about a particular design.
MR. APOSTOLAKIS: PWRs as we know them today.
MR. BUDNITZ: Yes.
MR. APOSTOLAKIS: You are saying that, if I remove the
containment, I am not compromising overall performance?
MR. BUDNITZ: I am saying to you I believe that you can
still meet the overall safety goals for some designs.
Now, without arguing whether that's true or not -- I don't
want to argue that. What I'm saying is it is surely true at Yucca
Mountain that you can't remove -- totally remove -- and the staff is not
talking about that. That's what we're going to come to.
You certainly can't remove the canister. You can't remove
the ground. So, we have to talk about what I'm going to come to in the
next slide, under-performance, rather than removal, and that's where the
details come in.
Let's not argue about what I said here about reactors. I'm
talking about Yucca Mountain.
Now, I'm going to go back and say, in practice, perhaps --
and I don't know, I'm guessing. Perhaps in practice, despite NRC's
words to the contrary, DOE will never actually flunk at Yucca Mountain,
but defense-in-depth will be used, instead, like ALARA.
Do what you can beyond meeting the thing -- you met the dose
in Amargosa -- do what you can beyond meeting the bare regulations
whenever it's cost-effective or whatever you mean by effective -- again,
some other parameter that you have to pay for.
I don't know that, but that's one possibility as to how it
will actually be used.
But if that's true, how does NRC conceive this would work in
practice?
I mean there's the classic. Might NRC ask for protection
from one or another barrier in the name of defense in depth even if the
overall performance is okay? In other words, you met it, but you still
got to have a containment.
I'm not arguing this is bad. I just want clarity. You need
clarity, just as when you say don't murder, you need clarity whether
abortion is murder or not.
And then there's the classic: What if one barrier provides
90 percent of the total protection? Maybe that's not enough of a mix.
Go read the Congressional legislation, which says you've got to have
multiple barriers. But what if one of them produces 90 percent of the
protection?
Maybe DOE can say, great, we can weaken that barrier so it
only produces 40 percent and we still meet the rules. None of us want
that. That's nuts.
MR. GARRICK: Bob, you're missing, I think, an extremely
fundamental point that the pioneers had the foresight to put in the
fundamental Atomic Energy Act, and that is the word "reasonable."
MR. BUDNITZ: Oh, no, I understand, of course.
MR. GARRICK: And I just think this is nonsense, these
arguments, because they're not reasonable.
MR. BUDNITZ: Exactly. But that's why we need specific
criteria so that people won't use unreasonable arguments one way or the
other, without specific criteria.
MR. GARRICK: You don't have specific criteria in an of the
reactor -- Part 50 -- along the lines that you're talking about.
MR. BUDNITZ: Yes, we do. We tell you what the containment
must do. We prescribe its performance.
MR. GARRICK: You don't prescribe the performance of the
safety injection systems.
MR. BUDNITZ: No. We prescribe the performance of the
containment.
MR. GARRICK: I think you're splitting hairs.
MR. BUDNITZ: Let me go on. The staff has gone further than
this, thank God, because if you didn't go further than this, we really
would be in the soup, and that's what I'm trying to say.
You can't have don't steal up here. You've got to have some
detail that they have to meet or they don't meet or they can analyze
against, that you can regulate against, that you can decide, and the
designers can use, and so on.
If all you had was the dose in Amargosa Valley, you know,
dose rate per year in Amargosa Valley, and that stuff, the designers
know what to do. They know what to do. But if they've got to do this,
too, the NRC has the obligation to tell them what to do, tell them what
they're going to test against, what the criteria will be. That's what
I'm arguing.
So, we're talking here about under-performance. That's a
phrase I've seen recently. So, perhaps the staff isn't thinking about
-- don't assume total failure.
We all know that's nonsense. I don't know what total
failure means. What do you mean, total failure? We're not saying the
can isn't there. The can might not behave as well. We're not saying
the earth isn't there. We're saying maybe we didn't understand travel
times or maybe the chemistry is different than we thought. It's at the
extremes of some state of knowledge uncertainty distribution unluckily,
even though we think it's over here but we think it's possibly over
there, but maybe it really is over there.
So, maybe we're talking about under-performance rather than
-- you know, to assume under-performance of barrier number two or
whatever and go analyze it again. Fine.
What does this mean? And that's the point. What does this
mean? What analysis requirements leading to some sort of decision
criterion will satisfy my three figures of merit? It has to be clear
and it has to be fair and it has to be logical, and I haven't seen that
yet, and short of seeing that, to be argued about amongst the technical
community and understood, short of seeing that, you still haven't told
the Yucca Mountain project what they should do in their design and in
their analysis so that they know where they're going, short of that.
You need that. You need to have the details.
Now, finally -- this is really a place where I am truly
stuck -- if NRC lets DOE decide what under-performance means -- and
there has been talk about in some of what I've seen -- if DOE decides
that under-performance means this and says bring me the rock, wrong
rock, late in the game -- remember, they're designing it now, they're
finalizing their design now, and then they're going to analyze for a
couple of years, and that's a terrible dilemma. You just don't want
that.
DOE will not assume so much under-performance that it will
flunk if, of course, it passes under the base case, you see, because
anybody can dream up a set of under-performances that will flunk.
I can do that, but in fact, isn't that just what NRC's
concern really is?
NRC ought to be concerned, as the regulator, in its
statutory role, to be sure -- they've got to look for combinations of
under-performance that might lead to serious compromises, whatever that
means, find out whether there -- what the probability is and the
consequences of those or how much we don't know or what the
uncertainties are or where we have to go get more knowledge and make
sure that's straight. That's NRC's regulatory job, as I see it, under
the philosophy of an independent regulator, right?
So, you just shouldn't ask DOE unless you ask them to
explore the whole face base, and then I don't quite know what to do with
that, because then it's the bring-me-the-rock thing.
So, perhaps NRC has to tell them how much to assume, and
that leads to the other problem, which I know the staff is wrestling
with, because I've seen discussions and so on, mainly NRC is trying not
to be overly prescriptive -- thank God, by the way -- in using the
philosophy of performance-based analysis and decision-making and so on.
So, this is the dilemma for defense-in-depth. The Yucca
Mountain project and the Department of Energy deserve specificity as
they're finalizing the design and doing the analysis.
MR. APOSTOLAKIS: What's under-performance again? I missed
that.
MR. BUDNITZ: Under-performance is the assumption that
barrier number two or whatever, instead of totally fails, only fails in
a certain way. Just as we say, in the reactor game, analyze as if you
had a loss of off-site power, even if the probability is low.
MR. APOSTOLAKIS: But why do I have to tell DOE how much
under-performance to assume? Aren't they going to do it as part of the
PA?
MR. BUDNITZ: Well, the face base is so vast.
MR. APOSTOLAKIS: But they have to do this, assign
probabilities to these things.
MR. BUDNITZ: No.
As I understand it, they are supposed to produce a base-case
performance assessment, with its uncertainties explored, but they don't
necessarily have to show what the dose in Amargosa Valley is if barrier
number two under-performs by X percent or fails at 1,000 years instead
of 10,000 or has more juvenile failures than they think is right or
whatever.
MR. APOSTOLAKIS: But if they assign probabilities to these
various scenarios, 1,000 years versus 5,000 years, then the performance
assessment will reflect all these.
MR. BUDNITZ: Only if they're asked to reveal it and if
they're told that that will be the thing against which they'll regulate,
George.
Let me just describe a possibility.
Suppose I said to you that the department believes that
juvenile failures of the canister will compromise X percent -- it might
be X-tenths of a percent -- of all the cans. That's their best
estimate, and they have a uncertainty distribution about that state of
knowledge.
NRC might say I don't care what you do with that. Put that
in the performance assessment, but I want to see an analysis that's
100-X percent.
In other words, instead of .02 percent, maybe 2 percent, as
a means of assuring that, gee, you know, I really don't know whether I
trust -- that's Dana Powers' argument.
That's a valid way to regulate, is to tell the licensee to
assume something that is unrealistically conservative and still show
you're okay, and that's not in the performance assessment.
MR. APOSTOLAKIS: Let me take an example of a PRA. Maybe I
misunderstand what you are saying.
Somebody brings me a PRA and I review it. That licensee
wants to use it in their process.
MR. BUDNITZ: Right.
MR. APOSTOLAKIS: The licensee cannot come to me and say I'm
not going to worry about common-cause failures, because you didn't --
MR. BUDNITZ: No. Let me make a postulate here that the
licensee, the applicant says we think that there are going to be five
juvenile failures of our canister in the first 5,000 years, and our
state of knowledge is such that we're very confident it's no more than
20.
It's not inappropriate for the regulator to say analyze for
400 and show me what that does, and if you still perform -- that's not
inappropriate. If you still perform, great. On that aspect, we're
going to give you your license.
MR. APOSTOLAKIS: This is not a performance assessment
anymore.
MR. BUDNITZ: We're regulating, George. That's just the
point. We're regulating. We're trying to regulate.
MR. APOSTOLAKIS: I'm playing devil's advocate.
MR. BUDNITZ: Of course. I understand.
MR. APOSTOLAKIS: So, DOE, the applicant, would like the
benefits of both performance-based regulation and the --
MR. BUDNITZ: No, no, no. Quite the opposite.
I can't speak for them, but they're probably thrilled with
just the single figure of the dose in Amargosa Valley, but if NRC is
going to say we're going to impose defense-in-depth by telling us that
we have to under-perform barrier number two as a means of exploring how
defense-in-depth actually works, somebody needs to write down what
under-performance means in detail so we'll know what to analyze, and the
under-performance is presumably outside of the realm --
MR. APOSTOLAKIS: You're really coming back to Tom Kress'
point that you have to have some sort of allocation.
MR. BUDNITZ: I'm not arguing that under-performance is the
way to go, but if they're going to do it that way, they need to
prescribe it, and it may be outside of the realm that DOE believes is
the real world, just as we said 2,300 degrees for the peak clad
temperature -- nobody thinks that's the right number, but if you meet
it, you get your license, and I'm worried that, absent -- and this is
early, soon, not five or 10 years from now -- I'm worried that, absent
specific criteria against which the department, Yucca Mountain, the
applicant can analyze and know that he passed or he didn't pass and can
change the design now, before it's too late, in order to, you know,
improve and meet, that it's an open-ended, unsatisfactory regulatory
arena.
MR. GARRICK: Bob, you seem to be strongly advocating an
allocation process.
MR. BUDNITZ: No.
MR. GARRICK: Well, you seem to be.
MR. BUDNITZ: No, no, no. I don't think defense-in-depth is
necessarily the principle that others do, but if they want to use it,
they've got to tell them how.
MR. GARRICK: The NRC has been very clear in telling them
that they want to know the role of the specific protection barriers, and
my whole point was that the only place that makes any sense is in
relationship to the bottom line.
MR. BUDNITZ: I quite agree.
MR. GARRICK: I think one of the things that's a problem
here is that -- the great thing about the PRA business is that we
established a measuring process through the PRA, and we got some
experience on it before we started fussing around too much and trying to
calibrate that measure, and I kind of see that here.
There are some fundamental principles that have been laid
down, and one of those principles is that all of the protection should
not come from just the engineered systems or just the natural setting.
MR. BUDNITZ: Sure.
MR. GARRICK: Now, it sounds like what you're saying is
that, if they say that, they need to say more about how much of it
should come from where.
MR. BUDNITZ: No, not necessarily how much of it should come
from where. I don't like that either.
They need to establish specific performance criteria or
analyses or outcomes or something like that that the department can
analyze to now, while they're still changing the design. Otherwise they
get the bring-me-the-rock problem.
They need to say under-performance of the canister means X
for juvenile failures, means Y for corrosion, means Z for when it will
happen, 1,000 years or 6,000 years. They need to tell them what
under-performance specifically means for those things, assuming, I
assume, that the under-performance they're going to tell them about is
outside of where the department believes is the true knowledge of the
performance.
Now, you know, I don't care whether you say it's this.
Analyze that anyway. That's not an illegitimate thing for regulators to
do, and they do that all the time.
MR. APOSTOLAKIS: Why would the NRC ask them to do that?
MR. BUDNITZ: Why don't you ask the NRC? But they're
talking about asking them to analyze under-performance of various of the
barriers, either one at a time or maybe in combinations, but absent
specific things, the applicant doesn't know what to do.
MR. APOSTOLAKIS: Are they doing sensitivity studies, then?
MR. BUDNITZ: Why don't you ask them?
MR. APOSTOLAKIS: It looks like you're saying the department
will come in here with a base case and what they think is likely and
this and that.
MR. BUDNITZ: Yes, sir, of course.
MR. APOSTOLAKIS: And then the NRC staff comes back and says
now do this, I would like you to do this, which is a sensitivity study.
MR. BUDNITZ: That's what I said. These are sensitivity
studies. They're always a good idea.
MR. APOSTOLAKIS: And is it because we feel that the
uncertainties -- that right now we cannot quantify them?
MR. BUDNITZ: Well, why don't you ask them? But here's what
I think, and I'm reading minds.
Apparently, somebody somewhere in this Commission and its
staff thinks that defense-in-depth needs to be invoked separate from the
TSPA, the performance assessment as a whole, taken with its state of
knowledge, and I'm not going to argue whether that's a good or a bad
philosophy, but if they want to do that, they need to tell the
department specifically, with specificity, what the things are to which
they're going to regulate, so they can change the design and show it's
okay now.
MR. APOSTOLAKIS: I'm not familiar with that particular
staff position, but if, indeed, they want to apply defense-in-depth
independently of the PA, then that's exactly what I'm against, and I
hope I learn more about it.
MR. KRESS: In fact, it sounds like a de facto way of
allocating, actually.
Bob, did you have a question?
MR. BERNERO: Bob Bernero.
I'd just like to add -- I was going to address it in my talk
-- there is a statutory difference here.
MR. BUDNITZ: Yes, there is.
MR. BERNERO: The 11th commandment, not out of the Book of
Exodus but out of the Nuclear Waste Policy Act, simply says the
repository must have multiple barriers. So, there is a regulatory need
to address how does one implement that commandment, and that's part of
this.
MR. BUDNITZ: Absolutely, but of course there's an easy way
to meet that.
The fact that there is engineered barrier design and the
earth is, by definition, a multiple barrier. If you really wanted to be
sloppy, you could say of course we've got that.
But if you want to go further -- and I agree with you, Bob
-- if the Congress wants to go further, got to go further, they've got
to go further specifically. They just can't let the applicant figure it
out.
MR. APOSTOLAKIS: The words "multiple barrier" are so fuzzy.
Anything is a multiple barrier.
MR. BUDNITZ: George, the statute has that language, though.
MR. APOSTOLAKIS: Well, then it must be right.
MR. KRESS: I think we have time for more discussion later.
MR. BUDNITZ: Without specificity, it's like don't murder.
Without specificity, you don't know how to regulate.
MR. APOSTOLAKIS: I find that very interesting, Bob, because
in reactors we see the same thing. People want the performance-based
regulation, and you give it to them, they come back and say, what, you
didn't tell me what you want me to do.
MR. KRESS: Okay.
[Recess.]
MR. KRESS: Will the meeting please come to order?
Thank you.
Now we're at the point on the agenda where we're going to
hear from Tom Murley.
You're up, Tom.
MR. MURLEY: Thank you, Tom and John. Thank you for the
invitation, also. I don't have view-graphs or slides, so I'll just sit
here and say my piece.
I should say at the outset that I am not sure just how much
I can help on your discussion on Yucca Mountain.
I've not kept current with all the latest policy statements
and SECY papers and ACRS letters and things, although I should say Jack
Sorenson did an excellent job, I think, in research this topic and
sending the material out, but I have given a good deal of thought over
the years to nuclear safety and defense-in-depth, and so, perhaps I can
discuss some philosophical issues, and if it helps you, fine.
The first point I guess I would like to make is that, in my
experience, defense-in-depth is not a regulatory requirement. It's not
a principle. It never was.
I would characterize defense-in-depth as an after-the-fact
explanation to Congress and to the public of how NRC achieves safety for
reactors.
That is, after regulations were developed and after the
staff implemented them through branch technical positions and reg guides
and things, there was an explanation of what it all meant, and one way
to do that -- and I think a very useful concept -- was the
defense-in-depth concept, and as I read Cliff Beck's 1967 explanation to
Congress, that's probably one of the early things I read when I joined
the AEC in 1968, but it was never used as something that the staff used
as a requirement, a hard-and-fast requirement, and I think -- I'll give
an example.
This was illustrated by the Three Mile Island 2 accident.
I recall a meeting some months after the accident where an
aerospace safety expert was giving his views of the accident.
He may have been from NASA, and I think he might have even
have been assisting the Kemeny Commission, and he observed that NRC
talks about defense-in-depth but they don't really enforce it, and he
said, for example, the plant was designed -- this particular plant,
Three Mile Island, was designed for the pressurizer relief valve to open
during a feedwater transient so that the high-quality primary system was
deliberately breached during a design basis transient, and of course, we
know that the relief valve stuck open in that case.
He continued by noting that the operators defeated the
safety systems by shutting off the ECCS, the high-pressure injection,
and his point was that one of the major fundamental barriers of
defense-in-depth was deliberately defeated by the operator action.
We now know, of course, that there were confusing indicators
and circumstances that led the operators to take those actions, and
finally, this observer noted that the containment was open during the
early part of the accident and that that fact permitted radioactivity to
be released directly to the auxiliary building and to the atmosphere.
Eventually, of course, the sump pumps were secured and the
containment was isolated in that accident, but his point was this
philosophy of defense-in-depth was something that the agency, back then,
at least, talked about but didn't really enforce, and it was not -- his
point was, of course, a negative point with regard to the NRC and the
staff, and this analysis -- I'm sitting there listening to it, and I
became very embarrassed as a NRC staff member, because he was right, and
it had a profound impact on my thinking about safety at the time, and
that was, if NRC has a regulatory requirement and one relies on that
requirement in this defense-in-depth argument, then you really have to
enforce it.
So, you've got to make sure that the containment is reliable
and so forth.
In other words, the barriers of each level of
defense-in-depth should be highly reliable. That's the message I took
from that discussion, and it did follow me, and I did use it and think
about it during my career, at least, in that term.
I sent the committees -- actually, to John Larkins -- an old
document dated April of 1989 on Shoreham emergency preparedness that I
had in my files, and insofar as that was what we relied on -- that's
what I relied on when I licensed Shoreham in 1989, and it is, thus,
official Commission policy as of 1989.
So, it is a discussion of how emergency preparedness fits
into the defense-in-depth safety philosophy, and so, there's an
introduction in the first page of where emergency preparedness fits in,
and we termed it, then, as effectively a fourth level of safety. I
think that's the phrase we used.
Now, the significance of that paper for this discussion, I
think, is that the topic of defense-in-depth was used only as a
philosophical introduction. It doesn't say that it's a requirement.
I then stopped the discussion of where it fits in and went
through a point-by-point discussion of how Shoreham met the actual
regulations, and so, there was never a use of defense-in-depth as a
requirement per se.
As I said, it's kind of an after-the-fact explanation of how
NRC achieves safety, and my explanation -- I should say the agency's
explanation then, at that time, was that emergency preparedness was, in
effect, a fourth level of safety, but it was not meant to be that it was
an absolute barrier, or there were no numerical guidelines or
requirements for each of those levels.
There were other instances where I recall falling back on
the defense-in-depth philosophy in my own thinking about specific safety
issues, and I'll give a couple of examples.
The staff -- and I'll speak for myself, because I can't
speak for the staff today, but I was always sensitive to conditions or
accident sequences that could breach multiple levels of defense-in-depth
through a common cause, and we always paid a lot of attention to those.
That's why steam generator tube integrity was always such an
important issue for the staff. We gave it high attention, because
multiple steam generator tube ruptures could lead to bypassing
containment either before or after core damage, and that -- one may
wonder why, I guess, steam generator tube -- maybe it's obvious, but it
was for that reason, at least in my own thinking, that this was a path
that could breach multiple barriers of defense-in-depth.
And then in the late 1980s, I recall thinking about safety
culture and what does it mean, where does it fit into the overall
picture of safety, and it slowly became clear that and I concluded that
it was extremely important, safety culture was extremely important,
because -- it was Chernobyl, actually, that showed that a poor safety
culture at a plant could lead to actions that could cut through all
levels of defense-in-depth.
In other words, it could be a common cause for breaching
multiple safety barriers. If you've got a poor culture, you can do
stupid things that initiate the accident. You can do a test that's not
properly planned. You can put the reactor in conditions it was never
designed for. You can shut off safety systems.
In other words, it is a means for slicing through the
defense-in-depth barriers, and it was that thinking that personally I
went through that caused me to conclude that safety culture was an
extremely important safety concept. To me, it's not an abstract concept
or idea, but it's an essential aspect of nuclear safety.
So, I hope I'm giving some examples of how one regulator, at
least, on the staff used and thought about defense-in-depth.
There are some questions that were posed in the material
that was handed out to us, and I know Bob Budnitz and Bob Bernero have
talked about some of them, and I'll aim at a couple that I think I can
contribute to.
One is, is there an over-arching philosophy of
defense-in-depth, or a discussion of it, and I have not spent a lot of
time on the definitions.
I know there are lots of them, but the philosophy, to my
mind, is fairly simple, and that is, there should be multiple barriers
for protecting public from radiation, such that single mistakes and
single failures, even of programs -- like emergency preparedness is
really a program, you can think of it, but in that sense, as George
said, it's a barrier.
It doesn't have to be a physical barrier, and insofar as
possible, these barriers should be independent, and I don't think that
should be an absolute requirement, but one should try to make them as
independent as possible. So, multiple independent barriers for
protecting the public from radiation.
It should be made a regulatory requirement, in my judgement,
but it should remain a guiding principle, because it is a good way to
think about safety, as I think I've tried to illustrate.
A second question, how is it used in materials -- and I'll
let Bob Bernero, who's thought about this a lot more than I have and
also speaks about it better -- give some examples, but there's one that
I've come across recently that seems to me a perfect example of how
defense-in-depth thinking is used, and that is in criticality safety.
There is this concept of single contingencies, multiple --
double contingencies, triple contingencies as protection against
criticality, and that, to my mind, is a perfect illustration of how one
thinks about multiple barriers of defense-in-depth.
Apparently there is -- well, I know there is a lot of
discussion of how should PRA be used in risk-informed regulation
consistent with defense-in-depth, what does that mean, and I guess I
don't have the answer to that, but I can tell you how I interpret it,
and that is it means don't use risk arguments solely to weaken or remove
levels of defense-in-depth.
I think that's how I would use it if I had to use that
language, and even though one has to, I guess, hold open the theoretical
possibility, George, that you could use risk arguments or numerical
arguments to remove containment, that comes very close -- well, it's a
regulatory requirement, so you probably can't do it, but it comes very
close, I think, to using defense-in-depth as close to a requirement.
MR. APOSTOLAKIS: I'm coming back to Bob's question of what
is murder? What is a risk argument? A risk argument, in my view,
includes all the engineering analysis and physics that is appropriate to
do.
So, in my mind, one could use risk arguments to reduce
defense-in-depth, as long as the uncertainties are handled properly and
convincingly.
So, a risk argument -- I mean PRA, in my mind, includes the
underlying physics, chemistry, and engineering that sometimes we call
traditional analysis.
So, I assume that's what you mean by risk argument?
MR. MURLEY: Yes. And I did not say and I certainly didn't
mean to imply that you cannot use risk arguments or engineering
analysis, the whole panoply of arguments, to reduce margins where
they're excessive and that sort of thing, but I think you would run
across some severe resistance if you pushed the argument to remove an
entire barrier of what people view as defense-in-depth.
For example, people have used the argument, risk arguments
-- and I've heard them -- to remove emergency planning, period, for
advanced reactors. I think that's going to run into some serious
programmatic, you know, policy problems.
I think it can be used to quantify the protection offered by
these levels, and I think John Garrick's paper -- I did skim it, and I
did listen to him carefully. I think it's a very good analysis, an
appropriate use of how to analyze and understand barriers.
If it's pushed to the level of using numerical goals for
those barriers, then I think that's maybe pushing things a little
further than people are ready for today, although in principle, one has
to hold open the possibility that it can be done.
There is the notion of safety goals. Are they clear for
regulatory use in the materials area or even the reactor area, for that
matter, and I must say the safety goals -- I found them to be not much
use at all.
The public health goals -- I'm sure you realize, of course,
there's a big gap -- there's an order of -- two orders of magnitude
difference between the public health goals and the plant performance
goals in terms of the protection that they offer to the public, and this
has always been a stumbling block for use by the staff.
The staff was told by the Commission -- they worked with the
ACRS for years to try to rationalize a large early release goal with the
public health goals, and it couldn't be done, because there's this
two-order-of-magnitude difference.
One can have a TMI-2 meltdown accident every year and still
meet the public health goals. You can work it out.
So, they were not very useful at all, and certainly, when I
was with the staff, we didn't use them in our day-to-day activities,
with one exception.
We found them -- we did -- in reviewing and certifying the
evolutionary advanced reactors, we used a conditional containment
failure probability of .1 as a guideline, and we found that very useful
as a guideline, but even there, we had to back off using a numerical
goal, because -- in this case, it was General Electric complained -- and
I think they were right.
They complained that, in some cases, by forcing that goal,
you're actually increasing the core damage frequency.
So, we did is tried to formulate an equivalent deterministic
requirement that we felt was equivalent to the 10-percent conditional
containment failure probability, but overall, I have to say I don't
think that we found the safety goals very useful.
Finally, there is a nexus in all this discussion of
defense-in-depth to risk-informed regulation, and I'm a big fan of
risk-informed regulation.
I wrote a paper about it five years ago or so supporting it,
and I think I am very pleased with the way the agency is moving in this
direction, but there is a troubling aspect, and maybe I don't see it
correctly, but I would like to at least tell the committees what's
troubling me, and that is that there is a whiff in all of this
discussion, more than whiff, an aroma of relaxing regulations and
reducing burdens, almost as if this is a deregulation exercise, and you
know, there is room for that, I agree with that, but people forget the
other side of the coin, and that is there is this role of risk-informed
operation, too, where the operators of reactors, in particular, can use
risk to improve safety, and you can do them at the same time.
You can have reduced burden and improved safety at the same
time if it's done wisely, but I don't hear any discussion of that coming
out of this committee or coming out of the staff these days, or the
Commission, and I think somebody needs to pay attention to this, because
if risk-informed regulation comes to be seen as just a code word for
deregulation, I think the whole thing is doomed, because I don't think
you will have public support in the long run for that.
Some conclusions, then.
I agree with, I guess, John Garrick's characterization that
there is fuzziness in this defense-in-depth concept and that it can
stand some clarification and even some numerical clarification, and I
commend the committees for shining some light on this subject.
I am very uneasy with any notion of pushing defense-in-depth
to the level of a principle or a requirement, and I am also uneasy if
there is a trend to allocate numerical goals to the levels of
defense-in-depth.
I think you'll run into trouble just like the safety goals
kind of ran into trouble, and ultimately, it would not be much use.
That concludes my remarks.
Tom?
MR. KRESS: Thank you.
That brings us to Bob Bernero.
MR. BERNERO: I, too, would like to thank you for the
opportunity to speak to the joint subcommittee, and as I will explain in
my remarks, I'm going to try to focus more on the material licensing and
the high-level waste arena, or waste management arena, than on the
reactor arena.
I would, however, like to start out with just an exposition
-- I used to tell people when I was here that the greatest conflict of
interest you'll face in your life is defending what you said yesterday,
and I feel a little bit of that now, because I'm going to go back to
statements I made in the past decades, when I was working in the NRC and
had the good fortune to be involved in safety goals and things like
that, regulatory philosophy.
A safety goal has practical use as a description of the
levels of safety or reliability that is sought by a regulatory system,
and similarly, a probabilistic risk assessment or any kind of risk
assessment has value as a description a display of your best knowledge
about the level of safety or reliability you are achieving but to
regulate to a safety goal, to define quantitative standards in a safety
goal as the formula for a safety decision on the acceptability of a
reactor or its features is not a wise move, and for years and years, as
safety goals were developed, there was a very strong philosophy that,
beware, don't regulate to safety goals, use safety goals in formulating
regulatory systems or approaches but don't regulate to the safety goal,
and of course, I will acknowledge that the high-level waste program,
from the very beginning, has as one, not the entire, but one basis of
acceptable judgement a safety goal.
That's what the performance assessment is calculating.
So, a word of caution on that, but talking here today about
defense-in-depth, as I will say shortly, defense-in-depth as an
approach, as a strategy for safety analysis, a strategy for design and
safety analysis, is a very good description of your caution in avoiding
undue reliance on any single feature, barrier, or thing or aspect, and
when you do that, your safety analysis should beware of a prescriptive
approach and the safety evaluation, with quantification where you can do
it, without quantification when necessary, or with very, very vague or
poor quantification, it still has to rely on reasoned judgement with the
best display of information before you and then make a decision.
Jack Sorenson gave us some questions. In the slides you
have, I slightly changed the questions, and I geared them so that I
could go through responses to the general questions and the specific
questions in the three specific areas of regulation, and that, of
course, would let me emphasize the ones I'm more familiar with.
I, too, would like to endorse the book -- I have it over
there -- that Jack compiled, the research on defense-in-depth. It's an
excellent compilation.
When I made the view-graphs, I consciously selected one of
the papers to quote from, and now I have forgotten which one, and I
don't think it's worth the research to go back, but the point is it's a
good description.
It's a good exposition not of a formula for adequate
protection but as a safety philosophy, and many of those definitions fit
this.
Cliff Beck's 1967 one -- I was very familiar with that,
because I came to the NRC in reactor licensing, and that was treated
sort of like a gospel, but I think it was Tom or somebody said it was
more a public exposition of what we're about rather than a formula for a
licensee to build a reactor to.
Now, if I go to the very first question, is there an
over-arching philosophy, my answer is yes, there is an over-arching
philosophy as a strategy of safety analysis but not as a formula, and
the key thing here is the undue reliance on any single factor, a rarity
of occurrence, a design feature, a barrier, a performance model.
An example comes to mind.
Many years ago -- in fact, right now, it's more than 25
years -- I had the fortunate experience to be the licensing project
manager for TMI-1, and a principle safety issue and contention in the
hearing was adequate protection against the crash of a large aircraft,
because that plant sits not far from the end of the runway of the
Harrisburg International Airport.
There was a great deal of analysis to make sure that the
standard review plan, which was just developing at that time and used a
screening probability for screening out aircraft, that there was not
undue reliance on low probability of crash, and it ended up with a very
detailed analysis that included what would happen if an aircraft less
than 200,000 pounds hit, what would happen if the aircraft greater than
200,000 pounds hit, and one of the good aspects of it all was the
licensee, or applicant in this case, recognized all along that the
responsibility for developing a persuasive case to show no undue
reliance on that factor -- that licensee had that responsibility and
fulfilled it, and the staff didn't prescribe what was the due reliance.
The applicant demonstrated that there was not undue
reliance.
Barriers are an issue peculiar to material licensing in many
ways.
Basically, as I've said, it's not a formula for defining
acceptability, and I would caution that simply because one has
defense-in-depth, that doesn't mean that there is acceptable safety.
You can have very frail defenses, and on those grounds, I
would suggest, when you move to the additional thought of risk-informed
regulation, that's going beyond defense-in-depth.
It is looking at barriers or dependencies or uncertainties
and seeking to achieve a sufficient margin of safety, not too much and
not too little, and it goes to the degree of knowledge that you can
have, or the degree of experience, in many cases, with material
regulation
MR. APOSTOLAKIS: Before you go on, Bob, I think one of the
issues before this subcommittee, I think, or maybe this meeting, is to
try to understand words like "undue reliance."
I'm trying to put it in the context of uncertainties.
Perhaps it would mean the same thing. When you say "undue reliance," I
would say I'm too uncertain about the effectiveness of these barriers
for some reason. Maybe I don't understand all the conditions under
which the barrier is supposed to function. I don't trust, perhaps, the
calculations that the event is really very rare and so on.
Would that be consistent with your thinking? Why is there
undue reliance?
MR. BERNERO: Undue reliance -- as an example, in the TMI-2
case -- or TMI-1, actually. TMI-2 adopted the analysis verbatim.
In the TMI-1 licensing case, based on the traffic that the
Harrisburg International Airport supported and was reasonably expected
to support, a screening criterion like 10 to the minus 6, 10 to the
minus 7 per year likelihood of impact, using a conservative footprint
for the reactor plant -- that screening criterion was relied upon only
with respect to jumbo jets.
Basically, it was concluded that it is a relative rarity for
a jumbo jet, something substantially in excess of 200,000 pounds loaded
weight, to be in this airport or to be using this airport.
That left the screening criterion having (a) some good
traffic analysis as a basis and (b) the margin of safety implicit in the
robustness of the plant given that it was designed for aircraft up to
200,000 pounds, and it had things like a condensate storage tank on each
side of the reactor, so that your decay heat removal wasn't compromised
by the aircraft crash immediately.
You know, condensate storage tanks are out in the open. You
know, they're unshielded.
So, you had two things. You had an extraordinary
robustness, and frankly, the applicant said I'll change sites if I have
to get a degree of crash resistance beyond the inherent robustness of a
dry containment.
You know, a large dry containment is a very robust
structure, and they said that's what we'll do. We're willing to expand
this facility to that degree of robustness.
So the uncertainty of a screening criterion of probability
had two factors to make an evaluation: Is this undue reliance or not?
But there's no formula for that evaluation.
Now, our current safety goals and objectives -- I said a few
words about safety goals to begin with, but of course, it goes without
saying -- you're all aware that the current safety goals and objectives
are very explicitly reactor-oriented, and there's years and years of
that dialogue, and if you go into the material regulation or especially
into waste regulation, the only thing you find is in high-level waste
disposal the criteria that originally derived from the EPA standard, 40
CFR 191, which is a performance assessment with a quantitative release
limit probabilistically set.
So, I say they're not clear, because first of all, the scope
is not clear.
There's a span of protection or a scope of protection
implicit in NRC regulation that includes public safety.
In reactor regulation, you're almost always talking about
off-site public safety and not talking much about the worker safety.
That's within the NRC jurisdiction but not quite so
robustly.
You know, look at the steam-line erosion/corrosion, that old
Surry incident, 1970-something, where a relief valve -- tail pipe came
out of the hole in the deck and scalded two workers to death.
Things like that -- NRC's jurisdiction for industrial safety
is not clear, and when you go into material regulation, you'll find that
ALARA for chronic exposure is an important aspect, but accidental safety
is dominated by chemical safety.
So, you have -- issues that are far more complex don't lend
themselves to formulation.
Go into medicine and there is serious challenge or question
about NRC's jurisdiction for patient safety -- you know, that is, the
person receiving nuclear medicine treatment, and of course,
environmental protection -- we have a congruence of NRC's
responsibilities and authority with EPA.
The practices at NRC, you're quite aware, has a very large
range, and I would just single out transportation, which I listed at the
bottom, as a very interesting example of lack of defense-in-depth.
Transportation relies on one barrier, a great big heavy,
bullet-proof, super-strong cask to hold spent fuel, and especially in
transport, you have one barrier, and the real question is not do I have
multiple barriers, but the real question is am I placing undue reliance
on that one barrier, and of course, here, you have a wealth of
experience, engineering, metallurgy, testing capability, quality
assurance. You have a variety of tools. But the test is, is there
undue reliance on a single factor or a single barrier?
Reactors -- I would just point out that, in reactor
technology, defense-in-depth discussions are, in my experience,
invariably associated with accidental releases, not chronic releases,
and that comes to be an important consideration in material regulation
and waste management, and of course, waste management is a chronic
release.
The very nature of it is you take the waste and you put it
somewhere and say it will stay there until it's gone or forever.
In the reactor regulation area, seismic safety, here again
you have a probabilistic screen, and you have behind it -- some of you
certainly had an experience in the seismic margin analyses that were
popular a long time ago, and my favorite term, "HCLPF," the
high-confidence of the low probability of failure, which is a very good
concept, but it's interesting, if you ever go through the DOE
regulations and safety analyses for seismic safety, they actually try to
quantify, specific a specific requirement for seismic safety that you go
up to your design basis, probabilistically set, and then you go beyond
it by some formula and show that this level of acceleration excedence
doesn't do some quantitative damage, rather interesting experiment.
But these are all, in my view, things where you're looking
at do I have undue reliance on a single thing, whether that single thing
is reactor vessel rupture or, as happened in TMI-2, a cognitive error by
the operators that bypassed the whole event tree.
MR. GARRICK: One of the things that is kind of important in
that point about having undue reliance on a single thing is that there's
never a single thing even when it appears to be single.
By that, I mean, if you're talking about a reactor vessel,
for example, you have lots of things that give you indications of the
condition of that reactor vessel in terms of monitoring, etcetera.
So, it seems that, in those cases -- and the fuel cask
transportation is another example -- you may not have multiple barriers
in the classical sense, but in most of those cases, you have a great
deal more information about the -- its behavior.
If a cask -- we have seen it in tests at Sandia under the
most severe circumstances you can possibly imagine, and absolutely
everything was destroyed but the cask.
So, I think that, sometimes, that may be an
oversimplification, just because from a phenomena standpoint or from a
process standpoint, it may have that pinch point, and we have to offset
the vulnerability of that pinch point by additional levels of protection
that come in the form of information-gathering, diagnosis, monitors,
transducers, etcetera.
MR. APOSTOLAKIS: And all that means less uncertainty,
right?
MR. GARRICK: Yes.
MR. BERNERO: Yes.
One could reformulate the whole system to say, rather than
undue reliance on a single barrier, you could have inadequate response
to a single challenge.
You know, you could restructure the whole thing logically to
do that.
MR. APOSTOLAKIS: We're interrupting you too much, Bob, but
counting the number of barriers has the same problem that in some
earlier times people were ranking minimal cut-sets according to the
number of events.
Ultimately, it has to come to the probabilities.
MR. BERNERO: Yes. And in reactor safety, I don't believe
you get there -- you have a regulatory system that gives you multiple
barriers rather prescriptively -- that is, reactor coolant pressure
boundary requirements, containment requirements.
It just doesn't give you the performance, and to resurrect
an old argument, you know, the regulations prescribe containment
performance predominantly as condensers for LOCAs rather than
respondents to loss-of-coolant accidents and core melts.
But anyway, one point I'd like to make on reactors is, when
you have a defense against some challenge, you need to have graded
goals.
You know, everything doesn't come out to the old PWR-1
release off-site, and I remember, years ago, in reactor licensing, we
used to have spent fuel handling accidents analyzed, and we consciously
used one-tenth of the Part 100 release guideline for analyzing a spent
fuel handling accident in the pool, which is almost a trivial analysis,
because you're under 20 feet of water and virtually nothing happens
off-site, and you have to look at that.
What are the consequences of the event?
When you get into material and waste, that becomes extremely
important.
In material regulation, the concept of accidental release is
certainly with you, but chronic release and even deliberate release has
to be considered.
Exempt products -- I list there -- if you're not familiar
with the terminology in material licensing, when you go home and look in
the ceiling of quite a few rooms in your house, you'll see a smoke
detector, and the agency had a major deliberating problem in regulation,
because a typical battery-powered smoke detector has one-half of a
micro-curie of a 500-year half-life alpha emitter, americium-241, stuck
in there to ionize the air so that the smoke can cause an electrical
phenomenon that will make the little buzzer go off or siren or whatever,
the horn, and in regulating such a thing, you have to recognize, you're
never going to get them back.
They're not going to end up in a low-level waste or
high-level waste repository.
They're going to be thrown in the garbage. They're going to
be picked open by people. And so, you have to look at what I would call
chronic release and uncontrolled, routine release for things like that.
In order to have graded goals, you have to think through
what are the potential consequences of the act which you would
authorize, or the procedure, the barriers, protective actions, if they
are possible, and evaluate, a balanced choice of defense.
You can't prescribe it. It's far too complex. But as you
know, a lot of experience -- and you can bound consequences practically.
There are knotty problems. That's really a jurisdictional
problem.
In 1975, when the agency became NRC, there was the Food and
Drug Act that transferred patient safety for nuclear medicine to the
Food and Drug Administration, and ever since then, the states have
authority over patient safety, which is clear, but the NRC does not, and
it's argumentative.
It's really aside from here, although we had a lethal
accident about 1991. In Indiana, Pennsylvania, a brachytherapy patient
was killed by radiation, and the NRC requirements which were imposed on
that brachytherapy treatment had a device which reeled out wire with, at
that time, a four-curie source on the end of it into the patient's body,
and that device said I am now safe because I reeled the wire up.
The NRC required on the wall an alarming radiation dosimeter
and a personnel requirement that you would use a hand-held radiation
dosimeter in supplement. That was the defense-in-depth.
The source broke off. The machine said I got the source
back in its shield.
The alarming dosimeter went off, or it had gone off, and
stayed on. It was judged to be a false alarm, and they didn't use the
hand-held, and the lady died a very horrible death.
In that practice, there is a serious question, what is due
reliance or undue reliance on any barrier? What is the defense-in-depth
appropriate to that?
MR. BERNERO: Now, in waste, it definitely applies to
release barriers. As I said earlier, interjecting, the Nuclear Waste
Policy Act requires multiple barriers. So somewhere in a licensing
finding, somewhere in the licensing exposition by DOE, they have to show
the statutory requirement is satisfied because we have multiple barriers
and this is our demonstration of the adequacy of those multiple
barriers, as well as our performance assessment.
I underline the word "one" because the fundamental basis of
acceptability is not simply the total system performance assessment.
That's only one basis. You don't license to the safety goal.
There are other considerations that must be taken into
account. Some of these uncertainties are readily quantified, many are
not readily quantified. So you have to look at the whole body of
information in order to do it.
There is often confusion because defense-in-depth or
multiple barrier analysis is just another form of uncertainty analysis
and in this particular case, the staff, in Part 63 and in their
intentions for their review plan, have talked about guidance on how one
might do -- what's a sensitivity analysis, really, in supplement to the
appropriate uncertainty analysis in the total system performance
assessment, and I think that's good.
The one thing, and I talked to the ACNW in November, the one
thing that I think still needs attention is graded goals for graded
uncertainties. See, in high level waste, you deliberately put it out.
It's out there and now you're talking about what uncertainties do I have
about the barriers that inhibit the release and exposure of the public.
And one of the difficulties that exists is everyone that
talks about it seems to say the performance standard for exposure of
someone so far in the future, 10,000 years, 30,000 years in the future,
is such that it would not be greater than we would accept today, and
they come out and they use licensing acceptance criteria, which are
clearly acceptable. They're very low, they're very conservative.
There is no gradation of objectives to say, okay, well, how
far from the edge of the cliff am I, and I suggest that one can put
grades on radiation exposures from waste releases; that you can have the
clearly acceptable level of exposure, an acceptable level of exposure,
clearly tolerable levels of exposure, tolerable level on counting orders
of magnitude, life- threatening, and then clearly unacceptable.
And I have included a chart that I used before in November
and I just penned in. This is counting -- this is chronic doses and
then when you get to the top of the scale, you're really talking about
accident doses. For instance, when you get up to 10 rem, the accident
dose that's acceptable and has been for years, in things like reactor
accidents, 25 rem whole body exposure, is really a clinically detectable
threshold.
What you're really saying is if you limit the accident dose
to 25 rem, that is a sufficiently harmless level because there are no
clinically detectable effects in the human body from that kind of an
exposure. You have to go up a factor of three or something like that.
I usually use 10 rem as that.
But when you get up in this high level we were discussing
earlier, you get up in cancer therapy, and you get doses like that. My
wife has just had very substantial doses.
So the whole point I'm trying to make, the focus is down
here. When you do the uncertainty analysis, it is nice if you meet your
clearly acceptable goal with your base case, but if you are depending on
some shaky uncertainty analyses, you should be looking for the edge of
the cliff; not only in uncertainty variation, but in objective or goal
variation, because you've got these orders of magnitude of tolerance
behind it.
So that completes what I would like to say.
MR. KRESS: Thank you very much. Any questions, before we
move on the agenda? Very good. We are now at a point in the agenda
that calls for a general discussion of the people at the table and
anyone in the audience who wants to join in, and we need to define the
issues for further consideration.
I don't know exactly how to approach this, except ask for
any volunteers that want to make additional points or question the
speakers.
MR. APOSTOLAKIS: If I could make a suggestion. Why don't
we start out by defining perhaps three or four or five points that need
some discussion, because otherwise we will be going in ten different
directions.
MR. KRESS: That's a good suggestion, George. Do you want
to make a stab and give us a couple of points?
MR. APOSTOLAKIS: Well, this issue of uncertainty that I
raised, I think, deserves some discussion and whether we want to place
defense-in-depth in that context. That's certainly something that I'm
interested in.
MR. KRESS: That's a good one. What I'm interested in, of
course, is the issue of should there be a specified allocation.
MR. APOSTOLAKIS: That's a good point.
MR. KRESS: That would be one.
MR. APOSTOLAKIS: And I must say I am still not comfortable
with my understanding of the issue of how to use defense-in-depth in the
high level waste repository. So maybe a summary of the issue and then a
discussion, a summary perhaps by John, would help me understand.
MR. GARRICK: One of the points I'd like to see on here,
too, we keep hearing this observation that licensing decisions should
not be based on PRA/TSPAs alone. I'd like to see us discuss that more.
MR. APOSTOLAKIS: Okay. That's a good point.
MR. KRESS: Yes, that is, particularly when we're talking
about entering into a mis-conformed regulatory system. That's four
pretty good items. Are there others people would like to add to the
list? I think those are a pretty good set of things.
I would like to add one more, and that is we have heard some
contrary and different opinions on this. Should we have -- well, we've
been calling them safety goals, but I've been calling them risk
acceptance criteria that we regulate to.
Should we have risk acceptance criteria that we regulate to?
MR. GARRICK: And I don't think, by the list here, that we
would want to bound up anybody from jumping the fence here.
MR. KRESS: Absolutely.
MR. GARRICK: If they have a burning issue that they think
is critical to the subject.
MR. KRESS: Okay. That's, I think, five pretty good issues.
How should we approach the discussion of these? George, do you have an
idea on that? Would you like to, say, take one and I take another one
and John take another one and --
MR. APOSTOLAKIS: Sure.
MR. KRESS: -- just throw out some thoughts and see what
kind of response we get?
MR. APOSTOLAKIS: We could do that, yes.
MR. KRESS: Why don't you start with the issue of
uncertainty?
MR. APOSTOLAKIS: Okay. Well, I tried to make a case
earlier today that the reason why we are revisiting the issue of
defense-in-depth is that we can now quantify a good part of the
uncertainties associated with the performance of the systems that we're
talking about that we could not quantify 15, 20, 30 years ago.
That includes identification, quantification,
characterization, all the words.
I also made the point that the language is extremely
important here. I was glad to hear Tom Murley say that, in his mind,
defense-in-depth has always been a philosophy and not a principle,
although the word principle is being kicked around. But I think Bob
Budnitz's point is well taken, that it ultimately comes down to what you
do.
I mean, what you call it is nice to have good terminology,
but what you actually do at the lower level, at the working level, is
what counts, and that's what I want to address.
I really think that for the uncertainties we have
quantified, defense-in-depth, the words don't belong there. You're
going to use the tools of defense-in-depth, barriers, diversity and so
on to manage your uncertainty and you have an excellent means, a
numerical standard against which you can decide how much is enough,
which is really a fundamental question today, how much defense-in-depth
is enough.
MR. KRESS: But, George, we don't have numerical standards
on how much is enough, unless you allocate --
MR. APOSTOLAKIS: Yes.
MR. KRESS: Now, if you would throw in this word allocate, I
would agree with you. But then, by my definition, that becomes
defense-in-depth in a regulatory sense, if you allocate.
MR. APOSTOLAKIS: But I would avoid the words defense-in-
depth, because they carry a certain baggage. Now, I understand where
you're coming from and in an ideal world, but I want to reserve the
words defense-in-depth to mean what they have meant all along; handling
unquantified uncertainty by using barriers, emergency plans.
MR. KRESS: Let me give you my problem with that. I
mentioned I my talk that I don't think we can live with unquantified
uncertainties in a defense-in-depth regulatory system. The reason I
said that is I don't know what to do, I don't know how to put limits on
defense-in-depth, I don't know how many barriers I need, I don't know
how good they have to be, I don't know where to put them.
And then when I do this, I don't know how well I have
compensated for the unknown uncertainties, and I'm saying you really do
have to have some knowledge of what that level of uncertainty is and how
putting barriers in different positions will compensate for it; how much
of that uncertainty will you get rid of or will you lower your achieved
risk to a level that that uncertainty is acceptable.
So I'm saying you really do need a quantification metric in
this, even for what we're calling unquantified uncertainty.
MR. APOSTOLAKIS: Okay. My response to that is, first of
all, the problems that you delete and the problems that you just gave
us, I would say that's the price you pay for not quantifying
uncertainties.
The second is, again, one of my bullets said that if we do
that, we will focus attention on unquantified uncertainty, and then my
hope is that by doing that, we will eventually do what you're saying,
because somebody might say, well, gee, is it really unquantified. Maybe
we can have an estimate of the probability that all this is wrong, but
right now we don't do that.
Therefore, right now, you pay the price. You put the
barriers and you pay the price. I'm sorry, what?
MR. BERNERO: I'd like to interject on this. In the earlier
discussion, we talked about if you quantify the uncertainties, you could
make a case to eliminate the containment, say, on a class of reactor.
MR. APOSTOLAKIS: Right.
MR. BERNERO: Setting that aside, if, on the other hand, and
to Tom's point that I've got to know what to require, like some
prescription, consider, for the moment, if one would resurrect the
question of urban siting of reactors, because of the growth in the
United States and the availability of industrial property, getting close
to load centers, now, that is almost impossible to quantify the
uncertainty associated with that siting ramp.
And it's an interesting thought experiment to say what
quantification of uncertainties or what formulation would be appropriate
to reconsider that. I don't think you can do it by having a regulatory
agency invent a new siting policy, saying here exactly are the
population distribution criteria and everything that we would have to
set rational bounds on it.
If you go back to the 1980s, the late `70s and early `80s,
the agency was very heavily involved in a siting study or a series of
siting studies to attempt that.
MR. KRESS: I'm going to make a provocative, radical
statement, so everybody knows that that's what this is when I say it.
I basically think the Europeans have the right idea that
it's irrational to rely any at all on emergency response to meet risk
acceptance criteria. Now, that's a radical, provocative statement, but
I think it is irrational. I think it's part of the whole problem of why
there is lack of public acceptance in nuclear power.
And if you could design into the system to meet risk
acceptance criteria at an acceptable uncertainty level, without
requiring emergency response, then I think then emergency response
becomes a true defense-in-depth, because you're not relying on it to
meet your risk acceptance criteria. You're just saying suppose we're
wrong, let's have it anyway.
MR. BERNERO: But you aren't now.
MR. KRESS: I know. You don't meet risk acceptance criteria
without emergency response in this country.
MR. BERNERO: I don't agree with you. Reactor siting
studies that were done in the late `70s and early `80s, it is there as
defense-in-depth, but you didn't have to meet it on emergency response.
MR. KRESS: I do not think you will meet the safety goals
without effective emergency response. This is a point we'll agree to
disagree on.
MR. BUDNITZ: I have a puzzle for you, staff and ACRS, that
I can put in a pretty stark context. I want you to imagine you're
running a reactor in one of the former Soviet countries. Soviet's gone,
but there were, of course, several countries, Lithuania, Armenia,
Russia, Ukraine, that are running reactors, and a lot of those don't
have a containment at all. The old 442- 30s certainly are BMKs.
The United States Government, as a matter of policy,
implemented through the Department of Energy and the State Department,
has, as a policy, that we are trying to get those governments to shut
down all of those reactors as a matter of our policy. We have stated
that to them at the highest levels and it's part of our detailed policy,
too, I know, because I work in this arena a lot.
So that, for example, Richardson is going to go to Lithuania
in February. He is likely to tell them that we continue to oppose
running Ignolena and RBMK because it's not safe enough.
Now, suppose a government there says we've done a PRA.
Suppose a water reactor, not an RBMK, where the PRAs are more reliable,
and the core damage frequency is several times ten-to- the-minus-four,
but considering our desperate economic situation, we need that reactor
and that's safe enough for us.
The U.S. Government policy position today is no containment,
shut them down. By the way, it's not the only reason, but no matter
what else you do, no containment, let's say for the 442- 30s, whatever,
now.
What do you think of that? Knowing as much as we, everybody
around this table that knows reactors knows about them, about what those
probabilities mean, knows what -- and you understand the government says
we're going to take a bigger risk than you would be willing to take in
the United States because we need the power, that's their prerogative,
as a matter of sovereignty, and they say we know it's not contained, we
know that the consequences were we to have one of these would be greater
than they would be in the United States for a water reactor of the same
size.
They have said that one crucial element that we invoke of
our defense-in-depth philosophy, as implemented through the containment,
is absent and is still acceptable.
Now, I'm not arguing about their right to make that, that
they're sovereign, but what about that here, what would you say?
MR. APOSTOLAKIS: It's a different objective.
MR. BUDNITZ: I understand that, but what do you think --
MR. APOSTOLAKIS: So it's not an issue of defense-in-depth.
MR. BUDNITZ: But what do you think about whether -- suppose
they were three-times-ten-to-the-minus-seven and 440 megawatts, would
that be acceptable in the United States without a containment? No, not
today in the regulations. But what do you think about that as a matter
of whether it should be?
MR. APOSTOLAKIS: There's nothing we can do about it.
MR. BUDNITZ: No, no. But in other words, we're at three-
times-ten-to-the-minus-seven core damage frequency in the United States,
440 megawatts, would that be acceptable here to you?
MR. KRESS: The question would it be acceptable or not is a
tough question to ask, because it's a judgment to be made on --
MR. APOSTOLAKIS: It's a policy issue.
MR. KRESS: The question is whether it's a rational position
to take, a different question, and I think it's entirely rational to say
that that's a reasonable position to take. As long as you state your
goals on what risk acceptance criteria you're willing to live with in
terms of the uncertainty and its determination.
If you meet that ten-to-the-minus-whatever at a level of
uncertainty that's acceptable, then it's a perfectly rational position,
and that would be the rationalist view of defense-in- depth.
MR. BUDNITZ: I heard you expound that, and George saying.
On the other hand, I heard my close friend Tom Murley say, and I think
I'm with you here --
MR. APOSTOLAKIS: Unlike me, you mean?
MR. BUDNITZ: No, no. You're another close friend. But Tom
said, and he's sitting here, so maybe he -- he's two meters to my left,
so he'll say what it he wants for himself; that no, no, in the United
States, we wouldn't like a reactor without a containment, just totally
uncontained.
MR. KRESS: That's another question. I think it's probably
true, we wouldn't like it.
MR. BUDNITZ: I'm not saying whether we wouldn't, not
whether we wouldn't, but whether we should.
MR. GARRICK: I think it's a bit irrelevant. I think it is
a policy question. First off, at these reactors you're talking about,
if I had to make that judgment, I would -- getting back to George's
topic -- I would really want to turn up the microscope on the
uncertainty of the core damage frequency.
MR. BUDNITZ: Of course. I wasn't arguing that case.
MR. GARRICK: And I think I would find the kind of
information that would suggest to me that the U.S. policy is sound.
MR. BUDNITZ: I'm not arguing that for a minute. I
subscribe to that policy.
MR. MURLEY: John, could I make a point, too?
MR. BUDNITZ: Of course.
MR. MURLEY: Coming from the outside now, there's almost an
air of unreality to this discussion, because you've got to take into
account the human safety culture issues, which do cut across a lot of
these sequences and stuff.
MR. BUDNITZ: Of course.
MR. MURLEY: So Bob's premise, I think, is unrealistic. I
agree if you could absolutely prove that you had five times or
four-times-ten-to-the-minus-seventh or something, but I don't think
anybody believes you can ever do that with humans.
So you just have to keep that in your discussion somehow. I
think I understand what you're saying and the premises and so forth, but
the public, listening to this, think that what were these guys -- what
do they own, what do they have.
MR. GARRICK: I would like to comment to the allocation
issue, because I think it's --
MR. APOSTOLAKIS: That's another issue.
MR. GARRICK: Well, we've drifted into it from talking about
uncertainty. I've got plenty to say about that, too.
I need to understand a lot better, Tom, what your bounds and
references are with respect to the issue of allocation. But on the
surface, it bothers me a great deal.
The reason it bothers me is that the risk assessment is, in
my view of a risk assessment, a set of scenarios and the performance of
a particular system that you may want to allocate some risk criteria to
is strongly dependent upon where that piece of equipment sets in what
scenario.
I'm sort of reminded about the situation following the Three
Mile Island accident, when there was all this fuss about maybe we should
add a third auxiliary feed water pump to all of the reactors.
So there was an analysis that was performed as to what
benefits you would get from adding that third auxiliary feed water pump.
The answer to the analysis was that, well, if you added, in the context
of what the NRC views as a safety grade auxiliary feed water, the
benefit is very marginal. But if you remove the NRC criteria and are
allowed to not have that auxiliary feed water system have to depend on a
coolant system, a chilled water system, get it out of a hard room, so to
speak, and put it in something like the turbine building, where you
don't have to rely on certain support systems, you get a heck of a lot
of benefit.
And I can point to hundreds of those kinds of examples in a
nuclear plant, and so I have a great deal of difficulty knowing how you
could possibly allocate risk criteria in a situation where you have
reactors and plants as different as they are, where you have accidents
extremely dependent upon -- or the performance of systems extremely
dependent upon where they fit in the accident sequence.
And that may not be what you're talking about, but it's
something that bothers me. And I think that one of the things that's
fundamental and crosses a lot of these issues is that we're still
learning and the safety goal issue only began to formulate some meaning
after we started to get some results of risk assessments.
I remember the Commissioners arguing about -- and it was a
ridiculous argument -- about whether it should be one-times-ten-
to-the-minus-four or five-times-ten-to-the-minus-four, on a parameter
where the uncertainty is a factor of ten.
That's why the uncertainty is so absolutely critically
important here. As one of my colleagues would say, the uncertainty is
the risk. That's where the ballgame should be played.
I've never been one to think in terms of uncertainty being
complimentary to risk, but rather uncertainty being an inherent element
of risk assessment, just as I would argue, and that brings me down to
the TSPA/PRA issue and how much we should depend on it, that if we can
think of something in addition to the TSPA or the PRA that's a basis for
decision-making on the safety of the plant, we damn well ought to be
bringing that into our risk assessment and our TSPA.
Expert opinion, for example, is not something that should be
outside the scope of a risk assessment. So we should be striving in
that regard to make the TSPA and the PRAs as encompassing as possible.
Now, when the NRC got into the PRA act and was trying to
respond to the criticisms of the industry that they were too expensive
and went to a highly simplified and limited scope, and as the image
started to develop, in people's minds, that a PRA was something much
less than what it might be, then I can understand why you would have to
conclude that you've got to consider things beyond what's in a PRA, if
by what's in a PRA is what the NRC meant by the old IPE, where there was
essentially no uncertainty, no external events, and not much scope.
So I think these are things that really make it very
difficult for me to imagine how we can get unduly specific with respect
to something like allocation.
MR. KRESS: Let me respond a little bit to that. You can
envision all sorts of levels of allocation. You could allocate system
reliability or even component reliability. That's not what I had in
mind. I think basically with defense-in-depth, we're dealing with
prevention versus mitigation. That's basically what we're doing.
The four elements of that I talked about. What I had in
mind here was let's take the case of nuclear reactors, power reactors.
We're talking about core damage frequency versus conditional containment
failure probability.
How are we going to allocate between those two to meet, say,
LERF, which is our overall thing. What I'm saying is that in decision
theory, you ask the question if a core damage manifests itself, what are
the consequences of that in terms of my loss function; how valuable is
it to me to prevent that from happening, as a regulatory agency.
You've got to make a decision theory process and you arrive
at a loss function that says that's so valuable to me that I want to
place goals on core damage frequency or risk acceptance criteria, and
there are probably going to be a lot more going into the prevention than
there is to the mitigation.
Then you also ask yourself, well, suppose you do the same
thing with the conditional core damage frequency. You take another loss
function. What is -- and it basically becomes what's remaining of LERF,
because you've already established the loss function with your CDF.
That's a level at which I would advocate the allocation.
MR. GARRICK: Well, that's what I said, I qualified my
comments with not knowing what you really meant by criteria.
MR. APOSTOLAKIS: But in this context, then, when you talk
about, first of all, prevention and mitigation, in this case, are terms
with respect to core damage.
MR. KRESS: Yes, absolutely.
MR. APOSTOLAKIS: Because you are preventing the release of
radioactivity to the environment. In this sense, then, there is no
prevention in performance assessments. It's all mitigation, isn't it?
It would be released from -- no? What are you preventing?
MR. BUDNITZ: If you can keep it inside the canisters, long
as it's inside the canisters --
MR. APOSTOLAKIS: For 10,000 years?
MR. BUDNITZ: If you can keep it inside the canister for
10,000 years, that's prevention. I would -- in other words, it hasn't
gone anywhere. That is, in fact, the case for canisters that we talked
about.
MR. GARRICK: If you can keep the water away, you can show
that.
MR. BUDNITZ: So, George, I see that break between
prevention and mitigation as very hazy for Yucca Mountain, but I
certainly know what prevention means. Prevention is keeping it from
going anywhere. It's just in the can.
MR. BERNERO: I beg to differ on prevention. The inherent
act of waste disposal is to place the material in the biosphere or
geosphere and from then on, the performance assessment is modeling what
happens.
MR. BUDNITZ: Right.
MR. BERNERO: Does it stay in place or does it ever so
slowly corrode, decay or whatever, and there are features in waste
disposal systems that can enhance, say, containment performance.
If Yucca Mountain adopted, as I wish they would, the
addition of depleted uranium filler in the container, I think that would
greatly enhance --
MR. KRESS: That would be a wonderful addition, I agree with
you.
MR. BERNERO: Yes. But, see, this is the thing. You're not
preventing something, you're inhibiting it.
MR. BUDNITZ: That's fair.
MR. BERNERO: And I think there's a danger -- it's really a
barrier, an inhibition to the movement of the waste, because that is the
measure of performance.
MR. BUDNITZ: Yes, but when we talk about prevention in a
reactor, we mean keeping it inside where it started. In that sense,
it's not a perfect analogy, but it's not such a bad one to say that
prevention is -- the earliest state -- keep it inside the can.
MR. KRESS: I also added -- in my definition of prevention,
I added the word intervention and you have lots of time and lots of
intervention strategies one could choose. So I would say there is --
MR. BUDNITZ: Except as a matter of public policy, the NRC
has said that they're not going to count on any human intervention 6,000
years hence.
MR. KRESS: I know, but that's a policy statement.
MR. BUDNITZ: I understand that.
MR. GARRICK: I think I can make one observation that covers
a lot of my concern here about issues of allocation and definitions and
what have you, and it has to do with I don't think we should do anything
that bounds our thinking about the safety of what we're dealing with, be
it a repository or a reactor plant.
We all know that we've had experience with this. When we
adopted the design basis philosophy of safety of nuclear power plants,
we, in a sense, bounded our thinking. The game became if you come
forward with a design basis accident and you convince everybody that
it's acceptable, then you're okay. It's the same thing. The other
language we've heard about is beyond Class 9 accidents.
There shouldn't be those kind of artificial thresholds and
boundaries, even though it made it more convenient, from a regulatory
standpoint. And allocations have a tendency to do that and subsystem
requirements have a tendency to do that. They have a tendency to narrow
the view of what we should be analyzing, what we should be designing
against, and what we should be analyzing, what we should be designing
against, and what we should be controlling.
Even core damage frequency is a limitation, because I can
think of scenarios in lots of plants that would decrease the core damage
frequency and increase the public risk, and I think we have to be very
open and clear about that, and I think that's the virtue of PRA.
MR. APOSTOLAKIS: I disagree, though. I think there is an
element that's missing here.
MR. GARRICK: You disagree?
MR. APOSTOLAKIS: No. It's not -- when we say allocation,
we should not take it only in the mathematical sense that you want to
have a certain -- meet certain goals and that you allocate the
performance of various systems. There is a more fundamental reason why
the staff wants to do some of that.
Even though there may be situations where you are -- you
know, a certain measure, as you just said, may decrease or increase the
core damage frequency, but the role is beneficial, the staff wouldn't go
for it, because core damage by itself is an undesirable event.
See, the assumption in what you said was that all I care
about is the QHO and the staff will tell you no, that's not all I care
about. In fact, the new oversight process makes it very clear in black
and white. The staff says we care about initiating events, we don't
want to see any of those. Why? Well, they aren't going to put it on
paper. They will tell you, though, that they don't want to be on the
front page of the newspapers. We don't want to see the primary system
being breached?
Why? It creates public outcry. We don't want that. So
there are more objectives that perhaps have not been spelled out in the
books until recently for which -- which you are trying to meet, and if
you look at it that way, then you are saying, well, maybe core damage
frequency is something I worry about, because it's not just a QHO.
The fundamental question is, though, whether you have
similar situations in the performance assessments and I think one of the
reasons why you don't is time.
In reactors, we can have a problem tomorrow with an
initiating event. In your case, you're talking about thousands of
years.
MR. GARRICK: Yes, the conditions are entirely different.
The real issue of risk probably in the waste field is the operational
risk and the handling and the way in which you do things.
MR. APOSTOLAKIS: But my point, John, is that maybe the word
allocation for reactors is not the right word, because they are not
allocating anything. They are saying I don't want this to happen, I
don't want the core damage event, I don't want an initiating event.
MR. KRESS: When I say allocation, I mean I don't want that
to happen at this frequency, with this uncertainty, with this confidence
level.
MR. APOSTOLAKIS: I understand that.
MR. KRESS: That's what I mean by allocation.
MR. APOSTOLAKIS: But there is a reason why they don't want
it to happen, because that by itself is bad; not only as a contributor
to core damage, but if I have a LOCA tomorrow, the agency doesn't look
good.
MR. GARRICK: But, George, you're not saying that the NRC
disallows the core damage. They can't do that. They can't do that.
Are you saying that -- what you seem to be suggesting is that the NRC
really doesn't think in terms of a ten-to-the-minus- four core damage
frequency, but a ten-to-the-minus-infinity.
MR. APOSTOLAKIS: When did I say that?
MR. GARRICK: Well, you made the point that they wouldn't
accept it. Well, what are they not accepting? They can't stop it.
They can't stop the fact that the core damage frequency has a likelihood
of occurrence.
MR. APOSTOLAKIS: What I'm saying is when we say allocation,
we have to be very clear what we mean. That comes back to what my
objectives are when I regulate. I got the sense from your earlier
comments that what you thought was the objective of the regulation for
Yucca Mountain or for reactors was the ultimate quantitative health
objectives or, in Yucca Mountain, the dose. The ultimate criteria, in
other words.
And then allocation, in that sense, means that some engineer
says, well, gee, you know, this is really my objective, but I would like
to see this performance here, that performance there, in the system.
What I'm saying is, no, there is a fundamentally different view of
regulation for reactors. It's not only the public health and safety.
That's how we start, but that's not our only objective. We
don't want to see core damage events by themselves, even though they
don't affect public health and safety, because they're contained.
But even more than that, in fact, the staff said it very
clearly, the initiating events, we don't want to see too many of those.
They create those sorts of headaches, other things. We don't want to
see -- whatever -- the four cornerstones they have. So what I'm saying
is that the decision problem is different in this case in the sense that
I have different objectives and I'm not allocating anything anymore.
All I'm telling you is I really don't want to see this.
MR. LEVINSON: But, George, I think historically we have
confirmation. The importance of TMI was not exposure of the public.
The importance of TMI was that it was core melt.
MR. APOSTOLAKIS: Yes. Yes. And we saw the reaction and so
on. So that supports, in fact, the staff's position. You may have -- I
mean, as Tom said earlier, you can have a TMI every year and you still
meet the goals. You tell me who at the NRC would accept that.
MR. GARRICK: And my only point is be careful about the
blinders you put on to support the staff's position, because we put
blinders on us to support the staff's position in the past and we
probably should have not. Be careful about that.
MR. APOSTOLAKIS: I'm not sure they're blinders.
MR. GARRICK: Well, you're the one that's suggesting that.
I think that all I'm suggesting, all I'm suggesting is that the real
virtue of the risk thought process, and by which I mean all these things
we've been talking about, quantification of uncertainty, complete set of
scenarios, doing the best possible job we can, is that we have not built
ourselves artificial thresholds, like safety-related systems.
I think that that's the thing that is an important virtue of
it that we should not lose by adding some constraints.
MR. APOSTOLAKIS: And I agree that they should not be
artificial. But look what happened at Northeast Utilities. Was that
artificial, was that a real reaction? Was public health and safety
threatened at any time?
So it's clear to me that for reactors, it's not just public
health and safety.
MR. GARRICK: Well, I agree with you and I want to stop
because I want to hear from a lot of people. I would say one of the
greatest advances we've made in the improved performance of the nuclear
plants in this country is not the business of the traditional safety
analysis and what have you, but it is the emphasis that the utilities
have been giving to human performance.
I am really impressed with what you will find at most
utilities today on evaluating human performance and how to motivate them
and how to challenge them and how to make them accountable for what
they're doing. And it's true, in the sense that it's outside our
database, which it isn't totally, we don't consider a lot of those kind
of things.
MR. LEVINSON: If I can make just one more comment, John. I
think these are not at all inconsistent. The value of good analysis to
reduce uncertainty, PRAs, et cetera, certainly is something we should
all strive for, but I think the point is what we get from it is not just
a single number, like dose to some person in the population.
It can also be used to achieve other objectives, like
reduced core melt. So the fact that you might have multiple objectives
for the PRA is not inconsistent with depending on PRAs and proving them.
MR. BUDNITZ: Let's go to Yucca Mountain for a minute. When
Part 60 was under development, I was on the staff 20 years ago when we
were thinking hard about it, and at that time, nobody had confidence
that what we now call performance assessment could be good enough to be
relied on as a principal means for understanding. And because of that,
the staff, at the time, wrote the subsystem performance requirements,
the canister lifetime and some canister leakage rate per year and the
thousand year travel time and so on into the regulation.
Notwithstanding everything else you did, you had to show
this thousand year travel time, for example. The staff explicitly, in
the statement of considerations of Part 63, just this year, said 15-18
years have passed; we now, says the staff, and I agree with this
entirely fully, we now have the confidence in the analysis methods and
the data that we didn't have them, we the same staff or the different
folks of the same staff, and, therefore, we feel that those things have
been superseded by this new technology and its use and our confidence in
it.
So they have come to the stage where they used to have what
you'd call barrier -- the concept of these multiple, whatever else you
do, you've got to do barriers or something, performance, they've
abandoned it for the moment.
I mean, there's still this other thing, and I think that's
completely correct. When evolution of knowledge enables you to say I
now don't have uncertain values to have, I now can do certain analyses
and I can have confidence in them at a certain level, I no longer need
what I used to need 18 years ago. That is completely rational.
MR. APOSTOLAKIS: But your objective is still to meet the
dose criteria. I fully agree with that approach. You don't have any
intermediate objectives. So what I'm saying is that in reactors, it's
--
MR. BUDNITZ: No, no. I'm not -- of course, I'm not arguing
with you for a minute, but then all of a sudden, in the same statement
of consideration, Part 63, they say but besides the dose objective in
Amergosa, we have this defense-in-depth. My slide showed, I asked the
question, well, if we're going to invoke it, can they flunk on
defense-in-depth, even if they meet that other thing with lots of
margin, and apparently the answer is yes.
The staff has said yes, they could flunk on defense-in-depth
and then you have to ask, well, what does that mean. I was trying to
probe in my slides what that might mean in terms of some sort of
allocation or in some sort of a figure or in some sort of a do it
analysis of a degraded or under-performing barrier and tell us what it
means and whatever it means, are we going to flunk you on that one.
If Yucca Mountain can flunk on one of these, even though
they meet the overall thing with lots of margin, then you have to figure
out what does it mean, what sort of allocation have you come up with,
you see.
MR. APOSTOLAKIS: You just said that now we have confidence
that we can calculate these.
MR. BUDNITZ: It's not a perfect tool.
MR. APOSTOLAKIS: But let me ask you this. What are the
major unquantified uncertainties in performance assessment?
MR. BUDNITZ: Unquantified uncertainties.
MR. APOSTOLAKIS: Yes.
MR. BUDNITZ: I suppose they'd be some of the models that we
still haven't tested well enough.
MR. APOSTOLAKIS: This is not something people talk about?
MR. BUDNITZ: Of course, we talk about it every day.
MR. APOSTOLAKIS: So models --
MR. BUDNITZ: It's at the center of what we talk about.
MR. APOSTOLAKIS: Are these uncertainties large enough to
invalidate the performance assessment itself?
MR. BUDNITZ: Well, my personal view is that Yucca Mountain
is very likely to meet that dose criterion out there in Amergosa with
lots of margin, including these.
MR. APOSTOLAKIS: Including the unquantified.
MR. BUDNITZ: Including -- I mean, there is some judgment
about the models. You always have to bring some judgment in the end,
because not everything has been tested, especially with those long
timeframes and that's certainly true of the metallurgy of the can.
But it is my view that in the end, that will be the case.
I'm still holding open judgment because the final design isn't here and
certainly analyses haven't been done on that. But if that's true, if it
turns out that there's lots of margin against the dose, the staff says
but you can still flunk because you flunk something about
defense-in-depth, what is that?
I'm struggling with it, because it isn't the same as what
you're saying, well, a core melt is bad. You know, Millstone was bad.
It's not the same sort of thing.
MR. APOSTOLAKIS: I understand that. That's what I keep
saying for the last ten minutes. They are two different things. If you
guys knew, if the Commission believed that by building Yucca Mountain,
you will have a major incident five years later, I'd bet you there is
going to be an objective there in order to have it.
MR. BUDNITZ: Yes, of course, or --
MR. APOSTOLAKIS: If it's a thousand years --
MR. BUDNITZ: Or even if it's a thousand years, because they
have a 10,000 year criteria. So I think it's a challenge. I'm looking
at Ray and John from the ACNW and all of us that have thought about this
hard. It's a big challenge to figure out what you mean and what you do.
MR. APOSTOLAKIS: I see there are two different variables.
MR. BERNERO: Tom, I'd like to interject here. The
discussion of an incident in the near term against a waste disposal and
also a remark that John made earlier about if you've got some
significant uncertainties, get them into the performance assessment,
which is an admirable objective.
First of all, there has to be not an allocation, in my mind,
but a recognition that in waste management, and I will use low level,
near surface waste disposal, as an example, there is a sequence of
allocated allowances or decisions; is this site acceptable, is this
emplacement design going to be an acceptable compliment with the site,
and, of course, taking the whole system into account, is it going to
satisfy the performance assessment requirements, the dose limits
off-site and so forth, taking account of the uncertainties and climate
and flow and intrusion and so forth.
Now, if you go, as a practical matter, in Part 61, there are
explicit site criteria and there is an extensive body of guidance on
performance assessment, but there is not a good way to analyze, to do
the uncertainty analysis of emplacement techniques.
Basically, what any new site that was going to be built east
of the Rocky Mountains, what they did is just adopt the French approach,
and the French approach is select the site that's proper, build it with
dual liner leachate collection system caps and all the bells and
whistles, and do your level best to make sure it never leaks.
And you don't quantify that in the performance assessment.
You have uncertainties and you live with those uncertainties. Take the
item 129, if you go to a low level waste disposal site, all these
shipments that come in, and you're talking 100,000 shipments, big
numbers, they all have item 129 is less than or equal to X.
It's detectability limit and if you take 100,000 times less
than or equal to X, it's five orders of magnitude higher than that.
I've had the authority for the French low level waste site at Loeb tell
me that halfway through, we're going to hit the limit on item 129, and
he doesn't have a performance assessment technique to get out of that.
He doesn't have an analytical detection technique. He's got to use some
judgment.
And ultimately I think they will get out of it. They're not
going to stop and say this is the limit for this site, because it's not
real and it's also not a real threat, item 129.
So there are many things in waste disposal that you cannot
firmly quantify. You've got to evaluate and make a judgment. It's very
difficult. And the decisions, right now the staff is heavily involved,
and the Commission, too, in advising or concurring in what DOE is doing
to clean up its waste tanks and a high level waste tank, when you
extract that waste, the Commission promulgated criteria on how can you
stand up and say the high level waste is out, when you know there is
residue there.
The residue isn't well quantified, it isn't well located,
and it's the difference between two very large numbers and it's very
difficult to do uncertainty analysis on it.
You can't characterize it, you can't sample it. And so your
performance assessment for that site is going to say I'm satisfied that
you've extracted enough, DOE, and that you have made a persuasive case
about how you grouted it, how much grout there was, how much residue you
estimated it to be, and so forth, and then you're going to do a very
elementary or simple performance assessment that doesn't take any real
credit for the grout and the can and many of the barriers.
MR. KRESS: This is an interesting discussion, Bob, because
I think what you're saying is here is a circumstance where we just have
uncertainties that we can't quantify, so what we do we do in that case,
in a risk-informed regulatory world.
MR. BUDNITZ: That will be true at Yucca Mountain. There
will be some uncertainties we can't quantify.
MR. KRESS: So it's an interesting question, what do you do
when you can't quantify the uncertainties. I think you fall back on
arbitrary defense-in-depth. Arbitrary in the sense that you put the
best you can here and there.
MR. GARRICK: You fall back on a combination of some sort of
judgment, too.
MR. KRESS: I want to just introduce a conceptual note here,
because what you're really saying, Tom, is that it's not so much you
can't quantify it, but you just don't like the result, because the
principals ought to be there, that you can always quantify it. It just
may be that you have ten orders of magnitude of uncertainty when you
would like to have two.
And in the presence of that level of uncertainty, then you
have to do something. But I think that the whole discipline that we're
talking about here is to be able to assign values to parameters based on
the evidence that you have, and you always have some, but in the
problems we're dealing with, there are too many areas where we have much
less than we'd like.
One of the things I would like to do here before the break
is look to my colleague, Ray Wymer, on the performance assessment angle,
who has been doing a lot of thinking lately about some of the key
uncertainties associated with one aspect of performance assessment
that's critical to improving the models, and I suspect, Ray, you could
identify some examples of areas of uncertainty on the chemical side and
offer opinion about the likelihood and what needs to be done to resolve
those.
Would you comment on those and kind of against the
background?
MR. WYMER: I suspect you think I've been too quiet for too
long.
MR. KRESS: Yes. I know you have a lot to say and I hope
that there is an opportunity for you to do so.
MR. WYMER: I'll say a little bit about chemical
uncertainties, which is fairly specific, and then I think tomorrow, when
we adjourn discussion, I want to make some general comments that I've
noted down here that are not necessarily appropriate to this specific
discussion we're having right now.
But there are a lot of chemical uncertainties with respect
to Yucca Mountain and the repository. For example, there still is
uncertainty about the corrosion behavior of alloy C-22 and while there
is a lot being done, it still remains that you can't take a couple of
years worth of studies and extrapolate them for 10,000 years very well,
although the more basic understanding you have, the better off you are
in your extrapolations.
So the primary line of defense, which somebody mentioned,
maybe Bob Budnitz, that the waste package, the waste container is really
the principal reliance, which is true, for containing the waste and
preventing exposure, there is uncertainty remaining there, which people
are working trying to narrow, both in the NRC and in the Department of
Energy.
In addition, there's a good deal of uncertainty about the --
once you breach containment and you get into the fuel material itself,
there is a lot of uncertainty with respect to the formation of secondary
precipitates, materials that would tend to provide another line of
defense against release of radioactivity.
People don't really know what these second phases are. They
are extraordinarily complex because of the complexity of the nature of
the fuel and the nature of the corrosion products that meet that fuel
and the complexity of the water that's coming in.
So there may be additional barriers to release. There's a
lot of uncertainty there, though, and there's been no real attempt, no
real concerted attempt to quantify those processes that may limit
release of radioactivity in a significant way.
It's been mentioned briefly here that you can put in
backfill materials, like UO2, into drift, or you can actually put those
inside the waste package, which, by a saturation effect, can reduce the
rate and extent of dissolution of the fuel, and also lead to additional
secondary phase formation.
These are all uncertainties. Most of what I mentioned, with
the exception of corrosion, is an uncertainty at the direction of
greater containment of the radioactivity to make the waste environment
more retentive than the analyses are currently showing.
But without belaboring the point too much, there are
chemical uncertainties which are, in my view, large. There are a number
of mitigating things that could be explored, like backfill materials,
that could enhance the safety of the repository and could decrease
somewhat the uncertainty in the analysis, and all of these things, in
the best of all possible words, would be examined.
The time constraints that we have with respect to the
license application would seem to pretty severely limit the amount of
investigation you could make of some of these potentially very important
chemical thought processes. However, if, for some reason, we get into
the bring-me-another-rock mode, there may be more time available to
solve some of these problems.
MR. APOSTOLAKIS: Are these uncertainties in the PA's now?
MR. WYMER: Only in a very general way, George. There is
practically nothing that I could think of or that anybody could think of
that hasn't been mentioned in the performance assessment, but mentioning
them is one thing and dealing with them competently and comprehensively
is quite another thing, and I think it's that latter that's weak.
MR. GARRICK: One of the things that's very interesting
about these problems, I'm always looking for comparisons. The key to
the reactor safety problem is water. The key to the safety problem is
the absence of water. Also, it turns out that one of the attractions of
using core damage frequency as a measure of performance in the reactor
is because of the step change in uncertainties that occur once the melt
occurs, and you try to quantify the accident progression.
But we're kind of in that position in the waste field. We
have a problem that's not too dissimilar in terms of the bounding of the
problem and what have you. Fortunately, the time constants are much
longer and that's to our advantage, but the problem in the waste field
is once you get the material mobilized, coming up with models that do a
rational, reasonable job of defining the mobilization, the retardation,
the dilution and the transport of the radioactive material.
It's a problem not too unlike the accident progression
following core melt, although the thermodynamic conditions are clearly
very, very different and the concentrations of materials are clearly
very different. But there are some interesting analogies.
MR. LEVINSON: I'd like to make a couple of comments. One,
I want to emphasize something that Ray said that slid by very quickly,
because it was one of the points I had before, and that is everybody is
talking as though uncertainties were all negative.
In fact, that's not true at all. There is a substantial
number of uncertainties which are positive, that reduce dispersion of
materials, et cetera, and we just have to remember that not all
uncertainties are negative in any sense of the word.
MR. GARRICK: What you're saying is that an uncertainty
distribution has a negative side and a positive side.
MR. LEVINSON: Absolutely, absolutely. But we talk about it
as though all uncertainties were bad. As I sit here and listen, I hear
more and more reasons for why the waste issue and the reactor issue
really are very, very different sorts of things. For instance, in the
waste thing, after you start out, the potential risk steadily
deteriorates as stuff decays away.
At a reactor site, the potential risk increases, as over the
life of the reactor, you continually increase the inventory of fission
products on the site. Thing after thing.
Bob showed his dose curves out at one MR or ten MR, it
doesn't make any difference. When you get to the top of the chart, rate
is probably at least as important as dose. Bob has, on his chart, a
thousand rad is certain death, but both his wife and mine, in the last
couple of years, have received significantly more than that in treatment
of cancer.
The dose effect -- now, in a reactor accident, the dose rate
basically, from a prompt criticality, it's an instantaneous thing.
There is no way, in a waste disposal, that anybody is going to get a
high rate of dose. So I just think these things are completely
different.
On history, I want to throw in one comment, since I'm
probably the oldest person here. The NRC may have invented the words
defense-in-depth, but they didn't invent the philosophy. When I joined
the project in 1944, DuPont -- and it wasn't the chemical part of the
company, it was the explosives division of DuPont that was in the
Manhattan Project, and they brought that concept.
It was the first lesson I got when I went to work there.
It's been around a long, long time and I don't know that we're going to
define it or cage it in. It's been a very useful device for designers
and builders, and it's been there a long, long time.
Just one other comment. There was a comment by Bob Budnitz
about U.S. policy for shutting down reactors without containment.
Clearly, that's not a technical based issue at all. But the Soviets
have very, very limited -- now, they have more because we've given it to
them, but they had very, very limited ability to do analysis. I
probably know about as much about it as anybody in this room, since I
spent eight years on the board of directors of the Soviet Nuclear
Society.
They did do an analysis in regard to shutting down the RBMKs
at Chernobyl and in a very basic way, in one of the discussions I had
with them, they said maybe our risk of duplication of the Chernobyl
accident is ten-to-the-minus-third, and I said is that acceptable to
you, and they said, wait, we haven't finished telling you the analysis.
If we duplicate the Chernobyl accident, we'll kill 30-some
people. If we shut it down tomorrow, probably ten times that many will
die this first winter. And in this country, we have the luxury of being
able to say you can shut down a reactor without major consequences. In
other parts of the world, that's not the case at all.
Their analysis -- it isn't that they have different values
for what's an acceptable number; they have other considerations.
MR. MURLEY: Tom, could I ask a question that occurred to me
about your concept of allocation? I guess I have different reaction, if
you want to impose it as a requirement or if it's a target.
If it's a kind of aiming goal or target, I think that's a
very good concept. But if you're suggesting that it become embedded in
regulations or something, I have a different reaction about it.
MR. KRESS: And I'm sorry to tell you I had the second, the
latter. The reason I have that is I think in a risk-based regulatory --
risk-informed regulatory system, you can no longer have targets for
individual plants. You have to have risk acceptance criteria for
individual plants.
If you have to have those, then they have to be part of the
regulation. So I really did mean the latter, which I know gives you
heartburn.
With that, I think this is a good time for us to break for
lunch until 1:00, at which time we will hear some interesting comments
from the staff. We're recessed until 1:00.
[Whereupon, at 12:05 p.m., the meeting was recessed, to
reconvene at 1:00 p.m., this same day.]. A F T E R N O O N S E S S I O N [1:00 p.m.]
MR. KRESS: The meeting will come back to order, please.
Before we get started, there's just a very minor change in
the agenda I'd like to point out to people. We were up to item five on
the agenda, which was NRC staff presentations by Gary Holahan and Tom
King.
Instead, we're going to interchange that with item six,
because of some problems, and we're going to have the NRC staff
presentations on the defense-in-depth in high level waste first, and
then move to the defense-in-depth in reactor regulation.
So with that, I will turn the floor over to John Greeves.
MR. GREEVES: My name is John Greeves. I'm Director of the
Division of Waste Management in the Office of Nuclear Materials Safety
and Safeguards. Mr. Chairman, let me thank you for making a schedule
change. Norm Eisenberg, the principal brief, is coming down with
something. He's been coming down with it for days and I think he's sort
of running out of energy. So we thank you for your discretion in
leaving the schedule a little bit.
We also apologize to the audience for moving the time around
a little bit, but for the sake of Norm being able to deliver his
presentation, I think it was the best thing to do.
Again, I am the Director of the Division of Waste
Management. I have spent a fair amount of time interacting with the
Advisory Committee on Nuclear Waste. So obviously this is a time for us
to comment and bring some of our ideas to the process.
I appreciate the difficulty which people were addressing
this issue this morning. Defense-in-depth for materials and waste
licensing actions presents a number of challenges, and you bumped into a
number of those this morning.
Unlike reactors, we have the full spectrum of activities
within NMSS, from exempt sources, which you discussed this morning,
medical activities, sealed sources, fuel fabrication facilities,
transportation, low level waste, high level waste.
It's really a family of different types of licensing
activities. So I think a lot of that was brought out this morning. I
was pleased to see that. I was also heartened by some of the views
expressed. I can tell you there's a number of views within the staff on
these issues, also.
The topics, depending on what type of a licensing activity
you're talking about, have different time spans, have different radio
activity, have different human action, have different criteria, and have
different rates. You touched on all that this morning.
I would like to just punctuate that the staff certainly
looks at the Commission policy statement on risk-informed
performance-based regulation, and I think it's probably in your package
and it has a definition on defense-in-depth, and the staff, in its
efforts, is looking to make sure we stay consistent with that particular
policy statement. It's on the web and is available to people.
As I said, Norm Eisenberg, Dr. Eisenberg is walking this
way. I'll try and not get too close to him. Norm is going to do the
principal presentation. He's going to try and set the context for all
the materials types of activities and a couple of things about Norm.
One, this may be your last chance. He's retiring this
month. He's moving on. The second thing is I think he's a
defense-in-depth expert. This is a gentleman that lives defense-
in-depth. When he gets up, you will notice that he has belts and
suspenders. I've heard statements that people thought they were the
best at certain things. Norm lives this issue.
The second presentation will be by Christiana Lui, to my
left, and that's more focused on Yucca Mountain specifically. I will
have some wrap-up statements regarding that.
As I said, we keep in mind the Commission policy statement
and what we are expressing are our preliminary considerations on a
number of these issues.
With that, I'm going to stop and ask Norm to go through what
I think is a thoughtful presentation. I think it's a bit
thought-provoking, as some of you put forth earlier.
MR. BERNERO: Do you have slides handed out?
MR. GREEVES: There are slides, should be. Norm, you
concentrate on the presentation. We'll get the slides to Bob.
With that, Norm, take over.
MR. EISENBERG: Thank you. I appreciate the subcommittee
letting me go ahead and do this. I am feeling under the weather and I
feel confident that if I start to become incoherent, nobody will notice.
They'll just figure it's me acting normally.
I should say that I'm going to talk about a provisional NMSS
perspective on defense-in-depth for risk-informed performance- based
regulation. These are some staff ideas that have been circulating
around and a lot of them were sharpened by considering the case for high
level waste regulation.
So you have to understand that these are provisional ideas
and they are subject to change.
So what I intend to talk about are what are some of the
motivations for defense-in-depth in NMSS; what are some of the current
things that are causing us to focus on it; what is it, which, of course,
we've heard a lot of discussion about that this morning; how does
defense-in-depth differ from margin and other safety concepts, which I
think is a very important issue; what are some provisional conclusions;
what are some things that we have to determine if we're going to follow
this path; and then I'd like to make a summary.
So NMSS has been engaged in a number of activities that
prompt a focus on defense-in-depth and a risk-informed performance-based
regulatory environment.
One of the first things is SECY 99-100, which was approved
by the Commission, which is an activity to develop a framework for
materials regulation similar to the framework for reactor regulation
that was developed by the Offices of Research and Nuclear Reactor
Regulation for risk-informing selected NMSS activities.
So this certainly has brought the subject up, certainly the
consideration of refining the approach on high level waste regulation,
as indicated in the proposed Part 63, is another area where
defense-in-depth needed to be considered, and we got a fair number of
public comments on that aspect of the proposed rule.
There are other activities in specific areas, interim spent
fuel storage facilities are being risk-informed. We have ISAs, which is
a type of risk assessment for fuel cycle facilities, and we are
risk-informing the transportation regulation. So there is a lot of
current interest in this.
Let me just say that the performance-based aspect of risk-
informed performance-based regulation places an emphasis on the overall
system performance and the risk-informed aspect considers the
uncertainties and the sources of those uncertainties.
All right. So what's the regulatory environment in NMSS
that we have to deal with? First of all, we have a lot of diversity.
We have a wide range of licensees and systems regulated. They have
varying degrees of complexity, everything from gaseous diffusion plants,
which are complex, to smoke detectors, which are not.
Different systems have different degrees of human
interaction or are dominated by human interaction. We have certainly
different levels of hazard. Some things are not very hazardous at all.
This gives rise to general licenses. Other things are, frankly,
hazardous.
There's diverse capabilities among our licensees for being
able to do analyses of any kind and especially risk analyses, and
there's many different tradeoffs in the need for risk-informed
regulation, the benefits and the costs in different areas that we
regulate.
We also need to consider, if you will, the taxonomy of the
risks, and Bob Bernero alluded to this earlier, that we have individual
risk to workers and we have the individual risk to members of the
public. We have normal risks and accident risks. We have perceived
risks and actual risks and we have a variety of initiators, mechanical
failures, external events and human error are some of the things.
MR. APOSTOLAKIS: Why do you have perceived risk?
MR. EISENBERG: Because we have to consider the
communication with the public and even though the actual risk in
quantitative terms may be small, the public reaction may be great. So
there will be a response. So we have to consider not just the actual
risks, but, to some degree, the perception of risk by the public, by
policy-makers, and others.
MR. APOSTOLAKIS: But I realize that communication is
important and so on, but surely you're not implying that you will take
actions based on perceived risk rather than actual, as actual meaning
technical. We are not regulating based on perceived risk, are we?
MR. EISENBERG: The agency may have to respond to some
things with an effort which is not in proportion to the actual risk
involved.
MR. APOSTOLAKIS: That I agree with and I think, in fact,
the cornerstones that we have on the reactor side are the result of
perceived perceptions.
MR. EISENBERG: I'm just trying to lay this out as the
environment in which we work. Now, how we actually treat it is another
issue, but it is a factor and it does influence what goes on.
MR. APOSTOLAKIS: I agree that it is a factor.
MR. EISENBERG: Well, I'm glad you agree with me. So kind
of moving to the next step, what are the factors for defense-in- depth
in NMSS, what's the current status?
Well, it's the nature of the licensees and the activities
regulated. We have to recognize that NMSS, by and large, regulates
systems with less hazard than nuclear power reactors. NMSS regulations
are a mix of performance-based and risk-informed regulations versus
prescriptive and deterministic regulations.
This is a little bit different, from my understanding of the
reactor side, where things have been dominantly a deterministic
approach. And for some NMSS licensed activities, the hazard does not
warrant a very strong preventative measure of any type, whatever they
are, performance-based or prescriptive or anything. The risks are too
low. Once again, general licenses are not worth very much concern.
Okay. So what's the NMSS safety philosophy? Well, our
strategic plan says that we want reasonable assurance of protecting
public health and safety, common defense and security, and the
environment. Some concepts that assist in achieving defense-in-depth in
this context are safety margin, diversity, redundancy, no single point
of failure, and quality assurance. There is a whole spectrum of things
we do to try to achieve reasonable assurance.
And in this context, defense-in-depth is a component of a
risk management strategy. This does not imply that we do risk
management, all the risk management that a licensee might want to do.
They have other reason to do risk management, but we are obligated to do
risk management in the public health and safety context.
MR. KRESS: When you say risk management, what exactly do
you mean there, Norm?
MR. EISENBERG: In other words, putting forward a structure
of regulations makes certain things less likely and other things more
likely and it is a way of determining what the risks are and how large
they might be allowed to become.
If you take the Kaplan-Garrick definition of risk as the
risk tripled, then regulations provide one constraint on the risk,
meaning that whole aggregate of points.
MR. KRESS: I think I know what you mean now.
MR. EISENBERG: Okay. All right. So if we're going to use
defense-in-depth to help achieve our top level goals of public health
and safety, what is it? Well, this is what was taken, and I forget who
threw it up this morning, but this is from the Commission white paper on
risk-informed performance-based regulation, and this is a paraphrase of
the two key features for defense-in-depth, which are, one, safety is not
wholly dependent on any single element of the system and, two,
incorporation of defense-in-depth into a system produces a facility that
has greater tolerance of failures and external challenges.
MR. KRESS: That's a pretty loose definition.
MR. APOSTOLAKIS: It's, in fact, not a definition.
MR. GREEVES: This is right out of the Commission paper.
MR. APOSTOLAKIS: We realize that.
MR. KRESS: We realize that. Thank you.
MR. APOSTOLAKIS: I thought our comment at the time was that
this is still evolving.
MR. GREEVES: This is what the staff is looking at in terms
of guiding its efforts and being consistent with the Commission paper.
MR. EISENBERG: We took this as one of our starting points.
MR. BERNERO: This is the same thing I put up. This is just
a paraphrase of it.
MR. APOSTOLAKIS: It's what? I'm sorry.
MR. BERNERO: It's the paragraph I put up. The paragraph
that I put up on the screen, this is a paraphrase of it. It's one of
the attempts at defining defense-in-depth. You've got a whole book full
of them.
MR. EISENBERG: And here is the whole statement, which I
think -- okay. Well --
MR. GARRICK: I think if you put it in the context we were
discussing this morning as a way of doing business, as a way of how we
provide protection, it fits in that scheme.
MR. EISENBERG: So then the question is how do you do
defense-in-depth in a risk-informed performance-based context. Things
change when you get into a risk-informed performance-based context,
rather than a prescriptive deterministic context. This, I thought, was
stated very nicely in this paper by Sorenson, et al, in which there was
the structuralist and rationalist approach.
So this is, once again, a paraphrase and may not be complete
enough to satisfy everybody in the audience, but basically the
structuralist approach maintains that the need for and extent of
defense-in-depth is related to the system, structure. Many
manifestations are based on the novitant perspectives that were current
at the time that the systems were developed or they were first licensed
and some manifestations have an ad hoc basis.
The rationalist approach articulates a philosophy that says
defense-in-depth should be related to the residual uncertainties in the
system and the rationalist approach is just beginning to be adopted in
this risk-informed, performance-based environment.
And we have taken the structuralist -- I'm sorry -- the
rationalist approach as appropriate for risk-informed performance-based
regulation. But the question is how do you implement it and what are
those uncertainties that you need to address.
MR. APOSTOLAKIS: What do you mean by residual
uncertainties? Unquantified?
MR. EISENBERG: Yes.
MR. APOSTOLAKIS: Okay. There is something that --
MR. GREEVES: I'm going to talk more about this.
MR. APOSTOLAKIS: Is there something wrong with the word
unquantified or why are you avoiding it?
MR. GARRICK: Don't be so sensitive, George.
MR. APOSTOLAKIS: Residual is different, because some of the
residual uncertainties have been quantified.
MR. EISENBERG: Remember, what we're assuming here is that
you have a risk-informed performance-based approach. So you've already
folded into your compliance demonstration -- this is very much the case
with Part 63. You've already folded into your compliance demonstration
--
MR. APOSTOLAKIS: I understand.
MR. EISENBERG: -- consideration of the uncertainties that
you have quantified. They are in there.
MR. APOSTOLAKIS: Right.
MR. EISENBERG: And whatever the criterion is, and for Part
63, it's that the peak of the mean dose be less than 25 millirem, as
long as you meet that, you're okay.
MR. APOSTOLAKIS: But what I'm saying is that after I have
implemented the risk-informed system, yes, I will tolerate certain --
some uncertainty that things will go the wrong way. But that doesn't
mean I'm going to invoke defense-in-depth to handle those, because those
I have quantified.
It's the things that I have not included in my analysis. So
the word residual perhaps is not so fortunate.
MR. GREEVES: He's got some slides that are going to touch
on your issue.
MR. APOSTOLAKIS: I think conceptually we agree.
MR. GREEVES: I think he's going to hit another button here
shortly.
MR. EISENBERG: Just briefly. So what are the uncertainties
that we consider in these safety assessments, and there's
MR. BUDNITZ: Regulatory.
MR. EISENBERG: Well, there is that differentiation, but
there is also, for those of us that are doing the pragmatic, there's
parameter of data uncertainty, there's model uncertainty, there's
scenario uncertainties, which, for a lot of waste work, involves the
exposure scenario as opposed to some physical scenario, and, also,
programmatic factors; the safety culture, for example.
So this is one cut at uncertainty.
MR. GARRICK: And on way you could look at that, Norm, is I
might even view scenario uncertainty as an integral part of the modeling
uncertainty, given that the scenarios are usually a fundamental part of
the modeling process.
MR. EISENBERG: It's the model of the world or the model of
the system.
MR. GARRICK: And the programmatic factors, like QA, those
are there primarily because we don't normally address them explicitly.
In other words, it's not that they couldn't be, it's just that we don't.
MR. APOSTOLAKIS: In fact, the last three, I call them
modeling uncertainty, but if it makes you happy, that's fine.
MR. GARRICK: Well, we agree.
MR. APOSTOLAKIS: We don't want to make Norm unhappy. Not
yet.
MR. EISENBERG: Okay. So now, if we get back to the
residual uncertainties or the unquantified uncertainties, I would
suggest that there may be two types. The first type is if you have the
best available risk assessment, if you do the best possible job you
could do, there are still unquantified uncertainties and it's because
human knowledge is finite and you just can't put everything in there.
You don't know everything.
So that's one type of uncertainty. But there's another type
of uncertainty and that's got to do with there's practical realities and
we can't always get the best available risk assessment. Very often, in
the real world, we have to deal with a risk assessment that was done.
It may not be the best available one. There may be significant flaws.
And we also have to consider, in those cases, that there are
unquantified or residual uncertainties.
MR. BUDNITZ: Norm, as a distinction here, in the first one,
you characterize that you did the best you could. You said the reason
why it's not better still is because the state of knowledge is
incomplete. Now, that's epistemic.
I want to argue to you that there are also aliatory
uncertainties that you can't know well.
MR. APOSTOLAKIS: Like what?
MR. BUDNITZ: Like, for example, suppose you would really
like to characterize the environment below the repository horizon, but
above the saturated zone at Yucca Mountain down to the one meter scale,
but, frankly, we can't. So there is a variability naturally in the
system which is going to cause uncertainty in your performance
assessment, and that is certainly aliatory and not epistemic.
So I think that that's incomplete, as written, unless you
acknowledge that this isn't only the state of knowledge. Some of it has
to do with variability in the natural world, which we can't characterize
always.
MR. EISENBERG: I don't want to get into a semantic
argument.
MR. APOSTOLAKIS: We understand what you're saying, though.
MR. EISENBERG: And you can --
MR. BUDNITZ: But it's a crucial conceptual point.
MR. EISENBERG: But some people would argue that all
uncertainty is --
MR. BUDNITZ: We've been there.
MR. EISENBERG: -- epistemic. It's not worth talking about.
I mean, some people would argue what you're talking about is the
inability to characterize an aliatory uncertainty.
MR. APOSTOLAKIS: But it's not worth talking about it today.
MR. EISENBERG: Some other time.
MR. BUDNITZ: Except that when you define defense-in-depth,
you need to understand that distinction, I insist.
MR. APOSTOLAKIS: So the second one then would be something
like the IPEs.
MR. EISENBERG: Then I thought I would go into a little
further detail on what these things are, what are the limitations on
knowledge. Well, you may not have included all the failure modes
because you may not know them all and you haven't had enough experience
to learn them all.
You may not have included all the phenomena for the same
reason. The range of variability in the system parameters may be
under-estimated or biased, and this happens not infrequently that people
make an estimate, take data, and their uncertainty increases.
Well, it doesn't mean that the uncertainty increases. It
means that their original estimate of uncertainty was an under-
estimate. Probabilities and consequences for rare events are based on
sparse or non-existent data. Models can't be validated. For the waste
business, we cannot wait 10,000 years to see if our predictions are
correct.
Although the systematic analyses methods can give great
insights on how a new system might perform, some problems only come to
light with experience. In other words, the state of knowledge is
evolving. I think that is the bottom line, for one type of uncertainty.
And there is a similar litany for the other kind. Why are
these risk analyses as -- and this includes performance analysis -- why
aren't they as good as they could be. Well, not all failure modes are
included because of limitations on time and resources, because the
people that try to enumerate everything didn't do it right, because not
all the phenomena were included because it would cost too much to model
everything in that detail, because in some cases, only certain kinds of
uncertainty are explicitly represented in the risk assessment.
Parameter uncertainty may or may not be propagated in the
consequence models. Some people would use point estimates. Model
uncertainty may or may not be represented. Probabilities of varies
scenarios and the uncertainty in those probabilities may or may not be
included, and not all the uncertainties that could be quantified have
been quantified.
MR. APOSTOLAKIS: Where are you going with this?
MR. EISENBERG: I'm trying to lay a groundwork that if you
just look at the results of risk assessment and compare it to a safety
goal, that there are uncertainties that you haven't considered.
MR. APOSTOLAKIS: But there is a difference between somebody
saying I will not propagate the parameter uncertainty and somebody
saying I will not do model uncertainty calculations. I will be
extremely hostile to the first guy and very sympathetic to the second,
because it's inexcusable not to propagate parameter uncertainty in
reactors, at least. In your case, it's expensive, but you have other
means to do it.
MR. EISENBERG: But suppose the model uncertainties are the
thing that dominates the result.
MR. APOSTOLAKIS: I understand that, but -- of course. Of
course, model -- but, I mean, just to say real life tells us that some
people don't do parameter uncertainty propagation, I don't know where
that leads us, because that is not something that you can tolerate these
days.
MR. GARRICK: I think the other issue here that is a little
bit troublesome in this regard is this implies that there is an
alternative and if there is an alternative, why doesn't it become a part
of the risk assessment. That's something I'm always wrestling with.
MR. GREEVES: Let me ask you to keep in mind that as Norm
goes through this, this represents our whole program. It's not in Yucca
Mountain and it's not reactors. I think that some people can't afford
to carry these things so far and appropriately so.
So Norm's presentation was trying to give you a spectrum
across the problem that NMSS has.
MR. GARRICK: We'll let him continue.
MR. GREEVES: Okay.
MR. EISENBERG: I was trying to make the point that there
appears to be a case for doing something beyond merely demonstrating
that you meet the risk goal. So before I talk some more about
defense-in-depth, I'd like to try to differentiate between
defense-in-depth and margin, which I think is an important concept, and
I will see how much controversy this raises.
If you will, margin is the cushion between the required
performance of a system and the anticipated or predicted performance.
Defense-in-depth, if you take the quasi definition from the Commission
white paper, is the characteristic of the system not to rely on any
single element of the system and to be more robust to challenges.
Margin describes the expected performance of a system versus
the safety limit. Defense-in-depth describes the ability of the system
to compensate for unanticipated performance results from limitations on
knowledge.
For example, increasing the margin in a system that relies
on a single component doesn't necessarily increase defense-in- depth.
You're still relying on a single component. Defense-in- depth provides
that if any component under-performs, the rest of the system has enough
good qualities in it that it can compensate and provide that the
consequences are not unacceptable.
In going through this briefing for different audiences, some
of the other things that have been suggested is that defense-in- depth
is like a safety net. If you're walking on a high wire and you fall,
the safety net does not assure that you get to the other side. But it
means that you may not get killed. So this can be a good quality of the
system.
The same with seat belts and air bags. Neither one of them
keep you from getting into an automobile accident, but they both may
prevent -- they put a lid on the consequences.
So if I can follow this -- you're shaking your head, George.
MR. APOSTOLAKIS: Finish, and I will tell you why.
MR. BUDNITZ: He wants you to quantify those differences.
MR. EISENBERG: This is an example where there's two systems
and we're assuming that components A, B and C, on the left-hand one, are
diverse and they don't have common cause failures, and they both meet
the same risk goal, but the one on the left has the quality that if any
one component fails to perform as expected, you could still meet the
ten-to-the-minus-four risk goal.
On the system on the right, if that one component is off,
you may have had it.
MR. APOSTOLAKIS: But this is a very misleading example,
Norm. Where are the uncertainties in these numbers? You can't present
an example like this on the basis of point estimates. I would say that
the system on the left, if it's an engineered system, will have smaller
uncertainty about the ten-to-the-minus- six.
So it may be preferable that way.
MR. KRESS: Or it may not.
MR. APOSTOLAKIS: Or it may not. It could be. If we take
the vessel --
MR. KRESS: And you might want to elect it because it --
MR. APOSTOLAKIS: So giving examples like this on the basis
of point estimates doesn't really help.
MR. EISENBERG: Well, what is it that you're shooting for,
and when you say that the uncertainties on the left may be smaller,
you're talking about the quantified uncertainties.
MR. APOSTOLAKIS: Yes.
MR. EISENBERG: And I thought I had made it clear that I was
talking about the unquantified or the residual uncertainties.
MR. APOSTOLAKIS: But even for the original uncertainties, I
would expect them to be smaller on the left.
MR. EISENBERG: Why?
MR. APOSTOLAKIS: Because for systems, components that are
at the ten-to-the-minus-two, in the ten-to-the-minus-two range, I
wouldn't expect the residual uncertainties of the unquantified to be
significant.
Now, you might say but if you put them together, there might
be something. Still, I wouldn't expect the probability of a dependency
that would defeat three components to be so significant as to overwhelm
the probability that one component that I wanted to be so reliable at
the ten-to-the-minus-six level, you know, the uncertainties are
different.
The whole issue of defense-in-depth is an issue of
uncertainty in the frequencies, not to the point values. If we don't
accept that, then defense-in-depth doesn't make any sense or it will be
a principal forever.
MR. EISENBERG: I guess I don't understand how you would
fold in to this consideration the unquantified uncertainties.
MR. APOSTOLAKIS: Because if I had to have the discussion I
mentioned this morning, focusing on the unquantified uncertainties, I
would have a bunch of experts arguing why, how can a system with three
components, a particular way it's configured, first of all, that must be
an "and" gate, not an "or" gate.
MR. EISENBERG: Yes.
MR. APOSTOLAKIS: And/or, what does it matter, right? It's
an "and" gate. They would have to focus on these -- on the failure
modes of a three-component system that would defeat all three of them at
the same time and express whatever uncertainty they have about those,
and it seems to me that is something that -- that's the value of
defense-in-depth.
By spreading it over three components, this residual risk is
smaller than on the right, where you have one. Think about all - - if
you read the documents from the agency over the last 40 years, I think
that's the running philosophy and I had about ten quotations from SECY
98-225, where the issue of confidence, uncertainty comes up every other
paragraph.
Anyway, that's my view and we can continue.
MR. EISENBERG: I think you're agreeing with me.
MR. APOSTOLAKIS: I won't do it on the basis of point
values, because my basic thesis is that defense-in-depth deals with the
uncertainties in these probabilities, frequencies.
MR. EISENBERG: One way of thinking about defense-in-depth
in the NMSS context is there appear to be two things that you want to be
concerned about. One is the hazard level and the other is the
uncertainty in the performance of the safety system. Here, again, I'm
talking about the residual uncertainty or the unquantified uncertainty.
This is not necessarily related to the behavior of the
system as modeled. It's related to the experience with the system,
whether, in fact, it ever has been built and operated or tested. So
there's a qualitative scale. This is not intended to be quantitative.
There is a qualitative scale in the Y axis that relates to the degree of
uncertainty.
There is a qualitative scale on the horizontal axis that
relates to the hazard. Small hazard, you don't need much
defense-in-depth because the consequences are not great. High hazard,
you need more defense-in-depth. So this kind of outlines three bands of
degrees of defense-in-depth and depending upon where you fall on a chart
like this or, in practice, the way we have decided to regulate these
determines how much defense-in- depth you have in each area.
But this might be a semi-quantitative, but rational approach
to deciding how much defense-in-depth is needed based on these two
qualities.
Now, there may be other qualities that are important in
making those decisions, also. This is a suggestion of how we might
approach it on, let's say, an NMSS-wide basis.
MR. APOSTOLAKIS: I like it. I like it a lot as a first
step and I think pictorially it shows -- I mean, I would translate that,
again, to uncertainty language. What you're saying is that if the
hazard is high, I really have an interest in the consequences. If it's
small, I probably don't care. If it's high, I have an interest.
And then on the vertical scale, you have put it very well.
If I have data and experience, in my language, there is no residual
uncertainty, there is no need for defense-in-depth.
So this is great. And as you move up, you hit a brick wall.
MR. KRESS: I'm wondering why you chose to stair-step this
particular thing instead of straight lines.
MR. EISENBERG: I think it's easier with the graphics
program.
MR. KRESS: Okay.
MR. APOSTOLAKIS: I must say, though, that your presentation
up to now probably has nothing to do with this.
MR. EISENBERG: We thought it did.
MR. APOSTOLAKIS: I think you could have started with this.
That's not a criticism.
MR. GREEVES: I think this kind of conveys the spectrum of
issues that challenge NMSS. It's multiple licenses and we've got we've
got to think in this context.
MR. APOSTOLAKIS: But, see, the problem I had with your
earlier viewgraphs is -- and I don't -- I suspect you didn't mean that,
but I don't think we should regulate taking into account the fact that
people don't like to do a few things, like propagating parameter
uncertainties.
On the other hand, you may have a problem on your hands with
the medical uses, all this, and where do you draw the line? I don't
know myself. When do you say, no, you have to do this? Otherwise, we
will do such and such a thing to you.
And I have seen nothing in this diagram that is based on
that. That's what I meant, that it's independent of what you presented
before.
I take the vertical axis as meaning it's an objective axis.
It says it has never been analyzed. That's a statement of fact.
Analysis are confirmed by data. That's a statement of fact. It has
nothing to do with the choices that the licensee makes.
MR. EISENBERG: This is choices for us. This is choices for
us and the preceding material, I think, made two points. One is that
it's the unquantified or the residual uncertainty that should have an
effect on how much defense-in-depth you need and, also, that what you're
really concerned with is not what the risk is. It's with the hazard
level, because the potential there is that if you're relying heavily on
a single element of your system, if you didn't do something right and
something goes wrong, you can be in trouble.
So it's the hazard and the residual uncertainty that you
really want to think about, not necessarily risk. Risk we covered
because we already said we were operating in a risk- informed
performance-based context.
MR. GARRICK: You want to be a little careful with pushing
this too far, because if you're concerned about dose, let us say, and
you have ten-to-the-ninth curies of fission products in one mode versus
another mode, the problems are grossly different.
In the case of a reactor, where you have lots of stored
energy and you have lots of mechanisms to enhance the distribution of
this material, that's very much different than having ten-to-the-ninth
curies in an unstored energy environment.
So you really have to be careful about drawing too many
conclusions about risk from these kind of diagrams.
MR. EISENBERG: I agree with you, and you also do not want
to use this as an open-ended invitation to require more and more things.
You don't want to imagine totally impossible or extremely unrealistic
eventualities.
MR. APOSTOLAKIS: I think this is a good communication tool,
that's all it is. It really conveys the idea. I don't see how you can
make this practical. You're going to tell us later, right?
MR. EISENBERG: Yucca Mountain is somewhere on the graph. I
don't think it's got as much hazard as a power reactor, but I don't
think we have as much experience with it as we do for the power
reactors. We don't have it built and tested yet.
Christiana is going to answer your question, because she is
going to tell you how --
MR. APOSTOLAKIS: You're doing a pretty good job yourself of
that. Don't be so defensive.
MR. EISENBERG: But in terms of how it's being implemented,
we're still working on it and maybe the first thing out of the box is
Yucca Mountain and we haven't gotten all the way there on that yet
either.
Remember, the comment period is closed. We're working on
developing the position. We haven't gotten it up to the Commission yet.
So what are the conclusions about defense-in-depth, some
provisional conclusions? Well, it's related to, but different from
other safety concepts like margin. It's not equivalent to meeting a
safety goal or the margin to be associated with meeting the goal. It
can be implemented in a risk-informed performance-based context as a
system requirement rather than as a set of subsystem requirements.
So that what we would suggest is that you can look at the
uncertainty, the residual uncertainty related to any particular barrier
in your system or any particular feature of your system and demand a
degree of defense-in-depth that is proportional to the uncertainty.
More uncertainty, you want more defense-in- depth. And all this is
leavened by the amount of hazard.
MR. APOSTOLAKIS: Now, that's an interesting thought. You
say you would look at each element and the residual uncertainty and do
this. How about if I take another approach? I look at each element, I
look at the residual uncertainty in each one. But then I use a
convolution there to find the residual uncertainty regarding the
performance of the whole system and then I impose defense-in-depth.
What's wrong with that? Instead of doing it at each
element.
MR. EISENBERG: Let me be clear. If you do it on an
element-by-element basis, it's all pointing at the ultimate risk goal.
It's all pointing to the performance objective.
MR. GARRICK: So your answer is you agree with it.
MR. APOSTOLAKIS: You agree with me.
MR. EISENBERG: I think we agree again.
MR. APOSTOLAKIS: Or it could be a combination of the two.
MR. KRESS: Let me sort of rephrase what I heard. I've
heard that more the residual uncertainty, and George has qualified
residual to mean unquantified, the more the defense-in- depth you need
and then George says you use defense-in-depth where you have
unquantified uncertainties, so you don't know what the meaning of the
word more is, and I keep saying you do have to quantify it.
I'm a little confused. What are we talking about here?
MR. APOSTOLAKIS: Unquantified in the sense that I hadn't
put down a probability distribution. But there is something, in my
mind, I mean --
MR. KRESS: You mean, it's big or medium or small?
MR. APOSTOLAKIS: Yes. I could say --
MR. KRESS: Isn't that quantified? See, I'm saying you can
quantify it to some extent.
MR. APOSTOLAKIS: To some extent, I agree. Yes. You're
right.
MR. GARRICK: And I agree with you, Tom. It's a very
abstract concept. In fact, I still struggle with what we mean by
unquantified or residual uncertainty and if we can handle it by some
other means, why can't we fold it into the basic parameters.
MR. APOSTOLAKIS: We could. We could. We could.
MR. BUDNITZ: I don't understand why, George, it's the
unquantified uncertainty and only that that you're emphasizing. I can
conjure up a system where it's a quantified, but large aliatory
uncertainty and you invoke defense-in-depth to find a way to do it
anyway that's safe enough.
MR. APOSTOLAKIS: I would say, in that case, I would use the
uncertainty diversity and so on to manage that uncertainty.
MR. BUDNITZ: In other words, aliatory is something that's
random in nature.
MR. APOSTOLAKIS: That's fine.
MR. BUDNITZ: But large, but we don't know how to control
it. So we find another way using defense-in-depth. But in that sense
--
MR. APOSTOLAKIS: But it's not defense-in-depth anymore in
the sense that it's not arbitrary. If I postulate a barrier, I can
calculate it.
MR. BUDNITZ: Defense-in-depth isn't arbitrary here. He
said defense-in-depth involves -- we're now going back to the white
paper -- it involves assuring that there's -- you're not relying only on
one barrier.
MR. APOSTOLAKIS: But that's arbitrary.
MR. BUDNITZ: Well, wait. Whatever you say, however they
defined it, I insist that I think it is not only the unquantified
uncertainty, by any means, especially in some of their systems, where
they may have a very large -- by the way, aliatory, maybe they have 800
licensees and they're all different in the arena of some little thing
and in order to have one rule for them, they may have to do it another
way, with the defense-in-depth idea, but maybe two barriers or
something, rather than -- so that might be a variability in nature,
because all the hospitals are different or something.
MR. APOSTOLAKIS: Let me tell you --
MR. BUDNITZ: It's more than unquantified uncertainty, is my
point.
MR. EISENBERG: But remember, this is predicated on meeting
already the risk-informed performance-based goals.
MR. BUDNITZ: I understand that.
MR. EISENBERG: Your aliatory uncertainties, if you have
included them, have already been taken care of. You've already arrived
at a satisfactory performance of the system.
MR. BUDNITZ: I understand.
MR. APOSTOLAKIS: I want to give an example, John Garrick,
what is an unquantified uncertainty. If there is a fire in a nuclear
plant, we have now a methodology that calculates, to some extent anyway,
but it calculates the probabilities of failure of cables and so on due
to overheating.
We know that the fire creates smoke and we know smoke is
hazardous. Yet, right now, we are not quantifying -- this is not part
of my risk assessment. So I can say now, okay, that's not part of your
risk assessment, defense-in-depth, help. So I want you to have barriers
between compartments so that smoke doesn't propagate, I want you to have
smoke detectors, I want the people to have masks and oxygen and this and
that.
So I'm giving you a set of measures and you say, fine, I'll
implement them. This is a traditional way of regulating defense-
in-depth. Then tomorrow somebody does a calculation and he includes
smoke into this, into the fire risk assessment. Now I can see what the
impact on the frequencies of failure, for example, of core damage or
whatever is of having those barriers or having the oxygen masks and so
on, and I may very well decide that some of them are not needed.
So that's what I mean by unquantified, that you invoke then
the principle of traditional engineering and you say then put a few
barriers there that make sense.
In this particular case, I happen to believe that given
sufficient time and will, we can include it in the fire risk assessment.
It's not something -- it's not like safety culture, which is much more
difficult.
So that's what I mean by -- and then we will just have to do
-- and from the engineering perspective, does this make sense? Yes. To
contain the smoke and make sure that people are not hurt and so on, the
firefighters and so on. So you are invoking a series of measures to
manage this risk, which you have not quantified at this time, and it may
very well turn out in the future that some of these measures were not
the best or were not necessary, they contributed very little, after you
quantified it. It's very good.
MR. EISENBERG: I think we have two problems in our arena.
We have a diverse set of things we regulate. So for each arena, we have
to decide how much defense-in-depth should we have for this particular
set of licensees, how much should we have for the radiographers, how
much should we have for medical licensees.
Then once we decide that, within each system, we have to
decide how do we put in defense-in-depth appropriately to counter the
residual uncertainty. So it's a two-step question.
MR. APOSTOLAKIS: I agree.
MR. EISENBERG: So we think that defense-in-depth can be
used to address these residual uncertainties and we also think that it
should depend on the degree of residual uncertainty and the degree of
hazard.
But it's not easy. Regulatory life is not easy. So given
this, we still have to decide how to measure the degree of
defense-in-depth, how to measure the degree of uncertainty in the
performance of the safety system, encompassing both quantified and
unquantified uncertainty; how do we measure the potential hazard posed
by a system.
Some of these we've already discussed. How to implement
defense-in-depth when there is different uncertainties in different
parts of the system; how do you use the current state of knowledge to
make reasonable tests for the system to have an appropriate degree of
defense-in-depth when what you're trying to accommodate is imperfect
knowledge.
And then the real killer, how do you explain this to
stakeholders so that we can preserve the flexibility that's inherent in
a risk-informed performance-based approach to defense-in-depth, but also
provide for reasonable assurance of safety. This is not easy.
MR. KRESS: I think this is a good list of issues.
MR. EISENBERG: So in summary, we intend to consider
defense-in-depth in the context of risk-informed performance- based
regulation and a lot of ongoing activities and as part of the continuing
evolution of the risk-informed framework in NMSS.
As a general safety principle, the degree of defense-in-
depth needed to assure safety depends on several factors, including the
degree of residual uncertainty and the degree of hazard. We would like
to implement defense-in-depth as a system requirement, where feasible,
rather than by prescriptive subsystem requirements, and please remember,
NMSS needs flexibility in any overall approach to implementing
defense-in- depth to permit us to appropriately regulate the wide range
of systems and licensees that we have.
MR. APOSTOLAKIS: I think this is very good, Norm. You did
a good job.
MR. EISENBERG: Thank you.
MR. APOSTOLAKIS: Even if I sounded critical. The only
thing that bothers me a little bit is this degree of hazard. I'm sure
there is another way of putting it, but for this stage of development, I
guess it's okay.
I think it has probably to do with the goals, the risk
goals, that the degree of hazards affects the goals, the acceptance
criteria, and then that affects the residual uncertainty. So it's
really only one of the hollow bullets there that come at us.
MR. EISENBERG: I'm not sure I agree.
MR. APOSTOLAKIS: The degree of hazard, how you manage it is
a policy issue and the Commission says I have the quantitative health
objectives. Then trying to quantify now your actual system to compare
with your objectives, you end up with a residual uncertainty which is
driven by the Commission's health objectives.
If the Commission had told me that ten-to-the-minus-two is
the individual risk I will tolerate from nuclear reactors, I will need
to worry about residual uncertainty in nuclear power plants. Right?
The goal is so high that it's irrelevant.
So I think the goal itself is really the driver that
determines the residual uncertainty. But that's a technicality.
MR. EISENBERG: You're tending to look at uncertainties
strictly in terms of uncertainty in frequencies of events of failures.
MR. APOSTOLAKIS: Uncertainty about the occurrence of
something.
MR. EISENBERG: I think that's what I said. But there are a
lot of other ways that the uncertainty can come in.
MR. GARRICK: My concern with the statement, the bullet on
degree of hazard, is a little different. I think that I worry about the
non-linearity between hazard and risk. I wouldn't bank too much on the
degree of hazard being a particularly important factor on this.
MR. APOSTOLAKIS: I think there will be other things driven
by the degree of hazard that will have more direct impact.
MR. KRESS: I would like to see a statement of what is meant
by degree of hazard. I would have interpreted it to mean that if I
didn't have any of the protective systems around this piece of scrap,
whatever it is, the reactor or what, then what is the probability of
producing certain consequences.
If we just laid the fission products in the hole up there,
why, you can come up with it, or if you didn't have any protective
systems around a reactor, you would conclude that the degree of hazard
of the reactor is much, much greater than one of a repository.
I think you can quantify the degree of hazard, if you just
ask yourself what it means. And it would incorporate your comment about
driving forces and mobility and where it can go and that sort of thing.
MR. EISENBERG: One of the problems of just considering the
risk is that the risk is predicated upon things behaving as they have
been modeled, and one of the things you want to get to with
defense-in-depth is what if, what if they do not behave that way.
MR. GARRICK: Of course, you can even take into account that
by the way in which you assign uncertainty to your model parameters.
There is nothing that prevents you from even accounting for residual
risk at the parameter level or at the barrier level by how you assign
your uncertainty, as long as you've got a case for it, as long as you've
got a story behind it. And I would agree with George. That was a good
presentation.
MR. GREEVES: And I think we'll keep Norm up here.
Christiana, at this point, as I introduced, the challenge that we have
is thinking across all of the NMSS activities and Christiana Lui will
give you some insight of our current thinking in the Yucca Mountain
context.
So Norm will stick around, because I'm sure it's going to
cause some additional discussion. Christiana?
MS. LUI: As Norm is getting his act together. Thank you.
Good afternoon. My name is Christiana Lui and I work in the Division of
Waste Management in the High Level Waste Branch, and we heard a lot --
we heard a lot of interesting discussion this morning and hopefully in
my presentation I will be able to help answer some of the questions and
make some clarifications to some of the issues that have been raised
regarding the high level waste program this morning.
I just want to provide the context of where we are. The
extended public comment period on the proposed Part 63 ended on June 30,
1999. Staff is in the process of analyzing the public comments and
preparing responses to those public comments.
The current schedule is to have the final Part 63 go to
Commission by the end of March this year.
Again, to emphasize that this is still work in progress. So
the objective today is to share our best current thinking with the
committee, and the focus is going to be on the post-closure safety
evaluation, how the multiple barriers requirement is being addressed in
the post-closure safety evaluation.
For pre-closure, the defense-in-depth follows the approach
of prevention, mitigation, and if you want to put emergency planning, a
separate category, but basically it's the same concept as the operating
facilities that you are most -- you are definitely will hear from our
colleagues from NRR and Research in the next two presentations.
I'm going to go from pretty much the very top level and
provide more detail as the progression of the presentation. So we want
to clarify what is the intent of multiple barriers first.
Just a side note that we received approximately 20 sets of
public comments on the issue of multiple barriers during the public
comment period, including Dr. Budnitz's comment asking us to clarify
what we mean by the multiple barrier requirement in Part 63, and we
appreciate your comment.
As both John and Norm have mentioned, the intent of the
multiple barriers is we are going to -- we are using the Commission's
white paper on the risk-informed and performance- based regulation as
the guidance for our approach to clarify multiple barriers requirement.
We also are going to measure at this point. We are
targeting the multiple barrier requirement as an assurance requirement,
and I will say about -- I will provide you more detail on this a little
bit later.
The known certainties are all captured, appropriately
captured in the performance assessment to demonstrate compliance to an
individual protection standard.
MR. APOSTOLAKIS: Are the model uncertainties also
appropriately captured?
MS. LUI: Yes. I'm going to talk about that. I'm going to
give you a little bit more detail on that. So just be patient, bear
with me. Thank you.
MR. APOSTOLAKIS: You're asking for the impossible, be
patient.
MR. GARRICK: I'll help you, Christiana.
MS. LUI: Okay. And --
MR. APOSTOLAKIS: But wait a minute.
MS. LUI: And the repository system is sufficiently robust
to account for -- maybe imperfect is not the best word here. Maybe
incomplete is a more appropriate word here, the incomplete knowledge.
MR. APOSTOLAKIS: This is the second time that we hear this
today. The first one was from Dr. Budnitz. So it is the community's
view that even without imperfect knowledge and the uncertainties and so
on, we are meeting the goals of the Commission, that Yucca Mountain
meets the goals?
MR. BUDNITZ: We don't know.
MR. APOSTOLAKIS: So what does it mean then, that it's
sufficiently robust or accounts for imperfect knowledge? To do what?
This morning you were more explicit. You said, Bob, that even if I
include those uncertainties, I know that this thing is --
MR. BUDNITZ: I expressed an opinion, but of course, we
don't know, because we don't have a final design or analysis of it. I
was of the opinion that I think it's likely that when the final decision
is put in place and it's analyzed, I think and hope that it will meet
the dose limits in Amergosa with a lot of margin.
MR. GREEVES: In spite of imperfect knowledge.
MR. BUDNITZ: No, not in spite of, taking into account. Not
just in spite of. Taking into account. So that's a prediction, because
I don't know, the final design may have some more difficult analysis
problems than the things I've seen.
So this is still an evolving sort of judgment and I don't
want to preempt even my own final judgment there, but I was just sort of
expressing and I was stipulating that if that's true, then what.
MR. LEVINSON: Well, the slide identifies this as the
intent. It doesn't say they have achieved it.
MR. BUDNITZ: Yes, of course. That's there, yes.
MS. LUI: There will be a lot of discussion. Next slide.
Now I'm going to be a little bit more specific on what are the
considerations of the multiple barriers requirement in Part 63.
I'm going to take you step-by-step here. The reason why the
fourth bullet is in yellow is because that's one particular item not
included in the proposed Part 63, but is being -- but is under
consideration. That as part of the clarifying language for Part 63, we
are intending to add that part to the regulation.
The first thing is to assess all significant and negative
impacts on safety in a compliance demonstration calculation. This
morning -- or what I really mean by that, this morning we have heard
quite a bit about TSPA or that particular terminology being used.
Basically, what we asked DOE to do is in the total system
performance calculation, that they carefully consider all the data
obtained from site characterization program, consider all the applicable
natural analog experimental and field testing information and justify
the models for the total system performance assessment.
In that, they also have to quantify and incorporate the
uncertainty for all the input parameters that go into a calculation.
DOE also needs to take into consideration the alternative conceptual
models that are -- that basically fits all the information that we have
up-to-date, provide that particular description, and provide a
description of what conceptual models they have considered and what they
have chosen to include in the total system performance assessment.
They also have to provide support that a model output is
trustworthy.
MR. APOSTOLAKIS: Again, let me play devil's advocate here.
Suppose you hadn't told them that. Don't you think they would have done
all this? This is nothing special about what you are doing. I think
they would have identified the barriers, they would have described and
quantified the capabilities, they would have provided a technical basis.
There is nothing new here.
MS. LUI: But these are the requirements that are under
consideration in Part 63.
MR. APOSTOLAKIS: You mean under consideration that you may
decide not to demand some of this?
MS. LUI: No, because as what John has stated up front, that
we are still in the stage of preparing the final rule package to the
Commission.
MR. GREEVES: The staff is being a little careful here.
Recognize, we've got a proposed rule on the street. The period of
comment is closed. We're going through a deliberative process, which is
what is in the regulation. I wouldn't make any more than that of it.
MR. APOSTOLAKIS: But there is nothing special to Yucca
Mountain here. I mean, you would do that for any system.
MR. GREEVES: I don't think there is a trick question.
MR. APOSTOLAKIS: Now, this business of wholly dependent.
What does that mean? I can build a --
MR. GARRICK: I hope it doesn't mean that you would
discourage them from providing you a design where a single barrier could
do the job.
MR. APOSTOLAKIS: I think that's what it means.
MR. GREEVES: No, it doesn't mean that.
MS. LUI: No, it's not that.
MR. APOSTOLAKIS: What does it mean?
MR. GARRICK: That would be terrible.
MR. BERNERO: John, there is a statute that says you have to
have multiple barriers. That colored, the fourth bullet could be
interpreted as a way to verify that, but I would think it would be
worded something like unduly dependent, rather than wholly dependent.
MS. LUI: The reason these words are here, they are taken
directly out from the Commission's white paper. We may -- in terms of
the exact language in the rule, that's still being crafted.
MR. BERNERO: But, Christiana, there has to be a finding
somewhere down the road that the statute is satisfied. DOE has to make
that finding in their submittal, and I agree with George, all of these
things are appropriate to a reasonable total system performance
assessment, except that fourth one. That's a ringer in it, because
that's the implementation of multiple barriers, and, by inference, the
implication of defense-in-depth.
MS. LUI: Right.
MR. BERNERO: The statute requires multiple barriers.
MS. LUI: Right.
MR. BERNERO: I would argue that defense-in-depth is a
strategy, not a statutory requirement, and it says don't unduly depend
on one barrier.
But if you could have a state of knowledge and a state of
certainty that could support one barrier doing the job, then you would
have a statutory conflict but not a logical conflict.
MR. BUDNITZ: In fact, let me postulate something that isn't
true. Suppose --
MR. BERNERO: Are you going to tell us a lie?
MR. BUDNITZ: No, no.
[Laughter.]
MR. BUDNITZ: It is a "suppose" -- suppose DOE came with a
canister design that they had extremely high confidence in they could
back up and everybody agreed the last 20,000 years, all of them, for the
first cracks, just as, by the way, if they asserted that for one year we
would agree, so then I am just supposing.
Now let's suppose they also had a site in which anything
that leaked the travel time was 50,000 years and they had a 10,000 year
requirement. You're home free -- either is wholly dependent, but it's
not because either one can actually be -- you could have them use a
paper bag and still be there and you didn't have to have the earth,
you'd still be there -- and we want to encourage that. Nobody wants to
discourage them from doing as best they can.
MS. LUI: Right.
MR. BUDNITZ: But --
MR. APOSTOLAKIS: So it is a model of language.
MR. BUDNITZ: No, no, but then if that is the case, let me
stick to it -- just pretend -- suppose that was the case. Would the NRC
ask them to do more? I my prepared remarks this morning I asked that
question.
In other words, if you are there --
MR. APOSTOLAKIS: I think the question would be, Bob,
whether you are there. The NRC will ask them -- I mean if you
demonstrate you are there, I don't think the NRC would ask them to do
any more.
MR. BUDNITZ: No, no, no, no, no. Wait -- no, no, no. I
want to insist. I ask another question. Let's suppose that the total
system performance assessment they do next year, two years from now, for
the design they are putting together now shows the doses are met by
three orders of magnitude. I insist that as best I can tell the
Department could still flunk on defense-in-depth. It was all one item.
MR. APOSTOLAKIS: I don't know what all one means.
MR. BUDNITZ: Let me describe.
MR. APOSTOLAKIS: I think the paper background, a second
one?
MR. BERNERO: Now let me give you an example. If the
repository was chosen to be in a site that's subject to significantly --
subject to erosion such that the deposited waste could be exposed in the
long range and you did have a gorgeous package, you know, boy, this
package is marvelous, best can in the world, but it could flunk the test
because the erosion would shift you to be wholly dependent on the one as
against unduly dependent on it.
You know, the erosion might be very far-fetched.
MR. APOSTOLAKIS: I understand that.
MR. BERNERO: But your dependence is upon the package.
MR. GARRICK: Well, you have cited a weakness in the
defense-in-depth concept.
MR. BERNERO: I still argue there is a difference between
defense-in-depth as a strategy or safety philosophy and what the statute
requires the high level waste repository to have, multiple barriers.
MR. APOSTOLAKIS: No, but the point, I agree with John again
that you can't do these things by counting barriers.
MR. BERNERO: Of course.
MR. APOSTOLAKIS: You can't for the same reason that you
can't rank minimal cut sets in a fault tree by counting the number of
events. The probability of failure must play a role. We are not going
to go back 20 years now and I think, you know, I can restate what you
just said, Bob, in terms of uncertainty and probability and then I will
conclude that it relies unduly on one barrier. I can do that.
MR. BUDNITZ: I agree.
MR. APOSTOLAKIS: It all comes down to the probabilities of
failure of pathways and so on, so by saying, you know, multiple barriers
and count them and so on, this is a first step.
MS. LUI: I don't think we are suggesting counting the
barriers.
MR. APOSTOLAKIS: We were not criticizing you. We are
talking to each other. When we talk to each other --
MS. LUI: Okay.
MR. APOSTOLAKIS: It's best to change viewgraphs.
MS. LUI: Should we go on to the next slide?
MR. APOSTOLAKIS: Yes.
MS. LUI: Okay. On multiple barriers, some of the concepts
we tried to express on these particular slides has actually come out
during the discussion you just had. What I want to make sure is that
because of the uncertainty in the barriers' capabilities based on
current state of knowledge, there are uncertainties in the barriers'
capabilities over 10,000 years and as the regulator why we want to know
is what if all of these barriers do not perform as well as what we
currently know.
We want to make sure if that kind of situation happens the
public health and safety is still protected, so what we are going to be
aiming at is that the demonstration of multiple barriers is going to
show that the balance of the system has the ability to compensate for
that kind of "what if" situation.
MR. APOSTOLAKIS: Now the "what if" -- are you going to put
any probabilities on the "what if"?
MS. LUI: We do not plan to do that at this point because,
remember, the TSPA is as good knowledge as possible based on the current
state of knowledge. What we are doing here --
MR. APOSTOLAKIS: Sensitivity studies.
MS. LUI: Yes.
MR. APOSTOLAKIS: That is really what you are doing.
MS. LUI: Or it is similar to a stylized calculation like
human intrusion. You really cannot quantify the probability. If you
can, then it should be really part of your TSPA.
MR. APOSTOLAKIS: I would do it in a different way. I would
start with "what if" and let's say that in "what if" Number 5 I do not
protect public health and safety to my satisfaction. Before I do
anything else, I would ask myself whether "what if" Number 5 has a
probability that would really upset all the calculations and the
confidence that I have.
In other words, I would not rely on a "what if" analysis
without addressing the issue of how likely that is.
MR. EISENBERG: But if you are trying to look at your
imperfect state of knowledge, you are speculating about what you don't
know.
MR. APOSTOLAKIS: I am not speculating because --
MR. EISENBERG: Then how do you know --
MR. APOSTOLAKIS: Wait a minute, wait a minute. At some
point you draw the line. I mean there must be some sort of an upper
bound that you can put. I mean it comes down to Tom's point and John's
that you can always give a number or do something, you know? The
problem with "what if" calculations is the same one as defense-in-depth.
There is no control over it.
This committee 20 years ago, 25 years ago, the moment the
Reactor Safety Study hit the streets several members for years took
extra pleasure by taking a few parameters, multiplying by 10 and saying
my god, look what happens to the result, and everybody said yeah, look
at what happens to the result.
The question is can you multiply it by 10? Is that real?
And I think you are going that way. You can start playing games here
that have no bound.
MR. EISENBERG: The key thing here is that the
underperformance would be related to the degree of uncertainty in that
particular barrier, so if you have a very good case, if you have lots of
evidence, then you would underperform it very little. If you don't have
a whole lot of data, if you have a 20,000 year waste package and you
have two months of data, well, maybe we would want to see it
underperformed more, but it is not unbounded speculation and it is not
intended to be unbounded speculation.
MR. BUDNITZ: I have peeked ahead but --
[Laughter.]
MR. BUDNITZ: -- but it is a fair comment to say that
although I wasn't in Las Vegas in November I read the transcript and
your thinking here is the same as there and that's great because, you
know, it's only been a couple months and I understand what you are
doing.
I am still troubled by two things. Unless I peeked ahead
and didn't get it right, you are still asking the Department, the
Applicant, to select the amount of underperformance that they will
analyze, and I think that is not necessarily right.
MR. GREEVES: Well, why don't we move to the next one.
MR. BUDNITZ: Maybe we can go to that.
MR. GREEVES: I am not sure you read that slide right.
MR. BUDNITZ: Maybe I didn't get that one right, but the
second point is on this slide. Go back to this slide. It has to do
with the word "compensate."
The word "compensate," my plain English reading of that
convinces me it is the wrong word. You can't expect that if you
underperform a certain barrier that you would necessarily still meet the
dose limit at Amargosa Valley or maybe you do mean that. It's very
important to understand that.
MS. LUI: Right.
MR. APOSTOLAKIS: What did you say?
MS. LUI: If you look at it carefully, it's not fully
compensated. We are talking about compensate.
MR. BUDNITZ: So let me try to say this. Suppose the dose
limit at Yucca Mountain is "x" millirem per year and the base case
calculation shows one-hundredth of "x" and then they undercompensate
Barrier Number 2, underperform, excuse me, underperform Barrier Number
2, and instead of being .01 of "x," whatever the limit is, it's now 5x.
Do they get a license or don't they?
Now that depends on something that they haven't told us yet.
It's really a crucial point.
MR. APOSTOLAKIS: What is it that you haven't been told?
MR. BUDNITZ: They haven't told us whether or not they are
going to get a license or not.
DR. KRESS: And is that acceptable. You haven't defined an
acceptable performance --
MR. APOSTOLAKIS: Isn't the obvious thing to do to ask
yourself how likely this postulate we made was?
MR. BUDNITZ: That is a piece of it, of course.
MR. APOSTOLAKIS: That is the most important piece.
MR. BUDNITZ: I am not arguing the case, but you see, if in
fact something becomes 5x instead of .01x but "x" is the limit, right?
-- we may all judge that that is sufficiently unlikely that we will give
them the license, right? But they haven't told us, the public, and here
I am a member of the public because I am not under contract to anybody
right now, or certainly they haven't told the Applicant yet, unless I've
peeked ahead and haven't seen it, whether -- what the decision criterion
is and in my remarks I said it has to be fair and it has to be
technically sound and it's very, very important that that be clarified.
MR. APOSTOLAKIS: The weak calculations set a bad precedent
there. Look at the spaghetti curves.
MR. BUDNITZ: Well, we are not arguing the case.
MR. APOSTOLAKIS: All of them are below.
MR. BUDNITZ: You see what I'm saying? So keep going.
MR. GREEVES: I understand what you are saying and you are
not going to be satisfied.
MR. BUDNITZ: I know I am not going to be satisfied and I
want to say that if I was designing the repository and some of the guys
behind me are, and if I was trying to put it together now so that I
could analyze it next year, so I could bring you the thing in the year
after next and I didn't even know whether the design I am contemplating
freezing for this will do this, that is a real problem, that's a real
problem.
MR. GARRICK: I think that the more realistic issue here, it
seems to me, and I am reminded of an earlier working group where one of
our consultants said it's the water, stupid, the more realistic thing
that is likely to happen here is that the initial conditions that are
the basis for the TSPA may not be appropriately represented.
MR. BUDNITZ: That's a fair comment.
MR. GARRICK: Because the thing that really distinguishes
this from the reactor case is the fact that the peak dose may not occur
for 300,000 - 400,000 years.
MR. BUDNITZ: Well, they have a 10,000 year requirement.
MR. GARRICK: I don't care. I don't care. I'm a risk
analyst. I am not a regulator, and so the thing that drives that --
there is almost as much of a singularity in the waste disposal problem
as core damage is in the reactor problem in terms of the release, and so
I think that what is really where we are going to find the most
opportunity for having miscalled this thing is not so much with the
design of the barrier but with the initial conditions that are the basis
for the performance assessment in the first place.
MR. BUDNITZ: You could be right.
MS. LUI: Okay. Next slide. There are two technical issues
that we are wrestling with in terms of the multiple barriers analysis.
Basically we mentioned about underperformance of a barrier. What we can
do is we can prescribe what should be the degree of underperformance or
we can take a more performance-based approach. Let DOE look at the
amount of evidence that they have in terms of supporting the barriers'
capability they claim in the TSPA analysis and then they can make a
judgment of what should be the appropriate degree of underperformance
for that particular barrier in the barrier underperformance analysis.
Another issue we are looking at is how should NRC evaluate
the outcome of the underperformance analysis?
MR. APOSTOLAKIS: Which is what I have been saying. You
haven't said anything about the assumptions that the analysis makes. Is
that buried somewhere here?
I don't understand.
MS. LUI: The assumptions for the barriers underperformance
analysis?
MR. APOSTOLAKIS: Yes, for transport of radionuclides.
MS. LUI: It is all part of the total system performance
assessment.
MR. APOSTOLAKIS: I understand that.
MS. LUI: Right.
MR. APOSTOLAKIS: But where in this scheme of things do you
worry about the assumptions being wrong?
MR. GARRICK: That's what I mean by the initial conditions.
MR. APOSTOLAKIS: I know, but I don't see where it is.
MR. GREEVES: I think Dr. Garrick would say that that is
included in the original performance assessment. When you step off and
start doing these under performance evaluations, I think you would have
to talk about understanding what those assumptions were and try to
justify why you made those.
MR. APOSTOLAKIS: Right.
MR. GREEVES: The DOE could make a statement this is my
assumption, we think it's reasonable. The Staff could look at it and
say looks good but we have a little wider band. I think that is part of
what we are about.
MR. APOSTOLAKIS: But that brings me back to my earlier
question where I was told that I was impatient. How do you handle model
uncertainty then in the base case? You say known uncertainties are
appropriately captured. What does that mean?
MS. LUI: If part of the consideration of the alternative
conceptual models --
MR. APOSTOLAKIS: But do we know how to do that? Do we
understand the conceptual framework? Do we know how to do that?
MS. LUI: Okay. There are a couple -- there is a stepwise
process. Basically DOE will have to identify what are the alternative,
what are the conceptual models, what are the different conceptual models
that are consistent with all the information that we have up to date and
that they have to make a justification why they have included certain
ones and they have excluded certain ones from their consideration in the
total system performance assessment.
MR. APOSTOLAKIS: What if they take all 11 of them and give
them different weights?
MR. EISENBERG: They can do that, but we would also want to
see that information disaggregated and we would look to see to some
degree what the bounding one would be and we would probably want them to
show compliance with that one.
MR. APOSTOLAKIS: Which each one?
MR. EISENBERG: Yes.
MR. APOSTOLAKIS: With each of the 11?
MR. EISENBERG: No, with whatever the bounding one was.
DR. KRESS: That is each of them.
MR. APOSTOLAKIS: That is each of them, yes, if the bounding
one does it -- it's each of them.
Is that something that people have really thought about?
DR. KRESS: It is not clear to me where you are using
probabilities in this process at all.
MR. APOSTOLAKIS: They are not.
DR. KRESS: That seems to be the shortcoming in this whole
thing.
MR. APOSTOLAKIS: That's right.
MS. LUI: Probabilities fall into a total system performance
assessment.
DR. KRESS: It is part of the performance assessment, I
understand.
MS. LUI: Right.
MR. APOSTOLAKIS: Yes, but --
MS. LUI: There are disruptive scenarios that have the
equivalent of initiating events probability and then you have expected
evolution of the repository behavior.
MR. APOSTOLAKIS: We just agreed that maybe in one piece of
this evolution there are questions about the medium, for example, okay,
and we have transport through fissures, fissures or something else, and
I think I heard Dr. Eisenberg say that if there are questions like that
and you have 11 different ways you can go, you better meet the
regulations with each one of them.
I am asking whether this committee has discussed this issue,
because that sounds to me like a license to kill.
MR. GREEVES: I think that there has to be a qualification
on 11. It has to be something that is reasonable. You can come up with
something that is non-physical and that one should be discarded.
MR. APOSTOLAKIS: Well, physical I understand, but how about
likely?
MR. BERNERO: You know, I am sorry to hear Norm use the word
"compliance." The total system performance assessment which is supposed
to take due account of uncertainties is being used as a compliance tool,
is the result of it consistent with the objective, the safety isolation
objective as stated?
These are sensitivity analyses and these sensitivity
analyses, somewhat arbitrarily chosen, somewhat arbitrarily applied,
should explore how close to the edge of the cliff of unacceptability
they are or their results would be, and it is not compliance --
MR. EISENBERG: For a particular barrier --
MR. BERNERO: I mean it is license to kill if you say now
change that assumption to the worst case and show me you still comply.
You just made that your compliance case.
MR. EISENBERG: No, I think we are talking about two
different things. I think what George was talking about was how do we
consider conceptual model uncertainty in the performance assessment as a
whole, not how do we do these defense-in-depth calculations.
MR. APOSTOLAKIS: They are related though, Norman. They are
related, very much related.
MR. EISENBERG: I thought how the question was phrased I
thought the predicate for it was that you had 11 different conceptual
models and you had no information to be able to distinguish --
MR. APOSTOLAKIS: Yes.
MR. EISENBERG: -- between one and the other.
MR. APOSTOLAKIS: Well, I didn't say, the second part I
didn't say.
MR. EISENBERG: Well, then do you have a preferred model and
do you have evidence to support the preferred model?
MR. APOSTOLAKIS: I don't know. Maybe there are two or
three possibilities. I don't know. We may do what NUREG 1150 did,
collect a bunch of experts and try to assign weights. I don't know but
I would really question the wisdom of saying that I will do it for each
model and see what --
MR. EISENBERG: But that -- my answer was predicated on the
basis that there was nothing to distinguish between --
MR. APOSTOLAKIS: Okay.
MR. EISENBERG: -- between the different conceptual models.
Now you are telling me you have more information. Well, if you have
more information, you should use it.
MR. APOSTOLAKIS: But is it being used now?
MR. EISENBERG: Yes.
MR. APOSTOLAKIS: Yes?
MR. GREEVES: Both the Staff and DOE have done these
calculations and we have briefed the committee on them.
MR. BUDNITZ: But I am still stuck with, sorry, with my
question.
Let's suppose that we have a barrier and we have enough of a
quantification of our state of knowledge of its performance so that we
can say its performance is in a certain range -- just to be numerical
about it, without knowing quite what it means, it is between 4 and 400,
this is a completely arbitrary discussion, and 400 is worse than 4,
right, and let's suppose we knew nothing more than that. It was a
complete maximum entropy. We said we knew damn well it couldn't be
lower than 4 or greater than 400.
You would be saying, gee, you better assume 400 and show us
it works. I am not disagreeing with that, but if you have a state of
knowledge that says, well, I am sure that it is between 4 and 400, but I
actually have knowledge that tells me that there is a curve and
distribution and the probability it's at either end is really quite
small although it is possible, and we know it is bounded. It can't be
more than 400. Then it is not right -- by the way, if you use 400 and
you still pass, great. You do that every day of the week in every
analysis we know. That is the best way to show it, but it is not right
to insist that when, and I know you understand this, but now we come to
this question about underperformance and compensation.
Are you going to ask for that barrier -- now this is just
very conceptual -- that DOE decide which underperformance number to pick
and then they are going to come and bring you the rock, and the thing I
said, "Wrong rock" or are you going to tell them in advance what your
decision criterion can be so that they can spend more money on a better
design or spend more money on more analysis or something so that they
know going in what they can expect from you, because I think unless they
know that, this process is unsatisfactory for me as a citizen, and I
hope it ought to be unsatisfactory for the Commissioners as the
statutory authority because the Department needs to know the rules and
the speed limit before they submit the application.
MR. EISENBERG: The Department doesn't have its design
finalized yet and it doesn't have its safety strategy finalized yet, so
it can't tell us how much reliance it is placing on different components
of the system.
MR. BUDNITZ: I understand what you are saying.
MR. EISENBERG: I am too. We are understanding each other.
MR. BUDNITZ: It's iterative but those guys have to do --
they are the Applicant.
MR. GREEVES: And those guys did a viability assessment.
MR. BUDNITZ: Yes, I know it.
MR. GREEVES: So they are not without ability.
MR. BUDNITZ: We all know that. We all know that.
MR. BERNERO: But I have got to quarrel with you, Bob, on
the regulator can't take the burden of sharp prescription of what does
it take to prove safety. You can't do that. It is, like it or not, it
is a show me the rock. DOE has the primary responsibility and there has
to be some kind of guidance on what size rocks and what texture.
MR. BUDNITZ: The boundaries.
MR. BERNERO: But at the same time you can't get away from
the fact that DOE has far more capability and far more responsibility to
develop these arguments to show that there is not undue reliance --
MR. BUDNITZ: Bob, I agree with you absolutely, completely
about whose responsibility is where. What I was worried about was that
the amount of underperformance the Department will assume may be way
short of what you would have done and then they have got their design
they have frozen. They are in the licensing process and they could have
fixed it earlier.
MR. GARRICK: Bob, I suspect that if you calculated the
matrix I showed you this morning, the more detailed one, the answer
would be obvious.
MR. BUDNITZ: You may be right.
MR. GARRICK: Yes. If you have the performance of the
individual barriers with and without in context, that to me would be the
strongest piece of evidence you could possibly have for me to make a
judgment about the performance and I know you said in your talk that you
can't remove the barrier --
MR. BUDNITZ: Completely, of course.
MR. GARRICK: -- completely, but you can do variations on it
and, as a matter of fact, as you decompose it into more and more
detailed barriers you can increasingly remove it more easily.
MR. BUDNITZ: That's fair.
MR. GARRICK: And with increasing accuracy.
MR. BUDNITZ: Just as your microscope goes --
MR. LEVENSON: John, as I have been listening to this, I'm
thinking what would bother me about it if I were trying to conform and
this word "compensate" is a very loose end, that it would change
completely what needed to be done if you said adequately compensate as
opposed to totally compensate, and without a modifier there is an
implication of total.
I would give an example. In your base case maybe the dose
to the public is -- I will use Bob's one one-hundredth of what is
allowable, but you fail one barrier and now you are only one-tenth of
what is allowable. Clearly you are way under what is allowable but you
haven't fully compensated and so I think the choice of the word
"compensate" without a modifier is likely to cause all kinds of
problems.
MS. LUI: Yes, we agree with you basically. That is why
these are two key technical issues that the Staff is struggling with, to
make sure that the rule and the guidance is going to follow and be
consistent with the Commission's mandate on a risk-informed, performance
based regulatory approach and at the same time provide sufficient model
to the Department so that they will be able to submit a quality license
application.
I think we have kind of skipped over some of the points that
are discussed on the next slide, so proceed to Summary.
MR. GARRICK: Which number are you on, just for clarity's
sake?
MS. LUI: Slide Number 8. The multiple barrier requirements
go to be a system requirement.
We shied away from the subsystem -- qualitative subsystem
performance objective in Part 63, in the proposed Part 63 and we will
continue the track that we will keep the multiple barrier requirement as
a system requirement.
In other words, we will not set performance goals for
barriers such as waste package and natural settings.
In our evaluation of DOE's license application, the goal is
to look for that Both the engineered and geologic systems contribute to
safety. That goes back to safety that is not wholly dependent on a
single barrier concept.
I think we have pretty much beaten the second check-mark
here to death --
[Laughter.]
MS. LUI: -- and the last one is we not seeking complete
redundancy for the barriers.
The last remark is just to reiterate that the public comment
period is over and we are well underway in terms of analyzing the public
comments and providing and preparing the response, and whatever
information that we hear during these particular meetings that will be
available to us in terms of finishing up the final rule and drafting the
Yucca Mountain Review Plan. We intend to put the transcript of this
meeting on the website so that it will be available to the general
public.
MR. GARRICK: Let me postulate a situation. We have learned
a lot from the TSPA work. We have learned so much that where we used to
use the word frequently "geological isolation" we are using it less and
less, because we have pretty much learned that if we have a source term
and it is mobilized, it just delays the transport of that material into
the biosphere. It doesn't isolate it from the biosphere.
At least we haven't been able to characterize, we don't
think we are able to characterize any site where we could achieve
complete isolation in the absence of assistance from engineered systems.
Now supposing somebody came along and suppose they convinced
you that I have designed the one million year waste package and my
confidence in that containment capability is far greater than my
confidence in the containment and transport capability of the natural
setting. Obviously if you have a defense-in-depth philosophy like you
are stating here and that we are seeking balance, which I in principle
kind of agree to, you'd deny them the license.
MR. GREEVES: Why would you deny them the license? You lost
me.
MR. GARRICK: Well, what I am saying, if somebody comes
along with the perfect, with a million year waste package, and there's
engineers that believe they can do that, and yet the geologic setting
they couldn't convince you that if there was a source term that there
would be adequate containment, but with the waste package of course
there is adequate containment, so you don't have the defense-in-depth
but you have a waste package that convincingly will last a million
years.
With Part 6 could you license that?
MR. GREEVES: I think you have carried us too far of a
stretch.
MR. GARRICK: Well, I don't think it is so far a stretch.
Frankly, I think it is probably much easier to design a million year
waste package than it would be to characterize Yucca Mountain down to
the few meters.
MR. GREEVES: Your dialogue was saying that the site gives
you nothing is the way you --
MR. GARRICK: Eventually it doesn't give you anything. It
gives you dilution. It gives you something.
MR. GREEVES: I don't agree with that statement.
MR. GARRICK: But the one thing that the Nevadans are coming
to us very strong on is, and the NRC is agreeing with them, at least in
the public media, that we are now talking about delay, not isolation.
MR. GREEVES: Anybody that's been in this business, Bob
Bernero said it earlier, it's just a question of time whether it is high
level waste, low level waste. You cannot guarantee containment. There
will be some time when you have to --
MR. GARRICK: The argument being, John, that there's a lot
of people that believe I can do a much better job at building something
to a specification than I can at characterizing a mountain into a level
of detail necessary to give me the same output.
MR. GREEVES: I am aware there are people out there like
that. We are also aware that there is a piece of legislation that calls
for multiple barriers.
MR. GARRICK: That's all I am getting at. That's back to my
question --
MR. GREEVES: The simplest -- an engineered barrier and the
site --
MR. GARRICK: Are we ending up with a law, with a regulation
here where we couldn't license a repository that has overwhelming
evidence that it will retain its integrity for a million years?
MR. EISENBERG: Dr. Garrick, there is no intent to put a
roof on the quality of any barrier. DOE should make each barrier as
good as they can.
MR. GARRICK: That isn't my point. My point is -- MR.
EISENBERG: Well, it sounds like it is your point.
MR. BERNERO: I would like to interject on behalf of the
Staff, as if I was still there.
What you describe is a very good description of the Swedish
strategy.
MR. GARRICK: Yes.
MR. BERNERO: Which is the sole purpose of the repository
isolation is to maintain reducing chemical conditions so that this very
nicely designed million year package will live for a million years.
MR. GARRICK: Right.
MR. BERNERO: And besides that, that water down there is
fossil water It isn't going to move for a long, long time, and it is a
marvelous system.
They of course are a piece of granite that is rising up out
of the sea and you have a choice of granite, granite and granite for a
Swedish repository
[Laughter.]
MR. BERNERO: The United States has a system of laws which
gives us a statutory requirement that says you must have multiple
barriers. It also has a statutory requirement that DOE cannot look at
crystalline rock.
Now that is not a technically based requirement. It's an
entirely politically based requirement.
There is a system of laws and there is a distinction that
one has to make in what would constitute an acceptable repository as
against what would constitute a preferable or ideal repository. At one
time we had three sites to be simultaneously characterized, and we used
to call it "The Beauty Contest." Insanely expensive. Just imagine
doing Yucca Mountain in triplicate and trying to keep them on the same
schedule.
What we have to have in the United States is what is an
acceptable repository. It's been accomplished in the WIPP case, warts
and all, you know, and certainly we can talk for hours and hours on what
should have been done there, but it's been done and I am convinced it is
an acceptable repository and warts and all this Yucca Mountain thing --
MR. APOSTOLAKIS: I think it also comes back to the issue of
prevention versus mitigation. Maybe -- I really don't like, to
generalize a little bit, regulatory documents that talk in terms of
number of barriers. In fact, if this subcommittee writes a letter, that
would be a good thing to attack, because it is such a fuzzy concept that
can be misused and so on. I don't know what it means, multiple
barriers, to begin with, and I think a lot of the debates we are having
here come from the fact that the Staff naturally feels that they have to
comply with what the Commission says and the Commission says multiple
barriers, the legislation, I'm sorry. But this is an independent
advisory committee so we can write --
DR. KRESS: Did the Senate say how many barriers was
multiple?
MR. BERNERO: No.
MR. APOSTOLAKIS: Well, the more I think about it, it's
really the root cause of a lot of emotional debates, because I am not
even sure -- you gave us a good example with the reactor vessel.
Up until this morning I would call it one barrier. Now you
tell it is not one barrier. Now I have no basis of saying it's not or
it is or it is not. I think it's wrong to count barriers, to count
something you have not defined.
MR. LEVENSON: But John, in response to your question, I
think the answer is it could be licensed because the legislation, as I
understand it, does not say that each barrier has to be 100 percent
effective.
The legislation just says there must be more than one
barrier.
MR. APOSTOLAKIS: Which defeats the whole idea, of course.
DR. KRESS: I think at this point -- are you finished?
MR. GREEVES: Let me just summarize. We are finished.
[Laughter.]
MR. APOSTOLAKIS: Good.
MR. GREEVES: You think I should stop there? He said we
were finished. He didn't say we've had it. I think we have worn it
out, right?
Just to summarize, I think Norm did a good job of showing
you the spectrum of issues that face us across the licensees that NMSS
has. It is a difficult issue and I think we have learned something from
watching the process here, and I think some things are going to come out
in the future that will help us, and each one of those -- it is almost
like the chart that Norm showed. For each one of those arenas, we have
got to start making some decisions.
You spoke at length about the DOE issue, but each of those
we have got to sort of make some decisions. I know you all appreciate
that the Staff needs to be consistent with the Commission policy and the
legislation, so that is something that we will be holding in our minds
as we draft the regulations.
Something that has come out to me is listening to us all
talk around the room is transparency. I think we have got to find a way
to explain these things that is a bit more clear. I think we talked
past each other on occasion, so I challenge us to -- over time we are
going to have to make this more transparent to other stakeholders.
I do ask you to keep in mind what the Staff presented are
preliminary considerations. We are working under the requirements for
developing the rule process, and I know Bob is disappointed he didn't
see the number he was looking for, but that is something we are about.
MR. BUDNITZ: Doesn't have to be a number.
MR. GREEVES: Well, I think you raised some good points and
I agree with the need to do it one way or the other, and we didn't tell
you today.
MR. BUDNITZ: That's fine.
MR. GREEVES: And so those will be my closing remarks and I
assure you we are still considering these issues and we are going to
look at this transcript and I think it will be helpful. Thank you.
DR. KRESS: Thank you very much. At this point I'll take
another break for about fifteen minutes, and that would be be back at
ten minutes, by this clock, after 3:00.
[Recess.]
DR. KRESS: We are at the point on the agenda where we are
going to hear from Gary Holahan and Tom King. Our pleasure, gentlemen.
MR. HOLAHAN: Good afternoon. This is Gary Holahan. I am
the Director of the Division of Systems Safety and Analysis in the
Office of Nuclear Reactor Regulation, and Tom King and I are going to
discuss what defense-in-depth means to the reactor program. I think you
will hear a lot of things that you heard this morning, because I think
we are all playing from the same historical book, so some of what we
discuss will be historical, some of it is recent and ongoing activities,
and some of it is looking to the future, so I will start out with a bit
of the historical perspective and Tom is going to cover the future.
I think it is interesting the first point we are making is
that in fact there is no formal regulation or agency policy statement on
defense-in-depth and I think this goes back and is consistent with Tom
Murley's comments this morning about defense-in-depth isn't a rule or a
specific requirement, which I think leaves a little bit to a number of
comments this morning about are we talking about a philosophy or a
policy or a guidance or a rule or a requirement or a commandment?
I guess at that point I would have to agree with Dr. Budnitz
that what really matters is how you implement it, so in fact we have
called defense-in-depth a philosophy, not a specific regulatory
requirement, and in our recent guidance documents we have said that it
is one of our principles that we preserve that philosophy, so George
might be offended. We used the word principle and philosophy in the
same sentence, but luckily George and his subcommittee concurred in that
document, so we'll feel comfortable about it.
[Laughter.]
MR. HOLAHAN: But it was two or more years ago.
MR. APOSTOLAKIS: Nothing less is expected of Gary.
MR. HOLAHAN: The second point in fact is that as with the
materials program, the reactor program is really working with the same
philosophical concept of defense-in-depth. In fact, we are quoting the
same version that Bob Bernero mentioned this morning where
defense-in-depth, as was said earlier, has successive compensatory
measures and it has this element of not being wholly dependent upon any
single element of the design.
There have been previous definitions of defense-in-depth and
they have all been more or less consistent. I am going to show you a
couple of historical examples in just a minute.
The third point I would like to make on this introductory
slide is that what really counts is that this philosophy, the same
philosophy can be implemented in a number of different ways and what you
see in the reactor program is not necessarily the same thing as you see
in the materials program and I think the agency feels reasonably
comfortable calling both of those defense-in-depth philosophy.
In the reactor program I am going to discuss the regulations
themselves where defense-in-depth is included in the regulations even
though it isn't a specific regulation itself, also how the licensing
process and the license amendment process have dealt with the subject
and the new reactor oversight process, where oversight includes
inspection, enforcement, monitoring of licensee performance, where the
elements of defense-in-depth are embedded in that process as well. Next
viewgraph.
Well, you can see on this viewgraph Part 50 includes
defense-in-depth in a number of ways. The concepts of prevention,
mitigation, single failure, redundancy, diversity -- these are all
elements of defense-in-depth. When we talk about it, you can talk about
defense-in-depth in a number of ways. You can talk about physical
barriers. You can talk about functional barriers. You can talk about I
think Tom Kress has suggested a number of times risk allocation in fact
is a defense-in-depth concept. You can put numerical goals on things
like core damage frequency and large early release, and that in effect
is a way of providing defense-in-depth. Next viewgraph.
There are two viewgraphs that are used as part of a training
program that NRC has. It's called "Perspectives on Reactor Safety" and
it is sort of, in part it is a history book that Denny Ross and a number
of people worked on with Sandia to put together so that NRC's new Staff
members have an appreciation of not only what the requirements are but
how they got that way, and it covers sort of the history of the '60s and
'70s as the requirements were built.
As part of that, in fact there is a section on the concept
of defense-in-depth, what it means and how it was developed and I am
going to show you two viewgraphs from that material.
What you see here is one concept of defense-in-depth, which
I think I would call the functional definition. That is, you look at
prevention, mitigation in terms of having safety systems and
containment, and siting and emergency planning. In this particular
example you will see that accident management is also identified as a
level of defense-in-depth. Some people would push it a little bit into
a containment performance issue. Some people would talk about it as an
emergency response issue, but you see how the measures of
defense-in-depth basically show that public safety is protected by a
series of functional type barriers. Tom, can I see the other one?
I think especially years ago people generally talked about
defense-in-depth in terms of physical barriers, and in fact in the
training book these are two pages right together, and so these concepts
sort of grew up together over the years and the concepts of physical
barriers including the fuel pellet and the cladding, reactor coolant
system, containment, and then things like exclusion areas --these are
the physical barriers.
Now what we know is this is a defense-in-depth concept.
Each of these defense-in-depth concepts really has its own sort of
strengths and weaknesses. If physical barriers were the only
defense-in-depth concept, I think we would have come quickly to the
realization that common cause failures and interdependencies make this
an incomplete concept for defense-in-depth. In fact, the functional
concept in my mind is more complete and in a number of ways, using PRA
and whether you call it allocation or other ways of looking at core
damage frequency, even the concept of Level 1, 2 and 3 in PRA in my mind
are a form of defense-in-depth and probably a more complete form.
One of the ways in which the regulations call for
defense-in-depth, and this is just one example that I have picked out,
you could probably find dozens, if not hundreds, of places where a
concept is embedded in the regulations, right in the general design
criteria.
In fact, it is broken up into six sections. One of the
sections itself is called "Protection by Multiple Barriers" but in
addition to that, the other sections of the general design criteria,
which really play a strong role in determining what an acceptable
reactor design looks like, in fact call for a reactor core that behaves
well, a primary coolant system with low failure probability, and then
fluid systems, either normal ones or emergency ones, to handle failures
and the reactor containment and fuel and radioactivity control really
talks about fuel in the sense of fuel handling, and that doesn't mean
that when it is in the core, it means when it is a potential source, so
the very structure of the regulations down to the general design
criteria have embedded in them a defense-in-depth concept.
I think I said I would talk about licensing but I think I
skipped -- let me do the oversight program and then I'll talk about the
license amendment process because that is one that we have been changing
lately and it has a good kick-off point for Tom to get into our more
future activities.
The reactor oversight process was really given almost a 100
percent overhaul in the last year, where the inspection program, the
enforcement program have basically been totally rewritten, and they have
been rewritten with two concepts in mind. One is to be more
performance-based, to look at licensee performance and react to it, and
the other is to use more risk insights in the process, but in doing so
the defense-in-depth concept is being preserved by the use of what are
called cornerstones, and I am going to show you how the cornerstones fit
into the process.
Basically the message is that the cornerstones in the
oversight process are the ways of embedding defense-in-depth.
Cornerstones are defense-in-depth features and in fact if you read the
papers on the subject, the concept of defense-in-depth comes up in a
number of points.
This is a viewgraph that many of you may have seen before.
It is used in a lot of the presentations on the oversight process and if
I can lead you from the top down, public health and safety really means
that we worry about how the reactors behave and radiation safety, both
in terms of the public and workers. That's the Part 20, Part 100 type
issue, and safeguards, so the issues to the right are really in addition
to what we have talked about most of the day in terms of public health
and safety from unusual type of severe accidents.
If you will look at the way the program is structured,
reactor safety has four basic elements to it. They are called
cornerstones but you could have called them defense-in-depth elements if
you wanted to.
We look at initiating events, mitigating system performance,
barrier integrity, and emergency preparedness, and those are basically I
think a combination of functional and physical barriers.
The way the oversight process works, the licensee
performance, both in terms of performance indicators and inspection
results from our inspections staff are put into these categories, and
then we make judgments about the licensee performance in those areas.
If you go to the next slide, I can continue.
I am going through this kind of quickly, just not to explain
the whole process to you but just to show how the concepts,
defense-in-depth concepts, are built in here.
The performance indicators as used in the reactor oversight
process are in fact groups together depending on which of the
cornerstones they relate to, so things like reactor scrams or
significant initiating events and transients, they go into the
initiating event cornerstone, and things like the safety system
performance and unavailability, those go into the mitigation system, and
so the licensee performance in terms of performance indicators and
inspection findings are measured with respect to thresholds to identify
their significance and they are folded into these cornerstones. We can
go to the next one.
In fact, I am not going to discuss this viewgraph. Just for
completeness it shows how each of the cornerstones has indicator input
to it.
The next viewgraph is a little hard to follow, but the basic
concept is across the top you will see a spectrum of results in which
various levels of performance of increasing safety significance are
monitored, and so on the extreme left what you will see is everything is
pretty normal, and that is the inputs to the cornerstones, each of the
cornerstones, not just public health and safety sort of dose limit, but
each of the cornerstones is performing well.
If you will look down that column it says we have a routine
inspection program and licensee fixes issues on their own, and sort of
everything runs sort of normally and this is, you know, we use the
terminology of "green" -- this is normal green performance in terms of
for a licensee. As you move to the right, across the top columns, you
will see increasing level of concern, and that is indicated by degraded
performance in one or more cornerstones.
As you can see, as it sort of escalates, it is not
only that the total licensee performance seems to be unacceptable in
some way, but the NRC response will escalate when the performance in one
cornerstone area becomes of increasing concern to the level of being
warranting interactions at Regional Branch Chief level, Regional
Division Director, Regional Administrator, EDO and even getting to the
point of the Commission.
So what it says, and there's lot of detail on here that I am
not going to cover today, the basic message is we are looking at
licensee performance at the cornerstone, but that's basically at the
defense-in-depth functional levels and making judgments about how well
the licensees are doing, what level of interaction we ought to take with
them, whether their performance looks normal and we ought to sort of be
restrained and allow them to deal with their own issues, take corrective
action when problems occur, or whether a higher level of management
involvement and more extreme expectations are appropriate
Now the system is set up basically as an early warning
system. It is not so easy to go from green to red. Part of the
workings of the systems is you expect the licensees to know very well
what the rules of the game are. If their performance begins to degrade,
they know it early-on. We expect them to be dealing with it early. We
don't expect licensees to be in the yellow and red area because there's
plenty of warning for them to turn things around, but the scheme shows
how the Staff will be responsive to cornerstones or defense-in-depth
weakenings, and in fact potential failures. Tom?
I know that is kind of a lot to digest. The only point I
wanted to get across is that even though defense-in-depth is not written
as a regulatory requirement it has a value as a guiding philosophy and
it can be built into various programs in a practical and usable manner.
Now in the license amendment process we have developed
Regulatory Guide 1.174. Even though 1.174 has a lot of general safety
philosophy in it, it was really meant as a licensing amendment guidance
document and there are five safety principles associated with deciding
whether a license amendment change is acceptable or not.
I know the ACRS members are very familiar with that. We
spent a lot of time with the committee on these issues and if my memory
is accurate, and I think it is, even the concept of having five
relatively high level safety principles was a concept that came up at
this table in the interactions between the Staff and George, your ACRS
PRA Subcommittee.
One of those five principles is that there ought to be a
defense-in-depth philosophy and my recollection is we talked a long time
about this issue of should there be defense-in-depth, should there be
defense-in-depth philosophy where we are talking about never giving up
any measure of defense-in-depth, and I think it was an important issue.
I think in a number of ways it still is an important issue and I think
next month we will talk about ACRS has a session on impediments to
risk-informed regulation, and I know a lot of people are concerned that
this is a potential impediment, and I think we have certainly got it on
our list of one of the things we want to talk about.
Reg Guide 1.174, its corresponding Standard Review Plan, and
the related documents on how to do risk-informed regulation not only
mention that there should be a defense-in-depth philosophy but give you
some insights as to what that means and it identifies issues like
balance between prevention and mitigation, avoidance of over-reliance.
Now these are general concepts. They are not numerical values. I think
George has expressed the idea that you should be very careful about not
counting the numbers of defense-in-depth or try to quantify it too much,
and I think we recognize the danger in doing these things.
Those concepts are discussed in the guidance documents. I
think it clearly says we are not trying to assure that there is no
change in the level of defense-in-depth. What we are saying is there
should be no change in the philosophy. So if a licensee wants a license
amendment to remove the containment, they ought not to bother because we
are not going to pursue that.
MR. APOSTOLAKIS: One important point here, which I believe
is an assumption on your part and most people when they talk about these
things is you are talking about these issues for the current generation
of nuclear power plants.
There is a certain assumption here that -- in other words,
would you be as absolute in rejecting a request for no containment for
any future reactor? I doubt that, because you don't know what physical
pieces of those --
MR. HOLAHAN: I wouldn't reject it categorically.
MR. APOSTOLAKIS: So this is really for the current
generation, which is I think a reasonable thing to do.
MR. HOLAHAN: Well, for the current generation and I think
for the evolutionary and advanced reactors that we have seen.
MR. APOSTOLAKIS: Yes. I agree.
MR. HOLAHAN: But I think this ought to be left as a
relatively high hurdle.
MR. APOSTOLAKIS: I agree.
MR. HOLAHAN: Okay. By its nature, what we are trying to do
in the reactor area, and I recognize that in the materials area there
are some other considerations, we are providing a very high level of
protection, that is very low probabilities for high consequence events.
Almost by definition, if that is the arena that you are in, you are not
going to have a lot of experience to deal with and you are going to be
extrapolating from pieces of what you know, and issues like completeness
and modelling are going to be difficult ones.
One of the things that I sort of keep an eye on is the
accident sequence precursor program, previously in AEOD, now in the
Office of Research, and my recollection of if not the last but one of
the recent Commission papers on that program, maybe a year ago or so, I
think it said something like half of the accident sequence precursors,
the ones of some significance, were things that were not previously
modelled, and so the signal is we are still at a time in which there are
surprises to be had, and by its very nature, you know, you are going to
have to develop an awful lot of operating experience before you get to
the point in which you say my modelling and my completeness are minor
issues.
MR. APOSTOLAKIS: Well, again, I would put some qualifiers
to what you just said. What does it mean it's not modelled? I mean
maybe the exact sequence of events was not modelled but maybe it is a
subset of something bigger that was modelled.
MR. HOLAHAN: Well, I think --
MR. APOSTOLAKIS: I agree with that.
MR. HOLAHAN: I think it is worse than that.
MR. APOSTOLAKIS: I think in some instances it might be, but
the other, I mean in all fairness you should also mention then the very
important findings of the former AEOD people that the system
unavailabilities they find are either -- are within the range of values
of PRAs found --
MR. HOLAHAN: Yes.
MR. APOSTOLAKIS: -- which is really a very good
confirmatory piece of evidence that what we are doing is not off the
mark.
MR. HOLAHAN: And in general initiating event frequencies
are somewhat better and in fact in my mind, more important than either
of those is that common cause failures are lower than is generally
assumed.
MR. APOSTOLAKIS: Right. They are going down and they are
going down.
MR. BUDNITZ: You are looking under the lamppost some of the
time because half of the risk overall of the fleet comes from fires and
earthquakes and configuration compromises that would make you more
vulnerable to fires and earthquakes are not modelled in ASP today, as
George and I know, since we wrote a NUREG about it which hasn't been
implemented yet.
MR. KING: But there haven't been that many fires and we are
looking --
MR. BUDNITZ: Well, there haven't been fires or earthquakes,
but we are talking about configuration compromises that will make you
more vulnerable if you had one.
MR. KING: Yes.
MR. BUDNITZ: Those happen all the time.
MR. KING: There haven't been any earthquakes.
MR. HOLAHAN: And my recollection is isn't that issue number
one of twelve that we are dealing with in the risk-informed fire
protection?
MR. BUDNITZ: I hope so.
MR. HOLAHAN: I think it is on top of the list.
MR. BUDNITZ: I hope so.
MR. HOLAHAN: So the message I want to leave you with is in
the reactor area for the plants we are currently dealing with, which
basically are operating plants -- not so long ago we dealt with advanced
reactor designs -- but in this context I don't think they were all that
different.
Defense-in-depth has been an integral part of our decision
process, what we envision for risk-informing Part 50, and Tom is going
talk to Option 3, but certainly if I remember the ways the options are
set up for risk-informing Part 50.
Option 1 is just to continue with some of the rulemakings
that we have ongoing, 50.59 and maintenance rule and things like that.
Option 2 is to take those issues related to day to day
operational performance and parts of the plant that get special
treatment in operations, things like quality assurance and technical
specifications, and maintenance type activities, and to risk inform
those sort of operational type activities.
In doing so, we intend to preserve the current design basis
and that means that the level of defense-in-depth in the plant probably
is not going to be changed very much, and also the other important
characteristic is in deciding what is of safety significance, because in
effect what Option 2 is going to do, it's going to take the old model of
safety-related and not safety-related, something that John Garrick
mentioned this morning, that the PRA world, the risk analysts don't care
much about, and it's going to look at what is risk-significant and what
is not risk-significant.
It's going to overlay those two concepts but in deciding
what is risk-significant or not, we are going to use a concept somewhat
akin to the maintenance rule expert panels where not only are we going
to use the risk analysis numbers, whether it's bottom line numbers or
importance measures, we will use the insights from experienced plant
people who can bring some defense-in-depth and safety margin thoughts
into that process, and we are developing some guidance as to what sort
of things they ought to be thinking about in doing that.
So my message is we currently have defense-in-depth in the
reactor designs, it is in our programs, it is even in our, what I would
say is our most modern risk-informing programs have the concept of
defense-in-depth.
Tom is going to talk about Option 3.
If I look about where we are going with risk-informed
license amendments and those sort of changes, there is a challenge on
the table for us.
I don't think we are going to quantify how much
defense-in-depth you need but we may put some more guidance in place as
to how to deal with issues where maybe it looks like we are either doing
-- I mean I must say I haven't heard any "too littles" but maybe we are
doing too much to preserve more defense-in-depth than a more
risk-informed insight would tell us is necessary.
So the program is ongoing. Defense-in-depth is a -- call it
a philosophy or a guidance concept, and it's basically built into where
we are.
MR. APOSTOLAKIS: But the point though, Gary, is that it is
not whether one should have that philosophy and whether one should
ignore, for example, the items you have under Regulatory Guide 1.174.
The question is not what role the risk -- the PRA methods we
have should play here, and I would say, for example, if I took -- given
the evidence that I have including the AEOD evidence, that PRAs have
done a pretty good job modelling system unavailability for individual
safety systems, there is strong evidence that we have done a hell of a
job, then again from my point of view that means that maybe the issue of
unquantified uncertainty is not that important there, although you might
make the point that under severe accident conditions we haven't seen
those and so on but let's take that -- so I would say that now I have a
good tool in my hands to take the seven or eight items you have there
and optimize my operations, optimize my design, and I don't really have
to have a diverse train for example because I manage to achieve the
required levels or the inspected levels simply with redundant trains.
I can make a good case that I have handled common cause
failures and so on, so I suppose the heart of the matter here is is
there anything that will stop me from doing that, another input, another
principle, a philosophy that will say, yeah, you can do all these things
but boy, I really want all seven, and what I am saying is I am not
willing to drop all seven, but first of all if you try to drop them you
will never achieve the numbers you want.
MR. HOLAHAN: Yes, that's right.
MR. APOSTOLAKIS: And second, all I am saying is these are
guidelines. It is a philosophy that you would like to have at your
disposal and use it, but now you have this tool which is reliable in
this particular context, so, you know, I can afford maybe to drop one or
I can afford to minimize the role in one place versus another and so on
and I think that is really what we are doing with the case specific
risk-informed guides.
MR. HOLAHAN: Yes.
MR. APOSTOLAKIS: So this is a good example in fact of a
case where the PRA, it's almost risk-based here, where risk is the
unavailability.
MR. HOLAHAN: Well, I think what I would say is if you go
back and read the section on defense-in-depth in 1.174, I think it's
okay, but that does not mean that in implementing it we won't run into
some tough cases, okay?
MR. APOSTOLAKIS: Sure.
MR. HOLAHAN: And we may be better off just fighting over
those cases than trying to write a guidance document that avoids any
fights in the future.
It may not be possible to write the definitive set of
guidelines on defense-in-depth that never has a problem.
MR. APOSTOLAKIS: And I realize that but I think some sort
of a high level discussion of these issues probably would be beneficial
because I agree that we can't really be too specific at this point.
MR. KING: Reg Guide 1.174, if you recall, in the
defense-in-depth discussion does talk about using PRA, not to do away
with defense-in-depth but to optimize how you achieve it and in effect
in Option 2 and Option 3 risk informing Part 50 it is the same
philosophy, the same approach we are taking.
What I was going to talk about is Option 3 and what are we
doing in our technical study or study of the technical requirements, how
are we folding in defense-in-depth considerations and melding them with
PRA considerations, because for all the risk-informed activities what we
are talking about is not a risk-based approach but using PRA to
complement our traditional way of doing business, which includes
deterministic analysis and defense-in-depth considerations, so we are
trying to keep that approach in both Option 2 and 3, and I will talk to
you about what our thinking is today for doing that under Option 3.
The last piece of this viewgraph I am not going to talk
about. You are going to get a separate presentation on that at some
point in the next month or two from Joe Murphy, but again the reactor
safety goal policy discusses defense-in-depth and we had identified that
as an item for consideration for modifying the safety goal policy.
Perhaps it needs to be updated, expanded, and so forth, consistent with
the risk-informed regulation thought process that we have gone through
in discussion there.
Maybe I'll just take one more minute for background,
particularly for the folks from ACNW on what is Option 3, what are we
trying to do. As Gary mentioned, NRR is working on a rulemaking now
that's called Option 2 that is basically looking at the scope of what
ought to be regulated based upon risk insights and that is in the sense
of special treatment rules -- by special treatment, what should get QA,
what should get equipment qualification and so forth.
The functions would have to remain the same but maybe
depending upon the risk associated with -- the risk significance of the
various systems, structures and components, maybe they don't need the
pedigree that they are receiving today, but again the functions would
all have to be accomplished.
What we are doing under Option 3 is going in and looking at
the functions, the design requirements, what changes should be made
there based upon risk insights.
Maybe to put in context what you are going to hear, the
Option 3 study is going to take place during this calendar year,
calendar year 2000. We are in the initial stages of getting started.
What you are going to hear about is work in progress today. Some of the
details have to be worked out.
What you are going to hear about today we are also going to
put out for public comment fairly soon and we have a workshop, public
workshop, scheduled the end of February to talk about this as well as
the other things we have been working on in the Option 3 study, so this
is subject to a lot of comment and a lot of further discussion. This is
not cast in concrete at this point.
In trying to do the Option 3 study we did realize we had to
come up with what we call a working definition of defense-in-depth,
something that the folks looking at the regulations and the Reg Guides
and the SRPs can take and take the risk insights and sit down and make
some decisions on does what is in there look okay or are some changes
warranted?
So what we wanted to do was basically develop an approach
under this working definition that would consider defense-in-depth that
traditionally provides some multiple lines of defense -- are not calling
them barriers, we are not counting barriers -- provides some balance
between prevention and mitigation and provides a framework by which we
can address uncertainties in the various accident scenarios, so that is
sort of the scope of what we thought this working definition ought to
contain.
There are two elements to the working definition. One,
which is probably the structuralist element, that in our view there
ought to be some floor on defense-in-depth regardless of what your PRA
says, there are probably some things you want to retain, just call it
deterministic or engineering judgment, and then beyond that, there would
be the rationalist piece or implementation elements that can vary
depending on the uncertainty and the risk goals and so forth.
MR. APOSTOLAKIS: This is the pragmatic preliminary proposal
we have?
MR. KING: Yes.
MR. APOSTOLAKIS: Structuralist at the high level and
rationalist at lower levels?
MR. HOLAHAN: The rationalist-informed structuralist
approach.
[Laughter.]
MR. KING: It doesn't have to be one way or the other. They
each have some advantages.
MR. APOSTOLAKIS: No, but this is the compromise we came up
with, otherwise the paper would never have been published.
[Laughter.]
MR. APOSTOLAKIS: Isn't that right, Tom?
MR. KING: Yes.
MR. APOSTOLAKIS: This is the pragmatic.
DR. KRESS: That's pretty much we covered.
MR. APOSTOLAKIS: High level structuralist and --good.
MR. KING: On Slide 15, it talks about the fundamental
pieces or the structuralist pieces. We want to build upon the
cornerstone concept that Gary showed, particularly building upon the
first four cornerstones that are affected by reactor design, initiating
events, prevention and core melt, containment of fission products, and
emergency planning and response.
We feel that this working definition ought to address those
things. We feel that there ought to be some, in the prevention side
there ought to be some again I will call it a floor on design features
that prevent core melt and whether we call those -- we put back in the
single failure criteria or somehow specify some redundancy or diversity,
we haven't worked out exactly the wording of that, but we would not rely
strictly on a risk number to say I have got a highly reliable system,
therefore I don't need any redundancy, diversity, single failure
protection and so forth.
Again, other things you have to consider are how do you
factor the human in and the active versus passive failure, particularly
if we are into the single failure question which in the past has always
been limited to an active component.
We feel that we should retain the ability to contain fission
products given a core melt, that that ought to be a fundamental concept
of part of this working definition and emergency planning and response
ought to be retained. Clearly emergency planning and response is also
affected by siting criteria if you are talking about new plants, but for
existing plants it is pretty well fixed.
Now in addition to assuring the prevention and mitigation we
wanted to assure a balance between the prevention and mitigation and we
felt that we needed to be consistent with the subsidiary risk guidelines
that were developed and used in Reg Guide 1.174.
Those actually came from Commission guidance that we
received over the past years where they gave us a 10 to the minus fourth
core damage frequency damage goal to use and then we developed, as part
of developing Reg Guide 1.174 worked backwards from the safety goal
quantitative health objectives and came back and developed a 10 to the
minus fifth large early release frequency goal that we felt was a good
design objective that if it was met would ensure you would meet the
quantitative health objectives.
MR. GARRICK: In your use of a mitigation here, does it
reach to consequence limiting? In other words, if you are having a goal
with respect to a large early release, now you have material. What do
you mean by mitigation beyond the usual engineered safety features or do
you mean anything beyond that?
Do you include consequence limiting?
MR. KING: The large early release, the word "large" has no
limit on it. It can be a large release --
MR. GARRICK: You are not including --
MR. KING: No. It can lead to early fatalities offsite.
DR. KRESS: Yeah, but it does include emergency response
measures for --
MR. KING: Sure.
DR. KRESS: -- for this LERF to be equivalent to the early
fatalities so that is in there.
MR. KING: Credit is given -- yes -- credit is given for
emergency response.
DR. KRESS: Credit is given for emergency response.
MR. KING: But there is no limit on what large should be.
MR. GARRICK: Well, I am also thinking of fission product
cleanup, retention --
MR. KING: Well, maybe I ought to say a little bit about
large early release. It is not large if it is cleaned up.
DR. KRESS: Yes.
MR. KING: In other words, if it goes through the
suppression pool and scrubbed, it is not considered a large release
because not much gets out of --
DR. KRESS: Those things are inherent in the definition.
MR. GARRICK: Yes, but I am getting at the 10 to the minus
five number.
MR. KING: Yes. That is for unscrubbed stuff.
MR. GARRICK: Unscrubbed, yes.
MR. KING: And it can lead to early fatalities.
MR. APOSTOLAKIS: It is directly related to early
fatalities.
MR. HOLAHAN: In effect what happens is if you have a
scrubbed release or a late release or a minor release in fact core
damage frequency 10 to the minus four by default becomes its limit.
DR. KRESS: Yes.
MR. GARRICK: Yes.
MR. KING: Okay. The next thing we have done was say okay,
for this bottom piece what does that mean in terms of looking at the
cornerstones and some practical guidance when you want to go in and look
at the regulations?
We developed sort of a chart that works its way down from
the cornerstone concept and in fact I guess it is a high level
allocation.
It is not intended to get down to the individual component
or system level. This is to be looked at as fairly high level guidance
but the idea is the following, that you have got various initiating
events and they have various frequencies associated with them.
Some of them are things that you know are going to happen --
loss of offsite power, turbine trips and so forth. They are fairly
frequent and then there is the more infrequent initiators, the large
LOCAs, the large reactivity insertion accidents and so forth, and then
there's the rare events that today aren't included in the design -- the
vessel rupture, the steam generator rupture and so forth.
You can have a list of those, and you can have an estimate
of their frequency and their uncertainty distribution that goes with
that frequency.
And then you want to look at, for each of those, how does
the plant ensure that the core damage frequency and the large early
release frequency is met?
And the idea is that for the more frequent initiators, you
want to be able to have systems in the plant that respond with a high
degree of reliability; so that when those things happen, you're assured
you still meet your 10-4 core damage frequency, and you still have a
robust containment that will meet your LERF goals.
For the things that occur less frequently, maybe you don't
need as much in terms of highly-reliable systems, but the combination of
the two still ought to ensure that you meet your core damage frequency
goal, and you still want to be sure to have containment with the same
degree of protection.
And you still have emergency planning out here, for which
you get some credit.
MR. KRESS: That second line there, does that imply you have
different responses to those initiators, for example, shutting down the
power or the emergency cooling to prevent core damage? You'd have those
same initiators.
If you had to have them for the infrequent initiators, you'd
have to have them for the more frequent ones also. I don't understand
this allocation.
MR. HOLAHAN: It may turn out that way, but, in fact, for
example, you might find that for large loca you need, you know, low
pressure injection, and ECCS accumulators, but for small locas, you only
need the high pressure injection system.
MR. KRESS: I see what you mean.
MR. HOLAHAN: So that says redundancy in high pressure
injection is very important, valuable, but redundancy in those other
systems may not be so important.
MR. APOSTOLAKIS: One comment on this: This would work well
for the so-called internal events.
Now, if you have an earthquake, and possibly a fire, or any
external event that could affect elements of prevention and mitigation,
somehow we need to have maybe a different approach and rethink the
concept of mitigation versus prevention of those big, common-cause
failures.
MR. KING: Common-cause failures, yes, how you apply these
to common-cause failures, and how to you apply these to something like
steam generator tube rupture.
MR. APOSTOLAKIS: Although one could apply the same approach
to the sequences that are initiated, perhaps, by the fire, for example,
and have certain requirements in the initiator frequency, the systems
that will mitigate it.
But somehow these two dashed-line boxes come together when
you have those big --
MR. HOLAHAN: I think I agree with you for seismic, but for
fire and flood, I think you can deal with these. In fact, more modern
plants, and certainly evolutionary and advanced plants have dealt with
fire and flood in terms of separation, which allows this to work out
very nicely.
What we see is that fire protection for older plants,
barriers, fire barriers and things like that, are ways of getting
isolation, even though it's not as complete as you see in the modern
plants.
With seismic, everything shakes at the same time, and so you
have to deal with that maybe a little differently.
MR. GARRICK: An important part of the large scope PRAs were
the recovery models that were employed. Does the respond include that?
MR. APOSTOLAKIS: Yes. Human recovery actions --
MR. GARRICK: Are over on the right.
MR. APOSTOLAKIS: -- respond to prevent core damage.
MR. GARRICK: Well, also things like recovery of offsite
power, recovery of --
MR. KING: They're in both of these boxes here. And that's
when you go in and look at the --
MR. APOSTOLAKIS: Even prevent initiators. An initiator is
a complete blackout, and human actions to recover diesels and so on is
part of it.
MR. HOLAHAN: I think Dr. Kress made a good point this
morning. Some of these differentiations are a little bit arbitrary. And
whether you say mitigation is mitigation of an initiator, or whether it
is mitigation of core damage, you can break this into finer pieces if
you like, and so a little bit of it is terminology.
MR. KING: The other thing this will help you do is, when
you have something like a steam generator tube rupture where you now
have lost the containment barrier, you've got some frequency associated
with it, and this now becomes one.
That tells you I better have some fairly highly reliable
systems to be able to deal with that.
MR. APOSTOLAKIS: So the message you are sending here, Tom,
is that one cannot really have goals independently of the accident
sequence.
And what really matters here is really what you have there,
the basis.
MR. HOLAHAN: And defense-in-depth --
MR. APOSTOLAKIS: And the allocation issue, depending on
reality, on preferences --
MR. HOLAHAN: And defense-in-depth doesn't mean equal
allocation among cornerstones or defense levels. But it means you don't
skip them.
MR. APOSTOLAKIS: And even there is a seismic issue that
maybe doesn't even allow you to do this, right? So depending on the
sequences --
Now, why on the performance indicators that the oversight
process uses, sequence or site-specific?
MR. HOLAHAN: Are they are aren't they?
MR. APOSTOLAKIS: Why aren't they?
MR. HOLAHAN: Oh, they are.
MR. APOSTOLAKIS: They are not.
MR. KING: The data is site-specific. The indicators and
the thresholds are generic right now.
MR. APOSTOLAKIS: Yes. The thresholds are generic.
MR. HOLAHAN: The thresholds are generic.
MR. APOSTOLAKIS: Would it be consistent with this approach
to have site-specific thresholds?
MR. HOLAHAN: Well, I think that just -- that would be nice,
but it's complicated. What we've committed to is, in the process where
there are inspection findings or events, we will use as part of this
process, what's called the significance determination process.
MR. APOSTOLAKIS: Yes.
MR. HOLAHAN: And we've committed to that process basically
being site-specific.
MR. APOSTOLAKIS: But isn't it true that in the maintenance
rule, the licensees themselves set the goals?
MR. HOLAHAN: Yes.
MR. APOSTOLAKIS: Why can't we ask the licensees to set
goals for their plants for each of the performance indicators? What's
different? Why can't we do it? Somehow we are scared of it.
And then we review it and say fine, or we say change this
and that, and let them do the work. You don't want to do that for 140
units.
MR. HOLAHAN: Well, we did it once.
MR. APOSTOLAKIS: Well, in fact, why don't you build on the
maintenance rule, and say, you know, for a San Onofre, this is what
they're using now for the trains, and San Onofre can --
MR. HOLAHAN: I'm not sure that that level of refinement is
really --
MR. KRESS: I don't think you can justify that level of
refinement.
MR. APOSTOLAKIS: I think you can.
MR. HOLAHAN: If you think of the scarcity of data, if a
reactor has, you know, four reactor scrams in the same year, whether
it's this type of reactor or that type of reactor, or something, you
know, something funny is going on.
MR. APOSTOLAKIS: I'm willing to grant you that, yes, for
several indicators, probably a generic number would be good enough.
But what I'm questioning is the philosophical approach. I
mean, this is really great.
But when it comes down to actually regulating and
interacting with the licensees, we are switching and going to generic
numbers as a starting point.
MR. KRESS: This thing comes very, very close to what I had
in mind by the allocation process as meaning the defense-in-depth.
Let me ask you a strange questions, Gary: That fourth box
up there, emergency planning and response, with the .1, if that box
wasn't there, and you still had to meet a safety goal that was early
fatalities, your LERF would simply be 10-6 instead of 10-5, I think
, because that .1 is about the mitigation you get.
MR. HOLAHAN: Yes.
MR. KRESS: Do you think all of the plants out there could,
at their present time, meet a LERF of that value?
MR. HOLAHAN: This is a side discussion that Tom and I had
this morning while the discussion was going on. I think it came during
Bob Bernero's presentation.
MR. KRESS: Yes.
MR. HOLAHAN: In general, most of the studies we've seen --
and you've got to recognize that there is completeness and uncertainties
and all those sorts of issues.
Most studies show that current generations of plants meet
the safety goal. That's a little bit of a funny thing to say since we
don't have a safety goal for each plant, but if you extend the concept,
they meet it. And they usually meet it by a factor of more than 10.
So I would think that if you took out a factor of 10 or 20,
which is not unusual, right, for a credit in evacuation, you would be
close. Whether it would exceed the safety goal, maybe not on paper, but
in reality, it would be close enough so that maybe you would say you
couldn't -- you don't really know, right? That's about as close as I
could get.
MR. KING: The assumptions that went into NUREG 1150 where
they actually modeled emergency planning, they were based upon looking
at some historical information, chemical spills and so forth, how long
did it take to move people.
And they assumed some lag time from the time the accident
started and you notified people, till they actually moved. And people
moved at a pretty slow rate, and they assumed 95 percent effectiveness
of the evacuation. They didn't assume everybody got out.
And then you see the resulting QHO numbers that came out of
that.
MR. HOLAHAN: And basically, if I remember them correctly,
Tom, and you would know better than I do, my recollection is that if you
moved, you didn't get a lethal dose, right?
I mean, if there were any fatalities, it came from those
left behind, not from some fraction of the people that moved.
MR. BERNERO: I'd like to go back. This is long ago, and
the Sandia siting study in the early 80s had the large early release
PWR-1 or BWR-1 release postulated, and then looked at all the sites that
were proposed or actually selected.
And my recollection is that the site remoteness and
meteorology alone gave you, without -- and I don't remember what the
modeling of emergency response was, if any -- but it gave you .1 for all
sites but Limerick, Indian Point I, and Zion.
MR. APOSTOLAKIS: But wait. I thought the safety goal said
that you postulate the individual is just outside the boundary.
MR. HOLAHAN: No, it's the average.
MR. APOSTOLAKIS: So it doesn't matter how far you are.
MR. BERNERO: What I'm saying is, is there defense-in-depth
that comes from site remoteness?
MR. APOSTOLAKIS: No. The way we're calculating the risk
now, no.
MR. KRESS: If you had a societal goal.
MR. APOSTOLAKIS: If you had a societal goal --
MR. BERNERO: I'm not talking about goals; I'm talking about
actuality. Right there, there is a box, Emergency Planning and
Response, and it says .1, .1, and that is the defense-in-depth factor or
share that is provided by emergency planning and response.
And what I vaguely recollect is that there was a calculation
that said the site, the remoteness and the meteorology are such that the
typical reactor site provides you .05 or something like that, and only
Limerick was .25 or something.
MR. GARRICK: Well, another study that I recall indicates
something when there was all this debate about the exclusion zone and
what it should be and what was the technical basis for the 10-mile, of
which there wasn't one, some analyses were done, and it turned out that
on a couple of plant-specific cases that some 95 percent of the acute
fatalities occurred within a mile and a half of the site.
MR. HOLAHAN: There is also a quirk in the way that these
are calculated, and I think Dr. Kress, you had an ACRS staff member do
some calculations not so long ago.
And every one of these calculations basically shows that the
value is .06, which means a 1/16th sector around the plant, and it's
driven by a modeling of where does the plume go and who gets affected
and who doesn't.
MR. KRESS: Right.
MR. HOLAHAN: So, it's a little bit of an odd issue.
MR. APOSTOLAKIS: Go ahead. You've given me an idea for
now. I think you should make the last column 1.
MR. KRESS: That was the suggestion that I made this
morning.
MR. APOSTOLAKIS: Because you're supposed to postulate that
that individual is at the perimeter of the site. So emergency planning
should have nothing to do with risk calculations.
MR. KRESS: That was the suggestion I made this morning.
MR. HOLAHAN: That's not a PRA.
MR. KRESS: That's a --
MR. APOSTOLAKIS: You're saying, I don't care whether you
evacuate.
MR. HOLAHAN: That's not a PRA.
MR. APOSTOLAKIS: The Commission says, put this guy there,
and tell me what is the probability of death.
So we want it both ways. We don't want to have a societal
health objective, but we want to take advantage of it.
MR. KING: The meteorology still affects that.
MR. HOLAHAN: Those are PRA numbers.
MR. APOSTOLAKIS: But it's the way PRA calculates. PRA
takes the actual population, divides by the number.
MR. KRESS: George is saying we need other risk acceptance
criteria besides the --
MR. APOSTOLAKIS: How can evacuation affect individual risk?
MR. HOLAHAN: It' can't.
MR. APOSTOLAKIS: It can't.
MR. HOLAHAN: You can't evacuate 95 percent.
MR. KRESS: In reality, we do have implied other risk
acceptance criteria, and one of them is involved in that.
MR. APOSTOLAKIS: I think we should rethink this .1, without
individual risk.
MR. HOLAHAN: The problem is that you can't evacuate 95
percent of a person.
MR. APOSTOLAKIS: That's correct.
MR. HOLAHAN: They're either there or they're not.
MR. APOSTOLAKIS: If you read the statement from the
Commission, it very clearly says person within one mile. You can't say
I have an average in one mile.
MR. HOLAHAN: Well, average.
MR. APOSTOLAKIS: The definition of the individual risk is
the probability of death of a postulated individual someplace. But
somehow it has been modified over the years.
MR. BERNERO: It's a one-mile annulus.
MR. HOLAHAN: Yes.
MR. BERNERO: The point I'm concerned about is, if what is
looking for a balance between prevention and mitigation, considering the
cornerstones; that there is a part of the emergency planning and
response cornerstone that comes from just being there in Lower Alloways
Township, New Jersey or wherever the plant is, that even if you said you
don't have to have emergency planning anymore, or we'll just give you a
telephone call and do the best you can, that there is a level of
mitigation that comes from siting remoteness and low population.
MR. HOLAHAN: Yes, I mean, that's true.
MR. BERNERO: And in the future, that could change.
MR. HOLAHAN: Yes. As a matter of fact, my recollection is
the study done by Rick Sherry showed that the safest site in the country
was St. Lucy, and it had nothing to do with the population; it had to do
with it being on the ocean and which way the wind blew.
MR. BUDNITZ: I have two comments about earthquakes, and
they're really very different, and you have to listen to them both.
The first is that, for sure, the very large earthquake --
we're talking about the earthquakes that cause trouble for plants, which
are much bigger than any earthquakes we've even had in California.
They're very large earthquakes, and I hope everybody understands that.
The earthquakes at any site, not just California sites, that
are bigger than the 1906 San Francisco earthquake, that magnitude,
they're very large earthquakes.
And for sure, that last column has got to be one for those
earthquakes. You can't count on evacuation for them, so you have to be
very careful for earthquakes, what you do there, and be sure not to be
optimistic.
The second point, and this is from the PRAs:
If you look at the LERFs from the seismic PRAs -- and I have
probably studied that more than most of the people in this room, and I
plead guilty to that -- they come from two kinds of things:
Part of it comes from very large earthquakes, you know,
really, real large earthquakes where it basically knocks almost
everything out, you know, all -- enough is knocked out so that -- and,
by the way, some are recoverable, but it's just that things break.
And those are, you know, these real rare events. But
there's another piece; there is a piece where I will call -- they're not
10-6 earthquakes, they're 10-3 or 10-4 earthquakes. They're still
infrequent, but they're not 10-6 earthquakes, in which you get a 10-3 or
10-4 earthquake, and what causes it is the failure of something else.
If there are two failures of something else, some of them
are non-seismic failures. For example, and a crucial one, is
non-seismic failures of containment isolation, and the second is,
seismic containment isolation, all right?
That seismic loss of containment isolation leads to the
LERF, because you're open, and you know you have your core melt, but
you're -- so in order to make sure that that was not a big, big concern,
in the IPEEE -- I'm proud of having been part of making sure that got
done -- we -- and I was here helping the staff at the time --
We wrote guidance to make sure that every plant did a
specific evaluation of the seismic capacity of containment isolation.
Does everybody remember?
That was the one thing we asked them to do in containment,
separate from the rest. And to our delight, actually, the seismic
capacity experts who were telling us this, told us that, but, you know,
it was very strong.
What we found wasn't a single plant in which that was a
problem. That is containment isolation, the valves, you know, they
turned out to be extremely robust.
People were telling us that, but we found it. Nobody -- no
plant that I can remember found a seismic leak of containment isolation
capacity.
And that then provides you with the additional confidence
that for those infrequent initiators, the contained fission product, you
know, isn't really what I will call the common cause part.
There is still the other part, you know, which is that
earthquake you've got going by the earthquake, but then the rest of it
is an accident, you know, just the usual stuff that happens in an
accident -- the fact that the earthquake occurred 12 hours ago isn't
really what's driving the rest of that.
So that .1, you know, for the contained, is because of the
rest of it, not because of the earthquake, and that's a very important
thing that we've learned from these analyses.
MR. HOLAHAN: There is an analogous thing in fires that
we've found; that the risks are either driven by the very big fire, or a
smaller fire when other things are out of service for other reasons.
MR. BUDNITZ: You mean, a non-fire failure?
MR. HOLAHAN: Yes, right. Now, for CDF, as opposed to LERF,
about half of the seismic CDFs are seismic and non-seismic combinations,
and the other half are all sesimics.
But for LERF, they're dominated by something else; for LERF,
they're dominated by these large, all-seismic failures, and some of it
is seismic -- is random failures of containment isolation.
MR. APOSTOLAKIS: Just to move on, how can we convey the
thought that when we say .1, we really don't mean .1? It's not a speed
limit.
MR. KING: These are guidelines. I mean, this is not
intended to be a risk-based application.
MR. APOSTOLAKIS: I understand that. But if it really has
an excellent containment, modern and so on, and they say, look, mine is
really .4, would you let them raise the 10-4 to 10-3 in response to core
damage?
Is an order of magnitude too much, in other words?
MR. HOLAHAN: The answer is no. Give me a harder question.
[Laughter.]
MR. APOSTOLAKIS: I don't know why you would say no. I
mean, one in a thousand is not --
MR. HOLAHAN: For core melt?
MR. APOSTOLAKIS: I think that comes back to the discussion
this morning that it's not just that you're trying to optimize, you
really don't want to see core damage.
MR. HOLAHAN: Right, exactly. Yes.
MR. KRESS: There is some floor on core damage.
MR. APOSTOLAKIS: How do we send that message that maybe a
factor?
MR. HOLAHAN: We have a subsidiary numerical objective of --
MR. APOSTOLAKIS: These are supposed to be a means, mean
values, right?
MR. HOLAHAN: Yes, of 10-4 for core damage frequency, and we
have a safety goal that says prevention of core damage is one of our
objectives.
MR. APOSTOLAKIS: If we put this in a diagram form and put
shades of gray --
[Laughter.]
MR. APOSTOLAKIS: This is really misleading, .1. Actually,
we're going to the three-region regulatory scheme where there is an
unacceptable region, we talk about between that and the goal, and then
it's fine.
MR. HOLAHAN: That sounds like a speed limit.
MR. APOSTOLAKIS: Variability -- no, for the
unacceptability, yes. Oh, I bet they're going to give you a speed
limit.
Anybody who comes in here with a core damage frequency of
5-10-3, will be arrested. There is a speed limit.
MR. GARRICK: Is there a limitation on --
MR. HOLAHAN: If the term, arrested, means stop their
actions, that's probably correct, yes.
Is there a limitation on the distribution, as well as on the
mean value?
MR. APOSTOLAKIS: Not yet, not yet. They only have the mean
value. I know you guys have thought about -
MR. HOLAHAN: I think if you let Tom finish the discussion,
you'll find out that you're most likely not going to find these numbers
in the regulation.
MR. APOSTOLAKIS: No, no.
MR. KING: These will result in some deterministic
requirement.
MR. HOLAHAN: Right.
MR. KING: The way I envision this will be applied is that
you will take each initiator and you go through and you look at, you
know, given the system that's there are systems that are there, giving
the initiating event, concurrently -- these are sort of aggregate
numbers.
When you add them all up, you want to make sure you've got
the 10-4 CDF -- minus fifth -- LERF, and I wouldn't propose we require
each one to meet a tenth of that, so there could be some flexibility.
Maybe some would meet it very well, and some would be a
little higher. But when you add them all up, you want to have the
aggregate come out to the 10-4, 10-5.
If you go through and you find out the regulations today
don't assure that you can meet these kinds of numbers, that's when I
think you come in and start looking at, do I need additional redundancy,
diversity, you know, additional QA, additional inservice inspections,
inservice testing, EQ, whatever it is to increase the reliability.
And that sort of gets to the --
MR. HOLAHAN: Before you leave this, I think this is a good
exercise. Conceptually, I've gone into this with the expectation that
if you look at the way the requirements were written in the first place,
if there were credible events, whether it was one a year or one in a
million years, we required multiple gold-plated systems to deal with it.
The natural consequence of that is, we provided too much
protection for the relatively rare events, and not enough protection for
the frequent events, okay?
And so, you know, my expectation is that when it comes to
large loca plus loss of offsite power, and these relatively rare things,
you know, we have too many requirements.
When you look at things like reactor scram and aux
feedwater, you have to make sure that you have enough, okay?
And that's generally what I think this is going to -- this
sort of analysis is going to lead to.
MR. APOSTOLAKIS: I would suggest, Tom, that given the
discussion of a few minutes ago, in addition to a goal, given upper
limits, I think it's important information.
And, again, the upper limit can be interpreted the same way
the goal is interpreted, not as a crisp line, but --
MR. KING: You mean an upper limit like this?
MR. APOSTOLAKIS: No, no, that's on a different quantity.
Let's go back to the previous one.
MR. KING: That's on the total.
MR. APOSTOLAKIS: Like let's you talk about anticipated
initiators. My goal is for the event response to prevent core damage of
a 10-4 number.
But anything above 10-3 is unacceptable, too. Two numbers
instead of one, in other words. Because that's the reality today, and I
don't see why we can't reflect reality there.
And if you have a problem with interpretation of 10-3, I
suggest you have the same problem with the 10-4. So these numbers
should not be interpreted as being absolute speed limits.
But at least you send the message, and I think this idea of
an acceptable, tolerable, and don't care regions, is a good one.
MR. KING: I understand what you're saying. I'm not sure --
MR. APOSTOLAKIS: Whether it's 10-3 or something else, I
don't know. That's what we just threw out.
MR. KING: Clearly, if we were going to apply this in a
mandatory fashion to existing plants, what you said would probably have
to be done. But remember, this is a voluntary program.
MR. APOSTOLAKIS: Sure, but even in a voluntary situation,
or even guidelines, it helps to give to guidelines as much as you can,
so people know where they stand.
I mean, the truth of the matter is that the core damage
frequency right now, greater than 10-3 starts also some valid --
management of the attention and so on. And yet we don't say that
anywhere, we just act that way.
What I'm saying is, why don't we say it someplace? If you
have a goal of 10-4 for core damage frequency, but we don't say
anywhere, what we really do.
What we really do is we allow 19 units to be above the goal
and we do nothing, but if anyone comes in here with a calculation that
the core damage frequency is greater than 10-3, things do happen.
MR. KING: Remember what we're trying to do in Option 3;
we're trying to come up with some revised regulations and if the plant
volunteers to meet those, they will now have to have system structures
and components and an operation that does bring them in at 10-4, not
10-3.
MR. APOSTOLAKIS: I understand that, but what I'm saying is,
you will be giving them a more concrete guidance if you follow that
approach, because you're telling them, really what you expect them to
do.
And that's something to think about, or maybe Joe Murphy can
think about it.
MR. KRESS: Let me ask one more question about this table.
If you look at the conditional containment failure probability line, I
contend the lower that number gets, smaller the uncertainty is in the
LERF.
Do you reach a limit of the uncertainty in the bypass, but
you get rid of all the other uncertainties to the failure, early failure
in the mode and the location.
And if then they got down to a level of .01 instead of .1, I
think you're near that minimum in uncertainty in the LERF.
It seems to me like that's a desirable -- since the
defense-in-depth is to deal with uncertainties, unknown and known, it
seems to me like having that uncertainty at a minimum level would be a
desirable thing to shoot for.
MR. KING: I'm not sure why you say the uncertainty would go
down. I mean, you still may have a wide band of uncertainty about it,
even though it's small.
MR. KRESS: It would be minimum. I don't know how big it
would be, because you get rid of the uncertainties due to the failure --
design versus failure location, the location of the containment.
As that conditional containment failure goes down, it means
you've got a bigger, stronger containment with more reliable systems.
MR. KING: You get rid of scenarios that lead to failure.
MR. KRESS: Get rid of all the scenarios that lead to
failure, except the bypass.
MR. KING: But the ones that are left, well, if it's just
bypass, yes, that --
MR. KRESS: Yes, so I'm saying there is some reason to make
that number smaller, and that is because it minimizes the uncertainty in
LERF.
And I don't know if that's -- I just thought I'd throw that
out as a concept.
MR. KING: I hadn't thought about it.
MR. APOSTOLAKIS: Did you say you will think about it?
MR. KING: I said I had not thought about that aspect of it.
MR. KRESS: That was in my talk this morning. That was the
red herring.
MR. APOSTOLAKIS: Did you reject my suggestion, or you will
think about it?
MR. KING: I'll think about it.
MR. APOSTOLAKIS: Good.
MR. HOLAHAN: I believe he's thinking about it right now.
[Laughter.]
MR. KING: All right, I think we talked about most of this.
We would use mean values.
In the table we show that numbers is associated with full
power, but we'd also apply this similar concept to the shutdown
condition as well.
And then my last slide, okay, what do we do with this
working definition? As I said, the idea was to take each initiating
event and follow it through to see if you can meet those risk goals or
what you need to do to meet the risk goals.
We're also going to take a top-down look where you take
these four cornerstones and line up today, what's in the regulations,
Reg Guides, SRPs, under each of those and take a look at the balance in
terms of there are probably a lot of things that affect reliability and
availability and redundancy and diversity of systems to respond to
initiating events.
Do we need similar types of requirements when you talk about
containment? Is there more we should do under prevention? What's the
balance when you come down vertically at each of the cornerstones?
So, that's sort of the concept that we're going to apply in
the application of that table.
Again, I just want to say that in terms of wrapup, what
we're talking about this is the basis for looking at the regulations.
We're not talking about putting these numbers into regulations; we're
talking about using these to come up with some change in the
deterministic requirements.
And we're not talking about putting in the regulations, a
rule or a definition of defense-in-depth. I think it's a philosophy
behind everything that's going to end up going into the rules.
MR. KRESS: I think the table itself is almost a definition.
MR. APOSTOLAKIS: Yes. Okay.
MR. KRESS: I like the approach myself. It's pretty much
what I was advocating this morning, I think.
MR. APOSTOLAKIS: This is the pragmatic approach. Very
good.
MR. KRESS: Very good. We appreciate that very much.
MR. APOSTOLAKIS: Based on what we saw today, it's very
good.
MR. KRESS: I don't know how we'll apply that to Yucca
Mountain, but --
MR. APOSTOLAKIS: The staff refuses to take it seriously,
but maybe one of these years.
MR. KRESS: Well, it's a way to handle uncertainty. I'm not
sure how we apply this to Yucca Mountain, but --
MR. APOSTOLAKIS: I think it's a different beast.
MR. KRESS: I think it is, too.
MR. APOSTOLAKIS: I think the fundamental difference is
time, the time scale.
MR. KRESS: We're due for another break. Does anybody need
one?
MR. APOSTOLAKIS: Yes, we do.
MR. KRESS: Another 15-minute break.
[Recess.]
MR. KRESS: The next item on the agenda is to hear some
words from the NEI and the industry, and from Westinghouse, so I'll turn
the floor over to you, Alex, and let you introduce the subject and
introduce the people.
MR. MARION: Good afternoon. My name is Alex Marion, and
I'm the Director of Programs at the Nuclear Energy Institute. I
recognize the time is late, but I do have a few brief comments to talk
about some of the things I heard today relative to the application of
defense-in-depth philosophy to operating plants.
But I would like to introduce Rodney McCollum, who is the
Project Manager at NEI involved with high level waste management, and he
has a few comments he would like to make on the application of that
philosophy to the Yucca Mountain Project.
Rodney?
MR. McCOLLUM: Do you want me to go ahead and do that first?
MR. MARION: Yes, please.
MR. McCOLLUM: I've been working for NEI now for a little
more than a year, specifically to follow Yucca Mountain and related
issues, so I have been attending meetings such as this one, and hearing
discussions such as I heard today for most of that time.
I always find these discussions very interesting and very
intellectually challenging. I think this one was definitely no
exception and perhaps even a little bit too much so on the
intellectually challenging part, but that's how I learn things.
I also feel it's a very important discussion, and it's
certainly a very timely discussion because the nation is entering into a
critical window of decisionmaking opportunity here where over the next
18 months, our leaders are going to be called upon to make a decision
about the future of Yucca Mountain.
And one of the things that will weigh most heavily in that
decisionmaking process is the topic of uncertainty that's been discussed
a lot today.
How will the decisionmakers, relying on the Nuclear
Regulatory Commission, the ACNW, the TRB, and all of the political
forces that come to bear, how will they view uncertainty?
And uncertainties will exist; that's really the only thing
that is certain. In fact, if it's good enough science, every answer
will simply generate more questions, it will bring up more
uncertainties.
And, therefore, because these uncertainties will inevitably
exist, the decisionmakers need to have some tools in place that will
allow them to address this.
And we firmly believe that the DOE, in the viability
assessment, and the NRC in the draft Part 63, is giving them these
tools. We feel that in referring to what Christiana was talking about
earlier, the way multiple barriers are being interpreted, that it is a
qualitative and not a quantitative argument, and that it should be up to
DOE to make the safety case. We feel that's appropriate.
We are concerned to the extent to which at this point,
having seen what's been done by both the DOE and the NRC staff to
develop those tools, what could be gained by inserting knowledge on the
reactor side from the reactor notion of defense-in-depth into the
repository process?
We've had a lot of discussions along this line with our
friends in EPRI included, and perhaps the best way for me to relate what
might happen if we were to bring these things in is:
I, once a upon a time, was a Branch Chief of Nuclear Safety
for a DOE operations office that had responsibility for a lot of very
unique, one-of-a-kind, non-reactor nuclear facilities. We had a couple
small reactors. This was the Chicago operations office so we're talking
about the Brookhaven's, the Argon's , the Princeton's, et cetera.
And I was in that position at a time when DOE was coming out
of its post-Cold War cocoon of beginning to realizer that it needed to
have some credible nuclear safety requirements, a regulatory structure
in place that it didn't have before when it simply did what it knew was
right or thought was right.
And doing son, they naturally looked to the best source of
expertise for that kind of a regulatory structure, and that was the NRC.
So the DOE made a lot of requirements that were first under the guise of
DOE orders, and later became -- a couple of them became rules, DOE
rules, that basically took NRC regs that were intended for the reactor
world, and put DOE order numbers on them and they were to be applied to
these non-reactor nuclear facilities.
Once that happened, I found myself spending a lot of time
trying to fit square pegs into round holes, and trying to explain why
the square pegs wouldn't go in the round holes. That they just don't
fit, never quite seem to be enough of an answer.
I saw a lot of effort being made at the five National
Laboratories to address all those misfitting pegs that didn't contribute
to their safety cases, and, in fact, just detracted from it.
I was very appreciative to hear what Dr. Garrick said
earlier about arbitrary thresholds and subsystem requirements that
detract focus from risk. I know from experience that that that does,
indeed, happen.
And I think we have a pretty similar situation here with
Yucca Mountain, because Yucca Mountain would be a very unique,
one-of-a-kind, non-reactor nuclear facility.
I think that the differences between Yucca Mountain and
reactors are so fundamental, it really becomes almost impossible to try
and draw from reactor defense-in-depth to multiple barriers in the
repository site.
A couple of those things have been mentioned, a couple of
others, I would mention: Of course, obviously you have more active and
passive barriers at Yucca Mountain, whereas you have more active,
engineered and more engineered features at a reactor.
Yucca Mountain has one common failure mode, really, a
two-part failure mode. It's water and time. And it's really a question
of where you are on the radioactive decay curve when those things attack
each of your barriers.
There are different timeframes to be considered. In
reactors, fractions of a second can be important; in repositories,
millennia are what's important.
You have a safety case in reactors where you're trying to
figure out where to best apply PRA; in a repository, your safety case is
a PRA.
You rely on humans to operate reactors; your expectation for
the repository is that once you seal it up, except for potential human
intrusion, humans won't be involved at all.
And probably the most important distinction that allows you
to treat uncertainty in a fundamentally different way at a repository
would be that you have this performance confirmation period. You have
not a two, but a three-stage licensing process.
And this is a 50-year period where you have a chance to
constructively address those what-if-we-were-wrong questions.
You don't have that at a reactor, and I don't think any
utility would want that, although some time felt they were approaching
that.
But it does give you an opportunity, and it does give the
decisionmakers to say, when they are faced with uncertainties, that
here's what we know, and here's what what we know tells us, and then
here's what we need to know before we close the thing, and put in place
the right research program that can answer those questions.
But you can't do that in the reactor world. So, given that,
and having heard the discussions -- and this is another one of the
things we appreciate where the staff is going with multiple barriers. I
was very thankful to see Christiana's presentation entitled Multiple
Barriers and not defense-in-depth.
We wonder -- and this is kind of the conclusion of the
discussions we had internally -- whether defense-in-depth is even an
appropriate term; whether it would be more appropriate to call what
you're doing at Yucca Mountain multiple barriers and call what you're
doing in the reactor world, defense-in-depth, and not even try to mix
the terminology.
It could only lead to a confusion in expectations, and as I
mentioned before, you know, we think the expectations are evolving well
for Yucca Mountain. We think that Part 63 will answer that.
We think that from what we've seen at EA, and from DOE's
draft Environmental Impact Statement, they'll be able to say that when,
you know, the final dose, if it's 1.3 millirems, 10,000 years from now
or whatever it is, that that dose is a function of a performance
assessment that includes a dry climate and includes a thousand feet of
rock to keep the water out of the repository.
It includes a lot of things in the repository, some of which
are engineered, and includes another thousand feet between the
repository and the water, and it includes things in the water that
retard the movement of radionuclides.
And it includes a sparsely populated area that keeps people
away from even those moving radionuclides. And, of course, the DOe will
have looked at a certain amount of variations and been cautious and
reasonable in looking at each one of those barriers. It will assume a
somewhat wetter climate. It won't take credit for the features of the
rocks that it doesn't understand as well as it understands some others.
When I visit the folks -- and in the year, I've had three
tours of Yucca Mountain now, and I talked to the scientists in the
tunnels and hear them talk. I appreciate what Dr. Levinson mentioned
about some uncertainties are not bad.
They are tending to find out things about the rocks that are
good news. And they will do that during the performance confirmation
period.
But based on what they know, they can make a case that that
1.3 millirems or 13.2 millirems, or whatever number it is less than 15
or 25, is a function of a number of things. And those things all
contribute to it.
And in that respect, it need not be much more complicated
than that. They will have then answered what Congress has asked for in
terms of multiple barriers, and the NRC can and should, in accordance
with its regulations, look very hard at that and make sure it's
credible, that it's believable before the Commission says to the
decisionmakers, we think this is sufficient, which is the sufficiency
comment component of the site recommendation.
Then we go on to the next stages in the process, and we
continue to look at it, realizing that the scientists will never stop
asking questions, and that every one of those questions will bring into
the proces, more uncertainties, and that's not a bad thing.
So, you know, I'm very encouraged that these discussions are
occurring, and I learned a lot from them, and look forward to this going
forward.
MR. MARION: Are there any questions of Rodney before I make
a couple of comments?
MR. APOSTOLAKIS: I don't so much have a problem with the
regulations, the way Christiana presented them. It's really the quality
of the performance assessment that would be of concern to me, given the
time scales we're talking about and the uncertainties that are involved.
And I still don't believe that the model uncertainties are
completely addressed. Even in WHIP, you know, there was primarily
parameter uncertainties. At one point they had two different models for
something relatively minor. I don't remember what it was.
They said, okay, we'll put a weighting factor of 1/3 to
this, and 2/3 to the other, and just add them up. But I think the
uncertainty is a key issue here.
MR. McCOLLUM: Oh, they clearly are. As I mentioned,
they'll be the major thing weighing on the decisionmakers.
And that is why, in demonstrating multiple barriers, DOE
needs to talk about what each of those barriers mean to the safety case,
and what is the meaning of those uncertainties?
And they're starting. And every time I have heard DOE
present on this subject now, dozens of time, and the story gets better
every time, that the science was always there, I believe. It's been
there since the VA.
But it's being able to talk, and it can't be completely
quantified. It shouldn't be. But to be able to talk about the relative
importance, what does that uncertainty mean, what if the climate does
get wetter? Have we looked at that?
Have we been appropriately cautious in what we've assumed
the rocks do for us, and what we've assumed the rocks don't do for us?
And so that if some of those uncertainties turn out to be
bad, are there offsetting things? And it's really going to be a
challenge in the next 18 months when we have this decision before us,
for that to be discussed.
And I have also heard Dr. Garrick talk a lot about plain
english, and that's why that's so important. Because those things may
be buried in the performance assessment in any number of ways, but if we
can't bring them out and discuss them in plain english so people
understand that that's what this means, that's what that means.
And because we know what all these things mean to the safety
case, we can say this is a good place for a repository or not. And we
can make a decision.
MR. GARRICK: George, I think that the Committee kind of
shares the concern for the TSPA. We know that in the early days of the
PA for WHIP, there were many, many problems, and through another
Committee, I was directly involved in that.
And I saw a major change. The big difference there over
Yucca Mountain is that except for human intrusion, there was geologic
containment at WHIP.
And the only way WHIP could get in trouble was through some
rather arbitrary human intrusion scenarios. Of course, we don't have
that luxury on Yucca Mountain.
MR. APOSTOLAKIS: Right. The other thing that we did that
you guys may find disturbing is that later on, I believe, 60 hypercube
simulations. All 60 of them were below the goal, which brings us back
to your comment, what if it is 5X?
What if Yucca Mountain, 58 of them are below and two are
above? That will create an interesting interpretation of the
regulations.
And why should all 60 be below? Just because it happened
there?
Now if you think of the state of knowledge on uncertainty,
the whole distribution is below -- I mean, the two high percentile, so
that -- anyway, these are not directly related to the subject matter.
MR. GARRICK: It's a good comment.
MR. MARION: Thank you. I'd like to make a couple of
comments about the operating reactor side.
I found Dr. Murley's comments this morning kind of
interesting. Having worked at a nuclear utility for 15 years, it sure
felt like defense-in-depth was a regulatory requirement at times.
[Laughter.]
MR. MARION: But I decided not to challenge it.
MR. APOSTOLAKIS: It was a voluntary requirement. We have a
lot of those.
MR. MARION: But I thought he made an interesting comment
about -- or a caution, I should say, as I interpreted it, about applying
risk insights to remove or otherwise eliminate barriers.
I think we need to be very careful, and I think that's an
appropriate cautionary statement. However, I think with risk insights
and operating experience, we can better define what's important in the
implementation of the very elements, specific elements of those various
barriers of protection, specifically in the area of emergency planning.
I believe we're very close to the point of providing a case
to reduce the exclusion zone, based upon the robustness of the designs,
as well as the analysis supporting the advanced reactors.
And there are opportunities. We're not offering to get rid
of emergency planning as a concept, but better define it with the latest
intelligence and knowledge base we have.
And I think that's consistent with the comment that Dr.
Budnitz made about the evolution of knowledge to better focus on
barriers of protection, integrating operating experience and new
analytical techniques.
And I think we need to keep that in mind and take advantage
of those kinds of opportunities when they present themselves.
I think the example that Dr. Apostolakis used on the fire
analysis and the element of smoke and uncertainty associated with it is
an excellent one in terms of applying an engineered approach to address
uncertainties.
And then when knowledge comes to bear and the analytical
techniques improve to better reduce the uncertainty in the area of smoke
propagation, et cetera, then you can make adjustments along the way.
And I think those were excellent examples, and we're in full
agreement with those concepts and processes. And in NRC staff's
presentation this afternoon, I was sitting back there with Biff
Bradley's, the project manager at NEI directly involved in
risk-informing Part 50 and these PRA risk insights, applications, et
cetera.
And he indicated to me that we're in full agreement with the
approaches. And I think, between the industry and the NRC, we're in a
good position where we understand the importance of striking a balance
between the deterministic thinking that's made this industry very
successful within the defense-in-depth philosophy, and applying that in
some balanced way with probabilistic techniques and approaches that we
have today.
And from what everybody tells me, things are going well in
terms of the applications of risk-informed regulations, but we do have a
lot of work ahead of us.
And I just want to caution everybody that we want to be
careful not to limit our thinking or limit our approaches such that when
new knowledge or when new analytical techniques come to bear at some
time in the future, we can still take advantage of those and improve our
knowledge and understanding.
This is the defense-in-depth philosophy balanced with
risk-informed approaches, and is very fundamental to our thinking for
regulatory reform, more specifically in the area of risk-informing the
Part 50 regulations.
So we think it's very important to work hand-in-hand,
shoulder-to-shoulder, so to speak, in a complementary way with the NRC
staff, and to strike this balance and determine what we need to do with
future applications of the current state of knowledge.
And that completes the comments that I have. Are there any
questions about anything I said about operating plants, or that Rodney
said?
[No response.]
MR. MARION: Okay, with that, I'd like to introduce Gary
Vine from EPRI, who is going to take a few minutes and provide you with
a general overview of the defense-in-depth philosophy as it was applied
in the design requirements for advanced reactors.
I think you will find that informative and beneficial. And
he will be followed by Brian McIntyre from Westinghouse, who is going to
specifically discuss the application of that philosophy in the AP-600
designs.
MR. APOSTOLAKIS: One of the victims of defense-in-depth.
MR. MARION: We were going to bring that up a little later,
Dr. Apostolakis.
MR. APOSTOLAKIS: Perhaps the only one still alive.
[Laughter.]
MR. APOSTOLAKIS: While these are getting settled, somebody
said this morning that there may be a perception out there that we're
using risk-informed regulatory approaches to remove barriers, to remove
regulations and requirements.
I think it's important to say that where PRA indicated that
additional requirements were needed, the Agency acted immediately. And
in the last 20 years, in fact, the eagerness of the Agency to add
requirements based on PRA insights created a somewhat hostile view
within the industry towards PRA, because PRA was used only to add
requirements.
So the fact that now we are finally looking at removing
some, should not be misconstrued as the Agency using PRA to remove
requirements. We have already added a lot, okay. That's in case
anybody reads the transcript.
MR. KRESS: Thank you, George, I think that was well said.
MR. VINE: Good afternoon. I'm going to start off. My name
is Gary Vine. I'm from EPRI. Unfortunately, I didn't have the benefit
that Alex and Rodney and Brian did of all the prior discussions. I got
here about 4:00 from another meeting in Tower I.
But Alex does tell me that a number of the points that I
intended to cover have been covered in some way, and so I'm going to try
to focus only on either new material or kind of an industry perspective
on some of the things you have heard from the NRC side.
I'm going to probably skip over the first slide or two. The
only key point on the first slide is simply that we did in the ALWR
program, which goes back 10-15 years now, fully embrace the concept of
defense-in-depth.
And we did that in a traditional way. I think we didn't use
the terms that you've been discussing today, structuralist and
rationalist models, but we pretty much followed the traditional
structuralist approach.
I also have a slide on ALWR policy statements, and I
intended to go through two or three of them in some detail, and I'm
going to skip that as well.
I have a high-level brochure document that provides a two-
or three-sentence description of each of these policies, some of which
have a bearing on defense-in-depth, and I'll just leave that for you to
look at.
Moving on to Slide 4, just a couple of key points: It's
very important to recognize that public health and safety is important
to both the NRC and to the owner/operator of a plant. In fact, the
owner/operator has the primary responsibility of protecting public
health and safety.
So his interest in safety is just as high as that of the
regulatory. Where the difference lies in the way we fundamentally
approached establishing design requirements for advanced reactors is in
the investment protection side.
That is where the industry has an equally high interest in
preserving their investment. But the NRC doesn't have a comparable
interest.
And so what that forced us to do was to make a lot of
tradeoffs as we were trying to optimize prevention mitigation
decisionmaking where the industry's interest was naturally always to
achieve safety as early in a sequence as possible.
We always wanted to prevent an accident or actually have a
robust enough design so that we wouldn't even get into an accident
sequence before we had to get into questions of mitigation.
We also found when we had a fresh sheet of paper and we
could look at these decisions, that almost always -- not always, but
almost always, when you had a particular sequence you were trying to
drive down or improve the safety for and you had a mitigation option and
a prevention option to do that with, the prevention option was usually
less expensive.
So there were a lot of incentives on the industry side to
truly tackle areas where we wanted to achieve improved safety by doing
it on the prevention side. Of course, this, as you can tell, created
some friction between the industry and the NRC, on occasion on certain
issues where the thought was that we were maybe not maintaining the
proper balance in defense-in-depth.
We maintained a strong commitment to mitigation as well.
Requirements for containment, for example, are just as strong or
stronger for advanced reactors than they are for current plants.
But as we pressed to achieve improvements on the prevention
side, there came some questions about balance.
Explicit consideration of severe accidents via a safety
margin basis, that's a very important concept which I think is probably
worth some discussion. I think there were some understandings in kind
of a process way in the program with the NRC that have stood the test of
time.
We fundamentally committed to the licensing design basis as
it was captured in Part 50. And we did not, with just a very few
exceptions, try to make any changes to the regulations.
The only example on this schematic where we tried to make
some improvements in the regulatory basis in the licensing design basis
side was in improving the source term that was analyzed in the licensing
case.
But we pretty much bought into the entire licensing design
basis approach as, quote, the "formal speed limit" for design.
But we were very careful in defining very separate and
distinct from that licensing design basis, the way we would approach all
other safety questions and primarily all questions associated with
severe accidents.
In this area, there were differences in almost every aspect.
We approached it, first of all, from a standpoint of a much more
risk-informed evaluation of the plant's overall performance.
Second, we insisted that we use best estimate analysis
methods, models, and so forth in addressing those issues.
Third, we proposed and the NRC accepted, the concept of the
industry pretty much driving the specific design approaches to address
severe accidents, and get the NRC to provide an overall approval to the
approach that we took, as opposed to agreeing on detailed prescriptive
requirements that would then become part of the licensing design basis
or some formal regulatory requirement for this right side of the
equation.
So the industry really drove this. We decided how we wanted
to satisfy the Commission's concerns about severe accidents, all the
research findings, the Commission policy statements and everything else.
The NRC then provided an SER on these utility requirements,
and then the vendors had a clear picture of how they had to achieve
basically what they had to do to know that they would have regulatory
approval in this area.
There were a number of areas, even though we pretty much
approached things in a conventional way with regard to defense-in-depth,
where we kind of pushed the envelope, and what I'm going to cover now
are some areas where I suppose if you get to the definitions you're
using now, where we used a more rationalist model approach or a more
risk-informed approach to the way we did business.
First of all, let me jump back to Slide -- yes, this is the
right slide. I'm sorry.
Major alliance on PRA and the process: It drove our side,
the industry side, very significantly. We made major plant policy
decisions and major plant design decisions based on findings of the PRA.
The regulatory side used PRA much more just as a
confirmatory tool, as opposed to a decisionmaking tool. One exception
which Brian will get into is the way we dealt with the regulatory
treatment of non-safety systems for the passive plants.
But beyond that, the regulatory side was pretty much a
confirmatory process. We established quantitative safety requirements
on the industry side that well exceeded the regulatory requirements.
And the idea here was that we wanted assured license ability
by knowing we had significantly exceeded what the regulatory
requirements were going to be in the area of safety.
I list our two quantitative safety requirements, and these
were requirements; they weren't just targets: The designers had to have
a CDF much less than 10-5, and they had to address mitigation by meeting
a goal of ensuring that whole-body dose would be less than 25 rem at the
site boundary which is about at a half mile as we defined it for all
sequences with a cumulative frequency of greater than 10-6.
You will notice that these two prevention and mitigation
goals are not coupled; they are decoupled, which gets to my final point
on that slide:
We did oppose the concept of coupling these independent
layers of defense-in-depth. We opposed the concept of a CCFP. We
didn't win that argument, but we do believe that CCFP is not an
appropriate means of enforcing a defense-in-depth approach because it
couples things that should remain independent.
Because one is set by design, you end up forcing the
operator or the designer to make less than optimum, sometimes dumb
decisions in having to reduce the safety of the plant in order to
maintain this spread of a factor of ten between prevention and
containment performance.
And there are -- you can go through some scenarios down on
the low probability events where the imposition of a CCFP becomes even
more ridiculous.
So we felt that that was an inappropriate approach and still
do.
Regulatory stabilization: I already mentioned assured
licensability by exceeding the regulations wherever feasible. This was
an important concept to us, and we've faced some problems in dealing
with the NRC on this because we wanted to assure significant and visible
and demonstrable margin between the regulatory requirements and actual
design performance and operational performance.
And there is just a natural tendency on the part of the
regulator to say, well, gee, since you're that much better, let's just
change the speed limit so we're a lot closer to where you are.
Well, that creates huge problems for us, because it
eliminates that assured licensability. And so we think that the
regulatory requirements ought to be based on the first principle and the
bases upon which NRC makes its regulations on adequate protection and so
forth, and allow the user of those regulations to exceed them and not
have that difference gobbled up into regulation.
There were a few case where we attempted to change the
regulations. We would propose in some areas -- these are usually some
modest areas -- we didn't go after things like large break loca and so
forth.
We did propose some changes to the regulations, and some of
them were accepted and some of them were not. Some examples that were
talked about were: More realistic source term; elimination of the
operating basis earthquake and going only with the safe shutdown
earthquake; changes to hydrogen regulatory requirements.
This optimized or simplified emergency planning that Alex
mentioned earlier, and so forth.
And the last slide I think is more just personal views as we
look back over the ALWR program and how we approached defense-in-depth.
We think that looking forward, that risk-informed regulation and
specifically a more -- an approach to defense-in-depth that is closer to
the rationalist model is really important to the future.
We are going to have to find ways to reduce the capital
costs of ALWRs, and we believe that can be done easily and safely, and,
in fact, probably in many ways improve safety.
But it does require more flexibility on the regulatory side,
and a rationalist approach would allow for that.
Further, I don't see how the NRC will ever be able to
license a reactor design such as a high-temperature gas reactor, unless
there is a more flexible approach to defense-in-depth, including
something similar to the way you've characterized this rationalist
model.
I think the die is cast; the rationalist model is ultimately
going to become the future approach for regulation, and I don't think we
need to be afraid of that. I think there are really no downsides to
that model, if, in fact, it's done prudently and carefully and safely,
and done with the things that are already pretty much established in
regulatory policy, namely, that it's not going to be a risk-based
approach; it's going to be a risk-informed approach.
There will be a balance, there will be still consideration
of defense-in-depth, there will be clear use of engineering judgment and
care and so forth in how you approach risk insights.
And just finally one comment on U.S. leadership: The ACRS
paper on defense-in-depth mentions a couple of INSAG reports, and it's
true that in the international arena, there is a much more rigorous
definition, a much more traditional and formal approach to
defense-in-depth.
And I think there probably will be some resistance on moving
quickly toward, say, a rationalist model, internationally, and the
reason is that I think there is a concern by IAEA and probably some of
the industrialized world regulators that if you move too quickly, you're
going to find some countries, third-world countries, people who don't
have the maturity and infrastructure, safety culture, and so forth, that
if you move to quickly in optimizing defense-in-depth philosophies, that
you're going to remove some significant safety protection.
And so there will be some desire, I think, in the
international community to move slowly and to make sure that, especially
for those who define defense-in-depth very broadly -- and I've seen it
defined this way to include things like safety culture and your
infrastructure and your regulatory infrastructure and so forth -- that
those things still are not subsumed under a risk approach, and you don't
make them subservient, but you still keep them at a high level.
MR. APOSTOLAKIS: It's important, of course, to note that
terms like quickly and slowly are relative.
MR. VINE: Yes.
MR. APOSTOLAKIS: And that the first major risk assessment
in the United States was published a quarter of a century ago. So for
us, it's not too quickly.
MR. McINTYRE: My name is Brian McIntyre, and I'm the AP-600
License Manager. I'm two things: I'm the practical application of what
Gary just talked about; and I'm also, I think, the most recent example
of where the rubber has met the road with the staff on defense-in-depth.
And this is -- we have really talked at lot about this, I
think, earlier, that it's more than the three barriers that was
originally put in to deal with uncertainties.
What I had written down is that we are never sure exactly
what it was. And after sitting through this morning, I think it's that
everybody was more or less sure what it was, and it was whatever it
needed to be, and it was sort of a flag that we all wrapped ourselves
in, both sides, I mean, the industry and the regulators.
But we never quite knew when enough was enough, and I'll
talk about that at the very end of this. And now it's clear that we are
moving towards some sort of a balance between the things that are on the
top there and the risk-informed information.
In the AP-600 case, for us, I broke this down into two
things, something that I called the unquantifiable aspects -- and this
goes beyond just power reactors. For us, it was a design philosophy.
Now, at the bottom I have some things that are quantifiable.
We actually, since were starting from scratch, weren't
trying to figure out how good the plant was; we were more interested in
how good we could make the plant. And you really take a different
approach if that's what you're doing.
And our design philosophy looked at -- people have kind of
wondered about passive plants -- that we have multiple levels of
defense.
And the first thing that you see there is that it was
usually a non-safety, active feature. We have a passive plant, and that
made the staff -- these are my words -- made them a little bit crazy.
Because, as you're going to try to address your transients
by using non-safety systems, this is as a first shot, yes. And then
almost the backups would be the passive systems which were the safety
systems.
And if you want to look at what this looks like, the next
figure or the thing that actually is the figure, this is -- and we did
this for a number of transients where we went through and we looked.
On the left side is a current plant -- and I need to put my
glasses on to see this -- that what they would do, their SSAR safety
case is that they would automatically actuate their high-end safety
injection, their aux feed; they'd isolate the steam generator, and
they'd start to cool down and depressurize, and that was their safety
case.
And if that isolated the leak, that was great, and if not,
then they had a non-safety case which would be in their emergency
operating procedures someplace, and they had a couple of things that
they could do. If not, then they were at a core damage situation.
For the AP-600, if you take a look at our top block, which
is the non-safety case, really, it's the same things that in a
traditional plant would be their SSAR safety case, except we had now
made these systems non-safety-related, which was really a change.
And there were some long discussions we had with the staff.
Gary talked about regulatory treatment of non-safety systems, and I'll
talk a little bit at the end about how we did approach that.
And then we got to our safety case, all these passive
features of automatically actuating the core makeup tank; the PR/HR heat
exchanger, which was basically replacing the axillary feed or startup
feed system in the safety case; the CVCS.
We'd isolate the steam generator and start the passive
containment cooling system, and if that isolated the leak, then that was
our safety case. And that's what we basically met the safety
requirements with.
The important thing to look at in the AP-600 is that down
below it there were then two or three other options that the guy could
go through. And this was important because, you know, we could have
just really stopped at the top, at the safety case, and with the top
two.
For various reasons, because these features were in the
plant, that they all managed to work together, and as a result, we got
really some good PRA results. But this, to us, was what we considered
to be the defense-in-depth.
We also used the PRA as the design tool. And that's like a
lot different if you're trying to figure out how good you can make the
plant, as opposed to how good the plant is.
We did a total of seven PRAs on the AP-600. And we weren't
dong them just to make the PRA different; we were doing them because
we'd made the plant different.
We'd run the PRA, we'd find out where the weak spots were.
This is where you're looking for the unduly -- not unduly dependent on
one system, so we were looking if something really stuck out, and we'd
go back and we would make the system better.
There was a lot of design with arguments even between the
risk analysis people and the designers. We actually got better PRAs as
a result of that, because sometimes the PRA people didn't understand
exactly how the system should have worked.
In a lot of cases, the designer said, you mean that if this
fails, then that's the result you're going to get in the PRA space. And
we made some significant changes to the plant as a result of the PRA.
We went through a lot of just discussions, review,
understanding the results. We looked at some of the backup slides.
When we got to reviewing things to see how we would expect
the systems to work, this is just one example. This is the PR/HR heat
exchanger. How would it fail? We would then walk through the various
things and decide what we needed to either to try to fix or to model or
not model in the PRA.
We went through each one of the various items, for example,
for the inadequate IRWST water level, and then that was broken down to
look. Are there things that we could fix, are there things that we
needed to do better?
I mean, we really did chase this design down to look for
ways that you could improve the plant.
And this is a philosophy, so it's not just applicable to an
AP-600 or a BWR or something like that. But if you think like this and
you bring this approach to the design and bring whatever it is from a
design to actually a facility, this works.
This is another way to look at defense-in-depth, but there
is no way that we could put a specific number on what we got out of
this.
We also looked at shutdown operations. We looked at low
power operations. We pretty much covered the waterfront.
One of the bullets on the previous slide was that for
systems that were more -- or for events that were more likely,
initiating events, we had more backups.
For steam generator tube rupture, a reasonably likely event,
there are five or six different thing you can do. When you get down to
the more unlikely things like large loca, you don't have quite as many
options of things that you can do, so we tried to focus our efforts on
the things that are more likely going to happen.
Also, one of the big reasons we were doing this is the big
push from the industry was this investment protection concept. If
something is more likely to happen, then we don't want to lose the plant
as a result.
We want to have things that the guy can do. He might have
to clean the plant up, but he won't lose the plant as a result of it.
We looked at a much wider range. We didn't restrict
ourselves to the design basis transients. We really looked at multiple
steam generator tube rupture, not willingly, but we looked at multiple
tube rupture, because this was a case of the staff's concern which was,
okay, you guys met the design requirements, but do you fall off the
table somewhere?
And the staff went to the extent of, after we had completed
our testing at the Oregon State facility, which was a quarter scale
model of an AP-600, it was a low pressure facility, but they went out
and ran beyond design basis transients there to look to see if there was
someplace that we hadn't tested that they could look to see if we were
going to fall off the table.
And the conclusion was, no. It was a surprisingly robust
plant. I mean, we'd been telling them that for a long time, but
eventually, it became obvious.
We also looked at a broad range of initiating events. And
as I said, this was to look beyond where you would normally go.
And, again, we're trying to figure out how to make it
better, not how good it is. And it's almost like IPEE and IPEEE, except
we could make the changes, because it's quite easy really to make a
change.
If you look at the quantifiable aspects, we ended up with
really a nice low core damage frequency. I'll talk about the focused
PRA in a second.
For large releases, what we were required to do by NEPA was
to look -- and SAMDA, if you're not familiar with those, those are
severe accident mitigation design alternatives.
I look at it as we had to explain to the staff, why we
didn't do what we didn't do. It turns out we're not really good at
documenting that, so we went through and have to figure out, why didn't
you make these changes to the plant, and you have to look at that on a
cost basis, the cost/benefit basis.
And it turns out there was nothing that we had to add,
nothing that could be cost effective when we finished the design of the
plant.
Our PRA results: This is looking at two things, the core
damage frequency and the large release frequency. It's the at-power and
the shutdown events.
The baseline PRA is pretty much a traditional PRA. It has
the safety systems and the non-safety systems in it.
As part of our ongoing discussions with the staff and the
regulatory treatment of non-safety systems, we had an approach proposed
by the industry, accepted by the staff, that if this plant was so good
that we could go out and meet the safety goals to 10-4 and 10-6, with
only the safety-related systems, then these non-safety systems that were
in that top tier or the first thing that the operator might actually do
to the plant to mitigate an accident, then they wouldn't require any
additional treatment.
And it's a sensitivity study, but we went back and looked at
it, and we showed that without the safety systems, we still, in the core
damage frequency area, we quite handily met the safety goal. In the
large release, well, it was close.
And the staff's concern was, well, uncertainties in the PRA,
we're not so sure about this, and we went back and forth and back and
forth and back and forth and back and forth.
And finally, it just went forth, and we said, okay, to move
this forward, we would put some administrative controls on certain
systems. And so we actually have in the AP-600, safety-related,
non-safety-related, and then there are these RT&SS important systems
that we have availability controls. So we're actually --
I would actually look at this as beyond risk-informed. It's
almost risk-based, this sort of an approach that you have a milestone
that you're trying to meet, that if you do this then you will be okay,
and if not, then you'll have to do some things to make it so.
And at the time, this was quite novel. It was much for
discussion, but it certainly is, I think, a case of how defense-in-depth
can come and be played through and be applied to a facility.
One of the reasons that you're here -- and this is sort of
-- if you look at Tab 1 in Jack's book of defense-in-depth discussions,
it was that we had a long discussion with the staff on containment
spray. The AP-600 does not have a containment spray.
Well, it does have a containment spray; it didn't have a
containment spray. Let's put this in perspective and in the proper
tense here.
And we didn't think that we need it, and it got back into
arguing about the uncertainties and the PRA and the models. In the end,
we ended up, as I said, with a containment spray system.
If you look at it from a risk-informed perspective, the --
and this is a slide that was put together by an ACRS fellow at the time
back in June of 1997 when this discussion was going on.
It gives you an idea of where our risk contributors are.
And for this plant, if you look at what a containment spray would help
you with, it's not going to help you with the bypass events or with the
early containment failure. It might some -- it would help you with the
containment isolation failures.
A presentation that I made to the staff had -- and you
haven't seen this one, George, but it has the more quantified basis of
what we would expect to get out of the spray.
And the spray here where it says low flow, it's lower flow
than the spray that we actually ended up putting in the plant. This was
a study that we were doing at the time to figure out how much water we
needed to make -- this is like 400 gpm, and I think we have a thousand
gpm actually in the plant.
So the spray that we have in the plant would work better
than the spray that's on this. But it shows that for earlier failure,
it would reduce it by about a factor of two, and it would help the
intermediate failures, but those are really pretty low-risk events. The
isolation failure, it would help that a fair bit.
It doesn't help the bypass, so by putting the spray in, we
ended up reducing a very small number by a factor of two. And this is
the reason that it didn't make the cut, if you will, in putting it in
the plant from the SANDA category.
And we took this actually as far as the Commission. There
was a SECY paper, and I think it's really one of the reasons that we're
here, because defense-in-depth really got down -- this was one of the
harder arguments that we have had about what is defense-in-depth?
And I'm going to read from one of the vote sheets on this
SECY, just one paragraph, because I think this answers your question
about if you pass all the requirements, would they still make you put
something in? Yes.
And the argument was that in spite of the fact that the
proposed system cannot be justified under any of the rational
decisionmaking guidelines that we have established for ourselves, the
staff would require it anyway.
The ultimate reason seems to be that it is justified to
compensate for uncertainties in how the design will behave under severe
accident conditions. Even this reason is not well supported because we
have not established a relationship between the proposed spray and the
particular uncertainties it is supposed to address.
Defense-in-depth becomes the final justification. And then
it goes on to say that the Commission and the staff should not continue
ad hoc decisionmaking indefinitely, and here we are. That's why we're
here today.
But the answer to your question is, yes. And I think that
we've perhaps moved beyond this now, and I was glad to see Gary's and
Tom's presentations. I'm not too sure, but I can probably use that to
take the spray out.
[Laughter.]
MR. McINTYRE: Since it's not a Tier I requirement.
MR. HOLAHAN: We'd have to talk about that.
MR. McINTYRE: So that's the way that defense-in-depth
actually gets applied. If you make it a way of life, almost a mantra,
you pray to it, you decide and you think like that, and it can really
result in a lot of, I think, good things in the design. That should
answer your question that you asked about five times today.
MR. KRESS: Thank you very much. I'm not so sure that if we
had had Gary's risk-informed matrix table back then, whether or not we
would have come down on the side we came down on.
MR. McINTYRE: I think what's important is that they were
looking at the balance between prevention and mitigation, because my
argument or complaint -- complaint, that's fair -- at the time was, what
are the units on this balance?
And I think there's an attempt to do that, and I certainly
applaud that.
MR. KRESS: That is exactly right.
MR. GARRICK: What would be much more interesting than these
point estimates, which see -- would be the PDF stacked on top of each
other for these two cases.
MR. KRESS: Yes, that was one of our problems, too. We
didn't have any of the PDFs. And all we had were point estimates, and
that made the decision much more difficult.
Had we had those, it might have been a different story.
MR. HOLAHAN: My recollection is that you didn't have them
because they were never generated.
MR. KRESS: That's right. That's why we didn't have them.
MR. APOSTOLAKIS: That's a good reason.
MR. BUDNITZ: But the difference at Yucca Mountain is a
qualitative difference about the staff behavior, I believe. See, you
were having this argument about a theoretical plant that wasn't sited or
being built anyplace in particular, in a room in an office building like
this.
But in Yucca Mountain, it's going to be in an arena in which
the Governor, the Senators and almost the entire population of a real
state are using every political opportunity they can and every legal
opportunity they can, not only to get in the way, but to embarrass the
staff.
And the staff is acutely aware that that embarrassment has
to be avoided, if they can, and that's why they can't find themselves,
if they can avoid it, in a situation where they're backfitting a
positive decision on what would have been a negative decision by
changing their minds halfway through.
MR. KRESS: Yes.
MR. BUDNITZ: And so they really have a different dilemma
than you and the reactor staff and at that time. It's much more
difficult for them.
MR. APOSTOLAKIS: Good.
MR. KRESS: Very, very difficult. I'm going to ask if
anyone in the audience feels compelled to add anything to what they've
heard.
[No response.]
MR. KRESS: Seeing no rush to the front --
MR. APOSTOLAKIS: Are the experts going to be back tomorrow?
MR. KRESS: That's a good question.
MR. APOSTOLAKIS: Are they coming tomorrow?
MR. KRESS: Tomorrow, we're going to try to wrap some of
this up and see if we can reach some conclusions, and maybe spell out
what the remaining issues are, and things of that nature, and as many of
the experts as we could get would be nice.
MR. APOSTOLAKIS: So we lost Dr. Murley then?
MR. KRESS: Lost Dr. Murley.
MR. APOSTOLAKIS: Are you going to be here tomorrow,
Budnitz?
MR. BUDNITZ: Yes.
MR. KRESS: We'll quit at precisely noon or pretty close, or
maybe even before noon, but more around there. Okay, great. The staff,
will you be here?
MR. HOLAHAN: Yes.
MR. KRESS: So we'll try to wrap it up then tomorrow, and it
will be more of a roundtable discussion.
MR. APOSTOLAKIS: Is NEI going to be here tomorrow?
MR. KRESS: You're welcome to be here. So if there are no
other comments from --
MR. GARRICK: Let me remind the ACNW and the ACNW staff that
our meeting will start in ten minutes.
MR. APOSTOLAKIS: And go on for eight hours.
[Laughter.]
MR. KRESS: With that, I'm going to recess until tomorrow
morning at 8:30.
[Whereupon, at 5:40 p.m., the meeting was recessed, to be
reconvened at 8:30 a.m., on Friday, January 14, 2000.]
Page Last Reviewed/Updated Monday, October 02, 2017