Reliability and Probabilistic Risk Assessment - April 17, 2001
Official Transcript of Proceedings
NUCLEAR REGULATORY COMMISSION
Title: Advisory Committee on Reactor Safeguards
Reliability and Probabilistic Risk
Assessment Subcommittee
Docket Number: (not applicable)
Location: Rockville, Maryland
Date: Tuesday, April 17, 2001
Work Order No.: NRC-157 Pages 1-205
NEAL R. GROSS AND CO., INC.
Court Reporters and Transcribers
1323 Rhode Island Avenue, N.W.
Washington, D.C. 20005
(202) 234-4433. UNITED STATES OF AMERICA
NUCLEAR REGULATORY COMMISSION
+ + + + +
ADVISORY COMMITTEE ON REACTOR SAFEGUARDS
MEETING OF THE
SUBCOMMITTEE ON RELIABILITY AND PROBABILISTIC
RISK ASSESSMENT
+ + + + +
Tuesday,
April 17, 2001
+ + + + +
Rockville, Maryland
+ + + + +
The Subcommittee met at the Nuclear
Regulatory Commission, Two White Flint North, Room T-
2B3, 11545 Rockville Pike, at 8:30 a.m., Doctor George
E. Apostolakis, Chairman, presiding.
PRESENT:
GEORGE E. APOSTOLAKIS Chairman
MARIO V. BONACA Member
THOMAS S. KRESS Member
GRAHAM M. LEITCH Member
ROBERT E. UHRIG Member
ACRS STAFF PRESENT:
MICHAEL T. MARKLEY
ALSO PRESENT:
TOM BOYCE NRR
STEVE EIDE INEEL
ADEL EL-BASSIONI NRR
TOM HOUGHTON NEI
ROGER HUSTON Licensing Support Services
MICHAEL R. JOHNSON NRR
STEVEN E. MAYS NRR
DEANN RALEIGH LIS, Scientech
JENNY WEIL McGraw-Hill
TOM WOLE RES
BOB YOUNGBLOOD ISL
A-G-E-N-D-A
Agenda Item Page No.
Introduction
Review goals and objectives for this meeting;
past ACRS deliberations on risk-based
performance indicators (RBPIs), G. APOSTOLAKIS . . 5
NRC Staff Presentation
Background/Introduction, S. MAYS, RES. . . . . . . 6
Relations of RBPIs to Revised Reactor
Oversight Process (RROP), TOM BOYCE, NRR . . . . . 9
RBPI definitions/characteristics
Potential benefits. S. MAYS, RES . . . . . .50
RBPI development process, S. MAYS, RES
H. HAMZEHEE, RES
Summary of results, S. MAYS, RES . . . . . . . . .63
Initiating Events: full-power/internal,
H. HAMZEHEE, RES
Mitigating systems: full power/internal. . .98
Containment. . . . . . . . . . . . . . . . 108
Shutdown . . . . . . . . . . . . . . . . . 120
Fire events. . . . . . . . . . . . . . . . 132
Industry-wide trending . . . . . . . . . . 134
Risk coverage. . . . . . . . . . . . . . . 136
Verification and validation results. . . . 143
AGENDA - (Continued):
Agenda Item Page No.
NRC Staff Presentation - continued
Discussion of implementation issues, . . . . . . 150
S. MAYS, RES
H. HAMZEHEE, RES
Discussion of industry comments
S. MAYS, RES
H. HAMZEHEE, RES
Industry Comments
Industry perspectives on RBPIs,
T. HOUGHTON, NEI . . . . . . . . . . . . . 178
General Discussion and Adjournment
General discussion and comments by members . . . 191
of the Subcommittee; items for May 10-12,
2000 ACRS meeting, G. APOSTOLAKIS, ACRS
P-R-O-C-E-E-D-I-N-G-S
(8:30 a.m.)
CHAIRMAN APOSTOLAKIS: The meeting will now
come to order. This is a meeting of the Advisory
Committee on Reactor Safeguards Subcommittee on
Reliability and Probabilistic Risk Assessment. I am
George Apostolakis, Chairman of the Subcommittee.
Subcommittee Members in attendance are Tom
Kress, Graham Leitch, and Robert Uhrig, and Mario
Bonaca.
The purpose of this meeting is to discuss
the results of the staff's Phase 1 effort to develop
risk-based performance indicators. The Subcommittee
will gather information, analyze relevant issues and
facts, and formulate proposed positions and actions,
as appropriate, for deliberation by the full
Committee. Michael T. Markley is the Cognizant ACRS
Staff Engineer for this meeting.
The rules for participation in today's
meeting have been announced as part of the notice of
this meeting previously published in the Federal
Register on March 26, 2001.
A transcript of the meeting is being kept
and will be made available as stated in the Federal
Register Notice. It is requested that speakers first
identify themselves and speak with sufficient clarity
and volume so that they can be readily heard.
We have received no written comments or
requests for time to make oral statements from members
of the public regarding today's meeting.
We will now proceed with the meeting and
I call upon Mr. Steve Mays to begin.
MR. MAYS: Thank you, George. I'm Steve
Mays from the Office of Nuclear Regulatory Research.
With me today at the front is Hossein Hamzehee, who is
the Project Manager for working on risk-based
performance indicators, and with me also to my left is
Tom Boyce from the Office of Nuclear Reactor
Regulation, who will speak in a couple minutes about
the relationship that this work has to the Reactor
Oversight Process. Also here at the side table is
Mike Johnson, who is the Section Chief in NRR, who is
our technical counterpart in NRR and our liaison with
this work, and we have a couple of our contractors in
the audience if there's any questions that I can't
directly answer or Hossein can't answer, they can come
up and give additional information about what we've
done.
What we are trying to do today is give the
ACRS an opportunity to provide some comments and to
provide some information to you about what was in the
report that we issued in January. We've already held
one public meeting in February to kind of lay out
what's in the report, and kind of frame the discussion
of what we are trying to do, and how we went about
doing it, so that when we have our public meeting next
week we would have the opportunity to make sure that
was a well-focused meeting and directed towards the
kinds of things we need to know what the response from
the outside stakeholders is.
We extended our comment period at the
request of people in the February meeting to May, so
that people can come to the meeting next week, discuss
points, get the opportunity to hear some answers from
us if they do, and then be able to take that into
consideration as they give us their formal comments in
May. We are looking forward to that meeting, and part
of what we are going to see here today is some new
stuff that we've done that's not actually in the
report, and we are going to also present that at the
meeting next week, so that we can hopefully move this
process along.
So, we are looking for feedback from the
ACRS, we expect probably a letter of some kind with
respect to whether they believe that the work we are
doing is a potential benefit to the Reactor Oversight
Process, whether we've gone about that in a
technically sound manner, and also to get some
feedback on the alternate approaches that we're going
to present today, which are not in the report, that
we've gone off and developed in light of some of the
early comments we got, both internally from the NRC
review, as well as some comments we've had from
external stakeholders relating to the total number of
potential indicators and what that impact would be on
the oversight process.
So, we are going to have this briefing
broken up into several pieces. The first part is the
relationship of the RBPIs to the Reactor Oversight
Program. Tom Boyce from NRR to my left will be
discussing that. Then I will come back and the rest
of the presentation will be primarily from our part on
the technical aspects of what's been done, including
what we see as the potential benefits, what we
actually did in development, some results that we
have. We want to go over the key implementation
issues that are before us, because we think those tend
to be the ones that we have the biggest comments on
from both internal and external reviewers so far, and
to go over the alternate approaches that we're looking
at as a means of dealing with some of the issues that
have been raised.
So, with that, I would like to go to Tom
Boyce, who will discuss the relationships of the RBPIs
to the Reactor Oversight Process.
MR. BOYCE: Good morning.
As Steve said, I'm Tom Boyce, I'm in the
Inspection Program Branch in the Office of NRR, and
NRR requested that we have a short amount of time at
the beginning of Research's presentation to let you
know the relationship, as NRR sees it, of the risk-
based PI development program to the current
performance indicators in the Reactor Oversight
Process.
CHAIRMAN APOSTOLAKIS: Is your presentation
consistent with the memorandum from Mr. Dean to Mr.
King, of December 1, 2000?
MR. BOYCE: It is entirely consistent. I
was the author of that memo.
CHAIRMAN APOSTOLAKIS: Okay, good.
MR. BOYCE: By definition.
CHAIRMAN APOSTOLAKIS: You may have changed
your mind.
MR. BOYCE: I maybe shouldn't have stated
it quite so positively.
Before I talk about the Reactor Oversight
Process relationship to risk-based PIs, it's important
to understand the overall environment with which our
agency is now regulated, and some of the changes that
are impacting the nuclear industry.
The Commission has provided direction to
the staff that its intent is to better risk inform the
NRC's processes, and it's done this for several years
on a variety of fronts. The Reactor Oversight Process
was revised in 1999 to be more risk informed,
objective, understandable and more predictable than
the previous oversight process. The Reactor Oversight
Process was implemented on April 2, 2000, so we have
had one year of practice in the Reactor Oversight
Process under our belts.
Another backdrop for the industry is
continuing advances in the use of information
technology and data. Industry is getting better and
better at collecting data, processing it for its own
internal uses. We also are getting better at it. The
Reactor Oversight Process has got a web site that has
gathered a great deal of kudos for its ability to
present information.
The internet and PCS have allowed much
more free exchange of information than has previously
been allowed, and both NRC and industry are continuing
to expand their capabilities in this area.
We wrote about the bases for the Reactor
Oversight Process in two Commission papers in early
1999, and there we stated that the Reactor Oversight
Process would use a combination of inspection findings
and performance indicators to provide oversight of
industry. We conducted a pilot program in 1999, and
the results were articulated in SECY-00-049. In that
same Commission paper, we stated that while the future
success of the Reactor Oversight Process would not be
predicated on the risk-based PI program, we thought
that there were a couple of places where the risk-
based PI program could, in fact, enhance our current
set of performance indicators.
These areas are actually articulated in
the last bullet right here, the reliability
indicators, unavailability indicators, shutdown and
fire indicators, and containment indicators.
We also thought that the risk-based PI
program offered the potential to establish, perhaps,
plant specific thresholds for these PIs on the current
set of PIs. Because we thought this, we decided to
task research to develop in these areas, and we sent
them a user need. Research responded that they would
examine the feasibility of these selected risk-based
PIs as part of their Phase 1 report, and you'll be
hearing more about that in a little bit.
Even though the risk-based PI program is
moving forward, we thought that there were several key
implementation issues that needed to be addressed
prior to implementing the risk-based PIs
incorporating them into the Reactor Oversight Process.
One of the keys, in general, is data quality and
availability. Our experience in the Reactor Oversight
Process is, is that while data is being collected by
individual licensees there are a variety of ways that
you can collect that data. There is a variety of
quality for that data, and how you collect that data
and pull it together into a graph that is presentable,
we found surprising variation. So, we thought that we
needed to be happy with the way data was collected, so
that it was done uniformly and consistently, before we
are able to implement it in the Reactor Oversight
Process.
Second, we thought that the models used
for assessing the data needed to be developed and
validated by licensees and the NRC staff in the
regions, and you'll hear more about the status of
development of SPAR models from Steve, but those two
were the key areas that we thought needed to be fully
mature before it was ready to be incorporated in the
Reactor Oversight Process.
CHAIRMAN APOSTOLAKIS: I'm a bit confused
now. Isn't this, aren't these two applicable to the
existing revised Reactor Oversight Process? I mean,
you also need good data, you also need some sort of a
PRA, to assess the significance of a particular
performance indicator being above a number and so on.
So, I don't know, why are these two issues,
implementation issues, so important to risk-based
performance indicators, but not to the existing
oversight process?
MR. BOYCE: Well, in the case of the
existing performance indicators for the ROP, we had an
opportunity to go through a pilot program, licensees
submit the data directly to the NRC, using mutually
agreed upon guidelines in the NEI document, NEI 99
Tech 02. That was developed by mutual discussions
with industry, over an extended and intensive an
extended period of time in an intensive manner. The
current data for the risk-based PI program is drawn
from sources such as the EPIX database, that's, I
believe, under the auspices of INPO, and that same
sort of rigorous look at how the data is submitted has
not been applied yet.
CHAIRMAN APOSTOLAKIS: But, it could.
MR. BOYCE: It could.
CHAIRMAN APOSTOLAKIS: It could, I mean,
they could do the same first of all, I don't like
this idea of they and us, I mean, it's one agency, but
there is nothing in the methodology that says, you
know, you have to use EPIX.
MR. BOYCE: Correct.
CHAIRMAN APOSTOLAKIS: They can use the
data that you are using.
Now, they felt the need to go to other
sources of data, because for some reason the data that
we receive right now is not sufficient, is that the
idea, Steve?
MR. MAYS: Yes, George. Let me propose
we'll get into that in much more detail in the section
where we talk about implementation issues, but to
address it shortly, remember when the ROP was put in
place one of the key issues for getting indicators was
what information is readily available and can be put
together in a consistent way, and that data that's
reported into the ROP is reported under 50.9
requirements for licensee submittal of data. That's
one of the major issues with respect to implementation
of these PIs, and that data that's being submitted
under the ROP was not specifically tailored to certain
aspects, like reliability indicators, and the models
in the ROP were more generic with respect to the
thresholds.
So, there were several things of that
nature that I would put in the category of expediency,
that required that to be there, and as we move to more
detailed and more plant-specific data and thresholds,
we think it's important to make sure that we have an
understanding of what that data is and a common
agreement as to how the quality of the data needs to
be assured and how that stuff needs to be reported,
and that's an implementation issue we're going to have
to work out, but generically, as long as you have the
data that fulfills the model, then that's all you
really need from a modeling standpoint and a
calculational standpoint, but from a regulatory
standpoint there are other issues that have to be
addressed.
MR. JOHNSON: And, if I could add, Michael
Johnson, NRR, if I could add to what Steve has said,
and I think he's made some good points, remember the
challenges that we face with the ROP PIs, as we'll
talk about when we brief the ACRS in May, during the
first year of implementation have been challenges
associated with verification of the data, even with
the relatively simple PIs that we have now in the ROP,
it's a problem.
So, George, to go to your question, your
point, it's not that we don't face these challenges,
these similar challenges with the existing ROP, it's
that these challenges will certainly exist as we go
forward with RBPIs.
CHAIRMAN APOSTOLAKIS: Yes, they do exist.
Well, I read the memorandum that I
mentioned earlier, dated December 1st, from Mr. Dean
to Mr. King, and I must admit I was surprised at how
cool it is towards this effort, as if somebody is
trying to force this upon you and you are resisting.
The report did not demonstrate that the
proposed RBPIs will be more effective than the PIs
currently in place. I don't know what that means.
Licensees may be reluctant to voluntarily implement
the new RBPIs because of two reasons, there are many
more indicators to track, calculate and report, which
increases the effort licensees have to expend. So
what, if they have to do it, they have to do it.
Where is the technical argument? Is there any
justification for needing more indicators to track,
calculate and report? That should be our criterion,
that there is some information there that's useful to
us, not that it imposes burden on the licensees.
First, we have to decide whether it's unnecessary, and
if it's unnecessary then, of course, we don't impose
it. But, I didn't see any argument anywhere here that
says, no, these additional indicators are not needed
because we already cover them. It just says, you know,
the licensees will have to spend more time doing it,
and, boy, we really don't want to do that, and
licensees will be putting themselves in a position
where it is much more likely they will have to report
a non-green PI and subject themselves to the resulting
increased regulatory and public attention. Well, I'm
shocked, I'm shocked shocked. There are indicators
that are not green? I just don't understand this
memo.
You guys don't like something, but you
don't want to come out and say it to us. Obviously,
you don't like it.
MR. JOHNSON: Steve, when we get into
under the implementation issues, will we come back to
this topic?
MR. MAYS: I think we will.
In fairness to in fairness to Tom and
building shock, we'll put that in.
CHAIRMAN APOSTOLAKIS: I just didn't want
to put Tom on the spot, but he's the one here.
MR. MAYS: I know.
MR. BOYCE: Thank you, George. I don't
mind.
CHAIRMAN APOSTOLAKIS: And, he said he
wrote it, big mistake.
MR. BOYCE: I retract my earlier comments.
CHAIRMAN APOSTOLAKIS: Okay.
MR. MAYS: Tom hasn't been before the ACRS
as often as I and Mike have.
In fairness, I think we, in the RBPI
report, raise the issues of the implementation because
we recognized that if we were going to make an
improvement in the ROP it was going to be a voluntary
improvement that we decided we wanted and that we
negotiated with our external stakeholders to determine
it was a benefit to the agency, and I think what you
were seeing there was just a recognition of some of
the issues that we knew were going to be raised, and
that we knew had to be addressed, as opposed to saying
that they could not, or should not, be done here. I
read that letter when I saw it more as a confirmation
that we had identified the correct issue in the RBPI
document, and that we knew from our previous
interactions with external stakeholders that those
were going to be concerns that we had to address, and
that the Commission was the one who was eventually
going to adjudicate whether or not we were doing that
properly or not.
So, I don't think it was nearly as
negative as you might have portrayed it.
CHAIRMAN APOSTOLAKIS: I thought it was
cool, there is a certain coolness here, that maybe
what you guys are doing has some value, but you have
not demonstrated it to us, and what's worse, you may
ask the licensees to do more.
MR. JOHNSON: Well, there is an aspect of
that, and maybe Tom was going to get into that. Let
me just say a couple of words before Tom.
CHAIRMAN APOSTOLAKIS: You should let him
at some point defend it.
MR. JOHNSON: Yes, we should.
MR. BOYCE: No, go ahead, Mike.
MR. JOHNSON: As Steve sort of indicated,
there is an aspect of this, and the ROP, when we set
out to develop performance indicators, remember the
performance indicator aspect of the ROP is a voluntary
program, if you will. Even the document that endorses
the guidance is the guidance is an NEI document,
NEI 99-002, that provides the criteria, we endorse
that.
CHAIRMAN APOSTOLAKIS: How many licensees
have refused?
MR. JOHNSON: None of the licensees have.
CHAIRMAN APOSTOLAKIS: It is still
voluntary?
MR. JOHNSON: Yeah. All of the licensees
are reporting on their existing PIs.
And so, what we are talking about with
risk-based performance indicators, as Steve indicated,
is an enhancement to this PI reporting program that's
a piece of the ROP, and as such, I mean, I think we
do, in fact, need to be careful about things like, are
we increasing the burden without commensurate benefit?
In fact, in our formal change process for the
performance indicators, we look at, should we be
making reductions in the inspection program, or
changes in the inspection program, in areas where we
have information that we get readily from the
performance indicators, so all of those things have to
be worked out in the implementation stages. So, we
wouldn't just adopt a suit of PIs that would make us
happy, if you will, without regard to the impact that
they would have on licensees.
CHAIRMAN APOSTOLAKIS: And, I think that's
a very reasonable thing to do, as long as there is
also a technical discussion
MR. JOHNSON: That's right.
CHAIRMAN APOSTOLAKIS: as to, you know,
this indicator gives us information we already have,
or maybe expands on something, but by and large we
really understand what's going on and the additional
burden is not justified.
MR. JOHNSON: That's right.
CHAIRMAN APOSTOLAKIS: I can see arguments
like that, but just to say that, you know, this
imposes burden, without addressing the kind of
information you get, I find that a little odd.
MR. BOYCE: I think there was also a
meeting that followed that memo, that was held
between, I think, Sam Collins and members of the
Research staff, and I think there we were able to get
past some of the detailed discussion you saw at that
memo, and I think at that meeting we said that we
believed that this was a good technical effort, it did
have potential value, and we did want them to continue
development, which is not stated explicitly in that
memo, because the intent of that memo, as I recall,
was to convey technical comments on the report itself.
CHAIRMAN APOSTOLAKIS: Now, I'm just
curious, what would you expect them to do to
demonstrate that the proposed RBPIs will be more
effective than the PIs currently in place? The word
"effective," what does it mean in this context, I
mean, independently of the coolness of this, I mean,
technically, what would you expect them to do?
MR. BOYCE: Well, I'm not sure we've
established hard criteria for what we mean by more
effective, but, in general, the PIs that we have have
certain limitations. I mean, not all of them have
been well-founded and risk-informed principles. Some
were selected based on 95 percent performance of
industry, there's a word, but it's not
probabilistically-based, it was, we took a look at
histograms and said that 95 percent of the plants
operate below this threshold.
CHAIRMAN APOSTOLAKIS: So, this is
thresholds.
MR. BOYCE: Thresholds.
CHAIRMAN APOSTOLAKIS: Yes.
MR. BOYCE: So, we could certainly improve
on our technical basis for thresholds for individual
PIs, that's one area.
CHAIRMAN APOSTOLAKIS: And, there is a nice
criticism of that on page A-10 of Appendix A, very
nice. It says, "If I wrote ..."
MR. BOYCE: But, we couldn't do that,
George, that would be a conflict of interest.
MR. JOHNSON: Can I add one other thing to
that answer, is that, again, I'll allude to a change
of the formal change process. What happens at the end
of this effort, and what happens, in fact, when we go
to put in place any new PI, as we go through a formal
change process, and that process has astute things,
like we'll conduct a pilot, we set criteria up at the
beginning of that pilot for what we want to see in
terms of evaluating the efficacy of these proposed new
PIs, and so, it's in those criteria that we'll be very
specific about what we'll look for in terms of making
a decision about whether to go ahead.
And, there's something already we'll
talk to you again in May about two PIs that we already
are piloting under the existing ROP that are not risk-
based PIs, but they are going through the process. We
have a pilot in place, we are looking at the results
of that pilot. We are looking at the performance as
would be indicated by indicators reported against
those proposed PIs, balancing that against the
existing PIs to see if there are differences. So,
it's those kinds of things that you look at, and those
are built into this formal process that we enter into,
after this phase of this preliminary development of
the RBPIs is finished.
MR. BOYCE: And finally, one more comment
on the tone of that memo. We had gotten informal
feedback from some stakeholders that the risk-based PI
program had, through whatever means, been perceived as
a certainty, that it would, in fact, be implemented.
And, we wanted to make sure that that expectation
was, in fact, addressed so that it would be put in the
right context. In other words, the change process
that Mike just alluded to did need to be followed.
There are, I think, 30 some odd performance indicators
that are being proposed here in the Phase 1 report,
and the data collection requirements do, in fact, add
significant burden to licensees. So, licensees do
give us a feedback that, hey, if you are going to
implement the new program like that, you need to
consider cost benefit, and we had not even engaged in
terms of cost benefit at that point. So, to some
extent, the tone you see there was to try and address
the perception that had gone out in industry.
DOCTOR KRESS: If you implement this, would
you do it on a pilot basis with a number of volunteer
plants to start with?
MR. BOYCE: We would expect that to be the
case, that's what our Manual Chapter 0608 calls for
for PIs like this.
DOCTOR KRESS: And, to determine whether or
not it's useful, then those volunteer plants would
have had to been compared with the old program, and
would have had to have degraded performance somewhere,
otherwise you are proving a negative.
I'm not sure, you know, you might go on
for years, and years, and years, before you ever come
to some conclusion that the new process is useful to
you that way.
MR. BOYCE: Well, I think you'll see from
the report that Steve is going to go over that that's
not, in fact, what happened. I think they ran some
test data through and found out that there was, in
fact, degraded performance that came out for the set
of data that they looked at.
DOCTOR KRESS: Oh, you looked at highest
performance.
MR. BOYCE: Well, let me we are jumping
the gun a little bit.
DOCTOR KRESS: Okay.
MR. BOYCE: We looked at the performance
over the '99 time period, basically, '97 through '99
time period, which is a time period for which the ROP
pilot program and the ROP program already had some
data on plants. So, we have looked at that, and we do
know that because there are certain areas that we are
examining with PIs that the ROP doesn't have PIs that
we now have the opportunity to see things that weren't
there potentially as indicators in the ROP.
We are going to cover some of this stuff
when we talk about potential benefits and things of
that nature, and the examples you'll see when we ran
for the 23 plants that we did run, we do have a fairly
broad range of coverage of that.
DOCTOR KRESS: Okay.
Could you indicate what are the
implications when you say reporting under 50.9, I'm
not sure I'm exactly familiar, I have kind of an idea,
but exactly what does that mean?
MR. BOYCE: 50.9 requires that information
submitted to the NRC by the licensees will be
submitted under oath and affirmation, and that means
when a utility does that, that information, once
submitted, can be cited in violations for failure to
be accurate. So, when you have that level of rigor
applied to data, and the potential for being cited for
inaccuracies in that data, that process makes
utilities apply more effort to ensure that that same
data is correct than they may otherwise have to, and
that's one of the data issues we'll get to.
DOCTOR KRESS: Okay.
MR. BOYCE: And, I'll give you some
examples of things when we get to that area, as to how
that can be a potential problem.
And, the issue from our standpoint is,
what level of quality and rigor does the data have to
have, and if that quality and rigor is something
different from 50.9 then the question is, is, well,
why would we have to, or would we have to have data
submitted under 50.9. That's an implementation issue
that would have to get addressed through the formal
process in the ROP change process that Mike and Tom
just alluded to.
In fact, when the existing ROPs were first
being tried, one of the things that happened, in order
to make sure they understood what the quality issues
were and the difficulties were, was there was, I guess
the right word would be, a waiver
DOCTOR KRESS: A discretion.
MR. BOYCE: a discretion on enforcement
on those issues, as part of the initial program, to
make sure that that wasn't becoming an impediment to
testing the program out and understanding what levels
of things needed to be done.
So, those are all kinds of details of how
you would go about doing the implementation, which,
quite frankly, we're not really here to discuss
exactly how that will happen today. The process would
be more like this. We go through public comment, we
get ACRS comments on the technical quality of what
this program brings, and whether or not it looks like
it's beneficial to the ROP, and at that point we would
produce our Phase 1 report and NRR at that point would
be in a position of saying, do you want to take all of
these, some of these, none of these, and try to run
them through the ROP change process. And then, once
they would go into that process, they would lay out
the plans in accordance with that procedure, and get
together with the industry and our other external
stakeholders, and go through the process.
So, it's a little premature to tell you
what that all would have in it, or what all the
decisions would be made, and how they would all be
made, because that's a little bit ahead of where we
want to be right now. We want to
DOCTOR KRESS: I'm just trying to
understand industry's concern. I guess by extension
then, all the EPIX data then could theoretically be
subject to the requirements of 50.9.
MR. BOYCE: Could be. It's not certain
that they would, it's not certain that they would not,
that's an implementation issue we have to address, and
we have recognized that that was a significant issue
when they were doing the unavailability data for the
current ROP indicators, and we no reason why that
issue wouldn't also be an issue for reliability data,
which is what the EPIX data is being used for here.
So, we recognize that that's an issue that
has to get resolved. Industry recognizes it as an
issue, and we all think it's something that has to be
taken care of through the ROP change process.
CHAIRMAN APOSTOLAKIS: So, the current
oversight, revised oversight process, when it does
risk-related calculations what models is it using?
MR. BOYCE: Well, actually, it's not doing
risk calculations in the PIs. The calculations were
done initially to establish what the thresholds should
be on the PIs.
CHAIRMAN APOSTOLAKIS: And, that's it.
MR. BOYCE: Now, after that, all they do is
calculate the value that's coming in and compare it to
the threshold. There's no more risk modeling being
done in the ROP to get the current indicators.
CHAIRMAN APOSTOLAKIS: But, you are doing
the same, aren't you?
MR. BOYCE: We are applying the same
philosophy here.
CHAIRMAN APOSTOLAKIS: Okay, but you are
using the SPAR model.
MR. BOYCE: We are using plant-specific
SPAR models to set thresholds.
CHAIRMAN APOSTOLAKIS: What did they use?
MR. BOYCE: They used a combination of
licensee models, some SPAR model runs that we did for
them in the process, and they came to a consensus
opinion of how to set thresholds based on those
results. And, they tend to be generic for the
industry, as opposed to plant specific.
And so, that's what was done, and it's
documented in 99-007. I can't recall exactly which
appendix it's in, but I know it's in there.
CHAIRMAN APOSTOLAKIS: It was H, Appendix
H.
I'm trying to see whether I mean, the
sig I understand the significance of the first sub-
bullet, data quality and availability, the second one
is not so clear to me. Is it because this new effort
is intended to be plant specific?
MR. BOYCE: That's right.
CHAIRMAN APOSTOLAKIS: The PRA model is
more important than it was in the generic case.
MR. BOYCE: That's right.
You see, the thresholds for individual
plants would could be lowered, and so the plant's
margin to the green/white threshold, if we use the
same process under the current ROP, could be less.
And, any time you are talking about increased
regulatory attention licensees are very sensitive to
that sort of thing, and so we want to have good
quality models so we have confidence in the thresholds
and in the information that's being presented to us.
CHAIRMAN APOSTOLAKIS: Right.
And, my counter argument to that is that,
this is a good idea to worry about these things if the
starting point has some logic to it. And, I don't
think that 95th percentile you guys did can withstand
scrutiny.
MR. MAYS: Well, George, you'll notice that
one of the things
CHAIRMAN APOSTOLAKIS: The weakness of this
method is that it depends only on the number of plants
with less than acceptable performance, but not on how
much their performance exceeds the norm. Wonderful.
MR. MAYS: Well, George, we went back, as
part of the RBPI process, as we outlined in the RBPI
White Paper, and said we were going to look at how we
thought thresholds need to be set, and we went and
looked at that particular issue and we, in
recommendations in the program, concluded that we
thought it made more sense to have the green/white
threshold for performance indicators be based on a
risk change rather than on a deviation from norm
principle. And so, we have made that case that it's
more consistent with the significance determination
process for inspection findings, which all three color
layers are based on a risk metric, and so we are
making that recommendation, we provided the
information for how we would say what the distribution
of plants' performance was, where the 95th percentile
was on that, and we've made that recommendation. And,
so far, quite frankly, I haven't had any technical
comments come back from either inside NRC or outside
so far, that said, no, no, no, we want to stay with
the 95th. There may be some that will say that, but
I think there is a better logical connection for using
the green/white interface on PIs based on risk than
based on deviation from the norm.
CHAIRMAN APOSTOLAKIS: That's right.
MR. MAYS: So, I mean, I want to I think
it comes to my mind at this point to recall a
principle that Mike and I have talked about long and
often, with respect to this and other work, and we
have worked by that principle from the beginning, and
that's our principle, is progress, not perfection.
The idea is, we want to make incremental improvements
where we can, and we are not going to worry about the
fact that we don't have perfection either in what we
started with or what we end up with, because we don't
want to end up with what I loosely refer to as the
source term problem, where we start out with TID
14844, which was several people gathered around the
table thinking what they thought best, and then no
matter how much subsequent technical analysis gets
done, it becomes difficult to change, because the
other thing was already there.
So, we are trying to say, what we have is
what we started with. The ROP is there. I'm not here
to say whether the ROP is perfect or not, that's not
my job. My job here is to try to address things that
could make the ROP better, and that's the tone in
which we are trying to do this.
MR. JOHNSON: Yes. I would just add to
that two years ago we had what we had with respect to
the performance indicators, and we picked for targets
of opportunity for which we had data, and we set
thresholds as best we possibly could, and that
included for the green/white threshold that 95
percentile breakout for performance indicators.
Keep in mind that performance indicators,
some of the performance indicators were new, and were
in areas where you didn't have couldn't apply a
risk model, for example, the Security Equipment
Performance Index PI was a new PI, and there's no way
you can risk inform that, if you will.
So, but remember, in the broad context of
the ROP the green/while threshold is meant to be
indicative of an area where we need to go out and do
some additional inspection to look. So remember, don't
view the performance indicators in a vacuum. They are
a piece of an entire program, by which we provide
oversight on licensee performance.
MR. BOYCE: To try and get us back on
track, I was on the last bullet here, and I think
we've covered all the points, with the possible
exception of the last one, and that is, is that one of
the significant comments that we heard early on from
industry was, is that the risk-based PI program does
represent a large increase in the number of
performance indicators. And, any time you increase
the number of performance indicators you have the
opportunity for an increased opportunity for one of
the performance indicators to exceed a threshold.
Again, using the green/white threshold, if we go into
the white or above regulatory action is mandated under
our current ROP. And so, industry's comment was,
there's definitely an increased chance to regulatory
attention.
So, one of the things that we would
consider, if we were going to move forward with all of
the risk-based PIs, would be to, perhaps, modify the
algorithms on our Action Matrix for changing columns
from the licensee response column to the regulatory
response column, or other columns of the Action
Matrix. But, that is not, again, I don't to establish
premature expectations, that is not our current plan,
and I think Steve is looking at ways to, perhaps,
combine the performance indicators, so that there
would be fewer numbers, and I think you are going to
hear more about that.
But, we did want to say that we would
consider that sort of approach if necessary.
CHAIRMAN APOSTOLAKIS: Well oh, I'm
sorry, go ahead.
DOCTOR KRESS: If changing over from one
color to another across the threshold represents a
delta CDF, for example, then I don't see how I
mean, you have a total delta CDF you don't want to
exceed, you know, in your matrix, I don't see how
having more PIs changes that. If you set the change
for each one of the PIs to be a certain delta CDF or
related to it, then it doesn't matter
MR. MAYS: Actually, Tom, I think you are
correct, the issue would be, was the change in
performance reflected through the PIs or was the
change in performance reflected through inspection
finding, your point being that if you have had a
change in performance it should be reflected in one or
the other, and that change is the same regardless of
how it got found.
DOCTOR KRESS: Yes.
MR. MAYS: that's true, but I think it is
true also that there is what I would call an optics
issue, which is, if you have more direct PIs you have
maybe a faster responding optics with respect to the
fact that something has changed, and the Action Matrix
was set up on the basis of the limited number of PIs
that you have. So, the issue here was, the Action
Matrix was a little was defined in light of those
numbers of PIs, and so, therefore, it might be
something that would have to get looked at.
I think it's pretty clear that the more
PIs you have, the more opportunities you have to cross
a particular threshold, and that was basically the
concern that industry raised for us and
DOCTOR KRESS: Well, you know, that's what
I viewed as the good part, about adding more PI.
MR. MAYS: Well, it's the double-edged
sword.
DOCTOR KRESS: Right.
MR. MAYS: If you have more PIs you have
more opportunities to look green, but if you have more
PIs there's also another opportunity to have not
green, and the question is, how much of a value, I
guess, would be more greens as compared to more non-
greens, and I can't answer that question.
CHAIRMAN APOSTOLAKIS: That's an issue I
wanted to raise when I read the report. In Chapter 2
of the main report, January, 2001, there are four
steps that are listed in the RBPI development, assess
the potential risk input of degraded performance,
obtain performance data for risk-significant equipment
related elements, identify indicators capable of
detecting performance changes, and identify
performance thresholds consistent and so on.
It seems to me there is a major
consideration missing here, which is related to this
concern that we just discussed. When you come up with
a new indicator, shouldn't you be asking at some
point, is this information redundant with respect to
what I already have?
DOCTOR KRESS: That is the key question.
CHAIRMAN APOSTOLAKIS: See, you have to
constrain the number. If I look at these four and I
didn't know any better, because I'm sure you guys
thinks about it, but maybe you didn't state it, but if
I look at these four it's an open-ended process,
because it doesn't, at any moment trying to limit the
additional information that I'm getting from the RBPI,
and what's worse, at no point do you go back to the
baseline inspection and say, well, I've added this
performance indicator, therefore, I don't need to do
this now in the baseline inspection, and that I think
explains the concern from the licensees. All they see
is more PIs without anything else changing.
MR. BOYCE: I was going to say, you do
sound like an industry stakeholder at this point,
which explains the tone, perhaps, in that original
memo.
CHAIRMAN APOSTOLAKIS: But, there should be
something to limit the number, though. There should
be a tradeoff somewhere.
MR. MAYS: George, you're correct, and that
process is what the ROP change process is designed to
do. Our task from the RBPI development process was to
go and determine what was potentially possible to have
more direct measurement and indication of as
performance indicators for the ROP, in light of the
areas that NRR asked us to go look at. And, you are
right, this process does not limit the number.
However, we recognized, in coming up with
the number that we had, that that was a potential
issue, and NRR has recognized it, and the industry has
recognized it, and I think the judgment as to are more
indicators better, and are more indicators of value,
is something that the ROP change process has been
explicitly designed to try to answer.
So, I think that is something that we
expect will get dealt with through the ROP change
process as NRR looks at what we have technically
developed and determines whether or not it makes sense
to do.
CHAIRMAN APOSTOLAKIS: Could you add a step
like that to the
MR. MAYS: My point, George, is, that's
their step, that's not our step. Our step is to do
the feasibility to see what's technically feasible to
do.
CHAIRMAN APOSTOLAKIS: Ah, okay.
MR. MAYS: Their job is to determine, once
we've got that technically feasible product, whether
or not it makes additional benefit to the process, and
that's what the ROP change process is designed to do.
MR. JOHNSON: And, if I can add, I just
checked on the way over, George, this morning, and, in
fact, the Inspection Manual chapter that provides that
change process is Inspection Manual Chapter 0608, and
it was issued earlier this week. It's available on
the internal web, and it will be available shortly on
the external web, and it provides for considerations
of the very things that you mentioned, does it add new
data, new information, what, in fact, changes ought we
be considering with respect to the baseline as a
result of those changes.
CHAIRMAN APOSTOLAKIS: But, if I look at
the beautiful figures that Research has developed,
like Figure 2.1, RBPI development process, where the
diamonds say do statistics accumulate quickly enough
to support timely plant-specific evaluation? Yes/No.
Timely quantification. Yes/No. There should be a
diamond somewhere there that says is this additional
information useful? Yes/No. Has it already been
covered? See, it falls naturally there, I think.
Now, whether somebody else does it is a different
story, but I think this diagram can be the basis for
evaluating this additional information, and then
addressing the licensee concern, which I think is
legitimate the way we are doing it. I mean, we are
just adding things.
MR. JOHNSON: Yes, I guess the answer we
are trying to give you is that those considerations
are already built into the process, the change
process. It has diamonds with Yes/No and you advance
you don't advance based on the answers to the kinds
of questions that you are asking, and we see that.
Again, what Steve has said is, Research's effort has
been the feasibility study, based on the results of
that feasibility study as we go forward and take
candidate risk-based PIs, we run them through that
process, before implementation we have answered all of
those questions.
DOCTOR KRESS: George?
CHAIRMAN APOSTOLAKIS: Yes. I'm getting a
question.
DOCTOR KRESS: Unless degraded performance
manifests itself as a uniform change across, say,
systems and components that are risk significant, so
that when you have degraded performance they all
degrade to some extent, then I don't see how you can
think that there might be redundancy or things covered
already, because all they are adding is risk
significant components and systems.
Now, if they add systems they could be
redundancy to components, of course, that would be the
only place I would worry about it, but otherwise,
unless you are
CHAIRMAN APOSTOLAKIS: For the initiating
events what you are saying might be more value.
DOCTOR KRESS: I think it's true for
reliability and availability also.
CHAIRMAN APOSTOLAKIS: For the mitigating
systems, I'm not so sure, but even for the initiating
events, it's not just a redundancy of information, but
maybe you can consider like I think they are
already doing that, things for which you do have some
records, and others that are really so rare that you
can't build a construct a performance indicator,
and maybe if you look at this class, for example, you
can pick one that would be more or less
representative, rather than having all of them. I
mean, you can bring additional considerations into
this to try to limit the number of
MR. MAYS: George, I think you've been
reading the script again. If we can get to the point
of the things that we've tried to do to address this
issue of the number of indicators, what we've referred
to as an alternative approach, I think you are going
to see a lot of these questions or issues dealt with.
We have looked at things of that nature,
and so I'll make the suggestion that maybe we get into
the meat of it, and you'll see where that comes out.
DOCTOR KRESS: Well, let me ask one other
question before we get there, is when you developed
your thresholds, for example, your delta CDF related
thresholds, you did them one component at a time.
Now, somewhere along the line you may end up with a
number of these things degrading.
MR. MAYS: You're reading the script again.
DOCTOR KRESS: Is that in the script
somewhere?
MR. MAYS: That's in the script.
DOCTOR KRESS: Okay, well, I'll just wait.
CHAIRMAN APOSTOLAKIS: Is this the last
time we are talking about the NRR reaction today?
MR. MAYS: Unless something else comes up
as we discuss the implementation.
CHAIRMAN APOSTOLAKIS: I want to ask a
question on the memo. Is that appropriate at this
time?
MR. MAYS: You can ask anything you'd like,
George.
CHAIRMAN APOSTOLAKIS: On page 7, there's
something I don't understand, but it appears to be
related to something that Doctor Kress and I have been
discussing over the years, it has to do with shutdown
PIs and it says, "Using the current process of
comparing time and risk-significant configuration to
a year does not seem appropriate for shutdown
conditions, since the entire outage may not be a
significant time interval compared to a year," 14 days
to 365. "As a suggestion ...," this is now what I
don't understand, "... perhaps, time spent in the
risk-significant condition as a percentage of plant
outage time would be a way of quantifying this risk
significance." Can you explain that a little bit,
what the rationale is, percentage of plant outage
time.
MR. BOYCE: I'm not sure I can without
reading the memo. I can only offer to you that the
way that memo has developed, we sent around the Phase
1, draft Phase 1 risk-based PI report to several of
our technical branches, and we brought comments
together in that one memo. So, I cannot recall the
specifics of why that particular comment was written
the way it was.
CHAIRMAN APOSTOLAKIS: And, it says
DOCTOR KRESS: It sure sounds like a bad
idea, doesn't it?
CHAIRMAN APOSTOLAKIS: Yeah. First of all,
I'm trying to understand it. "Using that measure,
shorter outages would result in higher risk
significance." Now, that I just I would like to
understand a little better what the rationale for that
is, but, I mean, if you can't answer now, you can't
answer now.
MR. BOYCE: I can't answer it definitively
right here.
CHAIRMAN APOSTOLAKIS: Is there any way we
can find out, Mr. Markley?
MR. HAMZEHEE: George, I think in general
what they are trying to say is that if you shorten the
outage you end up doing a lot of maintenance
activities at the same time, during a short period.
As a result, you have more equipment out of service,
and if something goes wrong then the availability of
your safety systems are limited.
CHAIRMAN APOSTOLAKIS: Yes, but this is a
very qualitative statement that, you know, somebody
can come back and say, gee, I'm using my PRA, I'm
using OREM (phonetic), Sentinel (phonetic), and all
these things, and I'm controlling on these things, so
how can you, you know, speculate? And also, this
becomes more specific, it says, "We can compare the
time spent in the risk-significant condition as a
percentage of plant outage time," in other words the
plant outage time has some magic to it.
MR. MARKLEY: The k heat load would be the
primary thing if they go into reduced inventory, even
though they have done a lot of maintenance on line.
MR. MAYS: Let me suggest that rather than
us speculate, that if you would like to get an answer
to that we will try to determine who made the comment
and try to get something out to you.
DOCTOR BONACA: Yes, I'd rather differ, as
the comparing time and risk-significant configuration
to total outage time, in that sense if you are
attempting to shrink the whole outage time by, for
example, staying a longer time in a risk-significant
configuration, okay, versus staying with a longer
outage time, total outage, by reducing the time in
risk configuration to a shorter time, that's the
comparator I see there.
MR. MAYS: I read that as a more general
concern, quite frankly, George.
DOCTOR BONACA: Assume that you have we
are going to go through some configurations, some are
riskier than others, and you may find that you may be
able to shorten the whole outage by staying a longer
time into a risk-significant configuration. Okay?
That's the concept, it seems to me.
CHAIRMAN APOSTOLAKIS: Perhaps, what we can
do is, can we ask NRR to send us a little memo
explaining this? Is that
MR. JOHNSON: Yes. I would almost suggest,
if we could come over and I mean, I'm not sure what
your schedule is like, but we would certainly your
question is a good one, and we certainly look forward
to trying to provide
CHAIRMAN APOSTOLAKIS: No, we can address
it at the full committee, you can address it at the
full committee meeting. It can be an item to you
have plenty of time until then.
MR. JOHNSON: Sure. Sure. When is the
full committee scheduled to meet, I'm sorry?
MR. MAYS: I believe we're on Friday on the
7th.
CHAIRMAN APOSTOLAKIS: The first week of
May.
MR. MAYS: The first week of May, is it the
7th?
CHAIRMAN APOSTOLAKIS: So, you have two
weeks at least.
MR. JOHNSON: Yeah, let us come back at
that time with an answer to your specific question.
CHAIRMAN APOSTOLAKIS: Okay.
MR. MAYS: And, George, the way I read that
comment was a little less specific than you did. The
way I read that comment was as follow, as licensees go
to shorter and shorter outage times, a greater
percentage of their outage time is spent in high
relative to decay heat (phonetic) scenarios, and some
percentage of their time is spent more in mid-loop
operations, and the concern was, is that constitutes
a higher risk situation. And, the concern was whether
or not the indicators, as we've proposed in the RBPIs,
would be capable of dealing with that particular
situation.
Now, I think they do. I viewed that as a
challenge to me to get back to the commenter and
explain to them how these RBPIs will deal with the
fact that if they go to shorter and shorter outages,
and they involved greater risk scenarios, that these
would be capable of detecting them. That was the way
I took that comment.
And so, I think it's covered, but that's
part of the process we'll have to do to get back with
the people we've received comments on, as we go
through to make the final report.
CHAIRMAN APOSTOLAKIS: Okay.
DOCTOR KRESS: Certainly, it seems to me
like the appropriate thing is just what you've done,
and that's time in risk-significant configurations.
MR. MAYS: I think it addresses that
comment, but it wasn't clear to that person making the
comment that it does.
DOCTOR KRESS: Yeah, that ought to be the
appropriate way to look at it.
DOCTOR BONACA: Yes, I think here central
is the statement above in the title that says,
"Licensees are currently performing so refueling
outages are of very short durations," and that's the
focus of that. You can be, you know, more capable of
going shorter, but
DOCTOR KRESS: That ought to be covered
with what they've got.
CHAIRMAN APOSTOLAKIS: I understand it
qualitatively, but I think this goes beyond that, it
actually tells you how to do it, and I'm trying to see
what the implications would be to Regulatory Guide
1174, because you have been arguing for a long time
that it's the average over the year, and these guys
seem to be going away from that. So, I'd like to have
some further discussion. It's not just in this
context, okay, but this is something that has been of
concern to Doctor Kress and me for a while now.
MR. MAYS: Well, let me suggest that the
context that would be most appropriate for you is for
us to go back and discuss this with the person who
made the comment, and then when we have come up with
a solution, present what the solution is to you and if
you agree with it then it doesn't matter what the
comment was.
CHAIRMAN APOSTOLAKIS: That's fine.
Okay, so I don't know why Tom took so long
to finish just the
MR. BOYCE: I apologize for that.
CHAIRMAN APOSTOLAKIS: Apologies accepted.
MR. MAYS: Tom has difficulty not talking
a lot, and he really is
CHAIRMAN APOSTOLAKIS: So, we'll go back to
Steve now.
MR. BOYCE: Steve made the comment we
shouldn't send him comments.
CHAIRMAN APOSTOLAKIS: Okay, Steve.
MR. MAYS: Okay.
The rest of what we are going to present
today is primarily the results of our stuff. Mike and
Tom will be sitting over here at the side if there's
any other questions.
So, what I suggest we do here, if it's all
right with you, the first portion of this is
discussions of the potential benefits before we get
into the summary of the thing, so if you want to do
those first and then I didn't know what time you
wanted to take your first break. Okay, so let's go
through the benefits first.
What we have outlined in this report is
some of the things that we think are the benefits of
RBPIs, and the first one which answers part of the
question you raised, George, is why we even want to do
this. Well, one of the reasons is, we get a much
broader sample of risk performance with this set of
indicators than we do with the current ROP, and they
are a more objective indication because they are more
directly tied to plant-specific performance and with
a relationship to the plant-specific thresholds.
So, we believe that's a positive, that's
one of those progress versus perfection things, that's
one of those potential benefits that we think this
thing has.
Also, years ago, NEI submitted a document,
a white paper, to us, NEI 96-04, which was their paper
on risk-based and performance-based regulation, and
they wanted us to move in the direction where we had,
as just quoted here, a regulatory approach that more
directly considered operating experience and the risk
implications of it, and performance-based process
where we had measurables, and objective criteria, and
specified reactor or, specified activities that the
NRC would take and flexibility for the licensee as
long as they were performing in an appropriate band.
Well, I think the ROP process reflects those general
principles, and the RBPIs are an example of a more
direct approach to applying operating experience and
probabilistic safety assessment judgments as to how we
would go about doing that.
DOCTOR KRESS: Steve, I think the word
"sample" within that dot is a really key word.
MR. MAYS: That's an important word.
DOCTOR KRESS: Because you are not
measuring the full performance always, you are taking
a sample.
MR. MAYS: That's correct.
DOCTOR KRESS: And, you are going to infer
from that what the full performance is, and I think
that's a key concept in this whole thing.
MR. MAYS: I agree, that is a key concept.
The issue that's part of built into the Reactor
Oversight Process is that the indicators will be a
sample of performance, and the inspections will be the
process by which we go out and sample the rest of the
performance, as it relates to meeting cornerstone
objectives. So, again, this is a balance of how much
of your sampling you want to spend in the PIs, how
much of your sampling do you want to spend in
inspection, and remember, a key part of this Reactor
Oversight Process is not that the NRC does all of the
sampling, it's that the licensees do the sampling,
that their problem identification and corrective
action programs are the key behind all this, that they
are continuously sampling and looking for things, and
we have a smaller subset that we look at to provide us
with the assurance that they are doing their job
right. So, that's an important point, I think, to be
raised.
In doing the sample with the RBPIs that we
proposed, we've got more systems and more components
covered by more objective and more risk-informed
methods than the current ROP has. And, in the
indicator space, we have some indicators that go
across system boundaries and across the breadth of the
plant, and we believe that's an important piece
because one of the issues earlier raised was what
about crosscutting issues? Well, what if I have my
maintenance program degrading and I just don't happen
to see it in my diesel generator or my HPI indicator,
how will I know that my plant is getting worse? Well,
by having some of these indicators that go across
systems, we think that might help address some of
those issues from an indicator standpoint. The rest
of it has to be addressed through inspection.
CHAIRMAN APOSTOLAKIS: But, you are saying,
on page A-25, "Currently, there is no established
method of identifying changes in operator performance
and then feeding this information back into the SPAR
models. As a result, equipment performance is the
only mitigating system out there that will be
evaluated in this analysis." Are you saying there
that the crosscutting issue of safety conscious
environment, and the corrective action program, cannot
have performance indicators, we have to do something
else about them?
MR. MAYS: I'm saying I don't have anything
readily available now to do it. I'm not saying it's
impossible to develop it, but I'm saying I don't have
that capability right now. The capability I have
right now is to reflect whatever operator performance,
with respect to safety culture, with respect to
maintenance program, as to how they manifest
themselves in respect to the availability and
reliability of the equipment. So, I can't directly go
out right now and measure the safety culture at the
plant, but I can go out and measure whether the safety
culture of the plant has had an impact on the
availability and reliability and the frequency of
events.
CHAIRMAN APOSTOLAKIS: But, I thought
equipment performance was taken as a separate
attribute from the human performance. In other words,
if it's a valve, and it is left inadvertently closed,
would that be part of the indicator for the valve?
MR. MAYS: Yes.
CHAIRMAN APOSTOLAKIS: Because even though
it was not a fault of the valve itself?
MR. MAYS: Correct.
MR. HAMZEHEE: But, it wasn't available.
CHAIRMAN APOSTOLAKIS: Huh?
MR. HAMZEHEE: But, it wasn't available.
CHAIRMAN APOSTOLAKIS: It was unavailable.
MR. HAMZEHEE: It's reflected in the
unavailability of that equipment.
CHAIRMAN APOSTOLAKIS: And, you will keep
track of the fact that it was a human error?
MR. HAMZEHEE: The cause would show, yes.
MR. MAYS: Well, the RBPIs would not
reflect the fact that it was a human error. The basic
data that was going into the RBPIs would be available
to us, so that if we determined that somebody's
performance was requiring additional regulatory
attention, we could go back and look at the
information and say what was causing this to be a
problem, and then use that as part of our guidance for
how we go and look at the plant.
The issue that we were raising in that
particular point was that we don't have direct human
performance measures that we are going to have
indicators for.
CHAIRMAN APOSTOLAKIS: So, these then
crosscutting issues should be part of the baseline
inspection.
MR. MAYS: They are.
CHAIRMAN APOSTOLAKIS: Okay.
MR. MAYS: And, this would be a case where
we would have more direct objective indicators of some
of the impacts of that.
CHAIRMAN APOSTOLAKIS: I thought they were
not.
DOCTOR KRESS: I didn't think they were.
MR. MAYS: Well, the crosscutting issues
no, the crosscutting issues are dealt with in the ROP
through the problem identification and resolution
inspections, to determine whether or not the plant has
an appropriate process by which they can manage those
kinds of issues.
CHAIRMAN APOSTOLAKIS: Well, is that true,
Mike? I don't remember. Oh, it's not that he's
lying, but
MR. JOHNSON: I'm sure it was true,
although I've got to confess I was talking. I didn't
hear the total comment.
MR. MAYS: The additional benefits that we
alluded to earlier has to do with the fact that we
have a better understanding of plant-specific
implications using these than we necessarily had with
the current ROP. Our thresholds are set on the basis
of plant-specific models. We don't average diverse
systems together, which can potentially mask the
contribution. For example, in the ROP, the turbine-
drive pump trains and the motor-drive pump trains are
AFW, their unavailability is averaged, and that's the
value that's used in the PI. Well, turbine or diesel-
driven pumps have different risk significance than
motor-driven pumps because of the station blackout
risk issue, and so the RBPIs that we proposed allow us
to deal with that.
The other thing that I think has come up
on a couple of occasions in the current ROP that has
been dealt with in the RBPIs is whether the failures
affecting the reliability and availability indicator
that you might have, whether they are based on the
risk-significant functions or whether they are based
on design basis functions. The example that comes to
mind was the, I believe it was Quad Cities had a case
where they ran their once a cycle test of their HPCI
system to see it it would automatically actuate, and,
in fact, it wouldn't. There was a problem with the
automatic circuitry to start the HCPI system.
Now, over the period of the cycle, they
had been manually starting the system every month or
quarter or something like that, and it was working
just fine. So, what happened was, they determined
that they had a problem with the automatic feature for
this system, and the fact that they had not tested it
since the last outage meant that they had nine months
of fault exposure time to put into the indicator.
Well, that indicator had nine months of fault exposure
time, which only represents that it wouldn't have
performed its automatic start capability, while it's
manual capability was not degraded at all.
And, from a risk perspective, having the
manual ability to start HPCI is success, so one of the
things we've done in the RBPI program is to deal with
the difference between auto and manual and design-
basis requirements versus risk-significant
requirements for the equipment to operate.
We've also had a different way of treating
fault exposure time than was in the current ROP, which
we believe is more consistently accounted for and is
more consistent with the way risk assessments are
done. The issue there having to do with the fact that
in the current ROP there are no reliability indicators
per se. Fault exposure time was included in the
availability indicator as a sort of surrogate for
having a reliability indicator, and because of the
relatively short time period under which the
availability is gathered, and the fact that the fault
exposure time every time you do have one of these
failures can be fairly long depending on its nature,
you have a false positive/false negative problem which
goes back to the old thing that Hal Lewis always
talked about, trigger values. The RBPIs don't have
that same problem because we classify the failures as
either demand-related failures or not, and for those
we use a probability calculation and distribution for
reliability rather than use the fault exposure time.
For fault exposure times associated with discovered
events, for which there was no demand, those go into
the unavailability in the RBPIs. So, we have a more
consistent way of dealing with that, which we believe
tends to reduce the problems that were currently being
experienced in the RBPIs or in the oversight process
with fault exposure time.
DOCTOR KRESS: When you determine
unavailability, is it true that you count into that
unavailability time the time spent testing a piece of
equipment?
MR. MAYS: If it's out of service and not
capable of being used while that test is going on,
yes.
DOCTOR KRESS: I personally think that's a
mistake to do that, but we can discuss it later. It
does a lot of it has a lot of negative aspects to
counting that in there, one of which is, when they do
this testing, they are on a higher alert and the
operator error in doing some compensatory measure is
probably much less than it would be, so the risk is
not the same as it would be if it were just
unavailable because it was not functioning correctly.
And, not only that, it gives a negative
incentive to not test as often.
MR. MAYS: Well, only if you are doing a
lot of testing to the point where it might reach a
threshold to contribute to your
DOCTOR KRESS: Of course
MR. MAYS: this is the classic issue
from the maintenance rule.
DOCTOR KRESS: I'm being too general
with this, but then
MR. MAYS: This is the classic issue from
the maintenance rule, the balance between the time you
spend in testing and maintenance and the impact on
reliability and risk. So, that's a problem, I haven't
resolved that problem, I'm just trying to be
consistent with the current approach.
Additional benefits that we have, this
process was designed so that the RBPIs would look
similar to the current performance indicators, that we
would have similar color scheme, they'd be amenable to
similar kind of presentations on the web site, and
they could be updated in a similar fashion that the
current process has.
One of the things we've also noted is that
these don't have to be implemented, it's not an all or
nothing deal. In other words, portions of these can
be implemented, some of them can come along later as
data, and availability, and quality become better, so
this is not an all or nothing deal.
DOCTOR KRESS: The nice thing about these
performance indicators that you have now is, you could
actually calculate a delta CDF. You could take the
set of performance indicators at some time and stick
them in a plant-specific model and get a delta CDF.
MR. MAYS: You are reading the script
again, Tom. That's correct. One of the things we had
as part of the Phase 2 work that we had originally
proposed was to look at how we might develop an
integrated indicator.
DOCTOR KRESS: That could be the integrated
indicator.
MR. MAYS: So, that's part of what you are
going to see a little bit later on.
You were correct in stating earlier that
all of the indicators we have now in the report, and
the current Reactor Oversight Process indicators, are
all basically single variate sensitivity analysis on
a larger model.
DOCTOR KRESS: Right, but you could take
the whole shebang and put it in and calculate it.
MR. MAYS: The issue then is, are there
synergies among these things that would make them go
up, down, or sideways, if you had a more integrated
look. We'll talk some more about that as we get
further in.
The other thing I wanted to mention,
because this became a point of confusion with people
both internally and externally, that the RBPIs, while
we went back and did a lot of work looking at
statistical methods to determine what's the right time
intervals, what's the right method of calculating
these things, and what's the process for setting the
thresholds, that this isn't something dramatically
exotic. We are using off-the-shelf, readily-available
models. The analysis routines that we are planning to
use are fairly simple, and most of the data we've got
is from readily-available current databases, there's
not, with a couple exceptions, any new information
that really needs to be required to make this happen.
So, most of the stuff is fairly easy to get and to
use.
So, we can get into some of the results
now. We talked about the four elements, George
brought them up earlier, about how we were going about
doing that. We wanted to look for areas where there
would be risk impact of performance if the plant was
degraded, find out if we could get data on that
information, make sure that if we did that we could
the tech changes in a timely manner, and then be
consistent with the 99-007 general rule process.
Now, what that means in a practical sense
is that, in order to do that you have to have three
things. You have to have a model that reasonably
reflects the risk at the plant. Now, the word I want
you to concentrate on there is reasonably. We were
talking about the progress, not the detection mode
before, what we have to have is a model that has some
fidelity to the risk at the plant, in order for us to
believe that we have something that goes on, is real.
Then we have to have some baseline performance to put
in that model in order to be able to say, this is our
starting point, and then we can vary the model off of
the baseline to determine what the impact is of
changes in the performance.
And, the last thing you have to have in
order to be successful at doing this is, you have to
have an ongoing source of performance data for
assessing the plant-specific performance. And, what
you'll see as we go through the rest of these is,
there were some cases where we had all three of those
things and we've made proposals, and some cases where
we didn't have them, and so, therefore, we weren't
able to do performance indicators on those areas.
CHAIRMAN APOSTOLAKIS: Should we take a
break now?
MR. MAYS: Sure, if you want to take a
break now, that's no problem.
CHAIRMAN APOSTOLAKIS: Until 10:00.
(Whereupon, at 9:44 a.m., a recess until
10:00 a.m.)
CHAIRMAN APOSTOLAKIS: Back into session.
Mr. Mays, continue, please.
MR. MAYS: Okay.
The first thing we are going to talk about
from the results of the RBPIs is the work we did in
the initiating event cornerstone, which was for full
power internal events. We used three data sources for
putting this stuff together, new Reg 5750, which was
the initiating event report which we did a couple
years ago and you've seen. We used the Sequence
Coding and Search System, which has the LER
information, which is the source of information about
plant trips, and the Monthly Operating Reports, which
gives us the critical hours information for the
plants. All those sources, by the way, are publicly
available, there's no issues with respect to
availability and quality of that stuff as far as
implementation goes.
So, we went back and in going through the
process we just discussed we determined that there was
three RBPIs we could do for each plant, and the tables
are listed as to where they can be found in the main
report and the appendices. The important part about
here was how we came up with the calculations of the
frequencies. Now, the current ROP merely counts the
number of SCRAMS you have and goes on from there. We
were looking more at the classical PRA definition of
establishing a frequency which has distribution
associated with it, and so we were looking to see what
we could do in terms of prior distributions for
figuring these out, and we had three options that we
pursued.
One was to start with, basically, a non-
informative prior, kind of a classical statistical
approach, how many failures did you have in how many
years, and that's your estimate.
The next thing we looked at was taking an
industry prior, which would be to say you would take
the distribution of the industry population and update
that with the plant-specific information.
CHAIRMAN APOSTOLAKIS: Can you tell me,
Steve, where you did all this stuff?
MR. MAYS: It's in Appendix A, I believe.
CHAIRMAN APOSTOLAKIS: Appendix A, I don't
recall seeing prior distributions. Maybe I missed it.
MR. MAYS: Just a second, let me find it.
MR. HAMZEHEE: Steve, Appendix F.
MR. MAYS: F?
MR. HAMZEHEE: Statistical Methods and
Results, yes.
CHAIRMAN APOSTOLAKIS: Oh, F.
MR. HAMZEHEE: Yes.
CHAIRMAN APOSTOLAKIS: Okay, thanks.
MR. MAYS: So, and then the last one we
tried was one that you've seen before in reports that
we've given you on system and other studies on
constrained, non-informative prior.
CHAIRMAN APOSTOLAKIS: So, this appendix
will tell me how the choice of the interval
observation was made?
MR. MAYS: Yes, right.
CHAIRMAN APOSTOLAKIS: So, what is it,
between one and five years?
MR. MAYS: Well, that's the next bullet
down. Let me explain what we were doing. We tried
three different priors to see which one would give us
the best performance that we were looking for, in
terms of being able to give us timely indication, not
give us too many false positives or false negatives,
and to be amenable to being done with the ROP process.
So, as it turns out, we were looking at
the time intervals. What we wanted to do is take the
shortest time interval that would give us indication
of performance degradations for which we wouldn't have
a false positive or false negative rate that was
excessive. And, by a false positive rate, what I mean
is that there would be a significant chance that
performance hadn't degraded, but the way you calculate
it it would send you over the threshold.
Then, the false negative would be the
situation where if you had a significant degradation
in your performance that you would go over the period
of time that you were looking and wouldn't have enough
data collected to see the changes that occurred. So,
that's the simple basis of what we did.
So, when we looked at that for the
initiating events, we used one year as the time
interval for the category referred to as general
transients, that's trips, the plant trips, but the
safety systems needed for decay heat removal, for
activity control, that kind of thing, are not affected
by the trip itself, and we also came up with three
year intervals for loss of feedwater and loss of heat
sink events, which are trips that are a little more
complicated and the ability to remove decay heat is
impacted directly by the trip itself.
For other risk-significant initiators that
you typically find in PRAs, like losses of off-site
power, steam generator tube ruptures, small LOCAs and
other initiators, the problem we had there was that
the frequency of occurrence on a plant-specific basis
of those things was so infrequent that over a you'd
have to take more than a five-year period to be able
to even see that, and that didn't seem to be
consistent with the ROP philosophy, which was to go
back every year and see where the performance was
going so that we could see what we needed to do more
of with respect to those indicators.
CHAIRMAN APOSTOLAKIS: Well, that's a very
important point, though, and I must say that I haven't
really read Appendix F in detail, but I was doing my
own calculations and I used as an example Table A.1.4-
2C plant, two initiating event threshold summary. It
seems to me that the results of the aleatory part, the
randomness issue that we have addressed here, which is
something that the quality control people do, so we
have two thresholds here. One is green/white, which
is .8, right?
MR. MAYS: Which page are you on there,
George?
CHAIRMAN APOSTOLAKIS: A-17, it's just
numbers on here, but A-17. So, we have the
green/white 8E-1, right, and the baseline was 6.8E-2,
right, the same table?
MR. MAYS: Yeah, why don't you just flip to
the next page in your presentation, that particular
chart is right here.
CHAIRMAN APOSTOLAKIS: Okay, fine, but if
the question is now, how long should the interval
be, so that the calculation of the rate will be
meaningful, and I would have some sort of conclusion
that I'm near the baseline or the white region, and
for the numbers here I calculated that to be about ten
years, which is really too long, as you just pointed
out.
And, the problem is this, that because
these numbers are very low, if you see nothing, that
doesn't necessarily mean you are near the baseline,
you could be near the you could be in the white,
because it's .8. If your observation is one year
MR. MAYS: Well, in this case, loss of
feedwater is three years.
CHAIRMAN APOSTOLAKIS: Okay, the three
years I think we are beginning now to be a little
better, but I think the analysis maybe I should
send you a little memo with what I did so you can tell
me what I did wrong.
MR. MAYS: No problem. The transient
initiator we used one year, loss of feedwater, loss of
heat sink we used three years.
CHAIRMAN APOSTOLAKIS: It's three years,
yeah.
MR. MAYS: And, when we got to that point,
you still have the possibility, because this is a
distribution, we are calculating a frequency and it
has distribution, but we are comparing the mean of
that distribution to a specific value for the
threshold. So, there's always the possibility of
false positive/false negative with that.
CHAIRMAN APOSTOLAKIS: What's new here, I
think, not new, but in PRAs typically we deal with
systemic uncertainty, the uncertainty, failures rates
and the initiating event frequency. Here you have to
worry about the aleatory part, too, because you are
talking about real occurrences. So, the fact that
they have seen none in the last two years is that due
to chance, and my rate of occurrence is, in fact,
high, but I just happened not to see it, or is it
because the rate is low, and that's the key question
that the quality control people are asking.
MR. MAYS: Right, and what we've done there
is, we've asked a slightly different question. We
didn't ask the question, is the mean what we are
calculating here, the "correct mean." What we are
saying in this situation is, if there was substantial
degraded performance, is a three-year period enough so
that we would be able to detect that it wasn't green
anymore, and the answer to that question is yes.
Now, the
CHAIRMAN APOSTOLAKIS: With a certain
confidence, though.
MR. MAYS: Right.
The other side of that coin is, okay,
suppose I do have a few events in a one, or two, or
three year period, does that necessarily mean that my
frequency has, you know, gone up.
CHAIRMAN APOSTOLAKIS: Yes.
MR. MAYS: And, what I'm saying is, the way
we've dealt with that is that we've dealt with that
issue, the problem I think you may be looking at is
the classical issue is, if my frequency is around .07
or so, then in three years can I get enough faults
where X over three years tells me something. That's
the problem you get when you use the classical
approach. What we've done instead is, we've said the
industry average is this number, .068, we used a
constrained non-informative prior, and we've used a
Bayesian update of that to get the new distribution.
So, what we don't have is, we don't have the same
amount of problems with either the inability to detect
changes with the classical approach from a false
negative standpoint, or a positive standpoint, and we
don't get the false negative problem we have when you
use the industry prior by itself, which means that you
have to have an awful lot of data at the plant to
overwhelm the industry prior.
So, the constrained non-informative prior
seems to be the middle ground between those two
extremes that works best, and that's what we chose to
use because it had the lower it had the performance
characteristics because it's a competing interest.
False positives and false negatives are competing
interests, so that's the way we did that.
MR. HAMZEHEE: I think, George, maybe when
you were doing your calculation you did not use any
prior distribution.
CHAIRMAN APOSTOLAKIS: I didn't.
MR. HAMZEHEE: That's the reason.
MR. MAYS: That's the classical approach.
MR. HAMZEHEE: So, you just used a direct
number and you get ten or 20 years sometimes to get a
reasonable number.
MR. MAYS: Right.
CHAIRMAN APOSTOLAKIS: Well, actually, for
the one standard deviation the interval is only 2-1/2,
so it's not bad, 2-1/2 years, it's reasonable.
MR. MAYS: Yes, as it turns out
CHAIRMAN APOSTOLAKIS: But, still, though,
there is a I mean, it's only one standard
deviation, so the probability that I'm wrong is not
negligible.
MR. HAMZEHEE: Oh, yeah.
CHAIRMAN APOSTOLAKIS: But, I'm going to
read Appendix F, so at the full committee meeting
we'll have a more meaningful discussion.
DOCTOR KRESS: What is your rationale,
justification, for using the industry distribution as
your prior?
CHAIRMAN APOSTOLAKIS: Yeah.
DOCTOR KRESS: Do you think that really has
the technical justification?
MR. MAYS: I think it does have a technical
justification. The issue there becomes one of and
this is a standard PRA issue that goes on you have
a limited number of a limited amount of data at any
one particular plant, and you have to go a long time
to collect data only at that plant. And, the example
I use for people is, go to Atlantic City, do I need to
go to every table on the roulette thing at every place
in Atlantic City and take infinite data on each one to
know something about their performance, or can I take
some data over a group and then go back to any
particular table and monitor its performance relative
to the group to see if it's different, and that's
basically the approach that we're taking here.
There's a ceratin amount of time that you
can take your sample for, to get enough information to
do what you need to do.
DOCTOR KRESS: I understand that
constraint, but I still don't believe
CHAIRMAN APOSTOLAKIS: I guess the
rationale is, my plant could be any one of these,
okay, and the plant-to-plant variability gives me the
fraction of plants that have this particular value.
It could be any one.
DOCTOR KRESS: Your assumption, though, is
that that distribution basically applies to the
distribution at that plant.
CHAIRMAN APOSTOLAKIS: Yeah, that it's one
of those.
MR. MAYS: No, not quite, Tom. If you go
out and you were to calculate, like we did in new Reg
5750, what the frequency was for either PWRs, or BWRs,
or the population of plants, when we did that
calculation we put a distribution on that that
represented the plant-to-plant variability in the
population. What we are using here is not that
distribution itself, that would be the industry prior
distribution, and that tends to give you a problem in
that you can have significant degradation and it takes
you a long time for your plant-specific performance to
overwhelm that initial prior.
What we did instead was, we took that
industry distribution, we took the mean out of that
distribution, and then we constructed a constrained
non-informative prior where we diffused the
distribution, so that what you see when you do the
update is the impact more of the plant performance
than of the industry performance, and that helps us
resolve, I think, the issue you are talking about.
DOCTOR KRESS: Yes, I think that would
help. I still think there's a problem with choosing
that mean also. I'll have to read it a little more
closely.
MR. MAYS: We checked in the if you'll
recall, the 5750 to determine whether or not we were
seeing plant-to-plant variability on those things, and
so there's a means of being able to deal with that.
DOCTOR KRESS: Well, you know, eventually,
though, you keep updating and the problem will go
away.
MR. MAYS: Correct.
DOCTOR KRESS: Eventually.
MR. MAYS: Given enough time.
DOCTOR KRESS: Given enough time, yeah.
MR. MAYS: So, that's what I was going to
do next, was turn to this page here with
CHAIRMAN APOSTOLAKIS: And, this 95th
percentile column is explained somewhere in the
appendices?
MR. MAYS: Yes, that's what the value for
the threshold would be if you took the industry prior
and
CHAIRMAN APOSTOLAKIS: Oh.
MR. MAYS: set the 95th percentile on
there.
DOCTOR KRESS: Just like they did in the
original ROP.
CHAIRMAN APOSTOLAKIS: So, the green/white
is on the second table .8, and the industry average
95th percentile would be .2, am I correct?
MR. MAYS: Correct.
CHAIRMAN APOSTOLAKIS: The industry would
be .2, l8, so it's higher?
MR. MAYS: In some cases.
CHAIRMAN APOSTOLAKIS: Plant-specific is
higher?
MR. MAYS: In some cases, in some cases
it's higher, and in some cases it's significantly
lower. It depends on how you go. There were examples
where that would happen, less so with the initiating
events, but more so with the availability and
reliability situation.
CHAIRMAN APOSTOLAKIS: It should be lower,
though, should it not, as a rule?
MR. HAMZEHEE: Well, usually yes, because
you are talking about 95 percent.
MR. MAYS: Just a second, George, I think
we may be making a difference. The value in the 95th
percentile column is the value that corresponds to the
95th percentile of the distribution.
CHAIRMAN APOSTOLAKIS: The industry.
MR. MAYS: Of the industry.
CHAIRMAN APOSTOLAKIS: So, that should be
higher than 95 percent of the plants, right?
MR. MAYS: Correct.
CHAIRMAN APOSTOLAKIS: Right.
MR. MAYS: Now, what we are saying is, for
this particular plant, and remember the baseline was
.07, the 95th percentile in this case was .2.
CHAIRMAN APOSTOLAKIS: Uh-huh.
MR. MAYS: All right. We are saying that
the risk contribution from changing from .68 to .8
gives us the delta risk increment, whereas the 95th
percentile just tells you how much it varied among the
plants.
CHAIRMAN APOSTOLAKIS: Oh, oh.
MR. MAYS: There's no direct relationship
between those two. We were showing where you might
set the threshold if you used the 95th percentile
approach, which is deviation from normal performance,
versus a risk threshold approach.
CHAIRMAN APOSTOLAKIS: So, what they should
really be comparing is the first two columns, and
there the baseline is lower.
MR. MAYS: Correct.
CHAIRMAN APOSTOLAKIS: Okay.
MR. MAYS: And, what we found was, is that
sometimes, in certain cases, as we tried to apply this
concept uniformly through the plants, sometimes you
would find cases where you wouldn't exceed the
green/white threshold until you were already up in the
yellow.
CHAIRMAN APOSTOLAKIS: Yes.
MR. MAYS: And, we said, that doesn't seem
to be a smart thing for us to do.
CHAIRMAN APOSTOLAKIS: So, how many plants
do you expect to see with 67 transients a year?
MR. MAYS: None.
MR. HAMZEHEE: That's just a number.
MR. MAYS: None, the point is, and this
goes back to Tom's earlier point, what we have
basically here
MR. HAMZEHEE: That was mine.
MR. MAYS: or your point, or whatever,
is that you have a single point variance analysis.
Now, what that tells you is that if everything else in
the plant were to stay the same, except for this
input, how high would it have to go to get me to an
increase in core damage frequency of E-4.
DOCTOR KRESS: But, you are never going to
see that. If you get that bad
MR. MAYS: The realities of I think
everybody will agree the realities are that other
things will go wrong before you get to that point, and
we'll find something and be able to deal with it
before it gets there.
CHAIRMAN APOSTOLAKIS: In other words, if
I go to the Action Matrix I will see enough whites and
greens, whites, way before I see any reds, unless it's
some sort of a major disaster.
MR. MAYS: Well, you know, I'm saying, I'm
not sure how anybody engineering-wise would be able to
trip their plant ten or 15 times a year without having
other problems in the plant that would manifest
themselves, too.
CHAIRMAN APOSTOLAKIS: Without having the
NRC
DOCTOR KRESS: That's one reason I question
the usefulness of that whole problem.
MR. MAYS: And, I understand that, that's
always going to be the case when you have, risk is a
function of multiple variables.
DOCTOR KRESS: Yes, absolutely.
MR. MAYS: And, if you have indicators that
are single variable sensitivity analysis, you always
have the issue of, is it realistic that this is the
only thing that will change?
CHAIRMAN APOSTOLAKIS: Wait a minute, now.
Isn't that dependent also on the baseline core damage
frequency?
MR. MAYS: Absolutely.
CHAIRMAN APOSTOLAKIS: For plants that are
already I mean, 19 units that are above, then you
shouldn't expect 67 to be in the red.
DOCTOR KRESS: No, you might
CHAIRMAN APOSTOLAKIS: In fact
DOCTOR KRESS: yeah, but
CHAIRMAN APOSTOLAKIS: as it should be.
DOCTOR KRESS: the question is, I'm not
sure where you see that reflected in the thresholds,
because the thresholds don't use the absolute value in
them. That's another thing that bothers me.
MR. HAMZEHEE: No, they use the impact on
the CDF.
MR. MAYS: Right.
DOCTOR KRESS: It's the delta, they just
use the delta.
MR. HAMZEHEE: Based on the delta CDF, you
set the value.
CHAIRMAN APOSTOLAKIS: Yeah, but if you are
already high.
DOCTOR KRESS: It doesn't matter.
CHAIRMAN APOSTOLAKIS: It doesn't matter?
MR. MAYS: Not quite.
DOCTOR KRESS: And, that bothers me a
little bit.
MR. HAMZEHEE: Well, but it shows the
importance of that.
MR. MAYS: Not quite.
CHAIRMAN APOSTOLAKIS: It does not, do you
agree with that?
MR. MAYS: Not quite. It depends on all
the other things that are in the model together. This
is the issue we are raising in the first place. You
have some baseline core damage frequency, depending on
the model of your plant.
DOCTOR KRESS: Yes.
MR. MAYS: And, depending on that, and the
relationship between the initiator frequency or the
diesel generator reliability, or whatever else is in
your model, you can vary that, and if you start at a
lower baseline you have to have greater changes in
order to get to a E-4 delta CDF. However, if you start
with a E-4 delta CDF you still have to have a certain
amount of change to go to 2E-4, which is what this
threshold would be measuring. So, this threshold
measures change in the CDF of the plant, it does not
measure directly the total absolute CDF. You can go
back and figure it out if you wanted to, but that's
another issue for the integrated indicator, which was
the thing we talked about before.
MR. HAMZEHEE: It also shows for that
specific plant that general transient by itself is not
very risk significant. In other words, you are never
going to change your CDF by greater than 1-4 unless
you go above 67 trips per year.
MR. MAYS: Yeah, that's the other thing it
tells you.
DOCTOR KRESS: Yes, and that's a
significant piece, an incite, I think. But, you know,
we are speaking in general when we talk about it, even
the other PIs, not just this one, that it seems like
the absolute value ought to be reflected in there
somewhere, and I don't think it really is.
CHAIRMAN APOSTOLAKIS: Somehow.
MR. MAYS: We chose as part of the ROP
philosophy that what we were going to do was, we
started with the basic assumption that the design and
operation of the plants was basically safe, and then
our job was to be able to detect changes in
performance in the plants that might be more risk
significant, so that we could engage them. So, that
philosophy is what determines this.
DOCTOR KRESS: Yeah, but you can turn that
around a little bit.
CHAIRMAN APOSTOLAKIS: That's a very good
point, actually.
DOCTOR KRESS: Yeah, but you can turn it
around and say there are some plants that are not just
basically safe, but, really, really good risk status.
CHAIRMAN APOSTOLAKIS: So, you are
penalizing those.
DOCTOR KRESS: And, you are penalizing
those.
MR. MAYS: No, actually, we are not,
because we are saying they have to demonstrate that
their change in performance is significant before we
go to them, and we're saying, what's the definition of
significant, it's consistent with the existing
philosophy, you've increased your change by a certain
amount.
DOCTOR KRESS: Yeah, but you could allow
those plants to degrade their performance without
worrying so much about it.
MR. MAYS: We're taking the same absolute
change in performance for all the plants.
DOCTOR KRESS: I understand.
CHAIRMAN APOSTOLAKIS: I think you made a
good point, Steve, but maybe we ought to think a
little more about Tom's point, too, but I think your
point is well taken.
DOCTOR KRESS: Yes, I think it's not a bad
point, I'm not totally disagreeing with you.
CHAIRMAN APOSTOLAKIS: But, here is the
place where I think we can revisit the question of
putting constraints on the proliferation of the number
of RBPIs. You state in the report that the loss of
feedwater and loss of heat sink are performance
indicators that are not in the existing Revised
Oversight Process, and they just talk about
transients.
MR. MAYS: Well, actually, they have two.
They have
CHAIRMAN APOSTOLAKIS: Unplanned SCRAMs.
MR. MAYS: they have three in the
initiating event cornerstone, they have unplanned
SCRAMs, which is just a count of all the SCRAMs.
CHAIRMAN APOSTOLAKIS: Yeah.
MR. MAYS: They have the number of
CHAIRMAN APOSTOLAKIS: Significant power
changes.
MR. MAYS: power changes, and they have
one that kind of represents feedwater and heat sink
combined.
CHAIRMAN APOSTOLAKIS: Right.
MR. MAYS: So, this one is
CHAIRMAN APOSTOLAKIS: But, the question
is, I think this is where we could ask the question,
is it worth treating them separately, so that the
number of performance indicators increases to the
dismay of the industry?
MR. MAYS: Actually, in this case the
number wouldn't change.
CHAIRMAN APOSTOLAKIS: But, why?
MR. MAYS: If you had three, you would have
three, so there wouldn't be any net change if you were
to make a complete swap out.
CHAIRMAN APOSTOLAKIS: In terms of
collecting data, it wouldn't make any difference, I
agree.
MR. MAYS: No, and it wouldn't make any
difference if
CHAIRMAN APOSTOLAKIS: But, in terms of
having more indicators it really does make a
difference. You have three now, they had only two.
MR. MAYS: They had three.
MR. HAMZEHEE: They have unplanned SCRAMs,
loss of normal heat removal pump and reactor power
changes.
CHAIRMAN APOSTOLAKIS: Where would you put
the unplanned SCRAMs, in general transient?
MR. HAMZEHEE: Yes, usually.
MR. MAYS: We would substitute general
transients for unplanned SCRAMs.
CHAIRMAN APOSTOLAKIS: Significant changes
in power?
MR. MAYS: We wouldn't use those because we
can't make a relationship between risk in that.
CHAIRMAN APOSTOLAKIS: So, in this
particular case you would preserve the number.
MR. MAYS: Well, that would be a decision
for NRR to say whether they were going to preserve it
or not preserve it.
CHAIRMAN APOSTOLAKIS: No, I understand
that, no, but let's not avoid the thing. I want to
get into the question of whether loss of feedwater and
heat sink, how can we scrutinize them? Let's say that
the other numbers don't change, or they change, what
kind of criteria would we be using to decide that,
yes, loss of feedwater deserves to be a PI by itself
because it gives me this information that I don't have
otherwise, or it does not because it doesn't really
add anything.
You know, this is, I think and we don't
necessarily have to have the answer today, but I think
it's an important question.
DOCTOR KRESS: But, I think the answer is,
is it by itself risk significant?
CHAIRMAN APOSTOLAKIS: Well, actually,
Hossein gave an answer, he said that their risk
implications are different.
MR. HAMZEHEE: And, that's the main reason
for this study to treat them separately.
CHAIRMAN APOSTOLAKIS: And, you should
emphasize that and tell the NRR folks that this is an
important consideration, that it's not just the number
of the PIs that matters.
MR. HAMZEHEE: Correct.
CHAIRMAN APOSTOLAKIS: But, in this
particular case you are also eliminating one or two,
but in others you might not, although I didn't see
again, Steve would say that that's for NRR to decide.
MR. LEITCH: But, are we losing some
significant piece of information by eliminating
unplanned power changes? Say it again, you can't draw
a connection between that and the risk?
MR. MAYS: I'm saying, I don't have to
go back to the three things I needed to be able to do
an RBPI, I have to have a model that reflects plant
risk, I have to have a baseline performance that
allows me to make changes to that model to set
thresholds, and then you have performance data. I
can't make, in my risk information, a link between
going from how often I go from 80 percent to 30
percent power at the plant to what that has to do with
risk. And so, therefore, I'm not able to make a risk-
based performance indicator from that.
But, whether or not that means that that
PI is useful for other reasons is something that NRR
would have to decide, as to whether or not they wanted
to keep it, or not keep it, and I'm not making that
judgment here. I'm saying what risk-based performance
indicators am I capable of putting into play.
MR. JOHNSON: Yeah, and that's I'm
sorry, Steve, I just was anxious to add to the point
that you were making. You know, if you look at some
of the PIs that we have now, we've said that they've
not been not all of them have been risk informed,
but, for example, we know that when you look back
historically at plants that have had a significant
number of power changes as a result of equipment
problems, to address those equipment problems, that's
indicative of a plant that's having problems. And so,
there may be a situation where you would have a PI,
even though from a risk-based PI perspective you
wouldn't have that PI, but because of what we are
trying to do with performance indicators, and
providing an indication of the old raw performance of
the plant, you might keep that performance indicator.
So, that's the kind of consideration that we'll go
through in the change process in deciding to what
extent we replace, or add, or whatever, check the PIs.
CHAIRMAN APOSTOLAKIS: But, this is where
we would like to see some more discussion of these
things, and limiting the number of PIs I think and
I think we already mentioned some very valid points.
So, out of curiosity, the number of
unplanned changes in power, significance changes in
power, what kind of an indication is that? I mean, if
it's not risk related, what is it then? Is it
sloppiness?
MR. JOHNSON: Well, it sort of gives an
indication, I see Tom from NEI, Tom Houghton from NEI
raising his hand, I guess you've got to get near a
mic, Tom, to speak, we believe it gives an indication
of, yeah, things that are not steady state at a plant.
If a plant is having situations that require it to
undergo a number of transients, again, setting aside
those things that are not induced by the performance
of the plant, that are not being generated from some
outside influence, but if a plant cannot maintain
stable operations because they are continuously having
to respond to things that are unplanned, that's
indicative of a plant that's beginning to have some
difficulty and, perhaps, warrants some follow-up.
CHAIRMAN APOSTOLAKIS: So, it really has to
do with the culture.
MR. HAMZEHEE: Well, it's such an indirect
indicator, you just don't know what it's coming from.
You can't conclude that that kind of condition may be
indicative of some problem, it may be culture, it may
be whatever, but it's actually, you know, an indirect
indication.
MR. HOUGHTON: Tom Houghton, NEI. We've
found that it is an indicator. It is more predictive
of future problems, and it did have a good
relationship with plants which were on the watch list,
okay, when the historical data was looked at. Okay.
So, it has face validity, I'd say, and it is somewhat
predictive, in that if the operations or maintenance
are not able to maintain the plant at the power level
that was intended in the management plan for operating
the plant, that there is a necessity of looking into
the problem further. Now, some of the cases have
involved bio fouling in condensers that weren't being
looked at as closely as before, or feedwater control
problems that weren't being looked at as clearly as
before, and partly because they weren't part of the
design basis or the tech specs of the plant, and so
this has led the plants to make a closer look at how
they are operating and maintaining beyond what's
absolutely required.
So, we see some value in that. Now, there
have been questions about why 20 percent, why 72
hours, et cetera, et cetera, and there are efforts in
piloting revisions to this indicator which NRR is
proposing to pilot and industry is looking at a
similar pilot to try and avoid some of these questions
that have arisen as to what was intent, because you
want to try and take what was the intent out of it.
But, we've found that it's been valuable in the
process.
CHAIRMAN APOSTOLAKIS: I guess what you are
saying is
MR. HAMZEHEE: One thing I would like to
note, however, on that example, dependency on watch
list, I looked through the data, too, and often times
a plant has a lot of power changes after it gets into
the watch list, which means the operators are
sensitive to regulatory observations that suddenly,
truly, I mean, it is so transparent, you know. So,
that's why I'm saying it's such an indirect indicator
that, you know, it's very hard to fathom what is
causing it, and if it is clearly there are
implications, because if you have a lot of power
changes they may initiate a transient of some type.
CHAIRMAN APOSTOLAKIS: So, what I gather
from this is that we just found a performance
indicator for the crosscutting issues. Well, that's
what you told us. So, if the maintenance department
doesn't do a good job
MR. JOHNSON: George, we actually think
that the full spectrum of performance indicators and
the inspectible area results provide a good indication
of crosscutting issues, in that
CHAIRMAN APOSTOLAKIS: I didn't say it's
the soul indicator, but it's an indicator.
MR. JOHNSON: in that problems
CHAIRMAN APOSTOLAKIS: Why are we so afraid
of this safety-conscious work environment, every time
I mention it I get no, ten nos. Why? Is there
something magical about it?
MR. MAYS: I've never given you a no on
that, George.
CHAIRMAN APOSTOLAKIS: Otherwise it would
have been 100 nos.
MR. MAYS: Right, I think you are seeing a
consistent situation here, George, and that is we
don't have anything that goes up and says this is our
direct indicator of safety-conscious work environment,
because we don't know how to measure it that way.
CHAIRMAN APOSTOLAKIS: I didn't say it was
direct.
MR. MAYS: But, what we have is
CHAIRMAN APOSTOLAKIS: I didn't say it was
the only one.
MR. MAYS: what we have is, multiple
ones, that's why we took the sample approach, and
that's why we are seeing that there are some cases
where we have things that help us in that area, and I
think that's appropriate. I don't think we should be
afraid to say, my personal opinion is, I don't think
we should be afraid to say that we have a sample of
things, and we have some that are more direct than
others, and giving us indication when that particular
area is having difficulty. I think we have those.
DOCTOR KRESS: I think you do have, and I
think the question of that bears on the question
of, is there an optimum number of PIs, and normally
when you ask that question, is there an optimum number
of PIs, when you relate to other statistical treatment
of things you are talking about a sample and how many
samples do I need to have the confidence level that
I'm measuring what I think I'm measuring.
And, in your case, I don't think you have
the capability of determining that optimum, and when
you can't determine an optimum in a statistical
manipulation or looking at the data, I think you have
to just fall back on take as many as you can. I hate
to say this, because the industry, you know, I can see
them now, but if you can't determine an optimum from
statistical analysis of the thing, then it seems to me
like the only other option you have. I'd be
interested in hearing your reaction to that.
MR. MAYS: Well, I think, I'm not sure your
taking as many as you can is necessarily the answer.
I think the problem is, you are trying to reach a
question of figuring out when you reach the point of
diminishing returns, and sometimes you can do that
because you have data and information on a model to do
that very precisely, and sometimes you have to do that
from a more judgmental approach.
DOCTOR KRESS: I count that in the phrase
as many as you can, I mean, that's part of the as you
can part.
MR. MAYS: I think the bottom line at the
end of the day is, do we have confidence that we have
a process by which we can detect when plant
performance is degrading from a safety standpoint, so
that the NRC can take appropriate action to intervene.
DOCTOR KRESS: How can you validate that?
MR. MAYS: Now, the question for that one
is, I don't know that you can do the kind of
statistical validation of that that might be desirable
to do if you could, but we have a philosophy that says
we want to try to have objective, measurable, risk-
informed information to do that, and I think, again,
this is part of that progress versus perfection
discussion, we will have more of it here, and then it
has to become a judgment as to whether or not we are
achieving much benefit when we do that. That's part
of what the ROP process has as their joyful task to
figure out, as part of the change process.
The next thing we wanted to show you was
the results of some of the work from the mitigating
systems. We had proposed in the RBPI report that we
could come up with 13 mitigating system component
class RBPIs for BWRs and 18 for PWRs. These were
using the SPAR models again for setting the baselines
and the thresholds. We used the system reliability
and component reliability studies that we've produced
in Research and formerly in AEOD as baseline
information to go into those SPAR models. We used the
EPIX data for calculating reliability parameters, and
we used the current information that's coming in
through the Reactor Oversight Process for putting the
unavailability data into these models.
So, the point here is, this EPIX data for
the reliability is the only part of this that is data
that isn't already reported to the NRC in some quality
fashion that we already know about, so this is the one
where we have the implementation issue.
And, we used a similar process that we did
for the initiating event indicators, for figuring out
what the time frame and the right prior was to do
that. And, when we did that, because we had and
when you get to reliability, if you look at
reliabilities of pumps, and diesels, and other things
which have generally mean reliabilities in the
vicinity of E-2 or potentially lower, we found that
even with a three-year type of time period we still
had situations where we would have false positive
rates that could potentially exceed the 20 percent
that we had set up as an initial basis. And, what we
decided to do with that, and you'll see it in the
tables as we flip back in a minute, is that whenever
we had a reliability indicator that crossed the
green/white threshold, we would also add an additional
piece of information which is, the probability that
the baseline value was still below the threshold. So,
basically, recognizing that the probability was a
distribution, sometimes the delta between the baseline
value and the green/white threshold was fairly small,
it would be easy for that distribution to cross the
threshold and we wanted to make sure we gave a little
more information to say, well, is it like really
across the threshold or is it just barely across, so
we gave a little more information because we couldn't
always meet the 20 percent false positive threshold
that we used.
DOCTOR KRESS: And, will that information
be used somehow in the overall plant assessment
somewhere?
MR. MAYS: Well, we thought that that was
appropriate to use because
DOCTOR KRESS: It's good information, I
guess.
MR. MAYS: we wanted to have some idea
of how sure we were that somebody had gone over that
threshold.
DOCTOR KRESS: I think some guidance needs
to be developed on how we use that.
MR. MAYS: I think that would have to be
done on how to use it, but we thought it might be
something that we talk to people about.
DOCTOR KRESS: I definitely think it's
useful additional information.
MR. MAYS: I'm going to skip on over to
page 18 now and show you what we had for the
CHAIRMAN APOSTOLAKIS: Oh, in general,
reading from the original report, I get the impression
that based on the numbers you got you can actually
have threshold values for classes of components or
classes of plants, that you don't necessarily have to
have a separate threshold value for each component at
each plant.
MR. MAYS: Well, we are looking we are
going to look at that.
CHAIRMAN APOSTOLAKIS: Is that correct?
MR. MAYS: Right now, we have only 23
plant-specific models for which we've done this, and
one of the things we said we would go back and look at
was, was the differences among the plants or within
groups so much that you had to do a plant-specific
value or whether it makes sense to make a group value.
We haven't got all that information to be able to do
that yet, but that's one of the things that was
proposed as a way to deal with the number of number
and types of PIs.
CHAIRMAN APOSTOLAKIS: Which is very
annoying here, I guess, but not the Maintenance Rule,
which is another mystery to me. Why in the Maintenance
Rule the licensee was asked to submit plant-specific
thresholds, and everybody thought it was great, but
when it came to the Revised Oversight Process it's
something that is like, you know, we don't want to
hear about.
MR. MAYS: One of the comments we've
received from industry was a concern that if we have
risk-based performance indicators set up this way that
there might be a potential conflict between thresholds
here and values set for the Maintenance Rule.
CHAIRMAN APOSTOLAKIS: And, that's
something we cannot resolve?
MR. MAYS: No, we could potentially resolve
that.
CHAIRMAN APOSTOLAKIS: Yeah.
MR. MAYS: The issue has to do, almost from
a technical standpoint, to do with the fact that in
this case we are doing a single-point variate
analysis, we take one thing, we hold everything else
constant, and we see what the impacts are.
Aside from the fact that they may be using
a slightly different risk model at the plant than we
were using, that's one of the bigger issues. One of
the things the plants were able to do, because they
were having a more integrated look at this, was to
say, for example, okay, suppose I desire to be able to
have a greater unavailability of my diesel generators,
because I have a financial or other reason for
conducting some on-line maintenance, well, what I will
do is, I will trade that off by making sure I have a
stricter standard for my reliability, so that in net
the risk hasn't changed significantly.
Well, if you have a single variable
analysis like we have here, you can't make that
tradeoff, because you are holding all the other things
constant, and what we will see when we get down a
little further in here, we talked about ways of
reducing this, one of the things we are proposing is
a more integrated way of looking at them, which can
allow for that kind of stuff to go on.
So, the Maintenance Rule was for the
licensees to set their own standards and for us to
monitor that they were doing those. So, I think
that's probably the answer why they didn't have a
problem at that level, because they were setting it on
their own standards, using their own risk information,
and being able to trade off back and forth where they
felt appropriate.
Anyway, coming to this example here, I
didn't want to go through all of the ones on each
case, I wanted to point out a couple things. One of
the things that you can see when you look at these
examples is that the case of the 95th percentile,
let's go down to emergency AC power unreliability,
what you'll end up finding there is a case, if you
take the 95th percentile, you get a value that's
almost up to the red value, if you were to take that
as your green/white threshold.
CHAIRMAN APOSTOLAKIS: Is that right? It's
close to where are you looking?
MR. MAYS: I'm looking at emergency AC
power unreliability, which is right here, this line.
CHAIRMAN APOSTOLAKIS: Oh, yeah, right,
because for the others that's not true, right?
MR. MAYS: Right, for the others it's not
necessarily true, so that was one of the reasons, an
example of a reason why we thought that the 95th
percentile approach may not be the most appropriate
way to do with these. So, that was an example.
CHAIRMAN APOSTOLAKIS: Well, I would say it
is not.
MR. MAYS: Another thing that we found when
we looked at some of this stuff is that sometimes,
because of the risk importance of a particular
component, even if its reliability or its availability
goes to one, it never producers the delta CDF
necessary to get to E-5 or E-4 to get you to the
yellow or red zones. So, that raises a question, is,
well, maybe we don't want to use that as an indicator,
or maybe we want to do something different.
We haven't come to a complete conclusion
on that, and sometimes what you'll see is, we'll find
that you can, in fact, reach those thresholds, but
only if you exceed te tech spec allowed outage times
for your equipment. So, the question is, do we want
to have an indicator that has a threshold that they
can only get to if they are violating the license.
I'm not sure that necessarily makes
CHAIRMAN APOSTOLAKIS: So, not reached
means not reachable.
MR. MAYS: Not reached has two things in
these tables. One of them has a footnote, I think,
which we eliminated the text on that, but the
footnote in the report, we have a not reached and we
have a not reached with a footnote, and we distinguish
between the ones that can't be reached because the
risk importance of the component isn't significant
enough, and those which it could be reached but it
would only be reached if you violated your tech specs,
in terms of operation. So, it's not really clear to
us which one makes the most sense here, we are just
laying out what the feasibility is of using an
indicator in that particular area, and what we were
trying to do is demonstrate that it's possible to set
thresholds for these particular values on a plant-
specific basis.
CHAIRMAN APOSTOLAKIS: So, what's the
difference between this and what the current process
has? Are you increasing the number of indicators?
MR. MAYS: Well, first off, we have
specific reliability indicators.
CHAIRMAN APOSTOLAKIS: That's correct.
MR. MAYS: We have availability indicators
and the reliability indicators have plant-specific
baselines and performance thresholds for them, and we
have, in another issue that we have, and we have a
broader coverage so we have more of them, and the
other thing we have, if you look at the bottom of that
page, we have three component class indicators for
air-operated valves, motor-operated valves and motor-
driven pumps, which go across systems. And, what we
have in that case is, we have a baseline value that we
get for the plant, and then what we have done is, we
said if we increased that value by a certain factor,
so, for example, the green/white threshold for AOVs
would be at 2.2 times increase in the baseline value
would get you to the green/white threshold, and what
we did there was, we took all the AOVs in the risk
assessment, said if we double them that's what gets us
to E-6. If we go up by a factor of 13, that's what
gets us to E-5.
CHAIRMAN APOSTOLAKIS: What's the point of
that? I mean, doesn't it go against what Hossein said
earlier, that not all AOVs have the same risk
significance?
MR. MAYS: Potentially, but what we are
trying to do here is say, if we had a broad
programmatic problem, if our AOV maintenance problem
was a problem, or general maintenance was a problem,
or we had a problem with our design and implementation
of motor-operated valves, if they were all to go have
a degradation, how much degradation would all of them
have to be going under in order to reach this
particular threshold.
CHAIRMAN APOSTOLAKIS: So, that would be a
useful incite to the Option 2, no?
MR. MAYS: I don't know enough about that
to be able to
CHAIRMAN APOSTOLAKIS: That's a good
answer.
MR. MAYS: to say.
CHAIRMAN APOSTOLAKIS: Very few of us know
enough.
MR. MAYS: Okay.
Moving on to the next thing that we were
asked to look at by the user need letter, had to do
with containment performance, because there was a
limited number of things that we had in the ROP to
deal with containment. Unfortunately, we were able to
identify things that could potentially be used as
risk-based performance indicators for containment,
mainly the performance of the drywall sprays in the
Mark I BWRs, and the performance of large containment
isolation valves in the others.
DOCTOR KRESS: This information came out of
older PRAs?
MR. MAYS: Right, these were the things
where it says what performance could have an
DOCTOR KRESS: Yeah, you don't deal with
those in SPAR.
MR. MAYS: well, not quite, when you
say SPAR, SPAR is a broad program, there is the level
1 SPAR models, there are LERF level 2 models, there
are shutdown models, and there are potential external
event stuff. So, SPAR represents that whole section.
DOCTOR KRESS: So, you are using the level
1 SPARs for this study.
MR. MAYS: We are using level 1 for the
initiating events and the mitigating systems, for the
containment we were looking to use LERF models.
DOCTOR KRESS: I see.
MR. MAYS: And, we are going to use LERF as
our metric, for containment related issues.
DOCTOR KRESS: And, that's another one of
my questions, but I'm sure you are going to discuss it
anyway.
MR. MAYS: So, we were planning on doing
that.
CHAIRMAN APOSTOLAKIS: Before we leave the
mitigating systems, there was a sentence in the
report, Appendix A, page A-25, "The same component
rate importance criteria were used to select class
indicators. However, the system level versus the
importance values were determined using the multi-
variable group function available in SAPHIRE." What
is this multi-variable group function available?
MR. MAYS: I think that's just a fancy way
of saying we changed all of the components to have the
same degradation at the same time, and in random model
again. Would that be correct, Steve?
CHAIRMAN APOSTOLAKIS: You have to come to
the microphone, please, and speak with sufficient
clarity and volume.
MR. MAYS: Fortunately, George, you and I
never have that problem.
MR. EIDE: Steve Eide from the INEEL, and
I believe Steve is correct. I don't know the
specifics of that actual
CHAIRMAN APOSTOLAKIS: Which Steve, this
Steve is correct?
MR. EIDE: Steve Mays, I don't the
specifics of that actual
CHAIRMAN APOSTOLAKIS: But, it sounds
better, right?
MR. EIDE: module in SAPHIRE.
CHAIRMAN APOSTOLAKIS: Multi-variable
function, this is really impressive.
MR. MAYS: We have the capability in the
SAPHIRE code to go over and change multiple components
at one time with a change set, and that's what we
basically did.
CHAIRMAN APOSTOLAKIS: Okay.
MR. MAYS: Moving to containment, the
problem we had, we were unable to develop containment
performance indicators, because we don't have the
models and the data currently available to be able to
do that on a broad enough either on a plant-
specific basis for sure, or on all the different
classes and types, so we were limited there by our
capability right now to be able to produce performance
indicators for containment.
DOCTOR KRESS: You do what you can, is that
it?
MR. MAYS: That's correct.
DOCTOR KRESS: But, I would like to you
are not necessarily going to limit to LERF when you
get around to doing it. You mentioned that
MR. MAYS: Our original intention was to
use LERF as the metric for the containment
performance, because that would be consistent with
what we have in Reg Guide 1174 and other applications.
It's potential that we could go to something different
from LERF if somebody thought that that was useful and
worthwhile, but right now that's what we were looking
at on the basis of what we have available.
DOCTOR KRESS: I know a few people who
think that it would be worthwhile to include LERF
is all right, but consistency, you know, is the
hobgoblin or something or other.
MR. MAYS: Foolish consistency is the
hobgoblin of small minds.
DOCTOR KRESS: But, I think one ought to be
concerned in the regulatory arena with late
containment failures and also
MR. HAMZEHEE: In Phase 2 we are going to
look into this, to see if large late failures are also
risk significant.
DOCTOR KRESS: and I think you could
probably deal then with just the conditional
containment failure problem then.
MR. MAYS: Right. The issue then again
comes to, do we have a set of models that reasonably
reflect some understanding of the risk that we can put
data through and do, and right now we are just not
there.
DOCTOR KRESS: Yeah, well, you know, my
I'm urging you not to think of risk just as prompt
fatalities, that's my point.
DOCTOR BONACA: Just one comment I have,
and probably I'm wrong, but because in many cases you
cannot really identify a meaningful RBPI, you simply
don't do that, and then you take the opportunity, you
know, to develop what you can get, but you want to
make sure that what you can get is meaningful, too,
right? I mean, what I'm trying to say is that, I'm
left with the impression that, you know, because of
that you are going to get a set of indicators that may
not be so significant after all, but the reason why
you got to those is because that's all you could get.
DOCTOR KRESS: I think one of their
criteria was, they have to be risk significant.
DOCTOR BONACA: They have to be, okay, but
I'm trying to understand the time, you know, how many
facets of this thing you are going to see, just maybe
two or three, and, you know, does that give you the
picture you want, or is it just all you can get. And,
I'm not sure they are the same thing.
CHAIRMAN APOSTOLAKIS: There is some of
that, this is a significant step forward, though.
DOCTOR BONACA: Oh, no, I'm not
CHAIRMAN APOSTOLAKIS: This could be never
sought perfection.
DOCTOR BONACA: I understand that.
DOCTOR KRESS: Progress is what we
CHAIRMAN APOSTOLAKIS: Progress, we work
with deltas.
MR. LEITCH: Wouldn't performance on
integrated leak-rate tests be a significant PI in
this?
MR. MAYS: I think that's been looked at
before. You could go back and look at performance
under leak-rate tests. The problem we've had in
looking at performance under leak-rate tests is, you
might be able to see that leak-rate test performance
has changed, but the question is, what's the risk
significance of that information? And, when you look
at the risk assessments and things that have been
done, the leak tightness of the containment in the
kinds of things that those things measure is rarely,
if I'm aware of, ever the dominant contributors to the
off-site releases.
DOCTOR KRESS: It's never even risk
significant.
MR. MAYS: It's not even close.
DOCTOR KRESS: But, that's when your risk
measure is prompted out.
MR. MAYS: That's correct.
DOCTOR KRESS: So, that's why I'm saying,
don't just focus on prompt fatalities.
MR. MAYS: Right.
DOCTOR KRESS: You might want that as one
thing.
MR. MAYS: But, even in the cases of when
you look at latent cancer deaths and risk significance
DOCTOR KRESS: It's not significant there.
MR. MAYS: it's not significant there
either.
DOCTOR KRESS: It's a risk of possible land
contamination, perhaps.
MR. MAYS: Maybe, but I'm saying, comparing
to the other things that would do land contamination
DOCTOR KRESS: That particular one is a low
risk.
MR. MAYS: it's pretty small, too.
DOCTOR KRESS: But, late containment
failure now is a different issue. It can be risk
significant from the standpoint of cancers and land
contamination. So, you know, but you are right on the
leak rate, unless it really gets bad.
MR. MAYS: The way it really gets bad is
somebody leave some major valve open, and that's what
I'm saying we would have
DOCTOR KRESS: You would capture that
anyway.
MR. MAYS: Right.
MR. HAMZEHEE: And, that was one of the PIs
on the Reactor Oversight Process, but they also
eliminated that from the list.
MR. LEITCH: Okay, thanks, I understand.
CHAIRMAN APOSTOLAKIS: I wonder whether it
would make sense to take the set of the performance
indicators we have, or we will have, and go to a real
accident or incident, and see whether, like Three Mile
Island, whether you would see any change in these
things before the incident occurred.
MR. MAYS: I'm going to show you something
that directly relates to that in a little bit.
CHAIRMAN APOSTOLAKIS: Good.
What did you say, Tom?
DOCTOR KRESS: We probably don't have the
data for Three Mile Island.
CHAIRMAN APOSTOLAKIS: Well, for something,
something, I mean, to validate that this process would
make sense.
MR. MAYS: Your point is, if there is
something that is dominant contributors to the risk,
are we having measures in our PIs that relate to
those, and I've got a particular slide that shows
that.
CHAIRMAN APOSTOLAKIS: Okay, so I'll wait
until we come to that then. Okay.
Is this a good place to take another short
break?
MR. MAYS: Sure.
CHAIRMAN APOSTOLAKIS: Okay. Then, we're
taking a seven-minute break.
(Whereupon, at 10:53 a.m., a recess until
11:03 a.m.)
CHAIRMAN APOSTOLAKIS: Okay.
MR. LEITCH: Can I ask just one question
for understand here before we get started again, or as
we get started again?
CHAIRMAN APOSTOLAKIS: Please, quiet.
MR. LEITCH: I'm looking at Appendix A, and
there's a number of pie charts
CHAIRMAN APOSTOLAKIS: Page?
MR. LEITCH: I guess it's actually
Appendix D, page 56, where the pie charts begin.
CHAIRMAN APOSTOLAKIS: Okay.
MR. LEITCH: And, I want to be sure I'm
correctly interpreting this information, just to pick
page 56 as an example, I think that's the first one,
it says areas not covered, 3 percent, indicators 2
percent, industry-wide trending 95 percent. Does that
mean, am I correctly interpreting that that 95 percent
of the issues are so infrequent that they are not
amenable to individual plant performance indicators,
that they have to be trended on an industry basis?
MR. MAYS: That's close.
MR. LEITCH: Okay.
MR. MAYS: When you look at the look at
what was in the IP database for the core damage
frequency associated with initiators for this
particular plant, what you find is that 95 percent of
the sequences involved an initiators other than the
ones we have direct indicators for, or the ones in
areas not covered. So, this might be a plant, for
example, that had really high contribution from loss
of off-site power events, or station blackout events,
since we don't have an indicator on a plant-specific
basis for that kind of thing, that would have to be
covered through the industry-wide trending. That's
how you would be able to see whether or not you
thought you had a problem, plus the plant-specific
inspections and the baseline inspections would go and
look at the areas that are not covered by indicators
to see if the performance that would impact were
changing. So, this is just to kind of give you you
are right, you are getting kind of the feel for which
portions of the initiating event indicators of the
risk are covered by the indicators, and which portion
would have to be either done by inspection and/or
trending.
DOCTOR KRESS: But, what is this a
percentage of?
MR. MAYS: Percentage of total CDF.
DOCTOR KRESS: Percentage of total CDF.
MR. MAYS: Right.
MR. LEITCH: Okay, thanks.
MR. MAYS: Okay.
Moving on to shutdown, this was an
important area because we didn't currently have in the
ROP any shutdown direct indicators. We looked at that
and we found that we couldn't do initiating event
indicators for shutdown because they just don't happen
frequently enough, but we did come up with some fairly
interesting things to do with respect to mitigation.
And, this has to do with several things.
We formulated a process by which we would
take into account the RCS conditions, vented, not
vented, open, not open, time after shutdown for decay
heat purposes, the availability of mitigating system
trains in those particular scenarios, and then we were
able to go back and try to set thresholds and
performances.
This one is a little different in picture
than what we had before, where we are actually going
out and calculating reliabilities and calculating
availabilities, now what we are doing is we are taking
a slightly different approach that says, if I have a
model that represents how a BWR or PWR responds can I
get groups of things, where if I spend time in those
scenarios I know those contribute more or less to
risk. So, we came up with for both the BWRs and
PWRs, were able to come up with thresholds. We put
together, started off actually with three categories,
low, medium and high, corresponding to an amount of
increase in CDF per day associated with being in those
conditions.
CHAIRMAN APOSTOLAKIS: You know the
question you are going to get from some of my
colleagues in May, how can you do this if you don't
have a good shutdown PRA?
MR. MAYS: This goes back to the first
point I made earlier, in order to do the risk-based
performance indicators I have to have a reasonable
model of how a plant responds.
CHAIRMAN APOSTOLAKIS: Do you think you
have it now?
MR. MAYS: I think I have it for these two,
because what I have in these two cases is a plant-
specific model from a representative PWR and BWR, that
they happened to use for doing their shutdown for
shutdown risk models. So, I think I have something
that's reasonable here that I can use. I don't have
something for every plant. I don't have the SPAR
models developed for every plant, or even for the
groups of plants yet, but I have this information
that's a starting point for progress, not perfection.
So, that was the basis for doing this.
So, when we looked at that, we said, all
right, let's go back to the baseline and say, how much
time do these people typically spend in various
configurations, because it's just necessary to go
through some of them in order to complete a shutdown.
And so, we would measure performance as being
deviations from the nominal performance that you have
to do, just to go and conduct a shutdown operation,
and if you spend more time in particular
configurations of higher, lower, or medium risk
significance, then that would be the basis for us
deciding what the thresholds would be.
And, when we came to that, we also
recognized that for PWRs there's a special category of
the early reduced inventory situations that they have
to go into in order to be able to do that, which
represents a higher risk significance than most of the
other configurations they go to, and because when they
are in that particular mode they are under the
shutdown guidance of NEI guidance on what was the
number of that, Tom, I can't remember?
MR. HAMZEHEE: It's 91-06.
MR. MAYS: 91-06? 91-06, which says, when
you are going into early reduced inventory modes you
have to take certain compensatory measures with
respect to availability of power, availability of
injection systems, so we said if you are complying
with that in your early inventory, and you don't spend
more than the nominal time, then we will do that. If
you spend more than nominal time in that one, then we
treat that as if it's a high. So, we treated that
category a little differently.
DOCTOR KRESS: That's a really different
concept than what you did for the others.
MR. MAYS: That's correct, because this is
all we were able to do with the information we had.
DOCTOR KRESS: And, it goes back to my
concern about whether the baseline, which in this case
is called nominal, is sufficiently good enough, and
whether or not you are penalizing some plants you
know, if I were a plant that took long times at high-
risk significant configurations earlier, then I would
be able to continue doing that on this and not get a
white reading, because you are basing it on that as a
starting point. This part worried me more than any of
it.
MR. MAYS: I understand your point. If we
had plant-specific history and plant-specific
DOCTOR KRESS: I understand.
MR. MAYS: values, that would be more
of a concern.
DOCTOR KRESS: Yeah.
MR. MAYS: I think what we are trying to do
here is say, what's typically representative of kinds
of times that industry generally spends in these
areas, so these baselines here were based on some
information
DOCTOR KRESS: They are basically industry
average lines.
MR. MAYS: industry information, so if
somebody were to go and then start spending more time
in risk-significant configurations, they would not be
benefitted by having done that, the particular
arrangement you are talking about.
So, again, we are talking the progress,
not perfection, situation here. We have nothing now,
and we are trying to do something that's a little
better and a little more risk informative.
CHAIRMAN APOSTOLAKIS: So, the categories
low, medium and so on, are determined by the condition
of core damage probability?
MR. MAYS: Right.
CHAIRMAN APOSTOLAKIS: And, what are the
values?
MR. MAYS: E-4, -5, or -6 per day, I
believe, that equate to core damage frequency for the
year of E-4, -5, or 06, if they were to spend their
time in that condition for a full day.
CHAIRMAN APOSTOLAKIS: Full day?
MR. MAYS: For a day. In other words, for
example, the high configuration would say, if you
stayed in that configuration for a day you would add
E-6, or E-4, to your core damage frequency associated
for that plant for that year.
CHAIRMAN APOSTOLAKIS: And, you would take
the day divided by 365 again, or that doesn't enter
into this?
MR. MAYS: We are saying that if you are in
a high configuration
CHAIRMAN APOSTOLAKIS: Yeah.
MR. MAYS: you are accumulating a
yearly increase of E-4 per day. That's the rate of
accumulation of the core damage frequency.
CHAIRMAN APOSTOLAKIS: Yes, I don't
understand that, but that's okay.
DOCTOR KRESS: It's like averaging it out
over the year.
MR. MAYS: That's right.
CHAIRMAN APOSTOLAKIS: Does the fraction of
one day over 365 enter anywhere?
MR. MAYS: Sure. What happens is, we base
all of our data gathering and our analysis on how many
days or hours you spend, and then the rate for the
high category is based on that translating to the
year. So, we do our calculations on the days, and the
rate for the threshold is based on the year.
CHAIRMAN APOSTOLAKIS: Right, and that's
wherein I offered the comment earlier I read, that no
matter how long you are there, if you divide by 365
you are effectively reducing its significance.
MR. MAYS: But, we weren't doing that.
That's what they didn't understand.
CHAIRMAN APOSTOLAKIS: Okay.
MR. BOYCE: It sounds like I don't have an
action item anymore.
CHAIRMAN APOSTOLAKIS: What?
MR. BOYCE: It sounds like it's understood
and we don't have an action item over on our side
anymore.
MR. MAYS: I wouldn't be that bold.
The next thing I wanted to show was, let's
put up the PWR chart, we can go through this one a
little quicker because we have others to do. So,
basically, what we were talking about here is, you
would start on the left-hand side and you would,
basically, start at the top of the chart and move
yourself down and you would see for various different
configurations, like whether you are pressurized,
whether you are in mode 4, whether your reactor core
system was in tact, how many days after shutdown we
were, those are the going in conditions, for which we
then went and evaluated configurations and
combinations of configurations that previous PRAs on
shutdown have said to be important.
So, we would go back and, for example,
let's take an example for the one diesel generator, if
one diesel generator is out of service when you are in
pressurized mode for hot shutdown with the RCS in
tact, that constitutes the low category. So, what we
would do is, we'd gather up the amount of time you
spent in that low category, compare that to the
thresholds.
CHAIRMAN APOSTOLAKIS: Can you point to us
where you are?
MR. MAYS: Okay. I am on this row right
here, pressurized cooldown, Mode 4, RCS in tact, and
I'm looking at the impact of having one diesel
generator out of service.
So, we went back and looked at several
other configurations that were found to be important
to shutdown risk, relating to power availability, RHR
availability, secondary cooling trains, the
availability of the RWST, other things of that nature,
and we laid them out and if we have no entry in the
block then that means that particular condition does
not present a significant increase in the rate, so
what we do is, the nice thing about this, although it
looks busy, is that before the outage even starts,
when you've done your outage plan, you can go into
this table and see what equipment you are having out
when under what states, and before you even start have
an idea about where you could be accumulating more
risk than others. So, this is a nice tool for both
utility and the inspectors to have before you even go
in.
CHAIRMAN APOSTOLAKIS: Isn't this similar
to what they are already doing with the various tables
they have risk configurations to avoid?
MR. MAYS: Correct.
CHAIRMAN APOSTOLAKIS: But, this is more
detailed, perhaps.
MR. MAYS: I'm not sure if it's more or
less detailed, but it has a similar concept. What we
are doing is saying, for a particular outage, we would
measure the time you spent in low conditions, the time
you spent in medium conditions, the time you spent in
high conditions, and we would compare those to the
thresholds, and that would tell us whether you were
spending excess time on those conditions, and then we
would know exactly what conditions we were in, we'd
know what to be able to go look for, so the idea here
is, is that you are going to be able to know in
advance what conditions to avoid. You are going to
know in advance what conditions you are planning to go
into, and the inspector, during an outage, if
something changes from the inspection plan, can go
right back to a table like this and say, all right,
now they are changing from this scenario to that
scenario, is that one I have to pay attention to and
worry about. And then, as we gather up the data as
you go through the outage, we can say at the end
whether your performance was basically nominal or
whether you accumulated enough risk in your off-
nominal conditions to warrant attention from the NRC.
That's the philosophy.
DOCTOR UHRIG: Would you get the same
information from one of these automated PRA
computerized systems?
MR. HAMZEHEE: Not exactly.
MR. MAYS: Potentially, yes, I mean, but
I'm not sure exactly how they are gathering and
putting the information in, and how they are profiling
out the outage, but conceptually it's a similar
scenario. We are saying, what's the risk associated
with being in particular scenarios as we go.
DOCTOR UHRIG: Yes.
MR. MAYS: And so, it has a similar
foundation as the shutdown risk monitors in principle,
and so we think this is something that's fairly easy.
It's fairly easy to tell how much time you spent in
each of these configurations, because you planned it
all out before you start, and you monitor what you did
when you went through it, so if we were to get
information on how much time they spent in each of
these categories it would be fairly easy to do a PI.
Now, going back to the three things we
talked about earlier, having a model, having
performance data, we think we have, you know,
reasonable general stuff, and it is generic based more
than plant specific in this case. But, right now we
don't have data reported to us on the amount of time
spent in these things, so we would either have to have
somebody go out and get it ourselves, or we'd have to
have the industry produce it for us, in order to be
able to have a PI with respect to shutdown.
MR. HAMZEHEE: And, I think currently the
risk assessment that they do during the shutdown is,
they input all the equipment availability and how much
time they spent, and then they get a risk profile on
a daily basis, but they don't question as to how long
they should or should not stay in certain
configurations.
MR. MAYS: I'm not sure whether they do
that or not.
MR. HAMZEHEE: But, they can do it if they
want to, they can go change some parameters and get
the results in the shutdown risk models.
DOCTOR KRESS: I'm particularly interested
in whether this will be part of the database you are
going to give back, because it's my view that you have
to have this information if you are going to do if
you are going to include shutdown risk within, say,
the 1.174 risk matrix, this doesn't do it by the way,
this information has to be fed back into some sort of
shutdown risk PRA in order to actually get the
contribution of shutdown risk to the total risk, and
also to determine the things like importance measures,
because this doesn't get reflected in importance
measures at all.
MR. MAYS: Indirectly it does. The
importance of the particular components, which is also
dependent on the particular condition that you are in,
is explicitly in the table.
DOCTOR KRESS: Oh, yeah, I'm sorry, it
doesn't show up in the importance measures you did for
the at power.
MR. MAYS: That's correct.
DOCTOR KRESS: Yes. I mean, you've got
some importance measures for components. This doesn't
reflect there in that.
MR. MAYS: Right.
DOCTOR KRESS: But, you do you are going
to have this kind of information for, you know, the
fleet of plants and for individual plants, if you are
really going to do a proper shutdown risk assessment.
So, I hope somebody starts developing a database on
this.
MR. MAYS: Well, that's what we've
proposed, that if we have that kind of data we can at
least put together some information that would give us
indication of when something might be changing
significantly.
DOCTOR KRESS: Uh-huh.
MR. MAYS: And, I think it's a good start.
DOCTOR KRESS: Yes.
MR. MAYS: Going on to the next thing,
which was fire events, this won't take very long at
all. Basically, the issue was from an initiating
event standpoint they don't happen often enough for us
to do plant-specific analysis of them. From the
mitigating system standpoint, we've identified what
systems would be important, which was the reliability
and availability of the fire suppression systems would
be the indicator we would try to put together, but we
really don't have the data to do that, to quantify
baseline and performance values, so
CHAIRMAN APOSTOLAKIS: Can we discuss a
little bit this issue of timely detection?
MR. MAYS: Right.
CHAIRMAN APOSTOLAKIS: And, I think it's
related to whether an indicator is leading or lagging,
is that correct?
MR. MAYS: No. Whether an indicator is
leading or lagging is what you are measuring and
comparing to. For example, all indicators, by
definition, which you gather from data are lagging the
data that you are getting, but they may be leading of
some higher order effect.
CHAIRMAN APOSTOLAKIS: Core damage
frequency.
MR. MAYS: Correct.
CHAIRMAN APOSTOLAKIS: Core damage, yeah.
MR. MAYS: So, the issue here is, does the
occurrence rate of information for this particular
thing happen so infrequently, if I have, for example,
losses of off-site power which only happens in the
ball park of once every 30 years or so at a plant,
then I'm not going to accumulate data in a sufficient
period of time to be used effectively in the Reactor
Oversight Process, to go year by year and say to
myself, where does this plant need more or less
attention. So, I can't make that kind of an
assessment directly with an indicator for things that
have a low frequency of occurrence.
DOCTOR KRESS: Is loss of off-site power
under the control of the licensees?
MR. MAYS: Some of it is and some of it
isn't.
DOCTOR KRESS: It could be a lack of, say,
a performance issue?
MR. MAYS: Right.
DOCTOR KRESS: Okay.
CHAIRMAN APOSTOLAKIS: But, is timely also
referring to the fact that if something happens it's
too late?
MR. MAYS: No, timely refers to the fact
that if there is changes in the performance, my sample
period is such that I can reflect and see that on an
ongoing basis and take action to deal with it on the
basis of that information.
So, if I have something like a LOCA
steam generator tube rupture frequency, or a LOCA
frequency on a plant-specific basis, there aren't
enough events coming along that allow me to trend how
that plant's performance is related to that particular
event. So, fire events comes in the same scenario
again, the frequency of fires at plants is low enough
that it's just not amenable to timely trending for
indicator purposes.
Now, we can do industry-wide trending on
that stuff, and we can cover the stuff that's not in
PIs through the inspection program, which is a little
more deterministic in some cases, the approach, but we
have a way to deal with them, but we don't have the
ability to do timely indicators of them, from an RBPI
standpoint.
DOCTOR KRESS: How would you get the data
in that middle bullet?
MR. MAYS: Data in?
DOCTOR KRESS: The fire suppression system.
MR. MAYS: Oh, if we were to be able to get
information from the plants on the number of times
that they find failures in the suppression systems, or
detection systems, the number of times they test them,
or demand them, those are the kinds of things the
same kinds of information we get for diesel
generators, or motor-operated valves, is not currently
reported.
DOCTOR KRESS: Do they test these fire
suppression systems?
MR. MAYS: Some of them have testing
information, some of them don't. We have to see what
they have in order to see whether we can make timely
indicators. And, availability of these things is
something else that could be tracked, but right now
that information isn't being tracked and reported in
EPIX or any of the other stuff that we have
availability to, so we are unable to do indicators
directly on those.
CHAIRMAN APOSTOLAKIS: So, the response
time of the fire brigade during drills, that would be
an indicator if it were reported? Is it reported?
MR. MAYS: I don't recall that when we
looked at the fire risk assessments that the response
time of the fire brigade was a really significant
factor in the risk. I think what we found was, the
probability of detection and suppression was generally
more important, and I think Bob Youngblood has some
comments on that.
CHAIRMAN APOSTOLAKIS: Yes, but the
probability for suppression is really a judgment that
comes from the fact that you are going to have the
fire brigade, you are going to have CO2 systems and
all that. The problem is that the models are not
detailed enough.
MR. YOUNGBLOOD: That's the point I was
going to mention, Bob Youngblood, ISL, we have
we've also had this report reviewed by fire PRA
people, and that's one of their comments. If we were
using IPEEEs in this, and they have a lot less detail
in that area, and one of our commenters said
specifically that he thought the fire brigade
performance was an interesting area, it's not, by the
way, equipment related necessarily, which would be
another desideratum, but he wasn't sure that the way
IPEEE has handled the whole thing that we were
necessarily getting the right perspective.
CHAIRMAN APOSTOLAKIS: Yes.
MR. MAYS: So, at any rate, we don't have
any fire initiators, fire initiator or mitigating
system PIs to propose to NRR, because we don't have
the feasibility to do them right now.
The next thing addresses, Tom, part of
what you just talked about in the pie charts. We
looked to see how much risk coverage the RBPIs were
giving us, and we took kind of two approaches, kind of
a Fussell-Vesely and a Risk Achievement Worth approach
to look at the thing. So, let me flip back to the
next table on page 27 here. What we went and said,
let's take a look at the information that's in the
SPAR models for the level 1 stuff that we were looking
at, how many events are actually in that model, and
relating to initiating events and cornerstones, and
how many of them are ones that we would be able to
cover using RBPIs, and you can see that we have a
percentage of those inputs into the total SPAR model
that would get covered by RBPIs.
But, the more important one, I think,
which addresses the question that George raised, is
the next chart, and that one is, this is one that I
think addresses the question that came out earlier,
and that is, we went back to the IP database and
pulled out the dominant sequences for each of the
plants that we were working on here, and looked at
what was the general things that were important to
those sequences. And, what we've done is, we've drawn
a box around all the pieces of the sequence for which
we either have an RBPI from initiating events or
mitigating systems, or we have an industry-wide trend
potential information.
So, what you can see when you go down this
list is that most of the dominant sequences have one
or more pieces of them covered by an RBPI in this set
that we have looked at. So, that's a pretty warm
feeling to have, to know that you don't have a lot of
dominant sequences for which you've got no coverage at
all of your indicators.
CHAIRMAN APOSTOLAKIS: On the right, the
things you have boxed are part of the sequence, and on
the left the initiating events, what's going on there,
everything is boxed, but you have bold faced.
MR. MAYS: Well, we have two things. We
have bold ones were the ones we are directly having in
the RBPI indicators, the dotted lines ones under the
initiating event are the ones for which we don't have
plant-specific RBPIs, but we have industry trending.
CHAIRMAN APOSTOLAKIS: okay.
DOCTOR KRESS: I think industry trending is
a really good idea. I just don't see how it fits into
assessing the performance of an individual plant.
Will you touch on that after a while?
MR. MAYS: Well, this has to do with
something yes, we will, we've got some stuff on
industry trending in a minute, but the short answer to
that is, if I have to go and determine what's
important at a particular plant, and I don't have a
plant-specific indicator for it, then I have to ask
myself, well, what do I know additional about it, and
one of the things I might know is, well, over the
industry this particular thing, which might be risk
important, has been going up over the industry, maybe
that's something I want to pay more attention to.
DOCTOR KRESS: It just raises your
awareness.
MR. MAYS: It raises your awareness.
CHAIRMAN APOSTOLAKIS: Increased monitor
attention.
MR. MAYS: And then also, if I have a
situation where I've seen a dramatic decrease in
something on an industry basis, then maybe I say to
myself, I don't need to spend as much time on my risk-
informed baseline inspection looking in those areas.
DOCTOR KRESS: But, what worries me there,
it's got compensatory errors, too, which bothers me,
some plants are going up and some are going down.
CHAIRMAN APOSTOLAKIS: See, that's his
concern.
MR. HAMZEHEE: But, we realize for this
event, though, Reactor Oversight Process, if it
happens once they are going to send a team to do a
root cause analysis, find out exactly what happened
and why it happened at a specific plant. So, this is
covered, but we don't have specific PI for it.
CHAIRMAN APOSTOLAKIS: But, you could also
do industry-wide trending for the stuff that you
monitor.
MR. HAMZEHEE: And, we are going to, yes.
CHAIRMAN APOSTOLAKIS: And, that's useful
information.
MR. MAYS: Right, we've got that in here,
too.
The next thing we talked about is
CHAIRMAN APOSTOLAKIS: So, let's pick one
there on
MR. MAYS: Okay. I'm trying to get you out
of here by 2:00, George.
CHAIRMAN APOSTOLAKIS: sequence nine.
MR. MAYS: Okay.
CHAIRMAN APOSTOLAKIS: All the way to the
right, it says "HUM," is that human?
MR. MAYS: Yes.
MR. HAMZEHEE: Yes.
CHAIRMAN APOSTOLAKIS: So, there is a human
action there, presumably, a dynamic thing, right?
MR. MAYS: Right.
CHAIRMAN APOSTOLAKIS: During to the
accident.
MR. MAYS: Right.
CHAIRMAN APOSTOLAKIS: And, there's nothing
we can do about it, right?
MR. MAYS: Well, that's not true. What we
are saying is that, the good thing about this table is
that these are the pieces of that sequence for which
I have direct performance indicators.
CHAIRMAN APOSTOLAKIS: So, the baseline
takes care of it, baseline inspection.
MR. MAYS: There you go, the baseline
inspection and any subsequent inspections should be
covering those areas for which I don't have a direct
indicator.
CHAIRMAN APOSTOLAKIS: And, this is,
perhaps, NRR folks, this table, or a table like this,
it could be the basis for these tradeoffs that we
discussed earlier. If I put an extra performance
indicator somewhere, then I should reduce the
activities in the baseline inspection.
MR. MAYS: This is a similar concept which
CHAIRMAN APOSTOLAKIS: That's very useful,
this table.
MR. MAYS: right, this is a similar
concept that was used for devising the baseline
inspections in the first place. They went back and
looked at some PRAs, some stuff that was and wasn't
covered in the
CHAIRMAN APOSTOLAKIS: Not in such detail,
Steve, come on, not in such detail. I mean, it was a
general
MR. MAYS: It wasn't in that detail, but
the concept is the same, and what this does is provide
more detail they could be able to use as a basis for
going back and potentially looking at the inspection
program.
MR. BOYCE: George, I think we agree
conceptually.
CHAIRMAN APOSTOLAKIS: Sure. No, I
understand.
MR. BOYCE: Well, the program has got to be
mature before we can really utilize the results with
any degree of confidence. We are not going to revise
our program based on preliminary results. I mean, we
are very interested in this sort of approach, and I
think in our initial comments, perhaps, even in that
aforementioned December 1st memo, we pointed out that
this was an area where we thought the program could be
very useful.
And, right now, there's a separate program
outside of risk-based PIs to utilize risk incites in
our inspection program, and we've got that initiative
going in parallel to this, but it's not as systematic,
it's not as robust and detailed as this program has
the potential to offer.
CHAIRMAN APOSTOLAKIS: Good. Good.
MR. MAYS: The next thing we had in the
report was, we did some what we called validation and
verification, and what we wanted to do was go back and
prove that we could actually do this thing and produce
PIs and evaluate against thresholds, and so we went
back and used the 23 plants that we had for the period
1997 through '99, and put the data to the test to see
what happened. And, when we did that on the next
page, what we found out when we looked at that, we
think we have a more precise accounting for the risk
significant design features of the plants. We know we
have more plant-specific thresholds, and we think we
have a better dealing of false exposure time. That
was one of the things we mentioned earlier, and we
have this kind of "face validity" approach that we are
taking to say, does this stuff make sense from a risk
perspective, once we've put this stuff through the
models and seen what comes out.
So, we've got some tables to show you, and
we do have one caveat that we want to make sure we put
in here. We haven't had all this data and these
models go through peer review, so if anybody were to
conclude that this is a definitive statement that some
plant is either green or not green, that would be a
bad conclusion, because that's not something we are
trying to do at this point in time.
So, under the initiating events, we take
the 23 plants that we had, we've gone through and
determined what the actual data shows, we've got the
values in there, along with in the parentheses what
the particular color would be for those initiators.
On the next
CHAIRMAN APOSTOLAKIS: So, there are a few
whites there.
MR. MAYS: Yes, there are.
DOCTOR KRESS: Is this a good argument that
George can use to say that the previous use of the
95th percentile was not an appropriate way to go?
MR. MAYS: Well, we've made that argument
generically in the report.
CHAIRMAN APOSTOLAKIS: They agree, I think.
MR. MAYS: And, that's the earlier tables
where we were showing the 95 and the other one was
based on more than that.
DOCTOR KRESS: Right.
CHAIRMAN APOSTOLAKIS: It's interesting,
though, if you look at I mean, there is a yellow
here.
MR. MAYS: Yes, that's correct.
CHAIRMAN APOSTOLAKIS: That's B&W Plant 5.
MR. MAYS: Uh-huh.
CHAIRMAN APOSTOLAKIS: Yellow on the
general transient, white on the loss of heat sink, and
green on the loss of feedwater flow.
MR. MAYS: Right.
CHAIRMAN APOSTOLAKIS: So, what would the
Action Matrix dictate now?
MR. MAYS: Well, again
CHAIRMAN APOSTOLAKIS: That's beyond what
you are doing, right?
MR. MAYS: that's beyond what we are
doing now, and in addition, for that particular plant,
we were going back, remember I said we were doing
"face validity," we were going back and checking
because that looked to be higher than what we are
seeing on other B&W plants, and we're going back to
see if there wasn't a modeling issue that was causing
that to be, and we are looking at that as well. So,
that was the reason for that caveat in the previous
slide.
When we go to the mitigating system
unavailabilities, we have a similar layout for the
plants there on the next two charts. The key thing
there was that, for example, on AFW/RCIC, depending on
which ones you are PWR, we broke out the motor-driven
pump and either the diesel-driven or turbine-driven
pump separately, because that was one of the things we
found, that currently we were averaging trains
together, and they have different risk implications.
So, when we looked at them this way, we saw that the
risk implications were different, and that gave us
part of that face validity that we think we are having
something that makes sense from our understanding of
risk.
We also have tables for the unreliability
of the plants, and in this case, just to make the
table a little more presentable, instead of going out
and calculating all the individual mean values for the
unreliabilities for those sections, we know that if
you go over a three-year period we have no failures,
and any number of demands, that the update is going to
be equal to or less than the baseline, and since the
baseline was below green there was no point doing
anything more for it. So, we just took a shortcut in
this table and put less than baseline for all the ones
where we had no failures.
Now, if you look at that, for example,
down at the bottom, the PWR list, the last one,
Westinghouse 4-Loop Plant 23, you'll notice that in
the AFW column there is a number in there, and the
value is 1.5E-2 for motor-driven pumps, and then it
has an indication of white. And then you notice again
down below it there's a number, .13, that was the case
where we had something that went over white, and so we
said there's only a 13 percent chance, based on how
far that distribution overlapped that threshold that
the actual value was still at the baseline.
So, if you were to have, you know, a high
number like .87 or something like that, then you'd
say, well, maybe this isn't quite a white threshold,
maybe this is just the uncertainty in the data, but
when you have a fairly low number there you are more
confident that you've crossed the threshold.
I want to skip the last one, unless you
have a particular question on it. We had the
component class scenario, and we did a similar thing,
if we had no failures we didn't calculate the actual
number, and when we did have a failure in that group,
we calculated a number and determined whether or not
it was green, white or yellow.
The thing we've kind of touched on
tangentially several times here has to do with
industry trending. When we originally started out
this program, we were considering industry trending as
part of an integral part of the PIs, and then as we
looked at it more we recognized that it was related
but not directly a risk-based performance indicator,
or at least not on a plant-specific basis. But, we
thought it was important to capture that there might
be risk-important events that occur, and risk-
important activities that occur, for which we can't do
a plant-specific indicator, and it needs to be
captured someplace, it just can't be left alone.
So, for those we looked at doing industry-
wide trending, and what we proposed for industry-wide
trending, both as an input to the ROP for those areas
for which you don't have direct indicators, well, if
you don't know the specific plants is the industry
getting better or worse, that's an important thing,
and also because we have a requirement in our
strategic plan to report to Congress whether or not
we've had any statistically significant adverse trends
in industry performance, so we viewed this information
as also being an important piece potentially to that
performance measure for the Agency.
So, what we proposed in the next table is
that we would have and develop, and they are in
Appendix A, I believe, is where most of them are, you
would be able to trend all of the proposed RBPIs that
we've already got in the report, as well as several
indicators that had frequencies that were less likely
to occur, and we grouped them in this table by the
cornerstones that they impact, and what we would be
able to produce in each of those areas. So, that's
what we've proposed as potential industry trends, and
some of that information is in the report.
I think we've spent a significant amount
of time so far already, by the way, talking about
several of these issues, and again, these are
implementation issues that this report and this
program is not going to directly address, because we
are really looking at the feasibility of putting
together indicators for the ROP, but we recognized
these were important things, so in our interactions
with the ACRS, and the public, and other people, as we
were going along, we wanted to raise these issues and
get people's thoughts going on them so that we could
know what the perspectives were before we got too far
down the road.
So, the first question was, well, do we
even need anymore indicators at all, or are we okay
with the set we've got now? Can we do everything we
need to do and still get by? I think the answer to
that is pretty clear. We believe that we can run the
concurrent Reactor Oversight Process and do an
adequate job. The question is, can we do better?
And, the stakeholders had different views.
The industry said, well, if we are going to get
greater sample and more PIs then we there needs to be
changes to the inspection program as well. So, our
position is that these are consistent with the policy
statement and the concept to try to use more objective
risk information in all areas possible, and the ROP
change process is where we are going to make an
assessment of whether or not they are worth it. I
can't tell you all the details of how that is going to
come about, but, I mean, that's where that I can
tell you that's where the question goes to get
answered.
On the next one, the key issue was how
many PIs do you have? We have, potentially, you
could have, if you made swaps for like indicators and
new indicators, you could have in the ball park of
about 30 indicators per plant compared to the 18 that
the ROP currently has, and people are questioning,
geez, is that really an appropriate level?
Our position in Research is that the total
number of performance indicators should be
commensurate with the amount of risk coverage you want
to do by objective performance indicators, and that
number hasn't been determined, and that's kind of a
there is no magic formula for calculating that,
there's no
DOCTOR KRESS: It's a policy.
MR. MAYS: it's a policy kind of thing,
and that's something that once we see how much
coverage the RBPIs have, with respect to what the
current ROP has from the indicator standpoint, and
what the desired mix is between the two, somebody can
come to that decision, but we believe that's the right
question.
DOCTOR KRESS: You just can't develop a
utility function for this, that's the problem.
MR. MAYS: Correct.
DOCTOR KRESS: And, that's what you need.
MR. MAYS: The next questions that come up
with implementation had to do with data sources, do we
have those data sources, do they have the required
quality in order to be used as part of the oversight
process? We believe that the key here is that the
data needs to be of sufficient quality so that if
there is an error in the data it's not going to change
your overall context of how you are going to view the
plant. So, for example, if somebody comes up and says,
well, I had reported 24.6 hours of unavailability in
my diesel and, in fact, you go back to the plant and
you found out that, wow, it was really 26, well, if 26
isn't sufficient to change your conclusion about the
plant we don't think that that's a level of precision
that needs to be part of the quality and the data to
do that. But again, this is another part of the
implementation issues that will have to get worked out
and we would expect that to probably get worked out
through a pilot program.
The next question that comes up has to do
with the next two, in fact, have to do with models.
I had pointed out earlier that one of the main things
was, you have to have a model that has a reasonable
representation of the risk. I chose that word
carefully, because we've developed level 1 Rev. 3 SPAR
models for about 30 of the plants now, and we've got
a program to develop them for the rest of them.
The number of models' needs depends on the
level of plant specificity that you want to have. We
may be able to group things, we may want to do them in
plant-specific, but that's something that we have to
eventually come to a decision by.
And, the external stakeholders recommended
that if we were going to use SPAR models in this kind
of a process that they be reviewed by the licensees.
We agree with that. We've already been in the process
of taking several of our SPAR models on on-site
visits, and we have got plans to try to do that for
all the rest of them, because we believe it's
important to make sure that we are not, you know, out
in left field compared to what the plants have.
Now, we've done ten direct on-site visits
to review SPAR models, and on some occasions we found
that there were either equipment or procedures to deal
or mitigate with certain sequences that we didn't know
about, and then once we found out about them we
included them, and sometimes we've been to plants
where we've gone and they've said, holy cow, we think
our model needs to be fixed, your's is a better
representation of what's going on here. So, there's
a difference in the way that things are done,
depending on how long it has been since the plant has
updated their IPE, and what their groups are, but the
key for us is that with the SPAR models we have a set
of models in which we have a consistent methodology
for examining the same kind of information across
plants.
CHAIRMAN APOSTOLAKIS: And, this is not
just for this project, I mean, this will
MR. MAYS: NO, this will apply to other
things in the agency, and I think that's an important
thing, also from a public confidence standpoint, that
I think we need to be able to say that we have
something that we look at that's independent of what
the licensees come and give us, so that we have the
ability to say we've done a critical look at what
they've presented to us.
The other advantage to us is that if we
have these models done this way, then we have the
ability when we have differences between their models
and our's to focus very quickly on what the basis for
the differences are rather than having to take a long
time to go over and review their model from complete
beginning to end.
CHAIRMAN APOSTOLAKIS: So, have the
licensees urged you to, in fact, have as the SPAR more
than their better IDs or PRAs? Some of the licensees
did a complete PRA.
MR. MAYS: That's correct.
CHAIRMAN APOSTOLAKIS: Have you seen any
desire on their part to have a SPAR model that you
have be the PRA?
MR. MAYS: Well, actually, what we found is
that, if we have significant differences between what
we have in our SPAR model when we go to a site,
between what they have, for example, in the core
damage frequency in our's, we sit down and say, what's
the differences, and if we find something in there
that is, well, we've put in this new system, or we
have installed these new procedures, or we've changed
the plant design from what you had here, we go back
and look at those things, gather that information, and
we make modifications to the SPAR models in light of
that.
CHAIRMAN APOSTOLAKIS: But, I understand
the SPAR models are sort of approximate, or is that a
wrong understanding?
MR. MAYS: I don't think approximate is the
right word to use.
CHAIRMAN APOSTOLAKIS: Can I put a complete
PRA, like full scope, the Seabrook PRA, can I put it
in a SPAR model?
MR. MAYS: Okay. The SPAR model stands for
Standardized Plant Analysis Risk, that's our
determination of the style and method of doing PRA
analysis and we apply it across all the plants.
However, the SAPHIRE suite, which is the
engine that allows you to run that, is capable of
taking a plant-specific PRA and putting it into it so
that you could do that.
Now, again, the problem there is, and we
have several plant-specific PRAs that are available,
Research has put those available in SAPHIRE, the
problem again there becomes, that just represents our
version, that gives us a model that represents their
PRA, and that one from the next one to the next one
may have different HRA assumptions, different CCF
assumptions, different modeling assumptions that they
put into the plant. So, while we have the actual
model in that case, we don't have a consistent basis
across them for examining what's happening. So, I
think SPAR models provide a different kind of benefit
to us, because we know that if we go and look at
Westinghouse 4-Loop Plant A and Westinghouse 4-Loop
Plant B, that if there's differences in the CDF
associated with those SPAR models it's because we've
determined something different about the plants, not
because we have different modeling assumptions.
So, we tried to limit the impact of
different modeling
CHAIRMAN APOSTOLAKIS: And, the
Significance Determination Process will be based on
the SPAR models at some point?
MR. MAYS: The Significance Determination
Process that currently exists is based on the SPAR
models now.
CHAIRMAN APOSTOLAKIS: It is?
MR. MAYS: It's based on the ASP and the
SPAR models, that's how it was developed, and the
Significance Determination Process for Phase 3, where
we go out and do a more detailed risk analysis out of
it than what's in Phase 2, which is the table lookups,
in Phase 3 we actually go and put together a model to
look at that, and in most cases that uses the most
recent updated SPAR models we have.
CHAIRMAN APOSTOLAKIS: So, the tables that
are being used in the SDP are based on the SPAR?
MR. MAYS: Absolutely.
CHAIRMAN APOSTOLAKIS: All right, let's go
on.
MR. MAYS: A similar question relates to
the LERF models. We only have a limited number of
those available, and we only have a limited capability
to develop those in the short term, so the issue with
the LERF has to do with the RBPIs that there may be
some mitigating system components whose threshold is
set based on CDF, that when you consider LERF might
actually get different thresholds. So, we haven't
been able to do that yet, but we recognize that that's
an issue with respect to whether or not what these
represent the public risk with respect to the
thresholds.
So, let me get to the stuff which I really
wanted to talk to you about today, which was
DOCTOR UHRIG: After lunch?
MR. MAYS: Maybe after lunch, if you want
to talk too.
CHAIRMAN APOSTOLAKIS: Maybe we should do
that after lunch.
MR. MAYS: Okay.
CHAIRMAN APOSTOLAKIS: We can't finish
everything before lunch.
MR. MAYS: No, so let me leave with you
with a taste of what it is. What we've done here
CHAIRMAN APOSTOLAKIS: I want to go eat.
DOCTOR UHRIG: Let's have a taste of lunch.
MR. MAYS: You want a taste of lunch?
Okay, no problem.
CHAIRMAN APOSTOLAKIS: Okay, we'll be back
at 1:00.
(Whereupon, the above-entitled matter was
recessed at 11:49 a.m., to resume at 1:00 p.m., this
same day.)
. A-F-T-E-R-N-O-O-N S-E-S-S-I-O-N
(1:03 p.m.)
CHAIRMAN APOSTOLAKIS: Back in session,
continuing with Mr. Mays and Mr. Hossein Hamzehee.
MR. HAMZEHEE: Yes, sir, correct.
MR. MAYS: Okay.
The thing we wanted to talk about next was
some alternate approaches we've looked at in light of
the comments that we had about the number of PIs being
excessive or too many, and so what we went to do was
relooked at what was the basis for doing these in the
first place, and what we did originally was we
devolved risks into smaller pieces, and we set all of
our thresholds for risk-based PIs that are in the port
at the level at which the data was being collected.
So, if I had data on reliability, I had a threshold on
reliability. If I had data on availability, I had a
threshold on unavailability. And then, we looked at
how much of an impact that changes in those values
would have on accident sequence frequencies when we
did that.
We took a slightly different approach,
which I'm going to go through in these next figures
and talk to you about, so let me just put the figures
up and go through them. What we did was, we took the
accident sequences and we devolved them down into risk
areas at a more functional level, rather than at
reliability and availability of components, and then
we looked to find out what data we could do within a
particular functional group and then reassessed that
against our criteria for whether or not it was a good
PI.
So, if you start from that's the wrong
title, it should be Industry Risks instead of
Individual, I'm sorry. But anyway, industry risk
comes from all the plants together and individual
plant risk comes from containment, core damage and
health effects things, and so underneath core damage,
which was where we were primarily looking, we looked
at what were the big pieces under initiators and
mitigating systems that might be amenable to a
slightly different approach, which would reduce the
number of PIs.
So, under initiators we said, well, we
might be able to group those into three groups,
transients, LOCAs and special initiators, and we list
some of the values, some of the types of initiators
that might go under that, for example, under LOCAs you
could have small, medium and large LOCAs, you could
have steam generator tube ruptures, you could
potentially have other ones like very small breaks or
inter-system LOCAs, things of that nature. And, under
mitigation we took the approach which is on the next
slide, which was we put together just kind of a very
basic functional event tree that's generally
applicable for anybody, for example, in a PWR where
you have an initiating event, your first issue is do
you establish reactivity control, then do you have
secondary heat removal. If you don't have that, do
you have feed and bleed, and then you have
recirculation, cooling. So, at that functional level
we were trying to see what we could do to do RBPIs.
And so, our concept is that what we would
do is we would develop functional impact models at one
of those levels and we would take the inputs for
reliability, availability and frequency that currently
apply to that functional level and use those as feeds
in together. Now, this is a case where we are having
a multi variate sensitivity study instead of single
variable sensitivity study. So, at the level, say, of
secondary heat removal, we take all the things that
impact secondary heat removal, put them into that
model, and see what that would change to the baseline
core damage frequencies that way.
So, when we did that, we came up with
three different kind of levels at which we could
potentially put indicators together. We could put
together an indicator, for example, at the cornerstone
level. So, if you went to the initiating event
cornerstone we could have an indicator that said this
is the impact of all the different inputs at this
cornerstone level together. We could do that also for
the mitigating systems, or we could go to the
functional level and have somewhere between three and
five indicators at a kind of higher order value, such
as heat removal, feed and bleed, those levels, or we
could go back down to the component and train level,
which is where we currently have stuff in the RBPI
report.
So, in looking at that, the way it would
look would be something like this. At the cornerstone
level, you would have, basically, two indicators. You
would have an indicator for initiating events, where
you would take the data associated with loss of
feedwater, loss of heat sink and general transients,
and you put them all together and run them through the
model and see what the output results was.
Now, what's different about this is that
in these cases, in all these functional cases, you
have the threshold being set at the output condition,
not on the input condition. So, currently in the
RBPIs we have a threshold for loss of feedwater, we
have a threshold for loss of heat sink, we have a
threshold for general transients, what we would do now
is take that data for all those data and say, what
would be the impact on the sequences for all of them
collectively. So, the threshold is now set on the
collective sequences, not any individual input.
So, you might, for example, have gotten
better on feedwater, or worse on heat sink, and
somewhere in between on transient, and you may or may
not get better or worse, depending on how that would
go. So, this is more like the integrated indicator
that we had talked about doing in Phase 2, but it's
not the complete total plant model version.
At the functional level, down from the
cornerstone level, we came up with two ways of
potentially doing this, and one was to take the
mitigating systems and group them by what initiator
they respond to. So, we would say, we'd take all that
data that we previously had in those 18 or 13 RBPIs
and we'd say, all right, which of that data when it
changes, how does that affect the transient sequences,
how does that affect the loop sequences, how does that
affect the LOCA sequences, and we just put them
through the entire model for all those things and see
what the impact would be.
DOCTOR KRESS: When you say how it affects
the sequence, do you mean how does it change the
sequence contribution to the CDF.
MR. MAYS: Right, collectively, together.
DOCTOR KRESS: Collectively, together.
MR. MAYS: Right. We take all, so in other
words we would take all the failure to start
information, all the failure to run information, all
the unavailability information for components that
affect loss of off-sight power sequences
DOCTOR KRESS: So, your threshold would be
a delta CDF.
MR. MAYS: a delta CDF for that
particular initiator.
DOCTOR KRESS: Uh-huh.
MR. MAYS: Or, that group of initiators.
So, that way we'd say, okay, the
mitigating system performance for transients is this,
it's green, white, yellow or red. The mitigating
system performance for loss of off-site power is this,
even though they'd be using some of the same data they
have different potential risk impacts. That's one way
to look at it. We'll show you some results of that in
a second.
The other way to look at it, which is a
little more like the current ROP, a little more like
the SSPIs and other stuff that INPO has, is to group
them by their function, their system functions. So,
for BWRs, for example, we would say we have RCIC and
HPCI systems that kind of do high pressure
performance. We have diesel generators. We have RHR
systems, we have what we've referred to as
crosscutting, which is those AOVs, MOVs and MDPs that
go across all the systems at the plant, we say we
could take that group and run them through the model
for all initiators, essentially, and see what the
combined impact of those is on the output.
So, we did that. We did a trial on that
to look and see what it looked like. So, let me show
you the first one we have, which is what would happen
if you did this stuff at the cornerstone level. We
took a BWR and a PWR plant that we previously have in
the report and we ran it through and said all right,
if we put, for the cornerstone level for the systems,
what we have here is that you take this particular
plant, and you take its data on diesels, on HPCI, RCIC
and RHR, all those together, and take all those
systems, all those inputs together and see how that
mitigation comes out, this particular plant comes out
to be white.
For the initiators, which is the next one
down, the initiator impact says it is green, so from
that particular plant we could come up with a white
for the mitigating systems and green for the
initiating event cornerstone, at that level.
And, we have a similar thing we've done
for Plant No. 23. Now, we didn't actually know these
were going to come out white and green, they could
have come out both green, or both white, or something
else, it just happened to be done that way. So, this
says I could actually come up with a value of what the
performance was at the cornerstone level, if I wanted
to do that.
Now, we'll talk in a minute about what the
advantages and disadvantages of doing that are, but
that's what we could have done at that level.
The next one I have to show you is if we
were to take the performance of the mitigating systems
and group their impacts by the initiating events for
which they are supposed to function. So, the first
case says, for the BWR plant, this says the front-line
systems, which is the RCIC, HPCI and RHR, as well as
the crosscutting component group, for LOCAs, for all
the LOCAs that would be applicable to that unit, the
performance for mitigating LOCAs is green. The
performance for mitigating losses of off-site power or
station blackouts is white, and performance for
mitigating transients is green. So, this gives you a
little more information than what you got a minute
ago. At the cornerstone level, you just knew
something was white, but you didn't know what. This
one gives you a little more detail. It says the thing
that's important at this plant is that this
combination of performance for all these systems is
most important in loss of off-site power sequences.
DOCTOR KRESS: This means that you take all
of your you have to take all of your input data on
liability and unavailability and run it through
MR. MAYS: Run it through the model.
DOCTOR KRESS: the model at that point.
MR. MAYS: Right.
So now, this is different from what we had
done before.
DOCTOR KRESS: You ran the model.
MR. MAYS: Now, I'm using the model to run
the entire thing through to get the impact.
DOCTOR KRESS: To get the impact.
MR. MAYS: Because I can't do it correctly
I mean, the advantage to the other ones
DOCTOR KRESS: This takes care of my
problem.
MR. MAYS: Right, but it creates another
problem.
DOCTOR KRESS: Yes.
MR. MAYS: And, the problem it creates is,
before what we had was, we would set the thresholds by
using the model and then we'd just collect data and
compare the data to the thresholds, we didn't have to
go back through the model again.
DOCTOR KRESS: Exactly right, now you have
to go through the model every time.
MR. MAYS: Now I have to go back through
the model every time to do this, so that's a
difference, because I can't take into effect the
combined effects without putting it through the model.
DOCTOR KRESS: Yes, that's right.
MR. MAYS: So, this is more like the
integrated model.
DOCTOR KRESS: It's almost like the
integrated indicator.
MR. MAYS: You are right.
So, that's how you could do it if you
wanted to group the mitigating systems in accordance
with, for example, the initiator they respond to.
Another way to do it, which we have on the
next slide, is to do it by kind of the high-level
functions that the systems perform. So, for example,
again, the same plants, same two plants, what I see
now is that the electric power system for the BWR
plant, which would be these reliability and
unavailability combined now, is green, the HPCI, which
is the reliability and availability combined, would be
white, the RCIC is green, the RHR is green, and the
component groups is green. So, now I have a different
kind of perspective about the performance.
DOCTOR KRESS: But, none of this changes
the amount of reporting requirements.
MR. MAYS: RIGHT. This is with the stuff
that we already have, for the existing level we were
using for the RBPIs in the report, so we are using the
exact same data that we had in the report to do the
indicator a little differently.
So, let's, you know, having done that, and
you can see now you see that the thing that's causing
the station blackout loop sequences in the BWR plant
here to be high are the ones associated with the HPCI
system, not with the emergency power system, so
there's kind of advantages to going both ways if you
want to do that. So, we looked at what the potential
benefits of these things would be, and at the
cornerstone level the biggest benefit is, I've got a
single indicator. It's just, did this plant's
initiating event information and mitigation systems
that I'm monitoring rise to the level of having
performance that I'm concerned about. One indicator,
one time, see it. That's not too bad, and the other
advantage of this is that it takes into account inter
and intra-system impacts of changing in performance in
different areas, and we actually went back and looked
at this and we found, as we looked at the plants,
sometimes you would have things that were, say, in an
individual indicator, it was green and white, you put
them together in a combined thing and they turn out
green, or sometimes we'd find it went the other way,
there would be a white and a white, and you'd put them
together and it turns out yellow, or you find things
that are green and green and they turn out white, or
two whites end of turning I mean, you see variation
depending on which sequences particular inputs are
involved with.
DOCTOR KRESS: Now the question I might
have is, what's the down side of doing all of these?
MR. MAYS: Of putting it all together?
DOCTOR KRESS: Doing all of them.
MR. MAYS: That's another that's another
thing you could potentially do. Let me get to that in
a minute.
DOCTOR KRESS: Okay.
MR. MAYS: The limitation, of course, is
that once you find something that's not green
performance, you don't really know directly what it is
that's causing it so that you can go out and find out
what you need to spend regulatory attention on. It's
not very precise in telling you what's the particular
area that needs to be dealt with.
If you go to the functional level, well,
the benefits are we have fewer indicators, instead of
18 or 13 we are now talking about three, four, five or
six, depending on how you want to use these. This
also accounts for intra and inter-system impacts, and
it can be grouped either by major types of initiators
or by system functions, or if you wanted some other
way of looking at it you could postulate one and we
could do that.
The limitations of this is, now when I
have them broken down into functional groups, I now
have to have some way of bringing them back together
to make my assessment of whether the cornerstone was
degraded or not, because they don't directly tell me
the entire cornerstone picture. And, I still have the
situation where if I have greater than green
performance I still have to do more work to figure out
why it was greater than green. I may know it's in the
HPCI system, but I don't know if it's the availability
or the reliability, I've got to go back and do some
more looking before I can figure that out.
DOCTOR KRESS: That's why I was asking why
not do all of them?
MR. MAYS: Well, you are getting to the
thing.
The last one I had was, if you do it at
the component or train level, like the current RBPIs,
the biggest advantage here is this is the broadest
evaluation of individual attributes, and the causes of
greater than green performance are pretty obvious once
you've got it at that level. I know it's diesel
generator reliability, or I know it's AFW diesel-
driven pump train reliability or its availability. I
know the area much more precisely when I have the most
broad individual indicators, and it's much more
similar to the current indicators because the
indicator data and the thresholds are set and I don't
have to do running through models anymore, I just pick
up the new data and compare it, away I go.
The limitations are, the inter and intra-
system dependencies aren't accounted for here. So,
sometimes you get worse and sometimes you get better,
depending on what the risk relationships are on the
accident sequences, and if you have them at an
individual variate level you don't see that.
Now, the disadvantage also is it nearly
doubles the number of current PIs that we have, and it
requires you to set an individual plant-specific
threshold on lots of different indicators.
DOCTOR KRESS: But, it doesn't double the
quantity of data that you need collected.
MR. MAYS: It's the same amount of data.
DOCTOR KRESS: Same amount of data.
MR. MAYS: Exact same data.
So, that's the kind of stuff that we've
looked at as potentials, so we are looking for
feedback.
Now, one of the things that you mentioned
that you could do is, you could say, well, if you are
going to take all that data and run it through the
model, either at some lower level, intermediate level,
or up to the cornerstone level, well, why not do it at
all of them and just have one of them be the one you
report out to the public, and utilities and everybody
else, and the other ones be the ones that you use as
subsidiary things to go back and see what was causing
it to be the way it is. That's a possibility. We
haven't really done much more than have some
preliminary discussions with NRR, because we just got
finished running some of these examples, as to what
one they think would be the best.
So, what we intend to do is try to get
your feedback on where you think we should go. We are
going to talk about this at the public meeting next
week, to see if people think that this is a good idea
that they would prefer or not, then once we get your
comments and their comments we are going to sit down
with NRR and we are going to say to ourselves
collectively, what makes the most sense to do and
publish in the Phase 1 report, which should be out in
November. So, we will have a kind of meeting of the
minds at that point, and we'll say, based on the
comments we heard, and our own internal discussion,
this is the way we think we should go. So, this
report could be dramatically changed if we decide to
go at a different level. If it's decided to throw
away the component level then this report would
certainly change, or if it's decided to do it at multi
levels this report will change, but we have to come to
that decision after we get some comment and feedback.
I think what I wanted to make sure you
understood today was we have the ability to do this in
different ways, and we are looking for feedback as to
what people think would be the best way to go.
And, we'll take those comments and we'll
go from there.
So, what we are looking for the ACRS to
give us feedback on, again, is whether they think
these represent potential benefits to the ROP or not,
whether they think this technically is an adequate job
of how you would go about defining and calculating
these things, and whether or not the alternate
approaches we just discussed here are something that's
worth pursuing or not, or whether it's something to be
not done until Phase 2, or done as part of Phase 1, or
where you think that kind of stuff should go.
So, I'm pretty confident that we have
information that uses readily off-the-shelf PRA tools
and methods, that uses readily available and
accessible data to us, that we can put together
broader and more plant specific performance indication
for potential use in the ROP. And, at that point if
we've got that, we'll hand this off to our friends
over at NRR, who can then go through the
implementation process, and where we will have to
answer the questions like can we actually get this in
the plants, what's it really going to cost to get this
data, are we willing to use these models, do we want
to instead use the plant-specific models from the
licensees by giving them some specification and using
those? I mean, those are all possible questions that
could get answered through the implementation, and I
don't want to minimize the fact that those are serious
questions and will require some serious work to make
it happen, but I think as long as we keep in the mind
set of progress, not perfection, and is this a valid
incremental improvement or potential improvement, then
that's where we want to be at the end of the day.
So, I guess the only other thing we have
is, if there's something I guess we have to hear
from Tom.
CHAIRMAN APOSTOLAKIS: Let's hear from Tom,
and then
MR. MAYS: About what you want us to
present to the full committee.
CHAIRMAN APOSTOLAKIS: we'll do that
after we hear from Tom.
MR. MAYS: Okay.
CHAIRMAN APOSTOLAKIS: So, Nuclear Energy
Institute, please go.
MR. HOUGHTON: The ball moved down the
field, so I decided to take notes and talk from them.
The first thing is, the Nuclear Energy
Institute and the industry is very interested in risk-
based performance indicators and trying to move
forward in having the process be more risk-based as we
can. Of course, the caveat always and I think
Steve has done a real good job in working through this
and raising a lot of the issues in the analysis he's
done, and a number of things he's done address
problems that are problems with the current program,
which I want to talk to you about.
But, the caveat, of course, is that it
needs to be considered in the context of the ROP, and
what the purpose of the ROP is, and the performance
indicators. The performance indicators are meant to
be used to help the NRC determine how much inspection
it's going to do above the baseline inspection, and
how it assesses plants and how it engages in
enforcement of plants.
Our feeling is, if there's no reduction or
efficiency improvement in inspection, that it's
difficult to understand why we would put additional
effort into generating performance indicators.
Now, one could say that a lot of this
information is gathered under the Maintenance Rule and
under internal performance indicator gathering, and it
is. The problem comes, is that we move from, gee, I
think that was a loss of feedwater initiating event
to, my inspector is coming in and he's looking in a
manual and he's making a decision for which I can be
cited for a violation of the regulations in my
reporting, or it involves a long-winded process of
trying to resolve weather the issue counts or not, or
whether the hours count or not.
We've had on the order of about 260
frequently asked questions over the course of the
program so far, and these involve sometimes matters of
15 minutes of unavailability. So, the devil is in
these details, and as we expand the number of PIs it's
not just we might have some of that data, the question
is, is it worthwhile the extra effort that has to go
into that. That was one point.
We support a stable, consistent and
improving system of performance indicators, which Mike
Johnson talked about and Steve talked about, in terms
of the change process. I suspect that some of your
questions that I heard today are, why aren't we
thinking about some of these implementation or use
issues at this stage, rather than finishing up the
whole Phase 1 effort and then turning it over to NRR.
Unfortunately, to determine that this indicator
doesn't provide us additional value or a P process, or
that it's too difficult to explain to the public what
you are working about, because you are talking a more
sophisticated type of indicator, particularly, as I
was hearing if we went to a cornerstone level
indicator I think that would be more difficult to
understand, as opposed to a shutdown or an
unavailability.
So, we would and we support piloting
these. There are a number of pilots going on now.
There's one going on about the SCRAMs and the loss of
normal heat removal. There's one that's going to go
on to try and revise problems with the power changes
that will be piloted soon. And, the results, the
purpose of these pilots, as you were asking, is,
really, is it easy to understand what the indicator
is, is it easy to report it without making errors, and
is it more efficient in terms of inspection, is it
more focused on risk-significant issues, those are the
sorts of things that we want to see coming out of a
performance indicator so that it's of value to both
the NRC and to the industry.
A couple of comments on the initiating
event PIs. I think plant-specific goals are a good
objective, however, when you look at the range of data
it may become difficult to explain to the public or to
understand as a licensee the fairness of one plant
getting extra inspection if they had two general
transients, two SCRAMs, and another had seven. Okay,
that just you know, a SCRAM is not a good thing,
and if you had such a disparity, even at the
green/white level, that's a question to raise as to,
what does that mean to the public, what does that mean
to the licensee.
Loss of heat sink, I may be wrong, but I
looked at added up over a three-year period, and for
one of the plants it was a .7 transients per three-
year period. That means if you had one loss of heat
sink you would go into the white category. I don't
understand that. I think it shows the limitations of
risk-based versus risk-informed and how we have to
look at what we mean in terms of implementation here,
not just what does it mean in the risk model, but what
does it mean in the implementation world.
Also, a yellow for three SCRAMs, general
transient SCRAMs in a year, would not be very
appropriate to do.
Mitigating systems, the biggest issue we
are having now is unavailability. That is a real
problem. It's causing a lot of gnashing of teeth
among system engineers who have to go out and do the
Maintenance Rule one way, and the INPO indicator the
other way, and the ROP another way, and their PRA
person has a different way. So, we are working on
that, and I think a lot of the things that Steve is
working towards, reliability, not counting fault
exposure, are good things that we want to work, and we
really want to work on them faster.
CHAIRMAN APOSTOLAKIS: But, there is
resistance of changing these things and coming up with
a uniform set of definitions, which is a mystery to
me. I mean, this is the third time that I recall this
committee facing this issue, of what is reliability,
what is availability, and so on, and every time we
recommend that we need a White Paper with consistent
definitions, and every time we get something, we'll
think about it.
It's a very thinking agency here.
MR. HOUGHTON: Well, we got started with
the kick-off meeting amongst key players a month ago.
We have another meeting in May, and anything you can
do to support putting focus on this
CHAIRMAN APOSTOLAKIS: Well, I don't know
what to do. What should we do, Steve, to give focus?
Do you guys want to come before the committee? You
don't have to, you are industry, but I wrote a long
memo with four or five definitions, when was this
MR. HOUGHTON: Of course, we have to
CHAIRMAN APOSTOLAKIS: was it A.D. or
B.C., I can't remember.
MR. HOUGHTON: we have to satisfy a
number of interested parties. There's the Maintenance
Rule, which has its set of rules and the way it's been
doing things. You know, and that's a rule.
We've got PRA practitioners and they way
they look at things. We've got the INPO/WANO system,
okay, which they've been very good, in that they will
defer to the ROP definition, because they feel it's
more conservative. Okay. And, we have the ROP.
We have a basic underlying issue, which
is, in unavailability it's to be used to help
inspectors decide how much to inspect. Okay.
Inspectors inspect design-basis tech specs, allowed
outage times. The best definition we can come up with
is going to be more risk-based and oriented, okay, so
there's an important issue there that needs to be
addressed. So, it's not trivial to do this.
CHAIRMAN APOSTOLAKIS: I don't know what
the best way is. I mean, we'll leave it up to you how
you want to involve the committee. We can ask the
staff to come here, we can't ask you to come here.
MR. HOUGHTON: Well, I mean, we are happy
to come and talk and participate. I'm just
CHAIRMAN APOSTOLAKIS: It's something maybe
you can coordinate with Mr. Markley.
MR. HOUGHTON: It might be appropriate for
someone from the staff from NCR
CHAIRMAN APOSTOLAKIS: Will you be ready in
May when we have the full committee meeting to address
this issue, or is too soon?
MR. HOUGHTON: Well, we can lay out some
parameters of what we think the definition ought to
move towards.
CHAIRMAN APOSTOLAKIS: Well, let's do that.
MR. HOUGHTON: Okay.
CHAIRMAN APOSTOLAKIS: In May, that will be
Friday, May 11th, at 2:30, we are discussing we have
an hour and a half on risk-based performance
indicators. Maybe that would be a good place to
start.
MR. MAYS: George, let me interject a
little bit here if I may about that. We've been
working for a long time with the folks down in INPO
and EPIX stuff, and there was an NEI industry task
group to try to deal with this problem of different
data being collected different ways, to be reported to
five or six different entities, and what the
implications of that were. And, this is something
we've seen along the way.
What we found is that, the definition of
unavailability isn't so much the problem. The problem
tends to be the definition of the unavailability
indicator that you are using, because you know as well
as I do the unavailability definition from a classical
PRA or reliability definition is not that big of a
deal, but what happens is, when you start taking into
account other factors, such as, well, could the person
have restarted it or realigned it very quickly? You
take into account the factors, well, are we talking
about the automatic or the manual feature? You take
into account, well, are we talking about meeting its
design-based intent or its risk-significant intent?
There really isn't a real problem with getting the
amount of hours a piece of equipment is taken out of
service is not available, the performance function, we
have that data in various different varieties all over
the industry. The trick is to try to gather that
information sort of in its lowest common denominator
form and then create a kind of expert systems or smart
systems that will take that and use that information
to do the kinds of indicators that you want, depending
on who and what you want to look at.
For example, INPO wants to give credit to
plants that have more trains than they need to have
from a regulatory standpoint, so they can take those
trains out of service. So, they want to let them have
more unavailability. So, the way they do it is, they
define the unavailability indicator that doesn't
include those unavailable hours.
But, if you are doing a PRA, and you are
saying, what's the likelihood that these three trains,
instead of the two that are required, are going to
work or not, you need to know the unavailability of
all the three trains.
So, I found, and what I've seen, is that
the problem is not so much the definition of the
unavailability per se
CHAIRMAN APOSTOLAKIS: Unavailability,
though, there is a problem with the definition.
MR. MAYS: The problem we've seen, and I
think the problem we've run into in the Reactor
Oversight Process, is whether we are talking about
risk significant or design-basis function, whether we
are talking about auto or manual, whether we are
talking about how much credit you can take for being
able to realign or automatically resume.
MR. HOUGHTON: And, what system cascades to
what system.
MR. MAYS: Right.
And so, those are, in my opinion,
indicator definition problems, more so than
unavailability definition problems. And, what we've
seen is, I think, that there's a way to get to that
through common terms and definitions from a database
standpoint that will help a lot of these out.
CHAIRMAN APOSTOLAKIS: Well, the concept of
reliability is defined differently too, but, fine, I
mean, if we have a single document that explains all
these things, and comes up with a set of consistent
definitions, says that certain things are really
indicator problems rather than definitions, but right
now there isn't such a thing. So, I am all for it, to
develop something like that.
MR. HOUGHTON: As Steve was saying, there
is an industry consolidation group that's looking at
having a virtual database, which you can pluck
different data elements, common data elements from.
So, we are working that.
We are also working to meet again, I
think, about a week after your meeting with the key
players again from both industry and NRC, both
Maintenance Rule, PRA, ROP type people, so that we can
work towards these common data elements, so that we
don't have to waste our time fighting that.
CHAIRMAN APOSTOLAKIS: Okay, great. So,
you can brief us next time on the activities.
MR. HOUGHTON: Okay.
MR. BOYCE: NRR is also on that working
group, so that we also agree working towards common
definitions is the correct goal. In our most recent
public workshop for the Reactor Oversight Program last
week of March, that was one of the things we tried to
work towards, and we got a lot of input, but it's hard
to bring all those different organizations to a common
definition for many of the reasons that Steve just
said, they have different purposes for the use of the
data, but we are working on it, we are trying to get
there.
CHAIRMAN APOSTOLAKIS: All right.
MR. HOUGHTON: Shutdown indicator, I think
it's a good effort, good start. However, at this
stage I don't think it passes the simple intuitive
capable of easy use that we need for a performance
indicator. It may be that it has greater value as a
Significance Determination Process, rather than as a
performance indicator per se.
I haven't had enough time to study the
details of it, but it looks a lot more difficult than
one would put on a public web site or that one would
base
CHAIRMAN APOSTOLAKIS: Which one is this
now?
MR. HOUGHTON: The shutdown indicator.
CHAIRMAN APOSTOLAKIS: Oh.
MR. HOUGHTON: Let's see, I guess the last
Steve brought up, this is the first time I saw the
alternate approach. Certainly think on it, but I
think one of the principles we started with was, is
that aggregating the information to higher levels was
really counter to what the concept of doing the
performance indicators was, rather than have
aggregation to cornerstones or some higher level we
feel that the indicators ought to be as close to the
reality of what's going on the plant and be
actionable, such that we would say that having a SCRAM
indicator pass a threshold is actionable. You can
look at your SCRAM reduction program. Having a
particular system exceed a threshold allows you to go
focus first on that system and then do your root cause
and extent of condition, and look and see if it
applies elsewhere in the program.
So, we really feel that the program is
best left at a more granular level, in terms of
actionable level, in terms of performance indicators.
Now, that might mean a few more additional performance
indicators, we certainly would be willing to trade off
something workable in reliability as opposed to the
fault exposure, which has caused a lot of problems.
I guess my last point is, we look forward
to making the program better. We know it can be
better. I think, as I said a few minutes ago, a very
important part of these performance indicators is the
interface at the inspector level and how they view the
design basis versus the risk basis, which I think
Steve has talked also talked about, okay, which is
not it's not a trivial thing to change that mind
set, and it's the whole mind set in terms of all of
risk-based regulation versus the deterministic that we
have now.
Thanks.
CHAIRMAN APOSTOLAKIS: Thank you very much.
Maybe we can discuss now for a few minutes
what the presentation in May will consist of. Should
we go around the table and see what the members are
interested in?
Graham?
MR. LEITCH: I have a question for Steve,
if you don't mind.
CHAIRMAN APOSTOLAKIS: Sure.
MR. LEITCH: Just before we get into that.
I'm coming away with the impression that
the risk-based performance indicators are almost by
definition, by the criteria used to determine whether
you can establish a risk-based performance indicator,
they are almost by definition a lagging indicator, and
that most of the leading indicators you can't really
draw a distinct correlation between those indicators
and risk.
I guess I thought you were going to tell
us at one point an example of a reactor that got into
trouble and going to try to back fit what the risk-
based performance indicators would look like and see
if it gives you any warning, any clue of impending
difficulties.
MR. MAYS: Those are both good questions.
Let me address the leading/lagging issue.
We tried to do that a little bit in the RBPI White
Paper discussion, and maybe we weren't as clear as we
need to be. The question, when you ask yourself about
leading and lagging indicators is leading and lagging
of what? I think you can see from the way we have
broken down risk from plant risk to the things
affecting containment, CDF and health effects, and
what are the things that affect CDF, I think you can
make the case that, for example, diesel generator
reliability, although that data is lagging of diesel
generator reliability, is leading of core damage
frequency, which is leading of public risk. So,
that's the perspective I have with respect to leading
and lagging.
Now, the issue about what are the causes
of those things to happen, I don't have really good
models right now to put in a risk perspective to say
the causes applied to reliability was getting worse or
availability was getting worse was this an aspect to
the way the plant is run, managed or operated? I
don't have that information. That would be even more
leading than what I have now.
But, I think we've made the case in the
White Paper that the combination of the fact that
we're looking at things that contribute to core damage
frequency, which contributes to public risk, makes
what we are doing leading, and, in fact, the
thresholds that we've chosen for those at the levels
we've chosen for them are significantly below the
existing public risk from all causes that relate to,
for example, early fatalities, that we have a pretty
good system of making sure we have a sufficient margin
built into the system so that even if we don't have it
completely down right we are not going to have gross
enough errors to really have a big impact on public
risk as compared to what we currently have for
accidental death rate.
So, from that standpoint I feel pretty
comfortable with the leading/lagging nature of what we
have.
As we've shown here, if you want to get
more leading, or you want to hit higher level
indications, you have to do more aggregation and you
have to do more work of that nature.
The other issue, I think, with respect to
those, is that when you go back and look at how you
are getting the data and where you are setting the
thresholds, whether you are setting it at the input
point, or whether you are setting the thresholds at a
higher level, also affects what your leading or
lagging perception was.
I'm not sure I answered both of your
questions or not.
MR. LEITCH: Well, I guess the second one
had to do with, is there any evidence that if you use
the risk-based performance indicators, and tried
somehow to go back and back fit that to any of the
nasty events that we've had, is there any correlation
at all? And, I guess as long as it hasn't been core
damaging
MR. MAYS: That's one of those good
news/bad news things. The bad news is, is we can't go
back and relate this to actual core damage events, the
good news is, we can't go back and actually relate
this to core damage events.
MR. LEITCH: Yeah, right.
MR. MAYS: So, no, but one of the things
that we've looked at, and one of the things we've done
in the Reactor Oversight Process, was the question
becomes is, what constitutes poor performance, and
really when we were working on the ROP there really
wasn't a standard that you could compare against as to
what constitutes poor performance, other than things
like the watch list, or people who are on the INPO
trouble list or whatever. So, the ROP process went
back and looked at the current sets of indicators and
said, do these have good or reasonable correlation
with the plants that we have historically known to
have bad performance, and they had some data, and they
went back and did some analysis to say, these look
like they are reasonable, the bad performers tend to
fall out when we go back and look at the historical
data.
The problem from risk-based performance
indicators is, I don't have data back into that realm
to make that I have two problems, one, I don't have
data on all these things back into the realms of the
1970s, '80s and early '90s, that I can compare these
to, to see whether they map out who were the "problem
plants," and I'm not even sure that the "problem
plants" were necessarily the worst ones from the risk
perspective either.
So, I have a problem on two levels. One
is the ground truth level and one is data to compare
it with.
MR. LEITCH: Yeah.
MR. MAYS: One of the things we did do, and
have done in looking at this, is we went back and
said, well, if the ROP was reasonable maybe we can go
back and take RBPIs over a similar period and look at
what the ROP did and see if we are coming up with
similar results or significantly different results, or
if we find differences do they make sense to us from
a risk perspective? And, that's what I meant by that
"face validity" comment.
DOCTOR KRESS: You still need to pick a
period you have the data for.
MR. MAYS: You need to pick a period where
you have comparable data for both processes, and the
best we can do right now is probably the '97 to '99
time frame.
DOCTOR KRESS: Right.
MR. MAYS: We've taken a brief look at
that, and I think we found that we do a pretty
reasonable job of correlating with some of the stuff
that was in the ROP. We have more information that
they don't have, so you can't really compare what they
don't have to what we do have.
But, we did, we were able to go and look
and see where we found differences, and if it made
sense to us that the differences should exist, and the
kinds of things we found were, we found sometimes that
the RBPIs would have whites or yellows where the ROP
currently has greens or whites, and we looked at why.
And, when we looked at that, the most common reason
was because we were using plant-specific thresholds,
as opposed to generic or group thresholds.
We also found some cases where the ROP
would have whites or yellows, and we've had either
greens or whites, and we went back and looked at those
cases and what we found in those cases generally had
to do with things associated with the false exposure
time, when you take the false exposure time into
account in more of the way you would normally do it in
a risk assessment, and take into account the
reliability indicator portion, we found some of those
problems tended to go away.
But, we have gone back and looked at all
of those, and the other thing we found was the design
basis thing. If you were reporting unavailable
because it couldn't do its design basis function by
automatically starting, but was still capable of
manually starting, our indicators would indicate that
that was not a degradation as severe as the current
ROP would.
So, we've looked at that, but we haven't
published a formal side-by-side comparison like that,
and I'm not sure that there's anything we could do
anymore rigorously than a general comparison like that
in the first place.
Now, maybe if we were to go through this
and pilot some of these, what you would do is you
would run through the pilot with RBPI portions, and
you would run through and see what the comparison
would have been with the ROP, and then you go back and
ask yourself that "face validity" question again which
says, does it make sense that I'm having differences,
and do I believe that the differences are risk
significant? If you find that, you find that to be
something, as you said earlier, George, of benefit
that you want to do as a regulatory agency, then that
might be what you would do there. But, I think that's
part of looking at the stuff through the
implementation process.
DOCTOR KRESS: I hate to say this, because
it goes against my grain, but I think this is one of
those cases where your technical process itself is so
sound that I don't think you need to validate it
through real experience. I hate to say that, because
that's contrary to my usual belief.
MR. MAYS: I think you need to make the
case why what you have makes sense.
DOCTOR KRESS: Makes sense, it makes such
good sense, I don't think anymore validation than
you've already done is much worthwhile, because you
are validating against things that are not validated
themselves against reality.
MR. MAYS: It's a problem of where do you
find ground truth to compare it to.
DOCTOR KRESS: Yeah, so, you know, I
wouldn't search too much for more validation.
MR. MAYS: Well, we haven't done anymore
than that.
CHAIRMAN APOSTOLAKIS: Can we address the
issue of what to do?
DOCTOR KRESS: Of what to do?
CHAIRMAN APOSTOLAKIS: Yeah.
DOCTOR KRESS: Do you want to go around the
table?
CHAIRMAN APOSTOLAKIS: Yeah, tell us if you
DOCTOR KRESS: Well, in the first place, I
think you need to tell us in general what the process
is, what you've done, but I would also be sure to get
to the three options that you talked about, because I
think it's very important that the full committee hear
about those.
I would talk about how you dealt with
shutdown, because it's significantly different than
the normal rest of the process, and I would go just a
little bit into the validation effort, comparing it to
the '97 data to '99 data, but not a lot. I wouldn't
spend a whole lot of time on that.
And then, I would point out this yeah,
I would point out this principle you are using,
progress versus perfection, and talk about things you
may improve in the future, because I think those are
questions that are going to come up.
So, that would be my opinion, George, on
what I think.
CHAIRMAN APOSTOLAKIS: Bob?
DOCTOR UHRIG: Well, I have the sense here
that what you are doing tends to validate the system
that is in place now. Am I stating that properly,
that you are getting comparable results to what you
are getting from the inspections that are going on
MR. MAYS: From the indicators that
currently exist.
DOCTOR UHRIG: yes.
MR. MAYS: I think we are getting
comparable readings in a number of areas. We are
getting more readings where they don't have readings
now, and where we have differences we know what the
basis for the differences are.
DOCTOR UHRIG: I think that should be
indicated, not on elaboration, but simply that's the
additional thing that I would add to what Tom has
suggested here.
CHAIRMAN APOSTOLAKIS: Graham?
MR. LEITCH: This last piece you covered
after lunch, the potential of the RBPIs went by me
awful fast. Frankly, I don't really understand what
was said there. I didn't have a chance to look at it
in advance, so I need some time to brush up on that,
but I think once more through that section, just a
little more slowly, might be helpful to the committee.
CHAIRMAN APOSTOLAKIS: Good.
Mario?
DOCTOR BONACA: Yes, I pretty much agree
with the other points. Just a couple of things. One
is, you know, this is really a good effort, a good
visibility study of RBPIs, I mean, and to stress the
fact that, you know, the ROP is something different,
and, ultimately, there may be changes to that
depending on how well some of these RBPIs compare with
the existing ones.
The second point, the one that Graham
pointed to, it went very fast, and yet there is a lot
of merit on some of the alternatives, although I'm not
saying that they are going to be the likely one.
And, the third one is just a point I would
like to make, is that I think there is a more
systematic approach than it shows in the way we went
about this. I got the impression at the beginning
that you were saying, well, you know, whatever is
feasible we choose, and whatever cannot be done we
just don't go with it. I don't think you said that,
and I think that somehow I got a message, and maybe
you can communicate, that you have a systematic
approach. You are looking at containment, you are
looking at all the functions, and you do believe the
two that you could possibly identify there are already
significant of themselves and compare with the ones
you have right now whatever you have in the other side
program. I think that's important, because I didn't
get that message at the beginning.
CHAIRMAN APOSTOLAKIS: Okay, and we can
have another, I guess, of a little bit like you did
today.
MR. BOYCE: It sounds like I'm on tap.
CHAIRMAN APOSTOLAKIS: Yeah.
MR. BOYCE: Can I just
CHAIRMAN APOSTOLAKIS: Maybe over some of
the issues that were raised regarding that memo.
MR. BOYCE: yes, I think comment number
seven is still down there, although I was still hoping
Steve had addressed your concern during the course of
the conversation.
CHAIRMAN APOSTOLAKIS: The full committee
probably needs to hear it.
MR. BOYCE: I was unsuccessful.
I did want to take, if I could, just a
second just to address some of the things that I heard
here, and give you a little bit bigger picture on, I
think, where NRR is coming from.
CHAIRMAN APOSTOLAKIS: Next time, not now.
MR. BOYCE: Well, I wanted to leave you
with just a general thought, if I could.
CHAIRMAN APOSTOLAKIS: Okay, go ahead.
MR. BOYCE: And, it relates to the tone in
that memo, as you pointed out, it was cool. The
approach I think we have is, is that the view we
have is that this project is ambitious, but it's
clearly in step with the Agency's direction to become
more risk informed, and so we support that, but we
have to be very, very cautious because we can't right
whole sail into this and then have some sort of
problem come up, like the SPAR models have a fatal
flaw, the licensees do not want to submit data to EPIX
anymore, and, therefore, the performance indicators
may not be valid anymore.
So, we are very conscious of the burden
that it places on licensees and the public acceptance
part of it. That's part of our performance goals, is
to enhance public confidence. And so, those sorts of
intangibles tend to get factored into technical
decisions on should we proceed with the risk-based PI
program, and that's why we are cautiously supportive
of this program.
CHAIRMAN APOSTOLAKIS: Okay, and that's
certainly something we want to discuss with the full
committee.
And, I also want to scrutinize Appendix F,
and discuss this issue of how the aleatory
uncertainties are handled, but other than that I think
we are in good shape. We had a good presentation
today, good discussion, we appreciate it. Thank you
very much, gentlemen, all of you.
DOCTOR KRESS: Once again, a very good
confident job and good presentation.
CHAIRMAN APOSTOLAKIS: Yes.
DOCTOR KRESS: Thank you very much.
CHAIRMAN APOSTOLAKIS: Nothing less is
expected of these guys.
DOCTOR KRESS: Yes, you know, we ought to
be raising the bar every time you guys come in,
because
CHAIRMAN APOSTOLAKIS: Yeah, it's over,
it's over, the subcommittee meeting is over.
(Whereupon, the above-entitled matter was
concluded at 1:55 p.m.)
Page Last Reviewed/Updated Tuesday, August 16, 2016