United States Nuclear Regulatory Commission - Protecting People and the Environment
Home > NRC Library > Document Collections > ACRS > Meeting Schedule and Related Documents > 1999 > Reliability and Probabilistic Risk Assessment - May 5, 1999

Reliability and Probabilistic Risk Assessment and Materials and Metallurgy

                       UNITED STATES OF AMERICA
                     NUCLEAR REGULATORY COMMISSION
               ADVISORY COMMITTEE ON REACTOR SAFEGUARDS
                                  ***
             MEETING:  RELIABILITY AND PROBABILISTIC RISK
                ASSESSMENT AND MATERIALS AND METALLURGY
                                  ***
                        U.S. NRC
                        Room T-2B3
                        11545 Rockville Pike
                        Rockville, Maryland
                        Wednesday, May 5, 1999
         The subcommittee met, pursuant to notice, at 8:30 a.m.
     MEMBERS PRESENT:
         WILLIAM L. SHACK, Chairman, ACRS
         JOHN BARTON, Member, ACRS
         MARIO BONACA, Member, ACRS
         MARIO FONTANA, Member, ACRS
         THOMAS S. KRESS, Member, ACRS
         DON W. MILLER, Member, ACRS
         DANA A. POWERS, Member, ACRS
         ROBERT L. SEALE, Member, ACRS
         ROBERT E. UHRIG, Member, ACRS
         GRAHAM B. WALLIS, Member, ACRS
                         P R O C E E D I N G S
                                               [8:30 a.m.]
         DR. SHACK:  The meeting will now come to order.  This is a
     meeting of the ACRS Subcommittees on Reliability and Probabilistic Risk
     Assessment and on Materials and Metallurgy.
         I am Dr. William Shack, Chairman of the Subcommittee on
     Materials and Metallurgy.  Dr. Apostolakis is the Chairman of the
     Subcommittee on Reliability and Probabilistic Risk Assessment.
         ACRS members in attendance are John Barton, Mario Bonaca,
     Mario Fontana, Tom Kress, Donald Miller, Dana Powers, Robert Seale,
     Robert Uhrig, Graham Wallis.
         The focus of this meeting is to review those topical reports
     prepared by the Electric Power Research Institute for risk-informed
     in-service inspection.
         The subcommittees will gather information, analyze relevant
     issues of facts, and formulate proposed positions and actions as
     appropriate for deliberation by the full committee.
         Michael T. Markley is the cognizant ACRS staff engineer for
     this meeting.
         The rules for participation in today's meeting have been
     announced as part of the notice of this meeting previously published in
     the Federal Register on April 14, 1999.
         A transcript of the meeting is being kept and will be made
     available as stated in the Federal Register notice.  It is requested
     that speakers first identify themselves and speak with sufficient
     clarity and volume so that they can readily be heard.
         We have received no written comments or requests for time to
     make oral statements from members of the public.
     This, again, is the second presentation on risk-informed in-service
     inspection that we've seen.  We've previously reviewed, a few months
     ago, a proposal from the ASME on one approach to risk-informed
     inspection.  This is another alternate.     
         We will now proceed with the meeting and I will call upon
     Dr. Jeff Mitman of EPRI to begin.
         MR. MITMAN:  Good morning.  My name is Jeff Mitman and I'm
     the EPRI project manager for risk-informed ISI.  I would like to turn it
     over initially to Ham Fish of the New York Power Authority and let him
     begin the introductory remarks.
         MR. FISH:  Good morning.  I'd like to introduce the members
     of our team.
         DR. SHACK:  Please use the microphone.
         MR. FISH:  Beginning at your left, we've got Carl Fleming,
     from ERIN Engineering.  Next in, Vesna Dimitrijevic, from Duke
     Engineering Services, a major contributor to this work on the project;
     Jeff Mitman, whom you've met, our EPRI project manager; myself, from the
     New York Power Authority, where I manage the research and development
     program for the nuclear effort.
         Next, Peter Riccardella, from Structural Integrity; and,
     Glen Smith, the New York Power Authority Project Manager at the
     Fitzpatrick plant.  Sitting in the back, Pat O'Regan, working with us on
     this project, with EPRI.
         Our objectives today are to facilitate your review and
     concurrence of the EPRI risk-informed in-service inspection methodology,
     in support of a safety evaluation report expected in September of this
     year.
         The agenda, the organization of our presentation today will
     be the status of the topical report, the status of the various pilot
     plants which are participating in this, an overview of the technical
     methodology for the risk-informed inspections, and, finally, our summary
     and conclusions of what we have drawn from where we are to date and what
     we look forward to.
         At this point, I'd like to turn it over to our EPRI
     technical manager, Jeff Mitman.
         MR. MITMAN:  Thank you, Ham.  What we're going to talk about
     first is the status of the topical, just a couple slides on -- or one
     slide on the topical, one slide on the pilot projects, and the vast
     majority of the discussion will be about the technical details of the
     methodology.
         The original topical was submitted to the NRC for review in
     June of 1996.  We reviewed extensive RAIs on that, due to the
     preliminary nature of that, and from there we've been working on RAI
     responses, which went into the topical or went into the NRC staff back
     in November of last year.
         We've also spent a lot of time revising the topical.  That
     came to the staff on April 15, copies of which were also supplied to the
     ACRS.
         The topical revision has included extensive enhancements of
     the procedure and basis description, quite a bit of lessons learned
     information from the pilot programs, resolutions of questions and
     comments from the NRC, and we've also incorporated quite a bit of
     related EPRI research information.
         The expectations are for a draft SER in June of this year,
     with a final SER in September of this year.
         A quick discussion of the related ASME code cases.  There
     are two that are applicable to the EPRI methodology.  There is a third
     that the WOG Westinghouse ASME research uses.  The one that's used there
     is N-577, but we won't be talking about that today.   
         N-560 was approved or code case N-560 was approved in 1996. 
     It applies to BJ welds in the Class I system only.  It excludes socket
     welds and it allows for inspection, ten percent of the inspection
     sampling of the Class I BJ welds.
         A revision to that code case was started this year and it's
     about halfway through the approval process and our expectations are for
     approval of that this year.
         N-578 is an alternate code case that also uses the EPRI
     methodology.  It was approved in 1997.  It applies to Class I, II, III
     piping and may include non-code piping.
         Inspection criteria there are 25 percent of the high risk
     welds, ten percent of the medium risk welds, and zero percent of the low
     risk welds.
         Likewise, a revision was started on that this year.  It's
     about halfway through the approval process and, again, we're expecting
     approval this year.
         The differences are a comparison of the two code cases, the
     risk assessment process is the same.  The consequence analysis is the
     same.  The degradation analysis is the same.  The distinction is in the
     element selection process.
         As noted on the slide, for the N-560, we do ten percent of
     the BJ welds.  For N-578, we do 25 percent of the high, ten percent of
     the medium, and zero percent of the low.  Both methodologies yield
     comparable risk results.
         DR. WALLIS:  Is it ever good to do zero percent of anything?
         MR. MITMAN:  The way we pick welds, and we'll get into that
     a little bit later, you'll see why we feel it's okay to pick zero
     percent of the low risk welds.
         DR. WALLIS:  In case you were wrong, you might want to have
     at least some sample from what you called low risk.
         MR. MITMAN:  As I said, let's go on with the degradation and
     consequence analysis and you'll see where that comes from.
         I want to talk about the pilot plants at this point.  There
     has been quite a bit of work that's been done already and the work
     continues.  As I said, there are two code cases, two different scopes,
     if you will, that we've applied the methodology to.  Vermont Yankee was
     the first plant to use the methodology and we received the first
     industry SER on that back in November of last year.  That's an N-560
     case and it was done on Class I piping only.
         ANO-2 is Combustion Engineering PWR.  We did a full plant
     evaluation of that under code case N-578.  Likewise, Entergy received an
     SER approving that in December of last year.
         ANO-1 is an N-560 application.  It was submitted last year
     and we're currently finalizing responses to RAIs that the staff sent us
     in the last month or so.
         Fitzpatrick is other full plant application, GE boiling
     water reactor.  We did 14 systems there and we're in final preparation
     of the submittal for that.
         The following plants are all Class I only applications and
     they're in various stages of completion.  Braidwood is about 80 percent
     complete, South Texas 70-75 percent complete, River Bend and Waterford
     haven't started yet, but we expect to do those later this year.
         At this point, I'd like to start to talk about the
     methodology itself.  We'll go through the -- initially, I just want to
     go through the process itself at a very high level and then we'll take a
     closer look at each of the steps.
         The first step is to determine the scope.  Do you want to do
     BJ welds only?  Do you want to do full plant evaluation?  How many
     systems do you want to include?  Where do you want to draw the system
     boundaries and stop?  That's determination of the scope.
         Then there are two independent analyses that need to be
     performed, the consequence analysis and the damage mechanism analysis. 
     They can be done in either order, they can be done in parallel.  The
     output of both of those is -- neither of those output is input to the
     other one.  So they can be done in parallel.  And probably 50 or 60
     percent of the total work is involved in the two analyses, the
     consequence analysis and the damage mechanism analysis.
         Once that's completed, we do a second check looking at plant
     history, what we call a service review of the plant.  From there, we do
     segment risk categorization, where we take the output from the
     consequence analysis and the damage mechanism analysis, put those
     together and categorize our welds into high, medium and low risk
     regions.
         From there, we decide which specific welds we want to --
     welds or elements we want to inspect and we also determine the
     inspection methods, what kind of NDE we want to perform.
         Afterwards, we go after the -- we do a risk impact
     assessment and then we finalize the program.  There is a feedback --
     there are two feedback loops, one coming out of the risk assessment,
     whereby if we decide that the risk impacts are unacceptable, we can go
     back and add welds, re-allocate elements, do what's necessary to ensure
     that we have acceptable risk impacts.
         Also, there is a feedback loop at the end of the project,
     where we do long-term performance monitoring, whereby we watch what's
     going on in the plant, what's going on in the industry, and, if
     necessary and as appropriate, we modify -- go back and modify the
     various analyses to make sure that we're looking in the right places and
     we're looking for the right things.
         The heart to the process is the risk matrix and I will show
     that in the next slide.  But before I bring up the matrix itself, I'd
     like to talk a little bit about the two -- some of the concepts behind
     the risk matrix.
         On one axis, we have the consequence evaluation, where we're
     looking at core damage in large early release and the input to that is
     both probabilistic insights and deterministic insights.  There are four
     impact groups that we're considering; initiating events, degradation of
     systems and/or trains, degradation of the containment, or combinations
     of the three above.  The output for that is a consequence ranking of
     high, medium, low or none.
         On the other side, we have the degradation analysis, where
     we're looking at pipe rupture potential and here we're looking at
     degradation mechanisms, which damage mechanisms apply or are potentially
     applicable to each piping segment.
         We do extensive service experience evaluations there and the
     output for that is a rupture potential ranking of high, medium and low.
         That goes into the risk evaluation matrix and here we have
     the matrix, we have the consequences across the X axis, degradation
     across the Y.  We end up with three consequence rankings, high, medium,
     low and none, and, again, the degradation analysis of high, medium and
     low.  Then we divide into three risk regions high, medium and none,
     again.
         DR. KRESS:  Your conditional probabilities, I assume, are
     weighted by the initiating events or added up for all initiating events.
         MS. DIMITRIJEVIC:  The conditional pipe failure.  So given
     the -- if we have a pipe failure, what would be the conditional core
     damage.
         DR. KRESS:  You do a conditional for each pipe segment, for
     example. 
         MS. DIMITRIJEVIC:  Right.
         DR. SHACK:  But then he has a frequency of challenge that
     determines whether it's high or medium.
         DR. KRESS:  So the frequency does enter into it.
         MR. MITMAN:  Frequency does enter into that classification.
         MS. DIMITRIJEVIC:  Frequency of challenge for the system,
     yes.  But actually frequency enters through the pipe failure frequency.
         DR. APOSTOLAKIS:  It's similar to the component, is it not?
         MS. DIMITRIJEVIC:  Yes.  If there is no need for surrogate
     component, because we are looking at actual pipe.
         DR. APOSTOLAKIS:  Jeff, what was the logic behind this
     matrix?  I mean, this is a decision matrix here.  You've given the
     categorization of the consequences of a pipe break and the degradation
     category, you decide that something is low, medium.  What was the logic
     of that?  For example, how did you decide that if the degradation
     category is high and the consequences are low, then it becomes a medium.
         MR. MITMAN:  The logic behind the matrix concept itself was
     driven by a desire to back away from having to do detailed probabilistic
     calculations or detailed calculations of the damage mechanisms.  We
     wanted to come up with a process that was easier to apply, yet was
     accurate and gave valid results.
         So that led us to the concept of a matrix where we can bend
     things easily.
         DR. APOSTOLAKIS:  Well, we can talk about easy easy is, but
     this -- the purpose of this risk evaluation, this matrix is to lead you
     to a decision how to treat a pipe segment, is that correct?
         MR. MITMAN:  Right.
         DR. APOSTOLAKIS:  So given now that you have done the
     consequence evaluation and all that, whether that's easy or not is a
     separate issue, the logic of this decision-making process, what was it? 
     I mean, why is medium degradation category and low consequence category
     leads to a low category six segment?
         MR. MITMAN:  Vesna?
         MS. DIMITRIJEVIC:  What we tried to decide in this category,
     we said, okay, well, let's assume the old pipe in the plant breaks and
     CCBP.  If they're going to go in danger of ten-to-the-minus-six, which
     is the danger where you're actually making the decision, if we're
     expecting that we can be above the ten-to-the-minus-six, because those
     CCBPs have the ranges between ten-to-the-minus-four and above, then we
     said this is a high.
         If we say we're expecting the piping in the plant breaks,
     it's still be under ten-to-the-minus-six, and then for this region
     between, we say, okay, between ten-to-the-minus-six, eight, and
     ten-to-the-minus six, if we don't really -- we're going to put the
     medium because we were not sure really which way because of all the
     answers in this.
         So we did this in order to cover our certainty ranges.
         DR. APOSTOLAKIS:  But that was the uncertainty on the
     consequences?
         MS. DIMITRIJEVIC:  Both, because when we estimate the pipe
     failure ranges, also, based on the presence of the degradation
     mechanism.  So this is built to cover uncertainties.  If we expect,
     based on our knowledge, that potential for the total CDF to be above
     ten-to-the-minus-six, we put it in high.
         DR. APOSTOLAKIS:  And then what does it mean?  If a segment
     is medium, what do you do to it?
         MS. DIMITRIJEVIC:  You're going to see this through the --
     it's coming.  We will have a different selection for this.  Carl, would
     you like to add something?  Did I miss something?
         MR. FLEMING:  Carl Fleming, from ERIN.  I think that part of
     George's questions could be answered that the logic is really risk
     equals frequency times consequence.  And if this were done in a fully
     quantitative way, we'd simply take the frequency of the pipe rupture,
     multiply it by the conditional probability of core damage, and then get
     the risk.
         In the EPRI risk matrix approach, uncertainties are
     addressed by looking at this in a very course, discreet way.  Instead of
     numerically quantifying the conditional core damage probability in each
     case and numerically quantifying the frequency of pipe rupture, they're
     put into broad categories.
         So the logic of high, medium and low stems from the concept
     of risk equals frequency times consequence.
         DR. APOSTOLAKIS:  I understand that.  But my question is
     this; presumably, there is a certain frequency of inspections and so on
     that corresponds to medium, high of these entries.
         MS. DIMITRIJEVIC:  Of risk.
         DR. APOSTOLAKIS:  Of risk.  Now, you don't have a
     quantitative estimate of the impact --
         MS. DIMITRIJEVIC:  Of inspection.
         DR. APOSTOLAKIS:  -- of the inspection on these two
     categories.
         MR. FLEMING:  No.
         DR. APOSTOLAKIS:  So this is purely judgment now.  You are
     hoping that --
         MS. DIMITRIJEVIC:  We're estimating total risk from those
     locations.
         DR. APOSTOLAKIS:  Yes, but --
         MS. DIMITRIJEVIC:  So let's say that not one of those
     locations is ever inspected, this is the total risk.
         DR. APOSTOLAKIS:  That's right.  So now the question is, if
     the degradation category is medium and the consequence category is
     medium, then you say medium.
         MS. DIMITRIJEVIC:  Right.
         DR. APOSTOLAKIS:  But you don't know now what the impact of
     what you're doing on risk is.  You just guess that this would be a
     reasonable thing to do.
         MS. DIMITRIJEVIC:  At this moment, the only thing which we
     are trying to do is to divide locations -- I mean, valves and the
     piping, based on the risk significance.  We are not looking at
     inspections yet at all.
         DR. APOSTOLAKIS:  Yes, but in order to decide that something
     is medium, high or low, you have to have some idea of what the impact
     is.
         MS. DIMITRIJEVIC:  Yes.  We have some idea of what
     probability of failure is and what is the consequence.
         DR. APOSTOLAKIS:  Without inspection.
         MS. DIMITRIJEVIC:  Without.
         DR. APOSTOLAKIS:  And you don't have any idea what happens
     to these probabilities with inspection.
         MS. DIMITRIJEVIC:  No.
         DR. APOSTOLAKIS:  No.
         MS. DIMITRIJEVIC:  Neither do we and neither does anybody. 
     I mean, this is a very difficult thing to estimate.  So basically, in
     this moment, we are not really trying.  When we try to calculate back to
     risk, we will try to do that.
         DR. APOSTOLAKIS:  So there is a judgment here that by doing
     high or medium, you know this is good enough.  That's really what it
     comes down to.
         DR. SEALE:  Are you going to show us a little later that all
     three mediums are comparable in risk and that all of the mediums are
     discernibly greater risk than low?
         MR. MITMAN:  Yes.
         DR. SEALE:  And conversely, mediums and high.
         MR. MITMAN:  Right.  We'll go into that a little bit later.
         DR. SEALE:  Okay.
         MR. MITMAN:  You wanted to add something?
         MR. RICCARDELLA:  Let me just make a comment.  There was a
     judgment, as you said, George, made in the beginning, when we set up
     these categories, and that judgment was based on looking at what we
     currently do under ASME code requirements and coming up with something
     which we think is at least equivalent or better.
         There was a judgment at that end.  But as you will see as we
     get on with the presentation, we have now implemented this in a number
     of pilot studies and when we complete those pilot studies, we've done
     delta risk calculations and quantified the difference between what we
     were doing before and what we're doing after.
         DR. APOSTOLAKIS:  Including the impact of inspections.
         MR. MITMAN:  Yes.
         MS. DIMITRIJEVIC:  Yes.
         MR. RICCARDELLA:  Yes.
         DR. APOSTOLAKIS:  And it's in this report?
         MS. DIMITRIJEVIC:  It's Section 3.7.
         MR. DINSMORE:  This is Stephen Dinsmore, from the staff. 
     I'd like to add something.  When EPRI first came in, they just had this
     matrix and ever since then, we've been pushing them to at least estimate
     the delta risk.  So at the end, they've come up with this methodology to
     estimate the delta risk.  So they use this process to select and to
     guide their inspection selections and then they use this delta risk
     calculation to make sure that things are okay.
         DR. SHACK:  Just a sort of related question.  You have sort
     of a conditional or damage probability associated with these things,
     that you don't really calculate.  You calculate from the guidelines that
     you've given based sort of on the number of backup trains and you
     associate a numerical value of the conditional core damage probability
     with those.
         How robust is that?  I mean, how many PRAs did you look at
     before you decided you could really bend them this way and have a
     conditional core damage probability that was in the range that you were
     assigning to each of those categories?
         MS. DIMITRIJEVIC:  It's plant-specific, so we look in the
     specific plant PRA when we do this.  Every plant, different events will
     fit into different categories.  It's not really so diverse.  You will
     see also the pipes which result in the LOCA lead to the high
     consequences and things like that.
         But this matrix is generally applicable to the methodology,
     but you apply it on specific plants, you look at specific events and
     specific conditional core damage probabilities.
         DR. SHACK:  Okay.  Maybe I misunderstood it.  I thought when
     you looked at the plant specific, you were really deciding, with this
     particular event, you had 2.5 backup trains.
         MS. DIMITRIJEVIC:  Yes.
         DR. SHACK:  And that was the plant-specific thing, but you
     really didn't calculate the conditional core damage probability.
         MS. DIMITRIJEVIC:  But that translates --
         DR. SHACK:  Yes.  That translates, but not from that plant's
     PSA.  Isn't that sort of arbitrarily assigned?
         MS. DIMITRIJEVIC:  How many backup trains is plant-specific. 
     What you can count as a train is plant-specific.  If some plant keeps a
     train with a probability of failure that's ten-to-the-minus-two, that
     counts as a train.  If that is ten-to-the-minus-one --
         DR. SHACK:  It doesn't count.
         MS. DIMITRIJEVIC:  -- it doesn't count as half-train.  So
     what do you count as a train is plant-specific.  So basically, it's a
     completely equivalent if as you were running the PRA.  Basically, this
     table was validated with a lot of PRA runs and always gives the same
     category, because it's basically principles of PRA are concentrated in
     this table, because this is what PRA sequences are, how many times you
     call for something, how many backup trains you have, and what are their
     values.
         DR. SHACK:  Okay.  And you say lots of PRA runs, and I guess
     my question is, is that lots of PRA runs on two plants or how many
     plants did we look at?
         MS. DIMITRIJEVIC:  We looked at four plants and it depended
     on the plant as to what was the number, between five and 20 PRA runs per
     plant.  We didn't have any disagreements.
         DR. APOSTOLAKIS:  Yes, I went through 3.7.  You say that
     while there can be a bounding analysis or if it doesn't work, then you
     do a more realistic quantitative analysis, but then you go on and say
     that this can be done, but then there is a whole paragraph why it cannot
     be done, because it's very difficult, and I agree.
         Then in the examples that you have in here somewhere from
     the actual plants, I didn't see the impact, the quantitative estimate of
     the impact of the in-service inspection.
         So is that really something that you have done or something
     that could be done in principal, but it's not really done?
         MR. MITMAN:  On VY ANO-2, ANO-1, we've done realistic
     analysis on some of the systems, as required, where we couldn't show
     qualitatively that risk was improving.
         On ANO-2, we've also done some Markovian analysis that
     further, in more detail, looked at several systems in the ANO-2 plant. 
     So those analyses have been done.
         DR. SHACK:  And when you say you did that, you actually
     computed the conditional core damage probability for that plant from the
     PSA, not from the binding value of the CCDP that you have.
         MR. MITMAN:  Vesna, correct me if I'm wrong, but if I
     remember correctly, we actually looked at each of the pipe segments on
     VY and ANO-2, calculated the conditional core damage probability, and
     then used those to validate the delta CDF calculations.
         MS. DIMITRIJEVIC:  But we do also use a bounding if you're
     in the medium range, because we can have a lot of else with the
     different CCDPs in that train and we used the upper value.  In every
     range, we will use the upper value in bounding, because then it's
     considered.  So we don't go to every specific valve that refers to CCDP. 
     We will take the highest value for this range.
         DR. APOSTOLAKIS:  So did you see a significant input?
         MS. DIMITRIJEVIC:  Of?
         DR. APOSTOLAKIS:  Of the inspection?
         MR. MITMAN:  No.
         MS. DIMITRIJEVIC:  We didn't see it.  We just saw
     ten-to-the-minus-six and ten-to-the-minus-nine.
         DR. APOSTOLAKIS:  So inspection did not help in any way.
         MR. MITMAN:  No.
         DR. APOSTOLAKIS:  So why are we doing all this?
         MR. RICCARDELLA:  It started out as negligible.
         DR. APOSTOLAKIS:  Why are we doing all this?
         MS. DIMITRIJEVIC:  We asked ourselves from the beginning
     this question a lot of times and basically the answer that I got most
     often is that so we take a look and we don't miss something.  But
     risk-wise, there is no justification for doing it.
         DR. KRESS:  It's defense-in-depth.
         DR. APOSTOLAKIS:  So in-service inspection can be dropped.
         DR. KRESS:  It's defense-in-depth.  There is large
     uncertainty in this area.
         DR. APOSTOLAKIS:  That's very good.
         MR. MITMAN:  They have a value impact analysis.
         DR. APOSTOLAKIS:  In the report, in several places, you say
     that the service experience to date provides a strong indication that
     the frequency of pipe ruptures is only very weakly correlated to the
     inspection processes.  So that's a general statement then.  Even if you
     use this more sophisticated approach, that's still true.
         MR. MITMAN:  That's true.
         MS. DIMITRIJEVIC:  Very true.
         MR. ALI:  This is Syed Ali from the staff.  I just wanted to
     clarify one thing.  When you say that the experience is that the
     in-service inspection has real flaws, that's basically related to the
     Section 11, ASME Section 11 inspections.
         Based on the actual experience that the plants had and the
     regulations actually followed, there are several augmented inspection
     programs, for example, for IGSCC and FAC, which have indeed found flaws.
         DR. APOSTOLAKIS:  Well, originally, I thought myself that
     this statement referred to experience, so ASME Code Section 11 didn't
     help us.  But now they're saying that even with this, they don't expect
     to see an impact, which is a very interesting conclusion.
         MR. MITMAN:  The methodology is there to help you decide
     where to do the ASME Section 11 inspections.  Now, we continue and we
     work with the augmented programs and there is a lot of value in
     particularly the FAC programs and the IGSCC programs.  But the ASME
     Section 11 required inspections, I think it's safe to say, we feel has
     very little impact on the safety of the plant.
         DR. APOSTOLAKIS:  But you are also saying that this thing
     does not have a significant impact, not just -- I want to make that
     clear, in my mind.  Not just the current Section 11.  But even if you do
     this methodology, which is more sophisticated, includes additional
     failure modes, it's hard to see an impact.
         MR. MITMAN:  That's correct.  Pat, you wanted to add
     something?
         MR. O'REGAN:  Pat O'Regan, from EPRI.  I just want to
     emphasize what Jeff said.  This methodology takes credit for and
     integrates the augmented inspection programs.  As Syed mentioned, IGSCC
     and FAC do actually have a substantial safety impact and we are taking
     credit for that.
         So the delta you're seeing is actually on the Section 11
     portion, not crediting the augmented programs.
         MR. MITMAN:  And Carl?
         MR. FLEMING:  George, with regard to the insights from
     service experience that back up this conclusion about the benefit of
     inspection, the underlying reason for that statement or that conclusion
     is that when we look at service experience, we find that we find piping,
     some small number of pipe failures and ruptures due to two general
     classes of failure mechanisms.
         The degradation mechanisms, some of which are actually
     amenable to inspection, and also we have loading conditions, like water
     hammer, vibration fatigue, over-pressurization, frozen pipes, things
     like that.  And if you look at the whole piping service experience, only
     a fraction of those failure mechanisms provide you enough warning time
     that a ten-year interval inspection process is going to give you an
     opportunity to prevent the failure.
         So that's one of the reasons for that conclusion.  A second
     reason --
         DR. APOSTOLAKIS:  Ten years is too long, you're saying?
         MR. O'REGAN:  No.  Some mechanisms don't give you warning
     times such that if you did an inspection today, you could prevent a
     failure tomorrow.  Then another basis for this insight is that a large
     fraction of the Class I pipe and Class II and III pipe out there isn't
     being subject to inspection.  The inspection programs that we do have
     from Section 11 are only looking at a small fraction, a small sample of
     the existing welds.
         So when we go out and look at the service experience, you
     find the vast majority of the pipe that produces the service experience
     has never been inspected for ASME Section 11 purposes, because the
     percentages are 25 percent for Class I, 7.5 percent for Class II, and so
     forth.
         So these are some of the underlying reasons to support that
     conclusion.
         DR. APOSTOLAKIS:  So just to close this.  In your opinion,
     is this a good idea or not?
         MS. DIMITRIJEVIC:  Is it a good idea to do these?
         DR. APOSTOLAKIS:  To inspect.  Not the EPRI approach.  I
     know this is a good idea.
         MS. DIMITRIJEVIC:  They don't inspect at all in the Section
     11.
         DR. APOSTOLAKIS:  What?
         MS. DIMITRIJEVIC:  Is your question, is it would they be
     able to inspect at all?
         DR. APOSTOLAKIS:  Yes.
         MS. DIMITRIJEVIC:  Well, we have a different opinion.  My
     opinion is no, but basically the people who -- because I am a PRA
     person.  The people who are actually engineers and who work on
     inspection, they think it's a good idea to take a look at what is going
     on.
         DR. APOSTOLAKIS:  Now, even for a PRA person, though, maybe
     your conclusion is based on the fact that you looked at core damage
     frequency and large early release frequency.  I mean, if one has other
     objectives, like I don't want to see any pipe breaks, that's my
     objective, then perhaps it would be a different conclusion, because we
     see that by the way the new oversight program, that the objectives now
     are spread all over the cornerstones.
         Would that change your conclusion?
         MS. DIMITRIJEVIC:  No.
         DR. APOSTOLAKIS:  Still it would not.
         MS. DIMITRIJEVIC:  No, because what Carl just mentioned, we
     don't claim the risk from pipe breaks is equal to zero or small.  What
     we are claiming what inspection does, that is negligible.  And where
     you're going to see the breaks, like vibration and everything,
     inspection doesn't do a thing.
         MR. MITMAN:  Pat, did you want to add something here?
         MR. O'REGAN:  Yes, just a point of clarification.  Pat
     O'Regan, from EPRI.  When we're talking inspections here, at least the
     discussion, it was just focusing on volumetric examinations.  We are
     still doing and still recommending doing leak testing and what service
     experience has showed us is that's where we find most of our leaks, not
     from the volumetric exams, but the leak testing.
         And even on a low risk location, that's still recommended to
     be done.
         DR. APOSTOLAKIS:  So if we were completely rational, we
     would keep doing those inspections, but we would drop this other stuff.
         MR. MITMAN:  I think rational, you'd probably want to go
     back and, at the very beginning of your design process, beef up your
     leak detection capability.
         DR. APOSTOLAKIS:  Okay.
         DR. SHACK:  I guess I'm not convinced that that would be
     true if you include the augmented inspections; that is, flow assisted
     corrosion and stress --
         MS. DIMITRIJEVIC:  No, we don't change anything.
         DR. SHACK:  You're from a PWR, where maybe, in your primary
     system, that probably is true.  That's a rather broad conclusion.
         MR. MITMAN:  You're absolutely right.  You want to continue
     to do the augmented programs.  It's very clear with the FAC program --
         DR. SHACK:  The augmented programs are, in many ways, much
     like this; that is, you're looking where you know you have a problem.
         MR. MITMAN:  Exactly.    
         DR. SHACK:  That's why it's augmented.
         MR. MITMAN:  And that's why this methodology blends very
     nicely with the currently existing augmenting programs, because of the
     similarities in the approach, where you're looking where you expect to
     find problems.
         DR. SHACK:  Without some random selection.
         MR. MITMAN:  Right.
         MR. RICCARDELLA:  Jeff, let me make a point on that. 
     Historically, what happened, and I think I said this like two years ago
     in this same group, is we had these ASME Section 11 requirements to do
     25 percent of Class I, ten percent of Class II, which we came up with
     those 20 years ago when we had very little operating experience.
         So as we went through the years and did that, those things
     didn't find much, but we did find problems and where we found problems,
     we implemented augmented programs.
         But what happened was the 25 percent type stuff never went
     away.  So we were inspecting more, but we're doing the augmented, plus
     we're doing this additional 25 percent.
         The attempt here is to integrate those into a single
     rational program that looks at the areas where we expect problems and
     doesn't look at the areas where we don't expect problems.
         MR. SULLIVAN:  This is Ted Sullivan.  I wonder if I could
     add a couple of points.  One of the reasons why I think traditionally
     in-service inspection has not found problems is because the inspection
     methods were not qualified until after a problem revealed itself.
         The Appendix 8 approaches, which we plan to adopt in the
     next revision of 50.55(a), should go a long way in resolving that.
         Another brief comment having to do with if we were
     completely rational.  I think being completely rational, I think we want
     to keep in mind defense-in-depth and that's part of the reason why the
     staff really wouldn't agree with a position that we should just look at
     this purely from a risk perspective and maybe only focus on the areas of
     augmented inspection.
         One final point.  I'm not sure that this came out explicitly
     in the presentations, I guess, last week to the ACRS on license renewal,
     but to some extent, license renewal is taking credit for Section 11 in
     the sense that there is a potential for degradation mechanisms to occur
     over the long haul and the inspection methods are supposed to be geared
     to using appropriately qualified techniques to be able to detect those
     degradation mechanisms should they arise.  That's another reason why I
     don't think we should take a view too strongly that we should just look
     at this from a risk perspective and possibly just rely on the numbers to
     say we don't need in-service inspection.
         DR. POWERS:  Could I understand better why you want to
     appeal the defense-in-depth here?  Why isn't this just completely and
     adequately resolved on a probabilistic basis?
         MR. SULLIVAN:  I guess there might be others that might care
     to answer from the staff, but I would say I don't think that we want to
     see failures occurring in piping just because the risk numbers say that
     we can handle those failures.
         DR. POWERS:  Thou shalt not have pipe failures.
         MR. SULLIVAN:  I think that's the whole defense-in-depth
     approach, which may sound circular.
         DR. SHACK:  The general design criteria call for the
     integrity of the reactor coolant system.
         DR. POWERS:  Yes, but it does not say thou shalt never have
     a pipe failure.
         DR. SHACK:  It's embedded in the whole regulatory system.
         MR. SULLIVAN:  There are places in the GDC that say
     extremely low probability.
         DR. POWERS:  I don't know, ten-to-the-minus-four,
     ten-to-the-minus-five is certainly what I would call extremely low.
         DR. APOSTOLAKIS:  But, also, let's make it clear that their
     conclusion is not based on the numbers alone.  I mean, it's the whole
     report and analysis that supports all that and the judgments and so on. 
     It's not that they got the number of ten-to-the-minus-five and say, my
     god, you know, this is insignificant.  It's the whole process that has
     led --
         DR. POWERS:  I will grant that.  What I want to understand
     is why -- I mean, very specific here.  We say we should look at and
     think about defense-in-depth, a very important concept in reactor
     safety.  Yet, this seems to be a topic that's susceptible to
     probabilistic analysis and I'm trying to understand why.
         MR. DINSMORE:  This is Steve Dinsmore, from the staff.  I
     guess I have to review a lot of this probabilistic analysis that come in
     regarding this stuff and the numbers are actually quite uncertain and
     one of the problems is when they say from experience data, they haven't
     seen much, there isn't that much experience to get
     ten-to-the-minus-seven, ten-to-the-minus-eight, statistical numbers from
     the experience data.  So they have to do some type of interpretations
     and Bayesian updates and engineering judgments to get the numbers which
     they are using to support this process.
         And even though you could say that the numbers indicate that
     if we stopped inspecting, there would be no great increase in CDF or
     LERF, the numbers that they're using, again, there is a lot of judgment
     in those numbers and there is not enough data to support them really at
     a statistical pace.
         DR. APOSTOLAKIS:  There are two things that come to mind
     from this comment.  First of all, I think if you want to defend
     defense-in-depth in this context, you have to tell us where the big
     uncertainties are that you just mentioned that perhaps could invalidate
     this conclusion.
         But second, I will come back to Carl's comment, that for
     most of these failures, you don't have enough warning time to catch them
     by inspection, which doesn't sound like a PRA argument, to me.
         That's really the physics of the problem.  So I inspect just
     to feel better.  That's really what is going on.  In the name of
     defense-in-depth.
         MR. ALI:  Syed Ali, from the staff.  Carl, correct me if I'm
     wrong, but I think the kind of mechanism that he was referring to that
     do not give you warning are loading type mechanisms, such as water
     hammer or fatigue.  But the mechanisms which are time-related,
     degradation type mechanisms, they do give you time and augmented, like
     IGSCC, erosion/corrosion.
         DR. APOSTOLAKIS:  But we're not referring to all the
     augmented programs.  Just this one.  We're not saying that we should
     drop other programs, right?
         MR. MITMAN:  First of all, I want to make something clear. 
     EPRI has not recommended that we drop all Section 11 inspections.
         DR. APOSTOLAKIS:  That's fine.  Other people might.
         MR. MITMAN:  You asked an opinion --
         MR. BARTON:  We'll deal with them when they come before us.
         MR. MITMAN:  You asked an opinion --
         DR. APOSTOLAKIS:  Let me tell you what's going on, because
     probably you think this is a discussion that -- we are facing a major
     problem now as to what the role of defense-in-depth should be in a
     risk-informed regulatory system.  That's why you're getting all this.
         That makes more sense now to you probably.
         MR. DINSMORE:  Could I just quickly respond to something? 
     When you asked me if the uncertainty doesn't or why we can't accept the
     results because of uncertainty, I would say we can accept the results of
     what EPRI is doing in spite of the uncertainty.  It's because of the
     uncertainty that we are -- that we can accept what they're trying to do
     and not say, well, you have to do even more or even less.
         DR. APOSTOLAKIS:  So what you're saying essentially is that
     their numerical assessment of the impact of this, which led them to the
     conclusion that it doesn't matter, is really not correct.
         MR. DINSMORE:  Probably not supportable.
         DR. SEALE:  They're not sure about that.
         MR. DINSMORE:  Not supportable from the statistical data.
         DR. APOSTOLAKIS:  But anyway, I think we've covered this
     enough.  At least for me.
         DR. POWERS:  I still haven't gotten an answer.  I may never
     get an answer.
         DR. APOSTOLAKIS:  That's my bet.
         MR. MITMAN:  I've put up a slide that's not in your package
     and it looks at each of the damage mechanisms and --
         DR. APOSTOLAKIS:  It's in the report, though.
         MR. MITMAN:  It's in the report, if I remember correctly.
         DR. APOSTOLAKIS:  Yes.
         MR. MITMAN:  But it's not in the presentation.  It shows
     each of the damage mechanisms and our calculated rupture frequency per
     the damage mechanisms.  There are two categories of those damage
     mechanisms, the dark ones and the light ones.  The light ones are those
     that we feel are not amenable to inspection, things like they include
     vibration, water hammer, unknown and other causes.
         DR. WALLIS:  Just for clarification.  Your rupture frequency
     is based on how many reactors?
         MR. MITMAN:  It's based on over 2,000 operating years of
     reactor.
         DR. WALLIS:  This is the total rupture frequency of all the
     reactor population.
         MR. MITMAN:  In the United States.  Yes.
         DR. WALLIS:  Otherwise, it looks pretty lousy.  If you
     multiply it by a hundred, it gets scary.
         MR. MITMAN:  Right.
         MR. FLEMING:  But these are on a per reactor year basis. 
     These are on a per reactor year basis.
         DR. WALLIS:  Per reactor year?
         MR. FLEMING:  Yes.
         DR. WALLIS:  No, it's not.  It's per a hundred reactor
     years.
         DR. APOSTOLAKIS:  Anywhere in the population.
         DR. WALLIS:  Anywhere in the population.  So I don't
     multiply by a hundred.
         DR. APOSTOLAKIS:  No, you shouldn't.  That's correct.  This
     is per calendar year anywhere in the United States.
         MR. FLEMING:  No.  This is per reactor year.
         DR. WALLIS:  So I multiply it by a hundred.
         MR. FLEMING:  This is for each reactor, on the average for
     each reactor, but it covers the entire plant.  It covers all the piping
     systems in the entire plant.  Only a small portion of this is in the
     safety-related systems.
         DR. APOSTOLAKIS:  You're right, because you had 1,511 total
     number.
         MR. FLEMING:  That's right.
         DR. APOSTOLAKIS:  Now it makes sense.
         DR. WALLIS:  So this is per reactor year.
         MR. FLEMING:  These are per reactor year, yes.  So you do
     multiply by a hundred.
         DR. WALLIS:  So I do multiply.
         MR. FLEMING:  Yes.
         DR. WALLIS:  Thank you.
         DR. APOSTOLAKIS:  But this is all pipes.
         MS. DIMITRIJEVIC:  All pipes, all sizes.
         DR. APOSTOLAKIS:  By the way, this number, which is in
     several places in the report, ultimately is not used.  Is that correct? 
     The ten-to-the-minus-two, I was looking very hard to find a place where
     you actually use it.  It's just an indication, but you don't really use
     it.
         MR. MITMAN:  It's part of the design basis of the
     degradation categorization and it is available to us if we need it to do
     the risk impact analysis.
         DR. APOSTOLAKIS:  Well, you can't really use it, though,
     because this is anywhere in the plant.  I mean, if you go to your pages
     3-1 and 3-2, then you have to specialize it.
         MR. MITMAN:  You're right.
         MR. FLEMING:  It requires further analysis.
         DR. APOSTOLAKIS:  Further analysis, right.  So right now
     it's not used.  In your calculations, ultimately, the
     ten-to-the-minus-two from there is not used and I think that should be
     made clear in the report.
         MR. FLEMING:  At this level, they are not used.
         DR. APOSTOLAKIS:  It is not.
         MR. FLEMING:  This is just a presentation of general
     experience.
         DR. APOSTOLAKIS:  I understand.  It took me a while to
     figure it out.  It's not stated in the report that it's not used.
         MR. FLEMING:  But I think it's worthwhile -- it's just an
     intermediate step along the way towards breaking the data down so that
     we could confirm that these high, medium and low categories that were
     developed on the deterministic degradation mechanism basis do correlate
     to order of magnitude estimates of pipe rupture frequency, and that's --
         DR. APOSTOLAKIS:  I think in your viewgraphs you have the
     equations.  So maybe you can go to them right now.  Equations 3-1 and
     3-2, on page 3-8.  You didn't expect that, Jeff?
         MR. MITMAN:  We're ready.
         DR. APOSTOLAKIS:  Okay.  It says pipe break frequency, is
     that what it is, PBF?
         MR. FLEMING:  Yes.
         DR. APOSTOLAKIS:  This is page 3-8.  This is what you would
     -- you would need this number, which is really now segment-specific, to
     do the calculations.
         MR. FLEMING:  Yes.
         DR. APOSTOLAKIS:  Do you actually get a number like that? 
     In principal, I know you can get something.
         MR. FLEMING:  Yes.
         CHAIRMAN JACKSON:  You do?
         MR. FLEMING:  We had developed a number, segment-specific.
         DR. APOSTOLAKIS:  And it's discussed in the report how you
     do that?
         MR. MITMAN:  It's not discussed in this report.  It's
     discussed in a supporting document.
         DR. APOSTOLAKIS:  Can I have a copy of the supporting
     document?
         MR. MITMAN:  The staff already has a copy of it.
         DR. APOSTOLAKIS:  Okay.  We might have it already.
         MR. MARKLEY:  Which document are you talking about, Jeff?
         MR. MITMAN:  TR111880, which is in the references.  It's
     currently in a final draft, but the staff does have copies of that.
         DR. APOSTOLAKIS:  So then what you do is you take the
     ten-to-the-minus-two and you can see that the number of systems and you
     judge subjectively then, like we specialize the fire frequencies for
     critical locations.  You start with a building and slowly go down to the
     location.
         MR. FLEMING:  Yes.  And very briefly, what we do is we break
     the whole population of piping failure statistics up into PWR and BWR
     vendor groups.  We break the system populations into several system
     sizes and then we look at all the different failure mechanisms.  So we
     can look at it on a conditional segment base type of analysis.
         DR. APOSTOLAKIS:  So this is one of the major uncertainties
     that Steve Dinsmore probably referred to.  Say yes, Steve.
         MR. DINSMORE:  Yes.
         MR. FLEMING:  And by the way, I wanted to respond to what
     Steve said earlier about the uncertainties, because I agree in his
     comment to some extent, but I also wanted to clarify that.  We have done
     a detailed Bayesian uncertainty analysis of the attempts to make pipe
     failure rates and rupture frequencies from the service data and when you
     look at -- if you look at these results on an order of magnitude basis,
     on a logarithmic basis, you have very, very broad distributions that
     characterize large uncertainties, and it's a true statement, what Steve
     said.
         But if you try to take that and then develop a conclusion
     about what do those uncertainties say about the impact on risk, it's
     still possible to develop very robust conclusions that in spite of the
     large uncertainty, it's still a very, very small fraction of the risk.
         But there is one other -- but I think coming around to
     support what I think the -- and the concern about defense-in-depth is
     that these PRA calculations that we can do today based on looking at
     historical data, I think, are good for making current estimates of what
     we think the rupture frequencies are today, but we'd be on very shaky
     ground if we tried to project these 20, 30 or 40 years out into the
     future, and I think therein would lie the difficulty in trying to make
     conclusions about life extension without having some kind of way to keep
     monitoring possible trends in performance.
         Because we can't see the future in yesterday's data.
         DR. APOSTOLAKIS:  Now, in the report, you say something
     that's not quite the same as what Carl said a few minutes ago.  Right
     under equations 3.1 and 3.2, you say based on the expressions, they're
     using a conservative estimate of the total PBF frequency of
     ten-to-the-minus-two, where it calculates the CCDP and so on.
         So that's why I raise the issue.  When I read this, I
     thought, my god, they're using something that is a frequency of type of
     break anywhere to make judgments, and so the report perhaps is not
     written very well on that point.
         MR. FLEMING:  The sections you're looking at right now
     explain the logic in deriving the original matrix to start with and then
     back in Section 2, with the benefit of more detailed analysis of the
     service data, we took another look at that question and confirmed that
     the order of magnitude assessments were --
         DR. APOSTOLAKIS:  That's not here.
         MR. FLEMING:  If you get to Section 3-7, for example, you
     finally -- if a pipe is classified as having a medium rupture potential,
     it's one-times-ten-to-the-minus-four per year.
         DR. APOSTOLAKIS:  Right.
         DR. SHACK:  They sort of assign conservative pipe break
     frequencies based on essentially degradation mechanisms.  If it's got a
     high degradation mechanism, medium or low.
         MR. FLEMING:  But in Section 2 of the report, this other
     more detailed look at applying the service data to look at these
     calculations is presented and it confirms the order of magnitude
     assumptions that were originally made.
         DR. APOSTOLAKIS:  I'm talking about the specialization. 
     Section 2 is very general.  It doesn't say how you go down the system. 
     Section 2 is an analysis of the existing failures.
         MR. FLEMING:  Right.
         DR. APOSTOLAKIS:  I would have covered that.  And there is
     equation 2-1 that bothers me, too, on page 2-8.  The probability of
     rupture given failure.  I don't know what that means.  Page 2-8.
         MR. FLEMING:  That was a simple model that we used to
     analyze the service data for those failure mechanisms that have a strong
     leak-before-break characteristic.  We define failure as the whole
     package of failure modes, including small leaks and ruptures.
         So the model basically says we have a rupture frequency
     which is equal to the failure frequency times the conditional
     probability that it's a large failure.
         DR. APOSTOLAKIS:  So that's what is missing from the report,
     though.  I didn't see that.  So it wasn't clear to me what the
     difference between the rupture and the failure.
         MR. FLEMING:  The way we defined it --
         DR. APOSTOLAKIS:  A leak is a failure?
         MR. FLEMING:  A failure is any fluid going through the
     boundary, including a leak.  A rupture we defined at 50 gpm and larger
     flow areas.
         DR. APOSTOLAKIS:  That would have saved me a lot of time.
         MR. FLEMING:  Sorry.
         DR. APOSTOLAKIS:  Back to your presentation.
         MR. MITMAN:  Next, the next slide is on the pipe service
     experience, which we've already talked a little bit about.  There is
     over 2,000 reactor operating years of experience --
         DR. WALLIS:  You better put in the word reactor there, too.
         MR. MITMAN:  Fair enough.  In that database, there is 1,145
     events, 1,145 failures.  Here, failures are defined as either leaks or
     ruptures.  The vast majority, 1,076 of those were leaks.  Most of those
     were less than five gpm and most are due to corrosion mechanisms.
         Out of the total database, there are 69 events that were
     categorized as ruptures.  The failure mechanisms are well understood and
     the conditions necessary to produce the failures are generally known.
         DR. APOSTOLAKIS:  Now, where are the numbers, on another
     table?
         MR. MITMAN:  Which?
         DR. APOSTOLAKIS:  Table 2-2 and 2-1.
         MR. FLEMING:  They're consistent.
         DR. APOSTOLAKIS:  You said 1,100-something total failures
     and I don't see that here.  I see 1,500.
         MR. MITMAN:  If I remember correctly --
         DR. APOSTOLAKIS:  Oh, 1,145.  I'm sorry.  In the database. 
     Okay.  You have to look elsewhere.
         MR. FLEMING:  It's consistent.  It's confusingly presented,
     but it's consistent.
         DR. APOSTOLAKIS:  Speaking of that, what's the difference
     between the degradation mechanism and severe loading?
     
         MR. FLEMING:  Degradation mechanisms are degradation
     mechanisms like thermal fatigue, stress corrosion cracking that occur
     over long periods of time due to physical degradation mechanisms, where
     severe loading conditions are the imposition of loads in excess of the
     capacity of the pipe due to water hammer, impact, external impact on the
     pipe, frozen pipes, over-pressurization of the pipe beyond its design
     capacity and things like that.
         DR. APOSTOLAKIS:  So what you're saying is that in the
     severe mechanism, there was no aging mechanism acting.  It's just that
     you had the load that exceeded the design capacity.
         MR. FLEMING:  In the vast majority of cases.  Now, in
     principal, you can have a failure due to a combination of degradation.
         DR. APOSTOLAKIS:  That was my next question.
         MR. FLEMING:  But in the analysis of the service data, this
     was not evident.
         DR. APOSTOLAKIS:  So when you had the degradation mechanism,
     you didn't see any failures because there was degradation and then there
     was a load.
         MR. FLEMING:  We didn't see any evidence of that, although
     in principal, we know it's possible.  We didn't see any evidence of
     that.  The severe loading condition failures were, according to the
     reports that we analyzed, the loads were sufficient to cause the
     failure.
         DR. APOSTOLAKIS:  So let's look at the rupture.  You say
     erosion/corrosion or flow accelerated corrosion, there were 18 ruptures,
     on table 2-1.
         MR. FLEMING:  Right, in large pipes.
         DR. APOSTOLAKIS:  So you're saying that these were due to
     the steady-state pressure in the pipe and you simply had deterioration.
         MR. FLEMING:  Yes.
         DR. APOSTOLAKIS:  Wow.
         MR. FLEMING:  Or if there was any variation --
         DR. SHACK:  EC will do that to you.
         MR. FLEMING:  Or if there was any variation to the loading,
     it was not of any -- it wasn't of any significance that was noted in the
     report.  There might have been some small pressure transient.
         MR. MITMAN:  Keep in mind this is Class I, II and III
     piping, not just Class I.
         MR. FLEMING:  It's the whole plant.
         DR. BONACA:  Do you make an analysis of the difference
     between the systems that normally run and systems that don't run?  The
     reason why that's an important question is that most safety systems are
     standby.  They don't run.
         MR. MITMAN:  Well, some of the damage mechanisms behave
     differently whether the system is running or not.  In our analysis, we
     go ahead and we look at operating conditions of the plant or of the
     system and of the portion of the system, and that helps us decide
     whether that segment is subject to the damage mechanism.
         DR. SEALE:  Could I ask a somewhat different question?  My
     impression is that one of the things that gets us off the hook on a lot
     of our concerns is the validity of the idea of leak-before-break.  Is
     that equally valid for what I will call degradation versus load-induced
     failures?
         MR. MITMAN:  It's true for some of the damage mechanisms. 
     IGSCC, it is.  It's not true for FAC.  So it depends upon the damage
     mechanism, whether leak-before-break is valid.
         DR. SEALE:  But in general, you can't make a generalization
     that would say that load-induced failures are less likely to exhibit
     leak-before-break.
         MR. FLEMING:  That's correct.
         MR. RICCARDELLA:  I think that's a true statement.
         DR. SEALE:  So the ones you're not -- that inspection
     doesn't help you with are the ones that are most likely to be severe
     immediately on occurrence.
         MR. FLEMING:  That's right.
         DR. SEALE:  Okay.
         MR. FLEMING:  And looking at your question from the service
     data standpoint is that we can look at this second parameter in that one
     equation, given the failure, what's the conditional probability that
     it's a rupture given a failure.  And looking at that parameter for the
     degradation mechanisms other than FAC, that number tends to be very low,
     on the order of a few percent, but for flow accelerated corrosion and
     the severe loading conditions, it tends to be much higher, water hammer
     events, over pressure events, which stands to reason, I think.
         DR. WALLIS:  I'm wondering if it's true that if inspection
     doesn't help you, I'm thinking of the fire-line break in Washington,
     where they had something like 17 water hammers they didn't pay any
     attention to and then the 18th broke the pipe, and probably this is
     because it's a loosened thing.  The other water hammers have loosened
     things up.  So paying attention to the history and inspecting would
     perhaps have prevented the failure due to sudden loading.
         MR. MITMAN:  I don't think we're trying to say that you
     should ignore what's going on in the system, but doing an ultrasonic
     inspection or volumetric inspection of a weld probably isn't going to
     keep you out of trouble with a water hammer.
         DR. WALLIS:  Something like a water hammer, minor water
     hammers have happened before in that component and no one has paid much
     attention maybe and then there is a big one.
         MR. MITMAN:  And that's one of the things we will discuss a
     little it later, is that we do take into consideration water hammer in
     our categorization of the degradation mechanisms.
         DR. BONACA:  Just to complete my thought.  I had a question
     before.  So you feel that that statement, failure mechanism, well
     understood, it's still applicable also for those piping by systems that
     don't run.
         MR. MITMAN:  Absolutely.
         DR. BONACA:  That's fully understood.
         MR. MITMAN:  Yes.
         DR. BONACA:  And for those, also, inspections doesn't have
     preferential value.
         MR. MITMAN:  That is correct.  In the database, there are 69
     rupture events where rupture is defined as greater than 50 gpm. Again,
     the failure mechanisms are understood.  Some of the mechanisms are not
     amenable to inspection.  There is only one event in the RCS and several
     events in steam water and feed water systems.
         EPRI has a program to periodically update the database, keep
     an eye on what's going on in the industry, and make sure that the
     methodology doesn't need to be revised.
         The next slide is an example of the deterministic criteria
     that we used to do the damage mechanism analysis.  In this particular
     example, we're looking at thermal --
         DR. APOSTOLAKIS:  Can I interrupt?  Because I have to go
     somewhere at 10:00 and I have a couple of questions that jump ahead and
     I wanted to ask them.
         You didn't really expect that we would let you finish your
     presentation the way you have prepared it.
         Would you go to your slide 19?  Consequence ranking
     criteria.  Now, the conditional core damage probability, this is
     calculated as shown in one of the equations here as the new CDF,
     assuming the system is down, minus the baseline, times the exposure of
     time, and the exposure of time can be anywhere from a year to maybe a
     few days.
         Now, as far as I know, the NRC does not have and nor does
     the industry criteria as to what's high or low with respect to the
     conditional core damage probability, except for temporary outages, where
     they give a ten-to-the-minus-seven number.
         So where did the ten-to-the-minus-four come from?  The
     ten-to-the-minus-four is used for the core damage frequency, not the
     core damage probability over a period of three months.
         MS. DIMITRIJEVIC:  This was this equation three you just
     asked us about, CDF.
         DR. APOSTOLAKIS:  Yes.
         MS. DIMITRIJEVIC:  Where as that equation?
         DR. APOSTOLAKIS:  There were two equations on page 3-8.
         MS. DIMITRIJEVIC:  Yes.
         DR. APOSTOLAKIS:  But it doesn't --
         MS. DIMITRIJEVIC:  It tells you that that was very
     conservative, assuming the --
         DR. APOSTOLAKIS:  But I don't know whether it's
     conservative, is it?
         MS. DIMITRIJEVIC:  If you look in the explanation after,
     because if you assume that we are going to look in every location and
     segment, if we assume that we're going to have a thousand segments at
     the top of a plant, it's ten-to-the-minus-two, which we think is a
     conservative estimate.  Then ten-to-the-minus-two for pipe failure
     frequency, times ten-to-the-minus-four, will give you
     ten-to-the-minus-six CDF, which corresponds to that failure for CDF.
         MR. FLEMING:  So it was derived from a CDF limit and working
     backwards to come up with a conservative estimate of what that would
     mean.
         DR. APOSTOLAKIS:  So if I don't do anything for a year, then
     you say the NRC's goal of ten-to-the-minus-four per year times one year
     becomes a probability.
         MR. FLEMING:  No, no, no, no, no.
         MS. DIMITRIJEVIC:  The frequency CCDP here is conditional
     given pipe failure.
         DR. APOSTOLAKIS:  I understand that.  That's why I'm --
         MS. DIMITRIJEVIC:  So we take the five FAC yearly frequency
     and we say the worst we can have is five yearly frequency of
     ten-to-the-minus-two if every segment is in the high risk area.
         DR. APOSTOLAKIS:  Okay.
         MS. DIMITRIJEVIC:  And then ten-to-the-minus-two per year
     times ten-to-the-minus-four is ten-to-the-minus-six per year.
         DR. APOSTOLAKIS:  But where did the ten-to-the-minus-four
     come from?  That's my question.
         MS. DIMITRIJEVIC:  Well, we selected that so that we get
     that ten-to-the-minus-six per year.
         DR. APOSTOLAKIS:  You selected it.
         DR. SHACK:  Took ten-to-the-minus-six and divided by
     ten-to-the-minus-two.
         MS. DIMITRIJEVIC:  That's it.  We took the
     ten-to-the-minus-six and divided it by ten --
         DR. APOSTOLAKIS:  And they did it correctly.  Where did the
     ten-to-the-minus-six come from?
         MS. DIMITRIJEVIC:  That came from the NRC criteria of
     ten-to-the-minus-six not allowing any changes and things like this, Reg
     Guide 1.174.
         MR. FLEMING:  Which we anticipated before it was published.
         MS. DIMITRIJEVIC:  It's very conservative because that means
     they actually accept changes, but we still have a medium region.
         DR. APOSTOLAKIS:  So the delta CDF -- actually, it
     ten-to-the-minus-six is negligible.  Ten-to-the-minus-five was even
     allowed.
         MS. DIMITRIJEVIC:  So that's why I say we are very
     conservative, because of uncertainty, we said, okay, everything which is
     ten-to-the-minus-six we're going to look at, that's our basis.  And we
     say even in ten-to-the-minus-six, we're still going to take a little
     look.
         DR. APOSTOLAKIS:  The report could have been written a much
     better way.  I really tried to understand it and, boy, every sentence is
     loaded with --
         MS. DIMITRIJEVIC:  Meaning.
         DR. APOSTOLAKIS:  With meaning, yes.  I mean, Carl had to
     explain a few things, you had to explain a few things.  Anyway, that
     makes sense.  What you just said makes sense.  But it's not easy to
     understand that from this.
         DR. WALLIS:  Are you suggesting a report in which the
     sentences were loaded with less meaning would be better?
         DR. APOSTOLAKIS:  Yes, because then you would have more of
     those and it would be easier.  The total sum would be the same.
         If I have to spend three-quarters of an hour understanding
     each sentence, that's a pretty thick report, that doesn't help.  There
     was a whole rationale that Vesna just gave us that is not evident from
     this paragraph.  That's what I'm saying.
         DR. SHACK:  Right.
         MR. MITMAN:  You wanted to ask additional questions?
         DR. APOSTOLAKIS:  Gee, now you're catching me.  Oh, yes.  A
     general question.  You used the word "easy" earlier.  Now, I must say
     this business of the trains and giving a worth of one-half and all that,
     why is that easier from having your PRA, your PC, and saying, you know,
     put this down, get the new number.
         I mean, we have the software now that does these
     calculations very quickly and get conditional core damage frequencies. 
     And to go through this exercise of deciding the worth of each train and
     going through -- in other words, you are redoing part of the PRA.  What
     is the reason for that?
         MS. DIMITRIJEVIC:  I will tell you why it was designed.  It
     was designed so that people who are not PRA analysts can come up with --
         DR. APOSTOLAKIS:  But they have to do PRA-type analysis,
     Vesna, anyway, because they have to look for trains that will save you.
         MS. DIMITRIJEVIC:  That's true, but they have to -- and they
     can be provided with them in the beginning and then they can do the
     analysis.  But if you really know PRA, which can be done fast, if you
     know something --
         DR. APOSTOLAKIS:  You don't really have to know PRA, you
     don't have to be a PRA analyst to use it.  You have your PC and you do
     your sensitivity calculations.  Otherwise, if you know nothing --
         MS. DIMITRIJEVIC:  Let me give you two reasons to let you
     even forgetting that PRA, in my opinion, really don't run them fast yet. 
     It always takes some time.
         But independent of that, let me tell you what's the main
     reason.  When you run the PRA, you come with a number.  It gives you a
     number, but it doesn't tell you what it is and why it is.
         When you look in this, you know exactly why and what's
     happening and it really helps you to understand the heart of this.  We
     went in the heart of the PRA.  So they know that this is -- if they have
     this pipe failure, that this is because they are HPSI is going to know
     how -- instead of getting the number three times putting
     four-times-ten-to-the-minus-four.
         We thought they would be more valuable for analysts to
     understand why are some failures and the other thing is also we have to
     rank the initiating events.  So we can really use importance measures
     for that.
         We thought that it will give more insight and will help
     people who are not PRA analysts do this estimate.
         DR. APOSTOLAKIS:  That's the right word.  But, Vesna, I
     think you have a valid point there, that you shouldn't really get a
     number alone.  But it seems to me you can easily get those insights from
     a good PRA, because a good PRA doesn't give you numbers only.
         MS. DIMITRIJEVIC:  You can look at sequences.
         DR. APOSTOLAKIS:  It gives you the sequences, the minimal
     cut sets and so on.
         MS. DIMITRIJEVIC:  Yes.
         DR. APOSTOLAKIS:  And maybe you can develop a small software
     package to compliment what you already have to get these insights.  But
     this reluctance on the part of the industry to say PRA is useful and we
     will do things using the PRA is a mystery to me.  It's mystifying.
         MS. DIMITRIJEVIC:  The one thing which I have to say is the
     misconception that this method doesn't use PRA, it uses it very well.
         DR. APOSTOLAKIS:  Of course it does.
         MS. DIMITRIJEVIC:  And uses it perfectly, in my opinion,
     because it's break is what actually PRA is all about and shows exactly. 
     So it shows the safety function, it shows you backup trains, it shows
     you the sequences.  So it's basically really uses PRA.
         It doesn't use it as being -- it gives you the importance
     measures and things like that, but that is what PRA is all about.  So,
     therefore, PRA is very -- and I did a lot of analysis.  You can believe
     me that this really makes it very easy and clean-cut. You look in the
     PRA, you see there the success criteria, you put the safety functions
     diagrams, you look in there, the ability of the trains and their
     initiating events and there you are.
         Then you have ten pages and you can just run through the
     events very nicely.
         DR. APOSTOLAKIS:  Who did that?  Did a guy with no
     experience with a PRA actually figure out what are the trains that are
     available and gave the weights?
         MS. DIMITRIJEVIC:  No.  That was -- that's done with the PRA
     specialist.  But once when you have them, everybody can do the analysis.
         DR. APOSTOLAKIS:  And this is an analysis you do only once,
     right?
         MS. DIMITRIJEVIC:  Yes.
         DR. APOSTOLAKIS:  It could be done off-line by somebody who
     knows something about the PRA.  It's not that you are asking your
     average --
         MS. DIMITRIJEVIC:  No.
         DR. APOSTOLAKIS:  -- doing this every month.
         MS. DIMITRIJEVIC:  It's always done with the people who know
     PRA.
         DR. APOSTOLAKIS:  I am really mystified by this reluctance
     to say we now have the PRA, here is what you can do with it, here is
     what you need for this kind of analysis, here is a way to get it. 
     Instead of doing that, we're saying we'll develop these mysterious
     tables, that give a wroth of .5.  I don't know how to do that.
         MS. DIMITRIJEVIC:  Well, because we are working in actually
     in the intervals, not with the actual number.  We do have a section on
     --
         DR. APOSTOLAKIS:  And I don't even know what the train is
     sometimes.  You're asking a poor guy to make all these judgments when
     the PRA has already provided you with answers.  I'm not saying that what
     you did was incorrect, but I'm just mystified by this reluctance on the
     part of the industry to say here is a tool that's useful, let's use it,
     and not try to dance around it all the time.
         And, my god, if we ever demand a PRA, it's a major crime.
         MS. DIMITRIJEVIC:  We added it --
         DR. APOSTOLAKIS:  I've made my case.
         MS. DIMITRIJEVIC:  But we did it add it in our report
     section 3.36, which tells you really how PRA --
         DR. APOSTOLAKIS:  Page?  I always give you the page.
         MS. DIMITRIJEVIC:  Page 3.32.
         DR. APOSTOLAKIS:  3.32.
         MS. DIMITRIJEVIC:  Yes.
         MR. FLEMING:  I want to just augment something, if I may, on
     what Vesna just said.  I think there may be a little bit of
     misunderstanding here.
         It is the role of the PSA analyst to figure out these train
     relationships and that is only for a qualified PRA analyst to do. But
     the idea here was to give the piping engineer a tool that he could put
     segments on the risk matrix without himself doing the PRA calculations
     to put them on the matrix.
         And since it didn't require him to come up with a numerical
     estimate, just to figure out which of the bins to put it in, which are
     four decades wide, there was a question of whether you want to use a
     surgical instrument or a screwdriver to do the appropriate tests.
         DR. APOSTOLAKIS:  Well, all right.
         DR. SEALE:  Can I ask a related question?
         DR. APOSTOLAKIS:  Am I going to say no?  Go ahead.
         DR. SEALE:  You have here an application of a PRA and it's
     clear in coming up with your simplified algorithms that consciously or
     unconsciously, you're responding to the fact that not all PRAs are
     equal.  There is another effort going on, quite separate from this, that
     the Commission has a commitment to, and that is the development of
     criteria for what constitutes an adequate PRA.
         I'm going to ask the staff.  Do you guys who find these
     kinds of applications for PRAs talk to those guys who are worried about
     what constitutes an adequate PRA?
         MR. DINSMORE:  This is Steve Dinsmore.
         DR. SEALE:  You've got your hat in both rings.  But you see
     the question.  I mean, clearly, there are other applications like this
     where people are using PRAs and there is not necessarily that nexus made
     to this other effort, which is supposedly graduating the level of the
     PRA, the product, to what we would hope would be the workable all things
     for all people version.  It's just a caution.
         MR. DINSMORE:  Okay.
         DR. APOSTOLAKIS:  Mr. Fleming, by the way, is intimately
     involved with that effort.
         DR. SEALE:  Okay, fine.  Fine.
         DR. APOSTOLAKIS:  But it's a more general comment that I
     wanted to make.  Every time we see something, for heavens sake, let's
     not ask them to use a PRA.  So we go out of our way to produce tables
     and things, like the five methodology, again, PRA, my god, no tables. 
     People are not stupid and finally --
         DR. SEALE:  You only get there, though, if you make the
     effort.
         MR. MITMAN:  Another thing to keep in mind is this
     methodology was started back in '92-'93 and the capabilities of the
     machines and the capabilities of the codes and the PRAs isn't what it is
     today.  And if we started today, we might do it differently.
         DR. APOSTOLAKIS:  Or you wouldn't do it at all, based on
     your conclusions.  If somebody took defense-in-depth away, you wouldn't
     do it at all.
         MR. MITMAN:  We would discuss it with the appropriate
     industry --
         DR. APOSTOLAKIS:  I understand that, yes.
         MR. MITMAN:  -- bodies and see what we would come up with.
         As we were saying at slide 14.  Slide 14 shows some of the
     damage mechanism attributes that we use to determine whether a weld
     segment is susceptible to those damage mechanisms.
         What we're looking at here is thermal fatigue and some of
     those damage mechanisms that are there.  Deterministic rules that we
     apply to determine whether the segments are susceptible or not.
         DR. WALLIS:  I was looking at the bottom there.  You don't
     often inject cold fluid in the hot pipes, so in infrequent even which
     will lead to fatigue, but there are sometimes situations where cold is
     somewhere close to hot and there are circulation patterns set up which
     are unstable and a piece of piping is bathed in hot, cold, hot, cold,
     over a long period of time, and that's a classical thermal fatigue
     mechanisms.  I don't see it here.
         MR. RICCARDELLA:  That's what we call tasks.
         DR. WALLIS:  That's what you call tasks?
         MR. RICCARDELLA:  Yes.  Tasks is thermal --
         DR. WALLIS:  Maybe I didn't -- okay.
         MR. RICCARDELLA:  -- and stratification --
         DR. WALLIS:  It's rather hard to figure out, as I remember,
     and sometimes it happens, sometimes it doesn't.
         MR. RICCARDELLA:  But we've looked -- EPRI has done a very,
     very large study on tasks.  They had a task force to put together a
     large thick report, and we've gone through that and we've attempted to
     come up with some conservative rules that if you have any of these
     conditions, we say it's susceptible to thermal fatigue.  All of these
     are very conservative.
         It's not often that a delta T of 200 or 150 is going to
     cause you a problem.  But if there is even the potential that a system
     could get a delta of that much, we say it's potentially thermal fatigue
     susceptible.
         Now, we could go in with a finer screen, and sometimes we
     do.  It's no good to prioritize if everything comes out.  So if we get
     too many locations, then we'll go in with a finer screen and find out
     what the actual delta T is.
         DR. WALLIS:  So this is a case where one could call
     understanding of the thermal hydraulics is important to risk assessment.
         MR. RICCARDELLA:  To selecting in-service inspection
     locations.
         DR. SHACK:  Whether that's important to the risk assessment
     is another question.
         DR. WALLIS:  Well, presumably it is.  If you don't -- if
     there is some physical mechanism which you completely ignored which can
     break a pipe, then it can seem that that's important to the risk and
     whether you assess it or not is perhaps up to you as a responsible
     professional.
         MR. RICCARDELLA:  Yes.  Well, part of what you're getting at
     is the reason for this performance monitoring feedback loop; that if
     there is some mechanism that we've just forgotten about or didn't know
     about and it occurs in one plant, then it's going to work its way into
     our methodology and will be picked up.
         MR. MITMAN:  The outcome of the damage mechanism assessment
     is categorization into three categories, high, medium and low.  Where we
     end up is to get into high, you have to have flow accelerated corrosion. 
     Medium is any other damage mechanism and if there are no damage
     mechanisms present, then that gets you into low.
         Now, there is one caveat with this.  It doesn't show on this
     slide.  It's discussed in the report.  That is, there is another way to
     get into the high category, which came up in earlier discussions, and
     that is if you have another damage mechanism, plus water hammer.
         So if you've got some degradation of a pipe due to MIC and
     you know that the system is subject to -- is susceptible to water
     hammer, that would also put you into the high category.
         In practice, what we see happening most of the times is if a
     system is susceptible to water hammer, most plants will go out and try
     and resolve the question on water hammer, which would drop it back down
     into the medium category.
         DR. SHACK:  I would assume that water hammer would -- if
     it's susceptible to water hammer, it's high, degradation or no
     degradation mechanism.
         MR. MITMAN:  It would be high, but you can inspect all you
     want with volumetric inspections and you're not going to find water
     hammer.
         DR. SEALE:  Okay.  These are for leaks then.
         MR. MITMAN:  This is leak --
         DR. SEALE:  For inspection, I beg your pardon.
         MR. MITMAN:  The whole methodology is to help you decide
     where to do your volumetric inspections per Section 11.
         DR. SEALE:  So in general, that's why load-induced failures
     don't show up here.
         MR. MITMAN:  That's right.
         DR. SHACK:  We were scheduled for a break at ten and since
     we're about to go to the consequence, maybe this is a good time to do
     it.  Does that seem reasonable?
         MR. MITMAN:  That sounds reasonable.
         DR. SHACK:  A 15-minute break then.
         [Recess.]
         DR. SHACK:  I guess we have enough members back, so that we
     can resume the meeting.
         MR. MITMAN:  Okay.  We want to start in on the consequence
     analysis at this point.  The first thing we do is there are four types
     of -- four considerations that we have when we go into the consequence
     analysis.  We're looking at initiating events, events that affect the
     mitigating ability of systems and trains, both from a loss of the system
     or train or degradation of the system or the train.
         We're looking at containment effects, both loss and
     degradation of containment, and then we also look at combination events,
     something that will affect both a mitigating system and containment or
     could be initiating event and affect a standby safety system also.
         We've seen this slide already.  The consequence ranking is
     -- there is a four-tier consequence ranking, high, medium, low, and not
     shown here is none.  There is -- none is the easiest to deal with. 
     There are a couple of abandoned in-place piping systems that we've found
     here and there that have no consequence whatsoever on the analysis, and
     that's why they show up on the matrix.
         The high category is conditional core damage probability
     greater than 1e-to-the-minus-four and conditional large early release
     probability of greater than 1e-to-the-minus-five.
         These are severe initiating events or severe loss of
     mitigation capability or high risk of containment bypass.  Medium
     category is conditional core damage probability, greater -- greater than
     1e-to-the-minus-six, but less than 1e-to-the-minus-four.  It should be
     less than or equal to.
         Also, conditional large early release probability between
     1e-to-the-minus-seven and 1e-to-the-minus-five, and these are moderate
     type events.
         Then on the low side is conditional core damage probability
     less than 1e-to-the-minus-six and conditional large early release
     probably less than 1e-to-the-minus-seven.
         These are for mild type events.
         DR. SHACK:  Again, I guess I asked this question before, I
     mean, you -- when I look at this guideline for assigning the consequence
     category, there is some sort of generic PRA that decides that when I
     have one unaffected backup train, that I'm in the
     ten-to-the-minus-four/ten-to-the-minus-five category, and it's high.
         The guy doesn't actually do that calculation.
         MR. MITMAN:  Initially, it's generic.  The tables were set
     up generically, but the expectation is that the plant will recalibrate
     those tables with the plant-specific PRA.
         DR. SHACK:  And they will bend it this way, then.
         MR. MITMAN:  This is the criteria for bending it and you may
     -- if you go to South Texas, with a three-train system, you'll probably
     move everything to the --
         MS. DIMITRIJEVIC:  Medium.
         MR. MITMAN:  Yes.  Everything will drop down to medium
     instead of having it in the high category.  But, yes, you do that
     plant-specific calibration.
         DR. SHACK:  What it could mean is that for some plants with
     one and a half backup trains, you could be in the medium instead of the
     high for some very plant-specific reason.
         MR. MITMAN:  Yes.  You could have some susceptibilities or
     some weaknesses in the plant design.
         MS. DIMITRIJEVIC:  You could have a different number of
     backup trains and the load would be different.
         MR. DINSMORE:  This is Steve Dinsmore, from the staff.  Our
     understanding is that the table won't change, because we're going to be
     moving the table.  What will change is the number of trains that they
     can take credit for.
         If they have two HPSI trains, but, again, if it was
     ten-to-the-minus-three, in one plant, they can only credit one and a
     half trains.  In another plant, maybe because it's better designed,
     they'd be able to credit it two trains.
         DR. SHACK:  So the table stays the same.
         MS. DIMITRIJEVIC:  The same.
         DR. SHACK:  It's how you credit --
         MR. MITMAN:  The number of trains.
         DR. SHACK:  -- the number of trains.
         MR. MITMAN:  The remaining tasks in the methodology, we'll
     go in and do segment risk categorization, selection of inspection
     locations, selection of the appropriate inspection techniques.  We do a
     risk impact assessment.  We document and finalize the project, put
     together the submittal, and then there is a performance monitoring where
     we have a long-term process to monitor the plant and also EPRI will
     continue to monitor the industry to make sure that there is no new
     mechanism that appears or an increase in the frequency of a degradation
     mechanism because of aging effects that we haven't seen yet.
         DR. SHACK:  Now, one of the things that the WOG ASME thing
     does that doesn't seem to come in here is they make an effort to
     estimate the leakage, also, as well as the failure, and their goal was
     essentially to maintain the leakage rates about what they've observed.
         That doesn't seem to be -- that's nothing you're addressing
     here.
         MR. MITMAN:  We are not trying to calculate any leakage
     rate.  What we essentially see, if you go back and look at the risk
     matrix, the medium category tends to be those damage mechanisms, medium
     category here tends to be those damage mechanisms that typically leak.
         The only way you're going to get into -- or getting into
     high, you tend to have a damage mechanism that can fail on your, can
     have a large rupture.  So that's probably the only place that the
     leakage rate figures into the analysis.  But we're not trying to --
         DR. SHACK:  You don't have a target leak rate for --
         MR. MITMAN:  No.
         DR. SHACK:  In a general sense, who ends up with more
     inspection locations following which process?
         MR. MITMAN:  I think we end up with --
         DR. SHACK:  With more.
         MR. MITMAN:  No.  We end up about the same.  About the same.
         MS. DIMITRIJEVIC:  We don't really know.
         MR. MITMAN:  Well, we've looked a little bit at Surry,
     comparing Surry to the work we've done, and I think if you look at
     what's done on Surry and what's done on -- what's coming out of the
     ANO-2 analysis and the Fitzpatrick analysis, that they're approximately
     the same order of magnitude of inspections.
         MR. ALI:  This is Syed Ali from the staff.  Just to clarify. 
     Although the comparison is really not valid because one is a PWR, Surry
     is a PWR, and some of the ones you have done are BWR.  But for Class I,
     this methodology, the EPRI methodology, the sample size went down from
     25 percent in ASME-11 to about ten percent, whereas the Westinghouse
     methodology is about six and a half or seven percent.  So that's the
     order of magnitude difference.
         MR. MITMAN:  Okay.  The next step in the methodology is to
     do the risk evaluation.  At this point, we have completed the
     degradation analysis and the conditional -- or the consequence
     evaluation and that allows us to bend the segments into the appropriate
     boxes on the matrix.
         DR. SHACK:  Another question.  You're also building 0313
     into this also, aren't you?  In the sense that that's how you bend
     things into high, medium and low.
         MR. MITMAN:  At this point, 0313 is -- we're only building
     in the category A 0313 welds for IGSCC.  We feel and it's part of the
     evaluation that's going on now with BWRs that with a category A weld,
     that we can justify that we need no additional inspection beyond what
     the risk-informed process is requesting.
         Now, there is a dialogue, a discussion and some analysis
     going on in the industry to revisit 0313 and that's not part of the
     methodology right now.  However, the methodology is set up so that if
     and when that happens, we'll be able to take advantage of it.
         One of the big concerns that we've addressed in the last
     year or so was to take a harder look at the risk impact of the analysis
     and that's -- well, once we've done the element, the risk-ranking and we
     go out and do element selection and then we go ahead and look at
     inspection for cause and decide what methodology should be looked at in
     the various segments.
         One of the questions that was asked earlier in the
     discussion that we have deferred was what do you do with the segments in
     the low categories and the methodology calls for no inspections in those
     low risk regions.
         You should keep in mind here that that's not to say we're
     not doing any inspections on the systems.  We continue to do inspections
     in the medium and the high risk regions of those systems and we haven't
     found any systems yet where we're recommending doing no inspection.
         In addition to that, we continue to do the augmented
     inspection for FAC, IGSCC and whatnot.  So it's not like we're walking
     away from large segments of the plant and not performing any assessments
     for those.
         DR. SHACK:  And, again, you're doing leak tests.  Everything
     but volumetric inspection.
         MR. MITMAN:  And we continue to do leak tests on all Class
     I, II and III.
         MR. RICCARDELLA:  Another aspect of that, too, is if you
     look at the lower right-hand corner, things with low or essentially no
     degradation mechanism identified, we still are inspecting some subset of
     those, because some of them do fall over into the medium category.
         If they have a high failure consequence potential, then they
     do get a medium and we see that we do a lot of category four -- there
     are category four inspections.  So it's not like we're ignoring things
     just because we haven't identified any mechanism.
         MR. MITMAN:  Now, in the earlier -- in the beginning slides,
     we talked about the two code cases that apply here, N-560 for Class I
     only and N-578, which applies to the full plant.  This is the one place
     where we come up with a slight distinction between how you select welds
     and elements between the two code cases.
         In N-578, we do 25 percent of the high risk category
     elements.  In the medium risk category, we do ten percent.  In the low,
     we do none, no volumetric inspections.
         In N-560, we start at the category one, category here, and
     start going and applying welds, looking at welds and adding them to our
     list.  But when we get approximately to ten percent, then we'll pull up
     and stop.
         In practice, what we see is we don't do all the high
     category welds.  We do -- for instance, if we have a BWR with four steam
     lines and we have four welds on each of the steam lines, we might take
     half of those welds, so we capture representative inspections for all of
     the steam lines and then save some of those other inspections for other
     places in the Class I system, so that we get a more representative and a
     better, broader perspective of what's happening in the Class I system.
         DR. SHACK:  Now, the description in the report actually says
     it's considered more prudent, but there is nothing that says you have to
     do it that way.
         MR. MITMAN:  To do it?
         DR. SHACK:  To split your ten percent.
         MR. MITMAN:  It's a very strong recommendation.  To date,
     EPRI has been involved with all the applications of this methodology and
     we very carefully ensure that it's not being misapplied.
         DR. SHACK:  Why isn't there just a step that says do some of
     them, 25 percent?
         MR. MITMAN:  Didn't want to get too prescriptive, wanted to
     be able to use engineering judgment and a little bit of flexibility to
     decide exactly where to get those ten percent.
         Pat, you wanted to add something?
         MR. O'REGAN:  Pat O'Regan, EPRI.  One of the reasons we
     didn't want to be too prescriptive is there are other considerations
     that go in when you select locations, such as high rad areas and
     accessibility and stuff like that.  We wanted to make sure that still
     carried weight when we went through the element selection process.
         MR. MITMAN:  One of the last steps we do is to do a risk
     impact assessment.  The process itself was designed as a risk-informed
     process.  So there are risk considerations taken into account from the
     beginning.  The risk impacts are -- we see risk impacts in three areas.
         Allocation of inspections at the high risk locations, we
     have inspection for cause impacts, and we have impacts from elimination
     of inspections in the low risk segments.
         The risk assessment process is a three-tiered process,
     essentially.  Initially, we want to eliminate as much quantitative
     analysis as is appropriate and we bring to bear some qualitative tools
     to do that.
         If we have regions that are -- if we have low risk regions,
     we can qualitatively show that it has no impact on -- no unacceptable
     impact on risk.  If we're doing -- if we're increasing the number of
     inspections, we can qualitatively show that we have a positive or an
     improvement in the risk consequences.
         So first of all, if we want to do the qualitative analysis,
     and we can do that in most cases.  Where we can't, we want to go ahead
     and apply some bounding estimates to help us decide that the risk
     impacts are acceptable.  If the bounding estimates are not sufficient,
     then we bring to bear realistic quantitative analysis to help us assure
     ourselves that we're having appropriate impact on risk
         DR. SHACK:  It seemed to me that it was almost a "gimme" if
     I believed your -- that one-times-ten-to-the-minus-four was really a
     bounding CCDP for a medium risk thing, that I was almost set up to get a
     controllable delta C-P, because I'm looking at the important stuff and
     I'm neglecting the unimportant stuff, and I'm just going to get the
     answer to come out.
         I guess the question is, can I really be confident that the
     one-times-ten-to-the-minus-four is a bounding CCDP for a guy with a
     medium classification, a two-train system with anticipated transients
         MR. MITMAN:  I think so, yes.
         DR. SHACK:  For all plants?  I guess that's the part I have
     a little trouble with.
         MS. DIMITRIJEVIC:  What's plant-specific is -- you want to
     --
         DR. SHACK:  For the plant-specific part is how I credit the
     two trains.
         MS. DIMITRIJEVIC:  How many trains you have, yes.  Yes.
         DR. SHACK:  And so I guess the answer is if I have two
     credited trains, am I always guaranteed that
     one-times-ten-to-the-minus-four is a bounding estimate.
         MS. DIMITRIJEVIC:  You're going to be -- yes.  You could be
     a little -- I mean, we cannot really guarantee always, because sometimes
     it happens that you're 1.1 because of these estimates,
     ten-to-the-minus-four or something in the medium.  But you are very
     inside this.  It is not going to be higher than 5e-minus-four.
         DR. SHACK:  Guarantee is a strong word to use in a PRA,
     right?
         MS. DIMITRIJEVIC:  That's true.  The best of our knowledge. 
     The best of our knowledge, yes.
         MR. MITMAN:  Pat, you want to add something?
         MR. O'REGAN:  Yes.  Maybe a point of clarification.  Pat
     O'Regan, from EPRI.  The methodology is set up that if you assign to
     medium, you'll always be less than ten-to-the-minus-four.  If you have
     two trains in the plant, there is no guarantee that you'll always be a
     medium.
         DR. SHACK:  If I have two qualified trains and the matrix is
     approved, then I'm going to be a medium.
         MR. O'REGAN:  If you have two qualified trains, that's true.
         MS. DIMITRIJEVIC:  Yes.
         MR. O'REGAN:  But if you have two turbine-driven pumps, that
     may not qualify.
         DR. SHACK:  That's different, right.
         MR. O'REGAN:  So you may be a high.  But the criteria always
     is set -- the criteria is set in stone.  It's just a question of whether
     the plant could meet the criteria or not.
         MR. MITMAN:  One of the discussions earlier today was a
     discussion about questions about supporting documentation and where more
     of the analysis is shown.  These are most of the published reports that
     we've published to date on the methodology.  Most of those are
     proprietary reports, but we can make the necessary -- we can make them
     available for the ACRS to look at.
         MR. FLEMING:  I wanted to reflect back on the last question. 
     With regard to the potential issue about how robust are these
     conclusions, the battle lines for this argument really are down at the
     boundary between medium and low, because the way in which we do our
     delta risk evaluations, we entertain the need to do some quantitative
     bounding estimates if we're in the medium or high risk regions and we're
     suggesting a reduction in the number of locations that are eliminated.
         So the question about robustness really gets down to between
     the medium and low region.  That's where the potential issue would exist
     and at that level, we're another two orders of magnitude down the risk
     scale.
         MS. DIMITRIJEVIC:  And an ideal medium region is to escape
     those boundaries.  This is why we introduced this medium region.  So we
     don't have to have a clear cut between high and low.
         MR. FLEMING:  So what I'm saying is that if, for example,
     for some error or uncertainty in analysis that a category that belonged
     in the high region was miscategorized in the medium region, our delta
     risk procedure would still catch that, because if we ended up in this
     hypothetical location suggesting a reduction in the number of locations,
     and, therefore, there's a potential risk increase, we would go and do a
     calculation where we actually go in and do a bounding estimate that
     would more than cover that uncertainty.
         DR. SHACK:  Is that clear to me where I have to do the --
     that was the point I couldn't quite figure out, is when you had to do
     the quantification.  There's the qualitative estimate and the
     quantitative.
         MR. FLEMING:  For any medium or high risk region in which --
     or segment in which we're actually suggesting a reduction in the number
     of locations, we would come in and do bounding estimates.  So we're only
     relying on the qualitative arguments for the low risk category
     components.
         So the robustness issue associated with how we put the
     things on the matrix really is the -- the central question is have we
     gotten the lows correct.  The low-medium issue is more important than
     the high issue.
         MR. MITMAN:  The second to the last slide that I have shows
     some results from the pilot studies.  We're showing the number of
     inspections that were done under the current ASME Section 11
     requirements and the number of inspections that we're proposing have
     been approved by SERs for the pilot plants.
         What we see is in the high risk regions, we have some
     increases in the number of inspections, some decreases, but essentially
     about the same number of inspections.
         For the medium risk regions, we show usually a decrease in
     the number of inspections.  The one exception to that is on ANO-2, we
     were actually showing an increase in the medium risks.  Then in the low
     risk regions, we're showing a decrease in the number of volumetric
     inspections.
         DR. SEALE:  A couple of questions.  First of all, what
     percentage of the sites that wind up in the high category are actually
     identified for an inspection, in the medium?
         MS. DIMITRIJEVIC:  In high?
         DR. SEALE:  What fraction of the highs --
         MS. DIMITRIJEVIC:  Twenty-five percent.
         DR. SEALE:  -- are -- what fraction of the ones that are
     qualified as high actually wind up candidates for inspection?
         MS. DIMITRIJEVIC:  Twenty-five percent.
         DR. SEALE:  Okay.
         MS. DIMITRIJEVIC:  Twenty-five and ten percent, and that's
     for the full scope.
         DR. SEALE:  For the full scope.
         MS. DIMITRIJEVIC:  Yes.
         DR. SEALE:  What about when you do it with your CI-1, Class
     I?
         MS. DIMITRIJEVIC:  Class I, we select ten percent total and
     most of them are from high risk.
         MR. RICCARDELLA:  It comes out to be about 25 percent of the
     high risk.
         MS. DIMITRIJEVIC:  The same.
         MR. RICCARDELLA:  It's just the way the numbers work out. 
     And all three of these examples, it's essentially -- in fact --
         DR. SEALE:  Old Arkansas-2 really inspects the devil out of
     things, don't they?
         DR. SHACK:  Just a question on Vermont Yankee.  What is
     their IGSCC fix?  Have they replaced piping?
         MR. RICCARDELLA:  Vermont Yankee replaced piping.
         DR. SHACK:  So basically all their recirc piping is Class A
     then.
         MR. RICCARDELLA:  Yes.
         MR. MITMAN:  We did put together a backup slide that you do
     not have and it is not in the report that shows comparisons between what
     we end up with or what we would end up with if we did -- depending if we
     did an N-578 or an N-560 analysis.  These are all --
         MS. DIMITRIJEVIC:  N-578 selection criteria.
         MR. MITMAN:  And this is for Class I only.
         MS. DIMITRIJEVIC:  This is just for Class I.  It shows you
     that actually when you apply N-578 criterion to Class I, you always are
     around ten percent total.  You are either 9.4, 10.6, 13 or 9.8.  So
     basically even if you use again 578 criteria on Class I, because of the
     division between high and medium categories, then it's around ten
     percent of total, which is the same as N-560 criteria.
         And we keep that in mind with this percentage and we were
     not surprised at that.
         DR. BONACA:  I have a question.  Maybe I should ask the NRC. 
     But if I look at the whole logic, you have some high degradation
     category, piping in the high degradation category, with no consequence. 
     For those, essentially, you're recommending elimination of inspection in
     many cases.
         MS. DIMITRIJEVIC:  Actually, ten percent of inspection in
     medium.
         MR. MITMAN:  No.  You're over-hearing the none.
         MS. DIMITRIJEVIC:  None.  Are you saying low or none?
         DR. BONACA:  None.  Just taking an extreme case.  That tells
     me that you are going to have failures at some point of some of the
     piping.
         MR. MITMAN:  The only thing that we've seen categorized with
     no consequence are those abandoned in place piping.
         MS. DIMITRIJEVIC:  So we never saw high degradation in none.
         MR. MITMAN:  Now, if you have low consequence, which could
     be extremely low, 1e-to-the-minus-ten or whatever, we're still bending
     that in a medium risk category.
         DR. BONACA:  And that is fine.  I just was worried more
     about the corner, because simply you're going to see more failure.  From
     a regulatory standpoint, a utility that commits t this program and then
     has failures in the field, you get a black eye anyway ultimately.
         So I understand now that that box on the top left corner is
     not really one that should be of concern.
         MR. MITMAN:  If we started to see a lot of pipe segments
     falling into that segment, we'd have to ask a lot of questions about
     whether that's appropriate or whether we should take a harder look at
     it.
         DR. BONACA:  Because although you have -- from a risk
     standpoint, it's fully convincing to me that there are other issues that
     have to do with how do you deal with actual failures there and how are
     they classified and is it part of corrective action programs and how is
     that going to fit back into the adequacy of a program of this nature
     that you implement.
         It will raise all kinds of issues within a power plant if
     you begin to have those kind of failures.   
         MR. RICCARDELLA:  I know for the Class I programs, which is
     most of the pilot studies we did, nothing has ever fallen in that
     category.  Every Class I -- all the Class I piping has either low,
     medium or high failure consequence.
         MR. MITMAN:  Carl?
         MR. FLEMING:  Carl Fleming, from ERIN.  With regard to the
     earlier discussion we had about the insights from service experience and
     the different kinds of loading mechanisms and degradation mechanisms
     responsible for service data, those insights are responsible for both
     leaks and ruptures.
         So the concern that reducing volumetric inspections might
     result in an increase in the frequency of leaks in piping does not seem
     to be supported by the insights from in-service data.  It indicates that
     they'll probably occur at about the same rate as we're seeing now,
     because they're due to mechanisms that are just not being captured by
     the inspection process.
         DR. BONACA:  Thank you.
         MR. MITMAN:  My summary and concluding slide.  The revised
     topical has been submitted recently.  It addresses questions and
     concerns, lessons learned that we've learned along the way in the
     application of the pilot process, the methodologies and compliance with
     Reg Guide 1.174 and Reg Guide 1.178.
         The methodology has been applied to a diverse and extensive
     group of plants, GE BWRs, Westinghouse, B&W, and CE PWRs, multiple AEs,
     and both full and partial scope.
         So we've got a very broad, diverse, I think a very good
     distribution of the pilots and there has been a lot of feedback into the
     methodology from that.  We are seeing significant rem reductions out of
     the application and the pilots and research support the conclusions of
     negligible risk impacts by application of the methodology.
         DR. SEALE:  You mentioned earlier that you had a process by
     which you did some follow-up to find out if there are any changes or
     wild trends that might show up in the results of the application of
     this.
         Is there any systematics to that follow-up and if so, what
     are they?
         MR. MITMAN:  The final step in the process is this
     performance monitoring that hangs out to the side here on the feedback
     loop.  EPRI has been doing database analysis.  We continue to watch the
     industry and the LERs and the reports from the industry.  That's all
     part of the process.  The process really isn't formalized at this point,
     but it's there.
         There is a discussion about setting up something that might
     be equivalent to a user's group that would be also fed into the
     monitoring process and we're also expecting each individual plant that
     applies the methodology to watch what's happening in their plant and in
     the industry and to go back and revise, change, correct their own
     application of the methodology.
         DR. SEALE:  I noticed you mentioned that the owners groups
     for the different vendors have been participants.
         MR. MITMAN:  Well, the CE owners group has not, per se, been
     part of this.  The vast majority of the CE plants have been part of an
     EPRI tailored collaboration, but not, per se, the CE owners group.
         DR. SEALE:  Okay.  But in the other cases, the owners groups
     may have their own impromptu groupings or reviews, as well, I guess.
         MR. MITMAN:  Obviously, Westinghouse and the WOG have their
     own methodology and their own process there.  There is no discussion at
     this point with either the BWR owners group or the B&W owners group
     about having any kind of feedback loop between the owners groups and
     EPRI.  There is -- it's on a plant by plant basis.
         That's all I had.
         DR. SHACK:  Any more questions?  They're not scheduled to
     start until 11:15.  Can we go ahead with them?
         MR. MARKLEY:  Yes, we can.  We're not changing subject here. 
     So that's fine.
         DR. SHACK:  Thank you.
         MR. ALI:  My name is Syed Ali, I am from the Division of
     Engineering in NRR.  With me is Steve Dinsmore, who is from the Division
     of DSSA in NRR.
         What we are going to try to do in the short time that we
     have is to basically give an overview of the status of the review of the
     topical report and also the associated pilot plants, and a little bit of
     the changes that EPRI has made in their topical report since we approved
     some of their pilots.
         Some of these items EPRI may have already discussed this
     morning or may have been discussed as a result of questions or just to
     complete the picture, we will go over the status of what we have done so
     far, again.
         EPRI submitted its draft topical report back in '96.  At
     that time, we were also in the process of developing the regulatory
     guide and the standard review plan for the risk-informed ISI.
         We issued some questions and requests for additional
     information the middle of '97.  Subsequent to that, EPRI was involved in
     developing the pilot applications and submitting the pilots, so their
     priority was more toward the pilot submittals. So they came back with
     their responses to the RAIs and the questions and comments that we had
     sent near the end of last year.
         That was also about the time that we had finished basically
     the regulatory guide and the standard review plan and also some of the
     pilots.  So they were able to utilize the lessons learned not only from
     the pilot study they had done, but also from the regulatory documents
     that we had issued.
         I think since then, this process has picked up and our
     interface with EPRI has become much more significant.  We had a meeting
     back in March to discuss the responses and there were a few additional
     issue.  EPRI responded to those additional issues and then submitted
     their current report or their new version of the topical report 11267,
     which is the one that you are reviewing and that we have been reviewing.
         Those are the items that have been completed.  We are having
     the ACRS subcommittee meeting today.  We intend to have a follow-up
     meeting with EPRI sometime in July.
         In addition to that, since right now we are actively
     reviewing the EPRI report and some of their backup reports, we have a
     weekly telephone sort of update as to the issues and the items that are
     ongoing.
         We plan another meeting or another presentation to ACRS in
     their September meeting and that's when we will be able to present our
     SER.  One thing that you see different here is that since the ACRS
     meeting is in September, we think that we will be able to issue the
     final report by the end of October and not the end of September, as you
     saw in some of the EPRI slides.
         That schedule was based on having an ACRS presentation in
     August, which is not on schedule at this time.
         DR. SHACK:  Are you still going to have a draft SER in June?
         MR. ALI:  We still plan to have a draft SER in June.  So we
     should be able to give you a draft SER.  Now, whether that draft SER
     will be a clean SER in the sense that it may have some open items, we
     don't know.  But probably it will have some open items, but we will have
     a draft SER.
         If we proceed the same way that we did with the
     Westinghouse, then we will probably have a draft SER with maybe a few
     open items and then have the meeting with EPRI to try to resolve those
     open items and then have a final SER by the time you have your September
     meeting.
         The next slide was just a listing of the pilots.  EPRI
     talked about those already.  Also, we have completed two pilots, Vermont
     Yankee, which is a GE BWR, and on that it was applied only to Class I
     utilizing the code case N-560.  We have also completed the review of
     ANO-1 Unit 2, which is a CE PWR, and we are currently reviewing the
     submittal for ANO Unit 1.
         DR. SHACK:  Now, those two pilots, the application basically
     followed this topical, although this topical didn't show up until April.
         MR. ALI:  It followed the topical that they initially
     submitted.
         DR. SHACK:  The old topical.
         MR. ALI:  Old topical, but then there were additional things
     done as a result of that review.  So a lot of that is reflected in the
     changes that were made in this topical.
         So basically this topical utilizes the changes or the
     lessons learned from the pilot applications and also from the reg guide
     and the standard review plan that we issued that was subsequently the
     first topical report.
         MR. MITMAN:  This is Jeff Mitman from EPRI.  Both the pilot
     plants that have been completed, Vermont Yankee and ANO Unit 2, are
     consistent with the new topical that you have in front of you.  There is
     nothing that they have done in those pilot plants that is different than
     what we have proposed in the revised topical.
         MR. DINSMORE:  This is Steve Dinsmore, from the staff. 
     There are more things in the topical than were approved in the pilot
     plants and I think the next two slides, we're going to go through the
     differences between what was approved and what's in the current topical. 
     So there are some differences.
         MR. ALI:  What we're going to go through are some of the
     major changes in the topical report rather than detailed item by item
     changes subsequent to the approval of the pilots.
         One change, I would say, an updating of the topical report,
     is that in looking at the first version of the EPRI topical report and
     also some of the discussions we had and the questions we had been
     sending, it was not clear as to how the augmented programs are
     integrated with the risk-informed ISI process.
         So one thing that they have done in this topical report is
     to clarify as to which of the augmented programs are, at this point,
     part of the risk-informed ISI process and which of the augmented
     programs are still being looked at as the way the licensees had made
     commitments to the NRC as a result of the degradations found either on
     an industry basis or on a plant-specific basis.
         DR. WALLIS:  This augmented part means it's increased,
     augment is to increase.
         MR. ALI:  What the topical report and the current EPRI
     position is that, for example, for IGSCC, category A will be most into
     the risk-informed ISI process, but for category B through G, the
     risk-informed ISI process at this time will not change the inspections
     that are being done for category B to G.
         They are not being increased or decreased.  They are just
     staying the same.
         DR. WALLIS:  Augmented implies increased.
         MR. ALI:  Augmented means in addition to what is in ASME
     Section 11.
         DR. WALLIS:  So that's the implication.  It's in addition to
     that.
         MR. ALI:  In addition to ASME-11 inspections.
         DR. WALLIS:  So on the face of it, it looks as if you're
     increasing some demands by augmenting inspection.
         MR. ALI:  No, that's not what is meant.  There were a number
     of -- there's a number of degradation mechanisms that were found in the
     plants over the last several years.  As a result of those, the staff had
     issued generic letters asking the licensees to address those issues and
     as a result of that, the licensee made certain commitments as to how
     they will monitor those degradation mechanisms and inspect for those.
         The second item is --
         DR. SHACK:  Syed?
         MR. ALI:  Yes.
         DR. SHACK:  An erosion/corrosion, is that augmented
     inspection program now going to be subsumed into the --
         MR. ALI:  No.  That, at this point, is going to remain as it
     is in the generic letter.  So those IGSCC B through G and the
     erosion/corrosion are the two programs that are not being changed as a
     result of this program at this time, and other programs are being
     changed.
         The second item is that as a result of our issuing the
     regulatory guide and standard review plan and also reviewing the pilots
     and issuing the SER, the staff and the industry felt that once we have
     reviewed the pilots and we have reviewed and approved the topical report
     for these two methodologies, then the industry and the staff should come
     up with a simplified submittal to the staff so that the review process
     can be expedited.  
         We had several meetings with the industry to try to come up
     with the contents of that template.  As a result of those meetings, we
     agreed to a table of contents and what will go into those templates.
         The only thing I think that's a little bit different is that
     as a result of our meetings with NEI and the industry, there were some
     changes that were made to that template, there were some attritions and
     the template that we see in this topical report does not reflect the
     latest.  So I think that's something that we will ask them to do and I
     don't think that should be a big change.  That can be done easily.
         DR. SHACK:  Do you have a template for the WOG ASME?
         MR. ALI:  Yes, we have for both.  Actually, the template is
     about the same because the high level aspects of the methodologies are
     really the same.  It's the details.  So the template is essentially the
     same.
         The third item on this slide is that so far, all of the
     pilots that we have looked at and approved have been either the whole
     plant or Class I.  So we have not been asked to review and we have not
     reviewed any submittals where there was a change in the inspection on a
     system by system basis.
         But the topical report presents that as one of the options
     that the ISI program may be changed on a system by system basis.  They
     did not talk about that this morning.  The criteria, the risk criteria,
     the risk acceptance criteria on a system by system basis is an order of
     magnitude more stringent than what it is for the plant basis.
         This is something that is different than what we have
     approved at this point and so we are still in the process of reviewing
     that.
         Steve will now go through the next slide.
         MR. DINSMORE:  This is Steve Dinsmore from the staff.  I
     will be talking about some of the differences in the -- I guess most of
     them are in the delta risk calculations.  Again, these differences are
     between what was approved in the pilots and what's in the current
     topical.
         We talked a lot about those tables, but we haven't been
     looking at the specific elements in those tables and what those rankings
     are and the word grade creep comes to mind.  Every pilot has been
     submitting their own table and the highs and the mediums are kind of
     drifting off to the left-hand side.
         So we haven't quite approved the table yet.  We're still
     looking at it.  Some of the elements in there are greater than
     ten-to-the-minus-four, which are supposed to be mediums, because they
     say it's a bounding calculation.  So we're still dealing with that.  So
     if you look in the table and you see numbers which don't seem to fit
     with the criteria, they don't, and we're working on that.
         The other big change is these delta risks.  Again, earlier,
     when they first started this process, they didn't really look at delta
     risks and we've been adding that slowly step by step.
         What's new in the topical, again, is there is a request to
     do no delta risk calculations if you only do Class I.  EPRI said that
     they had an argument why ten percent would not increase risk and that
     there -- if the pilot or if a utility could somehow show that their
     service experience is similar to generic service experience, then they
     wouldn't have to do the delta risk calculation.  So we're still looking
     at that.
         No delta risk contribution from low safety significant
     segments.  Carl kind of alluded to that this morning.  They don't want
     to calculate the change in risk due to low safety significant segments. 
     They have a reasonable looking argument in the topical where they use
     bounding calculations, but we're still looking at that, again.
         They added these screening criteria.  This is actually the
     first time I've seen those.  We've only had this thing for about two and
     a half weeks.  The ten-to-the-minus-seven CDF and the
     ten-to-the-minus-eight LERF, these are all reflected in a flow sheet on
     the last -- there is a flow sheet right in the back of the chapters. 
     Evidently, it's a system by system restriction which you use if you've
     done either by class or by -- or the whole plant.
         Then they added on this extra one where they say if you only
     do one system, then you decrease those by a factor of ten.  In general,
     with the risk-informed stuff, we've been -- we started out by saying you
     had to do full scope, everything, and we've been kind of slowly moving
     back.  Especially if they come in with the argument that it's a risk
     decrease, it's hard for us not to accept something that's not a risk
     decrease.
         Again, these are positive numbers here.  So we're still kind
     of looking at that.
         And the last one that they added was this Markov.  There was
     a lot of discussion this morning about this Markov and the data
     analysis.  We didn't use that in the pilots.  We used this bounding
     option, where they took the highest possible CDF for each category and
     the highest possible pipe failure frequency for each category and did
     some bounding calculations to show us that without crediting any
     increase in inspection efficiency, that there would be a very small risk
     increase.  If they credited some increase in the inspection frequency,
     then it would go down.
         So they've been coming in with minus numbers as their best
     estimate of the change in risk, which, again, I'm not quite sure what
     these positive ten-to-the-seven and ten-to-the-minus-eight boundary
     criteria are.
         A more mundane problem with the Markov stuff is from Dr.
     Apostolakis -- it's spread all over about ten reports and we're trying
     to pull it together.  So we don't really have a good feel for how the
     whole thing fits together with the data.
         But our current feeling is that the process seems
     reasonable.  It's just a matter that we have to really get and
     understand it and say that what we understand is okay.
         Those are essentially all the differences that we've
     identified.
         DR. SHACK:  Questions from the committee?  Do you need
     anything from us at the moment?  You're still working on your draft SER.
         MR. ALI:  We kind of talked about that before.  We usually
     ask for a letter when we write the SER, but we haven't written the SER. 
     So I think at this point, it's --
         DR. SHACK:  Unless we have a major problem.
         MR. ALI:  -- what you write as a result of your meeting or
     your deliberations.
         DR. SEALE:  Strictly information.
         MR. BARTON:  I don't see any big problem that would require
     a letter.
         DR. SHACK:  Okay.
         DR. SEALE:  So you're going to talk again this afternoon?
         MR. ALI:  Yes.  We are on schedule again this afternoon.
         DR. SHACK:  We had some discussion about that, since we've
     got a reasonable segment, we basically sort of thought we more or less
     had to do it, but we would do it in a fairly abbreviated fashion.
         DR. SEALE:  Yes.  I notice it's just a half-hour.
         DR. SHACK:  That was more we sort of felt we had to -- you
     know, since it was noticed, we would have to do it, but it would be an
     abbreviated session.
         With that, if there are no other questions or comments, we
     close the meeting of the subcommittee.
         [Whereupon, at 11:16 a.m., the meeting was concluded.]
Page Last Reviewed/Updated Wednesday, February 12, 2014