463rd Meeting - June 2, 1999
UNITED STATES OF AMERICA NUCLEAR REGULATORY COMMISSION *** ADVISORY COMMITTEE ON REACTOR SAFEGUARDS MEETING: 463RD ADVISORY COMMITTEE ON REACTOR SAFEGUARDS (ACRS) *** U.S. Nuclear Regulatory Commission 11545 Rockville Pike Conference Room 2B3 Two White Flint North Rockville, Maryland Wednesday, June 2, 1999 The subcommittee met, pursuant to notice, at 8:30 a.m. MEMBERS PRESENT: DANA A. POWERS, Chairman, ACRS GEORGE APOSTOLAKIS, Member, ACRS ROBERT L. SEALE, Member, ACRS GRAHAM B. WALLIS, Member, ACRS THOMAS S. KRESS, Member, ACRS MARIO V. BONACA, Member, ACRS ROBERT E. UHRIG, Member, ACRS WILLIAM J. SHACK, Member, ACRS P R O C E E D I N G S [8:30 a.m.] DR. POWERS: The meeting will now come to order. This is the first day of the 463rd meeting of the Advisory Committee on Reactor Safeguards. During the day's meeting the committee will consider the following: hydrogen control exemption request for the San Onofre Nuclear Generating Stations, Units 2 and 3; status of the pilot application of the revised Inspection Assessment Programs; proposed risk-based performance indicators, performance-based regulatory initiatives and related matters; proposed ACRS reports. The meeting is being conducted in accordance with the provisions of the Federal Advisory Committee Act. Dr. John T. Larkins is the Designated Federal Official for the initial portion of the meeting. We have received no written statements or requests for time to make oral statements from members of the public regarding today's session. A transcript of portions of the meeting is being kept and it is requested that speakers use one of the microphones, identify themselves and speak with sufficient clarity and volume so that they can be readily heard. We have a special event to note today. Michelle, could you stand up? Michelle has been selected to receive a Meritorious Service Award for support staff excellence. [Applause.] DR. POWERS: Many of us, of course, have worked with Michelle for several years and know that it's just overdue, that's all that is. DR. SEALE: Why so long? DR. POWERS: Michelle, congratulations. When do you actually receive this? MS. KELTON: The 16th. DR. POWERS: June the 16th. Very good. DR. SEALE: And exactly where is this party going to be? [Laughter.] MS. KELTON: Outside on the lawn. A picnic. DR. POWERS: And with the monetary award associated with that you get a cup of coffee in the cafeteria, is that -- [Laughter.] DR. POWERS: Well, very well deserved and congratulations. Members will notice that John Barton is not with us today and that is a problem for us, because as usual we have laden down John with more work than anyone possibly deserves, so Dr. Kress, can you take responsibility for the low power in shutdown program? DR. KRESS: Certainly. Sure. DR. POWERS: And Dr. Shack, if you can take care of the inspection and assessment program. I will also remind members that we have a reports and letter scheduling here where we indicate discussion times and preparation times. I think it is important that we do have draft letters available for the preparation. It's nice we have them for the discussion but not essential. We should have them available for the preparation time. In that regard, I'll call members' attention to the schedule, especially the time of four o'clock today and four o'clock on Thursday. There is a new approach towards scheduling that should make it possible for members to have draft reports available. That is an experiment in scheduling that we are attempting and we will see how well it works out. On additional items of interest, I'll call your attention to a speech by Shirley Jackson entitled, "Teamwork: Individual Commitment to a Group Effort." I think that will give you some idea of the range of accomplishment by the Staff and by the Commission during Dr. Jackson's tenure as the Chairman. With that, do any members have any additional comments they would like to make? [No response.] DR. POWERS: Then I think we can turn to the first item of business, which is the hydrogen control exemption request for the San Onofre Nuclear Generating Station, which has the nicest acronym of any of the nuclear power plants -- SONGS. I love that. Tom, I think you are the cognizant member on that subject? DR. KRESS: Yes. Thank you, Mr. Chairman. We did have a subcommittee meeting to hear about this last week. You will hear an abbreviated version of what the subcommittee heard. This is an interesting request because it is once again another test of the NRC resolve to grant exemptions based on risk criteria. It could very well have been granted under just the regular exemption rules. I think that is 10 CFR 50.12. But because it meets all the deterministic regulations, they also chose to show the risk-informed part and it's interesting because it is one of those cases where it doesn't affect CDF. The request is to remove all of the hydrogen control in the SONGS plant, except for there's four parts of that. There is a purge-and-vent, there's recombiners, there's a monitoring system, and there's also a mixing system to mix it up inside containment. Well, the request is to remove all that except the monitoring. I think they are going to decide to retain the monitoring so that they at least know when hydrogen is there. The interesting part about it is it doesn't affect CDF. If it has any effect at all, it is LERF and a technically appropriate way to do that would be to look at the uncertainties in the loads and the uncertainties in the failure of the containment and see whether they overlap and get kind of a probability of a delta LERF, but this is another example of -- you know, that's a real sophisticated PRA application and this is an example where qualitative analyses are good enough. You don't really need a PRA here. You can show that this, by just qualitative analysis that the recombiner systems or that overlap is so small that it's just hardly even worth considering, so with that as a -- DR. POWERS: Before we go on, let me ask this question. We consider CDF and we consider LERF in the probabilistic types of analyses, but there are a variety of other safety functions built in to the regulations that don't get addressed. Do we have any issues associated with those? DR. KRESS: You would have if you were just looking at CDF and LERF. I think there are other regulatory objectives that have to do with the fission product release in general and that LERF and CDF doesn't fully capture that. In this case though, I mentioned they meet all the deterministic rules and the deterministic rules are intended to capture some of those other regulatory objectives. As long as you meet those and don't really compromise defense-in-depth, then I think you are all right, so I think for this application where you have a large dry containment, I think those concerns are not there, but they could be for other applications, so we don't want to view this as a blanket approval of removing hydrogen control systems from all plants, but with that, why I think I will turn it over to the SONGS people. Ed Scherer I think will introduce the SONGS presentations, so the floor is yours -- up front, and please remember to identify yourself before you start for the benefit of our recorder. MR. SCHERER: Thank you and good morning. My name is Ed Scherer I am the Manager of Nuclear Regulatory Affairs at Southern California Edison, and I would like to open with a few remarks, most of which have been very effectively covered by Dr. Kress. I would like to briefly cover the scope of our application and then Tom Hook, who is the Manager of our Nuclear Safety Group, with the responsibility for the PRA analysis and our safety monitor operation at the plant will cover the basis and the details of our application. As indicated, the scope of our Hydrogen Control System are four parts: the hydrogen recombiner, the purge subsystem, the monitoring subsystem, and the mixing system inside the containment. The next slide shows the scope of our exemption request. At this point we are asking for the removal of the regulatory requirements on the hydrogen recombiners and the hydrogen purge. If approved, we would not remove the equipment from the plant. We would remove them from the FSAR and the technical specifications. We would reclassify the equipment as nonsafety-related. And while we would remove them from the EOIs, we would not remove them from our severe accident management guidelines, but would maintain them in the severe accident management guidelines with the caveat that as they are available, they would be utilized. So basically -- and we have also -- at the request of the staff we've agreed that we would notify the staff should we ever abandon our efforts to maintain the equipment that we already have. As part of another application, we -- in this application we had originally asked to remove the hydrogen monitors. We thought our risk-informed application had justified that. But as part of another application, we do intend to maintain the hydrogen monitors, so we've asked the staff not to take action on our request to eliminate hydrogen monitors as part of this application. DR. POWERS: What type of monitors do you have for hydrogen? MR. HOOK: The monitors are an electrochemical-type cell that utilizes an electrolyte in a permeable membrane through which the hydrogen from the containment atmosphere passes through. MR. SCHERER: And it requires some temperature compensation during operation, but -- and its range is 0 to 10 percent. The original scope or the original reason we made this application was not because it was the most expensive system to maintain, but because we believed it would serve as a basis for other exemption requests. It was considered a Task 0 of an NEI whole-plant study which had three applications, the first one of which the staff granted to Arkansas Nuclear 1, which was the approval to go from a 30-minute to a 90-minute requirement. Ours was the second part of that application. It is also an application that's consistent with the modifying of Part 50 to be more risk-informed, and we did go and look, under a risk-informed maintenance rule the hydrogen control system would not be risk-significant and would be outside the scope of any risk-informed maintenance rule. Basically the conclusions we reached in making our application, and Tom will cover these in some detail, is that we found the hydrogen control system is not needed for accidents based on realistic assumptions. In other words, when you use realistic assumptions, the hydrogen recombiners don't play a significant part in reducing the risk. In a severe accident, hydrogen is theoretically generated so fast that the hydrogen recombiners would not meaningfully keep up. The hydrogen purge is also not needed for a large dry containment such as ours, and when we say potentially detrimental, that's always the concern -- when you open a purge of the containment, there's always the finite chance that you can't reestablish that, even though we have dual isolation, and that would be of course a risk of creating a leak path that we didn't intend. We would also eliminate hydrogen control steps and activities from the emergency operating instructions, which we think will simplify those instructions, especially in the early days of trying to mitigate a severe accident. And we think that would be a risk-positive change, in that it would eliminate distractions which would otherwise exist. DR. POWERS: I'm fascinated by the term "risk-positive." I assume that you mean a reduction in risk. MR. SCHERER: Reduction in risk. DR. POWERS: There must surely be better nomenclature. MR. SCHERER: Thank you for the comment. It's always difficult when you're talking about risk to know which way is the positive direction and which one's the negative direction. DR. SEALE: There's a little oxymoron in there. MR. SCHERER: I think we bought that when we used the word "risk" instead of "safety." Let's say it would have been a safety-positive change. Finally, the reduction in regulatory burden would be a positive to us, but not that meaningful. We thought that the principles that were involved here were more important to us than the financial savings that may or may not accrue, but there are some safety benefits and financial benefits that would, and therefore we have asked the staff to proceed based on that basis. If there are no questions on my introduction, let me introduce Tom Hook, who will go through the basis we used for providing our exemption request. MR. HOOK: My name is Tom Hook from Southern California Edison Company, the manager of the Nuclear Safety Group, under which I have probabilistic risk assessment responsibilities as well as independent safety review. I'm going to discuss the basis for this exemption request and how we came up with the -- met the acceptance criteria in Regulatory Guide 1.174 for a risk-informed regulatory application. I'm going to describe three different areas in the reg guide and how we meet the criteria for those aspects of the exemption request: the defense in depth, the safety margins, and the risk impact. Turning to the next slide, slide 8, first of all in defense in depth in terms of fission product barriers, the cladding, the fuel cladding, the reactor coolant system pressure boundary and the containment pressure boundary, this exemption request clearly affects the containment pressure boundary as a part of the defense in depth. We concluded as a result of our evaluation of removal of the hydrogen control systems from the plant design that they do not challenge containment integrity. This is based upon an assessment that the hydrogen control system is not needed for realistic assumptions for accidents, and we're going to describe what those are. The hydrogen control system is not sized to mitigate severe accidents in terms of the amount of hydrogen that's generated in a core uncovery situation. The Zircaloy cladding oxidation clearly produces orders of magnitude more hydrogen than the hydrogen control systems are able to accommodate in terms of removal from the containment atmosphere. And the hydrogen control system is only valuable for the unlikely design basis accidents which are beyond the Appendix K design degradation of the core, but less than a severe accident that's typically evaluated in a probabilistic risk assessment, and IPE or IPEEE. DR. POWERS: When you do your design basis analysis, you're looking at hydrogen that comes from say 17 percent of the clad oxidizing? MR. SCHERER: Yes. DR. POWERS: And what other hydrogen sources do you consider? MR. HOOK: The hydrogen sources are the actual aluminum and zinc in the containment, and the radiolysis from the fission products in the containment sump and in the core. DR. POWERS: And when you do those analyses and you say it reaches some percentage, that must be some percentage at some particular time. MR. HOOK: That's correct. DR. POWERS: And what is that particular time? MR. HOOK: The particular time here is within 30 days. DR. POWERS: Thirty days. MR. HOLAHAN: This is Gary Holahan of the staff. I'd like to clarify, the design basis requirement, the 17-percent oxidation is for the peak. DR. POWERS: I understand. MR. HOOK: And 5 percent for the total. DR. POWERS: The point is, Gary, that we go on for some time. These things don't -- he has some continuing sources. His aluminum, his radiolysis, and his zinc oxidation are continuing in time, so, I mean, presumably if you go on long enough, those numbers get very high. If you were to allow me to go five years, I presume I could get well over 4 percent. I might even then get up to the downward flammability limit. And when the staff talks to us about catalytic recombiners and says they don't know what day to use, well, these gentlemen know what time period to evaluate those things. MR. SCHERER: And we'll show you some parametric studies that we've done to try to bound that. DR. POWERS: And the kinetics on oxidation of the metals, the aluminum and the zinc, is a challenging undertaking but probably not very important to your numbers. MR. HOOK: That's correct. They're a small fraction of the contribution. The long-term contribution predominantly comes from radiolysis in terms of the buildup, slow buildup for which the recombiners and the purge were designed to accommodate. DR. POWERS: When you do the radiolysis, you do it for pure water? MR. SCHERER: Do we know? MR. HOOK: I don't know. We utilize a Bechtel code called hydrogen which meets the requirements and the assumptions in the Reg Guide 1.7 in terms of the processes that are occurring. I'm not familiar with the specifics of how it -- what assumptions about pure water. DR. KRESS: The g factor in there is higher, bigger than you expect for pure water by almost a factor of 2. DR. POWERS: By a factor of 2. DR. KRESS: So I don't, you know, you couldn't say it's pure water. DR. POWERS: Okay. So, I mean, basically you use a g factor to calculate the release and assume that there's nothing in there that combines -- yet there would be. MR. SNODDERLY: Dana, I'd also like to -- this is Mike Snodderly from the staff -- that there is no 30-day requirement. Right, it's just that you do not -- DR. POWERS: I understand; 30 days is as good a number as any. MR. HOOK: Looking at the table at the bottom of slide 8 we show for a realistic, a design basis, and a beyond design basis the volume percent hydrogen, the first row without hydrogen systems -- or which hydrogen system's available in the containment, and we see for the realistic that the hydrogen generated in 30 days is less than 4 percent. For the design basis we see it's also less than 4 percent because of the removal of the hydrogen from the containment either utilizing either the recombiners or the purge, and beyond design basis it's much greater than 4 percent. And without hydrogen control systems, we see for the realistic that it is less than four percent for the realistic and greater than four percent for the design basis and much greater than four percent for the beyond design basis or severe accident. In terms of the -- DR. POWERS: Four percent is an interesting number if you have a well-mixed containment and part of your system is a mixing capability, but now you take that mixing capability out. Do you have a problem of stratification? DR. KRESS: I misspoke. The mixing system stays in. DR. POWERS: Stays in? DR. KRESS: Yes. MR. SCHERER: Yes, we are not intending to make any changes in our mixing system. DR. POWERS: So you can keep things relatively uniform? MR. SCHERER: Yes. MR. HOOK: And we have design features in the plant that I will describe in a little bit of detail that ensure that there is adequate mixing in the SONGS type containment, and in terms of characterization of the usefulness of the individual systems, the hydrogen recombiners in terms of realistic accidents are not useful because they are not needed. Hydrogen purge is also not useful but not detrimental because of the release it would create if it was used is small. In terms of the design basis, the recombiners are useful and hydrogen purge could be useful, but also would represent a filtered release path and in terms of beyond design basis, the recombiners would not be useful since the emergency operating instructions and the severe accident management guidelines instruct us to turn off the recombiners at 3.5 percent. The hydrogen purge would potentially be detrimental because it is not designed to operate at pressures above several psi and the ability of the valves to reclose upon a purge is not guaranteed and is not evaluated and is not engineered, so it could -- DR. SEALE: Excuse me. You just mentioned something that was kind of intriguing in the subcommittee meeting that we had, namely that at 3.5 percent you instructions are to turn the recombiners off and it was -- let's say the thought was, at least by some, it might be desirable to leave them on, because then they would serve as igniters while you were still in the low or non-explosive mixture range and purse the system in that way. Do you have any more thought on that? MR. HOOK: No. Since we had our discussion last Wednesday, we wanted to go back to Combustion Engineering and have them look at that and the benefits of leaving the recombiners on, and that is something that we plan to do, but we haven't had an opportunity to do that yet, and that is something we are taking under advisement because it does appear to be a beneficial use of the recombiners in a beyond design basis type event. DR. POWERS: I think I would approach that with a great deal of caution. DR. KRESS: Yes, that was going to be my comment. If you need igniters, you put in igniters. DR. POWERS: Yes. I look at it from the perspective of source terms and the idea of having hot spots in an atmosphere where I have iodine-carrying aerosols may cause us pause. MR. HOLAHAN: I think the Staff wants to take up this issue with the Owners Groups as well, and make sure we have a common understanding of what is recommended for accident management guidelines. DR. SEALE: My curiosity is assuaged for the moment. DR. APOSTOLAKIS: What exactly do the words "not useful" mean? MR. HOOK: "Not useful" means that they either are not able to mitigate the generation of hydrogen to that which is presumed to be acceptable, below four percent, or either they are not needed within 30 days to ensure that the hydrogen concentration in the containment is below four volume percent. DR. APOSTOLAKIS: I don't remember now -- in 1.174 what was the requirement regarding defense-in-depth? What were the words? Not be affected at all? Or the philosophy? DR. POWERS: Surely you remember that. DR. APOSTOLAKIS: The last few words there are philosophical, I think. MR. HOOK: And I will pick up in terms of the safety margin a discussion about the adequacy of the containment design to address a potential hydrogen burn in the containment. Moving on to Slide 9, we performed -- DR. APOSTOLAKIS: I guess the motivation for my question was what if it had some impact? Then what you have done? Let's say the measure -- I mean this is not for future applications. If you did affect defense-in-depth a little bit, does that mean we give up or you would try to argue that you still have sufficient defense-in-depth? MR. HOOK: I believe based upon the guidance in the Reg Guide we would look at the likelihood of the loss of that defense-in-depth, the likelihood of the loss of the containment cooling systems in concert with the generation of hydrogen and the likelihood that we would have a containment challenge. DR. APOSTOLAKIS: The only problem I see with that is that maybe the PRA would not be able to quantify small impacts. That would be lost in the noise there, so -- DR. KRESS: I think this is one case where it would be actually lost in the noise. MR. SCHERER: And philosophically on a risk-informed submittal we would not make an argument of the zero standard on anything -- defense-in-depth or risk. There is no zero standard that we would like to see us being held to, but a reasonable standard -- insignificant in terms of risk or insignificant in terms of compromising the philosophy of defense-in-depth. If we thought it was significant compromise then we would not make an application, but I don't think we would view this against any zero standard, no compromising of the philosophy of defense-in-depth. DR. APOSTOLAKIS: So in that sense you would argue that you still have sufficient defense-in-depth? MR. SCHERER: Sufficient, and that it is not a significant compromise. DR. WALLIS: Although your arguments for significance are entirely qualitative. There is no yardstick you are using to say whether or not it is significant. I don't want a zero standard, because that's ludicrous, but then you don't have a measure of how far from zero we are. MR. HOOK: Well, you will see later in the presentation that we looked at worst case events from a deterministic point of view and the pressures that you would see from a hydrogen burn at those concentrations of hydrogen, and concluded based upon a combined probabilistic, deterministic evaluation of the containment that the containment would be able to withstand those severe accidents. DR. POWERS: It's a big horse containment. DR. WALLIS: I know at the subcommittee meeting we talked about the 95th percentile and so on and there is some effect on risk. MR. HOOK: Of course. DR. WALLIS: But it is not really quantified in your submittal. MR. HOOK: It's very difficult. Moving on to Slide 9, we looked at a number of parametric evaluations of the hydrogen generation, considering first the design basis, utilizing the assumptions in Reg Guide 1.7 and the Standard Review Plan 6.2.5. And we looked at three cases where we revised the assumptions involving the hydrogen generation rate. We looked at the metal-water reaction conservatism that is included in Reg. Guide 1.7, a factor of 5, the radiolysis generation rate, the 20 percent conservatism that is in the Standard Review Plan, Appendix A, and we removed those from the calculation, the hydrogen code that we used that was developed by Bechtel, and reevaluated the hydrogen generation in the containment over a period of time. We also looked at Case 2 where we revised the sump radiolysis yield from .5 molecules hydrogen per 100 EV to .4, which is a mid-point between the Reg. Guide and a number of references from Westinghouse and Oak Ridge National Labs for the hydrogen generation. And in Case 3 we looked at the same parameters as Case 1 except revising the sump radiolysis yield from .5 to .3. So we evaluated, removed some of the conservatism from the Reg. Guide 1.7 and Standard Review Plan analysis and came up with the results in Slides 10 and 11. Slide 10 is a table showing the values for hydrogen concentration without any hydrogen control in the containment at 20 days, 25 days and 30 days. And we see for the design basis case that we exceed the lower flammability limit of 4 percent at approximately 13 days, and at 20 days we are 4.6 percent and so on. For Cases 1, 2 and 3, we see, regardless of the assumptions we make about the hydrogen generation rate and the conservatisms that were removed, that we stay at or below 4 percent in 30 days for all of these cases. And you see in slide -- DR. POWERS: When you do these evaluations, you are getting -- you have a background level of hydrogen coming from some assumptions that you have made about metal-water reactions in the core region. And then you have a continuing source coming from the water radiolysis primarily, low within the containment. And my recollection of the containments at SONGS is that you have a large open area up at the top, but down below the operating deck we have a warren of rooms and sumps and places like that where water accumulates. Do you have a problem of hydrogen concentration in these lower rooms? MR. HOOK: As a part of the individual plant examination, we conducted an extensive walk-through of the SONGS containment and concluded from that, and that was a walk-through that was attended by our contractors as well from Fauski Associates and Erin Engineering, and concluded that the SONGS containment, based upon its design, has vent paths for all of these lower compartments to the upper compartment, and what we would characterize as generous paths. As well as the mechanical moving mechanisms, the dome recirculators, the containment fan coolers that would assure that these lower areas are vented to the -- and there is good mixing to the large, open, upper area above the operating deck. So that was specifically looked at as a feature, as a part of the individual plant examination, and it is not a concern at San Onofre in terms of pocketing in the lower compartments. DR. KRESS: Dana, I raised this issue at the subcommittee meeting and my comment there was, once again, it would have been nice to have some of a code to calculate hydrogen distribution in containment. DR. POWERS: Wouldn't it be nice. DR. SEALE: Deja-vu. MR. HOOK: In Slide 11 we see the hydrogen generation rate over time, and you see that the design basis in the black line -- DR. WALLIS: Excuse me. This business of the walk-through about the venting, this is just a bunch of people looking at the room and saying, gee whiz, we think that is well enough vented? Isn't there some sort of analysis made of this? MR. HOOK: There is no quantitative analysis, it is a walk-through of experts who have viewed other containments in terms of their -- DR. WALLIS: I was wondering what their expertise is based on if they haven't done an analysis. Is it just a guess? MR. HOOK: It is engineering judgment. DR. POWERS: But the question is, where does the judgment come from? I mean if you have never seen something, you can't have any judgment about it. MR. HOOK: But the personnel that were involved had conducted walk-throughs of other plants, especially the experts that we had from Fauski & Associates who were our primary contractor for the containment analysis. DR. POWERS: But I mean you can walk through a hundred containments -- incidentally, I have walked through the SONGS containment for Unit 2, I believe, and memory fades, but I am reasonably familiar with it. But it is not evident to me that I can do a mental mass transport analysis and say, oh, yes, there is no pocketing here. And I could probably walk through a hundred containments -- I think I am at 30 or something like that, and I would be no better at doing a mental analysis. And the question -- DR. KRESS: Do you think Fauski could? DR. POWERS: But now Hans, I will grant Hans is far better at this than I am, but I am willing to question the accuracy of his mental analysis on this if he has not done an analysis of, say, comparing the HDR containment and hydrogen distribution in that, those experiments, to arrive at some judgment on these things. DR. APOSTOLAKIS: Surely, they must have done some analysis in the past. DR. POWERS: Well, that is what we are asking. DR. APOSTOLAKIS: Yes, I mean their judgment cannot be based only on the fact that they have walked through any containments. So now -- DR. KRESS: Very few people have done these hydrogen distribution analyses because they are hard to do. There are a few codes out there that will do it, COMMIX is one of them probably. And COMMIX is familiar to Fauski. HDR will probably -- DR. APOSTOLAKIS: COMMIX? DR. KRESS: COMMIX, it is a containment code. HDR will also probably do it. DR. POWERS: Well, I mean I think we have seen that some of the field codes, and even some of the codes that lack a momentum equation can, with skill, calculate -- skilled operators calculate the hydrogen distributions. DR. KRESS: Yes, I would be suspicious of it if it doesn't have the momentum equation. DR. POWERS: Well, they do it with tricks on the noding and things like that. What we have never been able to do is say, gee, these ideas work well, say, for the comparison of the HDR containment analysis -- experiments, but does it translate into something else, some other containment? I mean I don't know how you would go about proving to yourself that was the case. DR. KRESS: Right. MR. SCHERER: I am not familiar with what they did or did not do in the way of analysis, but I do recall back on the original licensing of SONGS that Bechtel had done analyses looking for pockets of hydrogen and making sure that the design had been modified to eliminate and facilitate the hydrogen -- the communication between those subcompartments so that pocketing did not exist, to the extent that they could at the time and analyze it. So I do recall that as part of the original design basis for the plant. MR. HOOK: And there is an evaluation described in NUREG/CR-5662 entitled, "Hydrogen Combustion Control and Value Impact Analysis for PWR Dry Containments" that indicates for a number of the large dry containments, including San Onofre, that the time required to process one containment volume utilizing the mixing system, the fans that pull air from the lower compartments up into the upper, is on the order of 10 to 30 minutes, and for San Onofre it is approximately 20 minutes. So there are large fans that would be operating if there was power available that would provide a tremendous amount of mixing in the containment through forced ventilation. But specifically regarding pocketing in a single compartment, we don't have any quantitative calculation to evaluate the potential for that. DR. KRESS: Where are your recombiners located in the containment? MR. HOOK: Now, the recombiners are located right above the operating deck. DR. KRESS: So they wouldn't be very useful with these pockets anyway. MR. HOOK: That's correct. And the mixing system actually is located at the operating deck with fans that pull the air from the lower subcompartments up into the large open area above the operating deck. MR. SCHERER: But, again, the original design was reviewed to eliminate pocketing and a recent walk-through certainly confirmed that nothing would have caused pocketing to exist. DR. KRESS: Dana, these curves illustrate your point, that they just keeping on. The only things that bends them over is the K-heat. DR. WALLIS: Are you saying the fans draw? But the fans are way up here somewhere. MR. HOOK: Right. DR. WALLIS: The fan coolers, the coolers are up here somewhere. The pockets we are talking about are down -- MR. HOOK: We understand. DR. WALLIS: It is not quite clear to me how the fans up here draw something. MR. HOOK: Well, there are actually fans at the operating deck that take suction from the lower containment volume. DR. WALLIS: So there are fans down there as well? MR. HOOK: That's correct. Looking on Slide 11, you will see that for the case, the Parametric Study Cases 1, 2 and 3 where we varied the assumptions, there is very little difference in terms of the hydrogen generation rate varying those assumptions, but we get to 30 days with the concentration less than 4 volume percent in the containment for the sensitivities. Moving on to Slide 12, in terms of safety margin and the impact of the hydrogen control exemption request on the containment, the SONGS 2&3 have large, dry containments with a design pressure of 60 psig. Large, dry containments are the least susceptible to damage from a hydrogen burn following severe accidents based upon a number of references and evaluations and actual TMI event. In the TMI event with approximately 8 percent hydrogen concentration in the containment from approximately a 45-percent metal-water reaction there was a resulting peak pressure of 28 psig. And the TMI 2 containment design pressure was the same as SONGS at 60 psig. Also, there were a number of other industry and NRC evaluations. An industry evaluation at NSAC 22 and NRC evaluations in the NUREG that I cited previously, 5662, show that large, dry containments can withstand hydrogen burns, including those from severe accidents up to a 75-percent metal-water reaction. And the SONGS 2&3 have been evaluated in a detailed deterministic and probabilistic containment analysis where we looked at individual penetrations and performed detailed stress calculations to determine the failure pressures for those penetrations. We concluded that the median leak pressure for the containment -- and this would be at the fuel transfer bellows into the fuel handling building -- is approximately 157 psig, and the median rupture pressure, which would be a major failure of a containment tendon and a venting of the containment, would occur at approximately 175 psig. The 95-percent-confident values, as we discussed in the subcommittee meeting last week, are 99 psig for the leak pressure and 139 psig for the rupture pressure. DR. POWERS: I guess you need to explain to me a little more what you mean by the leak pressure. The containment leaks all the time. MR. HOOK: This is a leak that has the potential to create a source term that could be characterized as a large early release, and for us that's an equivalent hole in the containment that is on the order of about two inches diameter versus the design leaks. DR. KRESS: Just out of curiosity, why did you decide to report medians instead of means here? MR. HOOK: That was just a -- DR. KRESS: Just a choice? MR. HOOK: Just a choice. DR. POWERS: Good sense. DR. KRESS: Good sense. That's what you would say, right? MR. HOOK: But the 95-percent confidence levels are the telling certainty we have in terms of the robustness of the containment, its ability to withstand beyond-design pressurization. DR. APOSTOLAKIS: So what are those? MR. HOOK: 99 psig for the leak and 139 psig for the rupture pressure. Those are 95-percent confident values. Turning to slide 13, the SONGS -- based upon the evaluation of the benefit of the hydrogen control system in severe accidents, we concluded that the hydrogen control systems consistent with other large, dry PWRs are ineffective in mitigating hydrogen in severe accidents. If you look at NUREG/CR-5567, it also supports that conclusion for the large, dry PWRs, and the NRC's post-TMI evaluations concluded that large, dry containments could withstand hydrogen burns following severe accidents. Slide 14 contains a worst-case evaluation that we performed with the assistance of our level 2 contractor, FAI, to evaluate the containment pressurization for a worst-case hydrogen burn, and this is at a containment hydrogen concentration of 11-1/2 percent. That deterministic hand calculation for the worst-case hydrogen generation accounted for both in-vessel and ex-vessel hydrogen. We've suppressed all the burning of the hydrogen up to a single point, letting it accumulate. We assumed adiabatic, isochoric complete combustion with no heat transfer to the containment structures and no failure or venting of the -- or loss of pressurization of the burning, the hydrogen that's burning. We looked at the worst-case water makeup to the core that would result in the greatest hydrogen generation rate, an unlikely case where we have three charging pumps operating, providing about 150 gallons per minute to the core but no high-pressure safety injection or low-pressure safety injection. So this optimizes the amount of Zircaloy oxidation providing just the right amount of water for that, and this is a very unlikely scenario, to have three charging pumps operating and no other ECCS operating. DR. WALLIS: There's also pressurization by the released steam; this is taken account of here, too? MR. HOOK: This calculation does not take into account the pressurization from the released steam. DR. WALLIS: So the actual pressure from that would be higher than the 142 if you considered the released steam as well? MR. HOOK: I'll talk about that in a second. But this assumes the burning of 2,433 pounds of hydrogen, and the post-peak pressure in the containment's 142 psig compared to the median rupture pressure of 175 psig, when we evaluate this scenario using MAPP and look at the benefit of containment spray, a single train of containment spray operating, and the actual heat removal from the containment structure as we calculate a peak containment pressure of 115 psig for this scenario. So this worst-case calculation here does not look at the steam contributions. DR. WALLIS: It's not quite worst-case. MR. HOOK: It's not worst-case. There are a number of situations where the containment can fail from a worst-case hydrogen burn combined with a loss of containment heat removal. DR. WALLIS: So if you had decided that this case should not be above the 95-percent value, we hadn't thought about this at the subcommittee, this actually is above the 95-percent confidence value for the -- MR. HOOK: For a leak. DR. WALLIS: For rupture pressure. Rupture pressure's 139. MR. HOOK: That's correct, but in terms of the realistic evaluation from MAPP, it's 115 psig, which is below the 95-percent confidence level -- DR. WALLIS: So if you're going to make this comparison and say it's worst-case, you should put in all the worst aspects and perhaps compare with something like a 95-percent value rather than the median, which makes it look better than it does, than it really is. MR. HOOK: Okay. Turning to Slide 15, in terms of the impact of removal from -- of the hydrogen recombiners and purge from the plant design, if they were unavailable and unable to operate, how would we accommodate the hydrogen generation utilizing the severe accident management guidelines? The guidelines provide direction to prevent a fission product release, to bottle up the containment structure and penetrations, to maintain a leaktight structure, to reduce containment temperature and minimize long-term hydrogen buildup from the metal-water reaction or corrosion of the zinc and aluminum utilizing the fan coolers and/or containment spray. They provide direction to operate both to reduce containment pressure and temperature, to scrub the fission products utilizing the -- operating the containment spray and the fan coolers. Obviously the containment spray is much more beneficial in terms of scrubbing fission products out of the containment atmosphere than the fan coolers. They also have some limited capability to provide long term clean-up of the containment after extensive scrubbing, and the potential exists then if the containment pressure is reduced to near-atmospheric that we vent the containment, utilizing the mini-purge or other temporary filtration systems if the hydrogen purge system was not available. DR. KRESS: I would like to go back to your previous slide just a second. For Graham Wallis's benefit, the hydrogen recombiners control system would have no effect on this at all. Whether it was there or not, you get the same numbers. MR. HOOK: That's correct. DR. WALLIS: For a reading you get 142? You don't get something like 139 with the -- MR. HOOK: No. DR. WALLIS: They don't -- don't they remove -- but they have been switched off. MR. HOOK: They have been switched off. In fact, they probably would never be turned on because we would go past the 3.5 percent cutoff limit before they had an opportunity to sample in the containment the containment hydrogen, and the purge -- DR. WALLIS: It's your strategy that is making this -- making them ineffective. MR. HOOK: That's correct, and that is the Owners Group strategy and consistent with the other -- DR. WALLIS: If the strategy had been to leave them on no matter what, then they would have had some effect on this worst case? MR. HOOK: Probably by creating a burn earlier, but not by actually removing any hydrogen, because there removal rate is on the order of a fraction of a percent of the hydrogen that is actually being generated. They have a design flow on the order of 100 CFM and we are talking -- MR. SCHERER: Very frankly, I think if you leave them on, it would reduce this number because this has been done so conservatively to maintain all the hydrogen until the worst possible time, you turn on that hydrogen recombiner and those numbers will go down but then again you are adding realism to a very conservative calculation. The calculation we tried to do as conservatively as possible. Any time you try to add realism the number will go down, but then again how did you get this in the first place? DR. WALLIS: I was just responding to Dr. Kress, who said they would have no effect. MR. SCHERER: They have no effect because we are trying to create the most conservative calculation. DR. KRESS: And there are probably remote sequences you can dream up where they would have some effect on the peak pressure, but not much. If the hydrogen generation is low enough that their rate can keep up with it, it's not going to be a severe problem anyway, and if the hydrogen generation rate is anywhere comparable to this, they are not going to be able to keep up, not going to have -- MR. SCHERER: Please don't confuse this with the strategy. We would not intend to turn off the high pressure safety injection either. We would not intend to turn off the low pressure safety injection. We would not intend to turn on the three charging pumps. DR. KRESS: This is just an illustration of the robustness of the large dry containment. MR. SCHERER: Exactly. DR. WALLIS: It's good to make the calculation suppose that you don't intend to but somebody does. MR. SCHERER: I understand. DR. WALLIS: To investigate that territory. MR. SCHERER: And that is why we did the calculation. MR. HOOK: And turning to the summary slide, we have indicated that the proposed change does not affect the SONGS 2, 3 defense-in-depth of the containment design. It does not challenge or increase significantly the likelihood of containment challenge of large, dry containments like San Onofre 2 and 3 that have the capability to withstand the worst case hydrogen burns as a result of the robustness of the containment design and the confidence we have in that robustness as a result of a combination of a probabilistic and deterministic evaluation. The hydrogen control system is not needed for accidents based on realistic assumptions. When we take out the conservatisms that were indicated in the Reg Guide and Standard Review Plan the hydrogen control system is not sized for severe accidents and could be potentially detrimental in the case of the hydrogen purge since the filter system was not designed for a high pressure release, the valves were not designed to open and close at elevated containment pressures. Elimination of the hydrogen control steps in the activities from the EOIs. [Pause.] DR. POWERS: Obviously there are some systems I would like to turn off here. [Laughter.] DR. APOSTOLAKIS: I have a question. This is not really a risk-informed approach, is it? DR. KRESS: It meets the rules for the normal exemption rather than a risk-informed -- but you are asking what is the CDF, what is the LERF, what is the delta. It is risk-informed from the standpoint of they decided that the change in risk was a decrease in risk and the risk-informed 1.174 says if you have got a decrease in risk, it is always -- if you don't compromise defense-in-depth then it is always granted. So it is, you know, risk-informed in a normal sense in that respect./ DR. APOSTOLAKIS: Is it really supporting risk-informing 10 CFR 50? DR. KRESS: I don't know how it does. It is consistent with what -- we don't have a risk-informed 10 CFR 50. DR. APOSTOLAKIS: We're trying to risk-inform it. DR. KRESS: Yes. DR. WALLIS: At the subcommittee meeting I guess we raised the same point and I thought we concluded that they're barely scratching the surface of risk-informed. It's actually quantitatively evaluating risk. It's not required for this application. Eventually it will be, but -- if you are really going to risk-inform 10 CFR 50. DR. APOSTOLAKIS: Because this was also part of the Task Zero of NEI so again it's not a very convincing case from that point of view. MR. SCHERER: It is not intended, and you point out, it is not a sophisticated approach to risk-informing the regulations. It was intended to be a starting point. DR. KRESS: Take the small stuff -- MR. SCHERER: -- and it was risk-informed to the extent that we believe the risk, the change in risk is very small. It was not a risk-significant system and to that extent we certainly feel it is a risk-informed submittal. Is it advancing the state of the art? I don't believe so. DR. APOSTOLAKIS: Well, the thing is -- DR. KRESS: It does illustrate, George, that you don't always need a full PRA. DR. APOSTOLAKIS: I was about to say that. DR. KRESS: Depends on the nature of the -- DR. APOSTOLAKIS: I suspect we are going to have a lot of cases in the future where one would do calculations like the ones you gentlemen did and perhaps advance qualitative arguments and ultimately reach the conclusion that, you know, risk is reduced for whatever -- for the increase is minimal. DR. SEALE: These are the folks that have the risk meter, and they did confirm that there would be no change in the risk meter reading as a result of this. DR. APOSTOLAKIS: Yes. MR. SCHERER: And we could be back here discussing some more applications that will. DR. APOSTOLAKIS: Again, since we are setting precedents here, it may be a minor point, but you know, communication sometimes is important. On the Slide 13 and 14, especially 14, I would delete the words "risk impact" from that heading of the slide because you say -- the analysis that you are doing is worst case so I don't want to associate worst case analysis with risk impact. DR. KRESS: That's a good point. DR. APOSTOLAKIS: Just delete the words "risk impact" -- I mean what you have done is fine. Okay? Because if we start creating a tradition again of jumping into conservatism and worst case analysis every chance we have, we will never become risk-informed. That is a minor comment. MR. HOLAHAN: One additional point I would like to add. One of the things that was really tested in this application is not so much the risk analysis as much as what values we place on other things, some of which can't be quantified or are sort of at the boundaries of what you think are valuable, things like hydrogen monitoring. I think the Staff was very reluctant to give up hydrogen monitoring because there's some value to knowing what is going on in containment. I think the Staff placed some value on the accident management guidelines as a way of dealing with things like multiple hydrogen burns, subsequent burns. Even though the analysis would show they are unlikely to cause containment failure, and so you wouldn't change your LERF values, as a matter of fact they occur so far out it's probably not LERF anyway -- but there's some value I think to not having additional burns in containment, so some of these things are not easily quantified but to the extent that they are tested in this case I think may be very typical of what future risk-informed applications do. MR. SCHERER: Why don't I just summarize then. Perhaps this is as good a point as any for me just to briefly summarize our position that for the four subsystems of the hydrogen control system. Obviously, we intend to make no change on the hydrogen mixing system inside the containment. We asked the Staff to hold aside any effort on the hydrogen monitoring system. We'll maintain that system as part of another application. As far as the hydrogen recombiner and the purge subsystems, if our application is granted we will remove them from the FSAR and the technical specifications. We will reclassify the existing systems as non-safety related systems, but we will not remove them from the plant. We will maintain them as they are available and we will maintain them in the severe accident management guidelines as the equipment is available, and we have agreed to notify the NRC should we ever abandon our attempts to continue to maintain either the hydrogen recombiners or the hydrogen purge system. By the way, there are two other purge systems in our plant -- mini-purge and the full containment purge. Those will continue to exist. Are there any questions? DR. KRESS: Since I see no further questions -- DR. APOSTOLAKIS: Just out of curiosity, do we ever calculate the Fussel-Vesley importance measure or RAW with respect to LERF? MR. HOOK: We do at San Onofre? DR. APOSTOLAKIS: You do? MR. HOOK: Yes. DR. APOSTOLAKIS: Do you have any importance measure for these equipment? MR. HOOK: Oh, it has not importance based upon our evaluation that the equipment plays no role in a severe accident. There was a conclusion made early in the analysis that as a result of the direction to turn off the recombiners at 3.5 percent and the fact that the hydrogen purge is not designed for a high pressure release that they would play no role in mitigating a severe accident, so their Fussel-Vesley role is basically inconsequential -- so there is no calculation. It's like looking at another system in the plant for which you conclude plays no role in the severe accident, and therefore you don't include it in the PRA model because of the lack of its contribution to mitigating the event. DR. APOSTOLAKIS: Isn't that an argument in itself? The thing is not modelled at all because it is irrelevant. That is essentially what you are saying. MR. HOOK: That's correct, and that is one of the reasons this was selected was because it was a clear-cut case for all the lightwater large, dry PWRs where this conclusion had been reached. DR. KRESS: George, earlier on when I was first looking at this, I had thought there may be sequences that went through early something like a design basis accident, so that the generation rate early-on in the accident would be something like what you got from the design basis and then -- and during that time the recombiners would be taking out some stuff at about the same rate it's generating until you turn them off and enter it into a severe accident, but you did reduce for those sequences some hydrogen and lowered the probability that you would fail the containment to some extent by this overlapping of the uncertainties. I searched and searched and I couldn't find any sequences like that for large drys. There may still be some for BWRs, I'm not sure. I didn't look at those. I just looked at the PWRs but I just couldn't find any, so they wouldn't show up in a PRA. DR. WALLIS: Now the folks that required that you install these in the first place, I take it, have all retired? [Laughter.] DR. WALLIS: And there is no one to testify I was wrong to insist that they were installed or -- MR. HOLAHAN: Professor Wallis, is that a question or a statement? [Laughter.] DR. SEALE: Just an observation. DR. WALLIS: I was just thinking it. This was a requirement that was insisted upon at some time and they have been in here all this time and now it's been decided they are not necessary, which may well be true. We don't have -- we don't get to question the people who required they be put in there in the first place. MR. HOLAHAN: As a matter of fact, some of those people still do exist. I think you could have spoken to them if you asked them -- DR. WALLIS: Did the Staff ask them? I mean I just wonder what their opinion is. MR. HOLAHAN: I think some of them are knowledgeable of this because they are in the same branch. DR. WALLIS: So they go along with it then? MR. HOLAHAN: Mike, do you want to speak to it? MR. BARANOWSKY: I think that part of the problem you run up against, and I spoke about this to the subcommittee, is that we're mixing the risk world and the emotional world. If you look at the original statement of considerations for why these things were put in, it was because of the adverse public reaction to the possibility of public -- I'm sorry, adverse public reaction to the possibility of venting at TMI, and then when you look at this from a risk perspective, the risk is acceptable, so that is really the basis for why the recombiners were put in, and there was so much uncertainty with the risk at the time of TMI and now that we have done a lot more research and we have a lot more -- we have quantified those numbers better and we have looked at it through 1150, which we are going to discuss here in a little while, then I think that argument, the emotional argument isn't as strong. DR. KRESS: That is the nice thing about risk-backed information. You can go back and re-evaluate the basis. DR. WALLIS: But if this was installed originally because of adverse public reaction, if that was the criterion, they I guess you would want to be reassured that there wouldn't be adverse public reaction to taking out recombiners -- if that was the reason why they were put in in the first place. It's the public that's concerned, and so you do have to make a case to them. DR. KRESS: We are going to have that issue every time we remove something based on risk-informed. MR. HOLAHAN: True. I think the other point I would like to make is in response to your comments at the subcommittee meeting, I did go back and look a little bit at the history of this issue, and I was reminded that in fact the Staff raised the hydrogen issue as part of the marginal to safety program in the early '90s as a candidate issue for being removed from the regulations. And at that time the industry indicated that it was not one of their priority considerations for removal. They had other priority issues such as graded quality assurance which we have dealt with in the last few years. So, I think, you know, this is not a surprising issue, we have known it is a candidate for removal or modification for a fair amount of time. DR. POWERS: Gary, I guess I am a little confused. I mean hydrogen recombiners have nothing to do with TMI. MR. HOLAHAN: Hydrogen had something to with TMI. DR. POWERS: Hydrogen has something to do with TMI, but the recombiners have been in there for a lot longer than TMI. MR. HOLAHAN: Yes. DR. POWERS: But that is not why they were originally put in. They were originally put in to handle the hydrogen that was produced by a combination of metal-water reactions, radiolysis and corrosion of metals in the containment. There was a presumption that those hydrogen sources could be larger than maybe they were anticipated to be based on deterministic analysis at the time, and so it was considered a defense-in-depth measurement. These gentlemen have come in and said, whoa, we are smarter now, know more, we have adequate defense-in-depth from other sources. They may in fact arguably make severe accidents worse. It has nothing to do with adverse public reaction to anything. MR. HOLAHAN: Right. MR. SNODDERLY: Dana, I was just, I guess -- you are absolutely right, and the recombiners were installed at TMI 2. And the only point I was trying to make was that the recombiner rule was reconsidered in 1981 as a result of TMI to see if that requirement was sufficient, or the design basis for the recombiners needed to be reconsidered. And in that Statement of Considerations is where I got the statement that I quoted. DR. POWERS: Sure, I understand. But I mean I just don't want people to think that the original reason had something to do with TMI. DR. BONACA: One thing I would like to just point out, I don't have a concern about having increased the challenge to the containment. In terms of what you, Gary, mentioned as your concept of defense-in-depth here, I agree with that. I am comfortable with that. The only thing that I see, however, in the presentation, we are being boxed in between design basis below 4 percent, severe accident above 10 percent. Life, it is different. I mean I am convinced if we have a design basis event and everything works as design basis says, we will have no hydrogen probably there, but there will be scenarios possibly, okay, in many different kind of LOCA events we may get, and, hopefully, zero, that will give you some hydrogen generation somewhere in the range of 4, 5, 6, 7 percent and I think that the recombiner could be useful because all the curves you showed us didn't all the benefit of the recombiner. If you use those, you will be tracking this curve down, you know, in those cases. So I mean I agree that there is no increased challenge to the containment and I am comfortable with that. However, we are removing equipment that could be potentially useful, and I think that is a realization that -- at least I would like to hear somebody telling me that is not true and I will be glad to agree. MR. HOLAHAN: No, I think we agree with you, and what I would characterize as the recombiners being equipment that is not essential, but might be useful in some cases. And what we did was we struggled with a place to put such a category of equipment. I think we are satisfied with the idea that the severe accident management guidelines and a voluntary commitment on the part of the utility to maintain that equipment is a reasonable way of dealing with equipment that is not essential but could be useful in some cases. So it is a kind of gentle middle ground for treating this sort of equipment. DR. SEALE: And in the subcommittee meeting, the concern was that, while this may be a no-brainer from a risk point of view, that there was a serious issue of precedent as to whether or not you might construe what you did in this case as a basis for dropping useful equipment that had previously been subsumed into a commitment for an issue that was judged to be safety significant but no longer had that status, and I think you probably addressed that issue. MR. HOLAHAN: Yes. MR. SCHERER: Can I make a comment? DR. KRESS: Yes. MR. SCHERER: One of the reasons we pursued this application wasn't, as I said, the cost, but a belief that this was a system that was not meaningful in terms of improving the overall risk of the plant and it, therefore, felt as an example of something that could be removed as a regulatory requirement. I was a licensed operator, and as an operator I would love to have as many levers as I could if I was facing a challenge. And as many systems as I could find a use for, I would want to see in that plant because it gives me more options as an operator, especially facing an unknown situation. But at some point we have to look at the risk benefits of the system, and adding systems upon systems upon systems also adds to the complexity. If I had a severe accident, I would be doing three simple things, I would be buttoning up that containment, I would be trying to cool it and I would be trying to clean it. And those are the things I want the operators concentrating on at the time of the accident. There is a lot of useful information to be gathered. There are a lot of useful tools to make sure they have them and, as we indicated, we have no intention of ripping it out. But if it had been the converse, could we have justified adding that piece of equipment now if it didn't exist? Well, should we spend assets that we would rather spend on risk significant issues to maintain this piece of equipment? I think this is a good approach and the test isn't the risk significance or the sophistication of our risk approach here but a test in our mind of whether the Commission, as the Chairman said, is able to let go of something if it is shown to be of very low risk significance. DR. BONACA: I appreciate your comment, and I agree with you. I am only pointing out that there is some issue there. And the reason why I raised it also during the subcommittee meeting, I pointed out, you know, I have a concern in general about piecemeal removal of equipment, because you already told us that you will come back and propose to remove PASS from your emergency action level. Now, at the time you also were planning to remove the hydrogen monitoring system. See, that is the point I am trying to make, is that I totally agree that some of this equipment shouldn't be there and is not needed, but there should be a comprehensive strategy to deal -- not boxed in design basis, severe accidents, but simply looking at possible scenarios, to look at what you have got and determine and decide what you want to keep and what you want to propose to eliminate. And there shouldn't be boxing of, you know, just -- you suggest hydrogen, and this is severe accidents and this is -- you know. That would be useful, I think. MR. SCHERER: We should not have started on a piecemeal approach, we agree, and that is why we backed off and took it as a holistic approach and have resolved that issue. DR. KRESS: I think we had better move on to the staff's presentation. We still have to get through the full agenda. So I will turn it over to you, Gary or Mike -- is it Mike? MR. SNODDERLY: Good morning, my name is Mike Snodderly. I am with the Safety Assessment Branch of the NRR and I have been assigned the task or reviewing this exemption request. And I guess what I want to discuss today is how I approached this review, and basically it was to start off by looking at how did we, the staff, model hydrogen combustion in NUREG-1150, and then to try to determine what value, if any, do the hydrogen recombiners have. And to do that, I requested and received assistance from the Office of Research and they helped us do some analysis and we are going to present those results today, too. The first thing I'd like to try to do is talk about different hydrogen source terms, and the first one I'd like to talk about is the source term in 50.44. And Dana alluded to it earlier. It basically says that you need to address three things: Zirc metal-water reaction, radiolysis, and corrosion. And 50.44 talks about dealing with an amount of hydrogen five times greater than what you would calculate in 50.46(b). And I'll call -- what I'll call the true -- what I'll say, let's consider the design basis LOCA. And that says that you won't exceed 1 percent metal-water reaction in accordance with 50.46(b). The amount of solid fission products that you would expect from that type of an accident would be much less than .1 percent. So radiolysis isn't a concern from the design-basis LOCA. And the corrosion rate, of course, you wouldn't expect much corrosion. 50.44 said we want you to address five times that amount. We want you to look at 1 percent of the solid fission products. And 1 percent of the solid fission products corresponds exactly with the amount of solid fission products that you needed to consider from a TID 14844 type of source term which I would consider and I believe the committee does too more of a severe accident or beyond design basis accident. So the point I'd like to try to make here is that recombiners are needed to address this type of accident. DR. KRESS: In these source terms, the iodine was not considered a solid fission product? MR. SNODDERLY: No, sir. It is considered, but not as a -- well, it's part of the solid fission products. It's considered -- you also need to consider iodine. But that's not what I'm -- DR. POWERS: I guess I don't really quite understand the technology here. I get radiolysis of water to produce hydrogen from radioactive material distributed or dissolved in the water. I get hydrogen also from the gaseous fission products radiolytically decomposing steam in the atmosphere. MR. SNODDERLY: Okay. DR. POWERS: Why does it make any difference? MR. SNODDERLY: The only -- I'm just -- DR. POWERS: You're just repeating what's in the particulars, the cited documents, but I just wondered why -- MR. SNODDERLY: Yes. I'm trying to do more than that. DR. POWERS: G values are about the same, I mean, roughly. MR. SNODDERLY: Yes. The main point I was trying to make here, what I was -- I think part of the thing I've been struggling with and that I've had difficulty communicating to the Committee and to others is that what amount of hydrogen do we really expect from a true design basis LOCA? What were we asked to design for in 50.44 which we're now saying is this kind of in-between accidents, not really severe accidents, not really a design basis accident, somewhere in between which says that you need recombiners, and then where is really the risk from hydrogen combustion. And the risk from hydrogen combustion I would argue would be this type of a source -- this type of a hydrogen source term, like a TID 14844 source term or a NUREG-1465 Early In-vessel or Late In-vessel. So what I want to do in this presentation is to talk about how NUREG-1150 looks at the risk from these types of accidents, and we're going to look at that, and then we're going to say what value are hydrogen recombiners, and really the hydrogen recombiners are only of value in dealing with mainly radiolysis. They're overwhelmed by this source term, and corrosion is about a 10-percent contributor to the overall hydrogen source term. So it's something that yes, the recombiners would help with, but what you're really concerned about is radiolysis. So the accident that we're going to look at or that we analyzed was to say okay, the hydrogen recombiners can't do anything about that initial burn, okay, but you're going to have a lot of radiolytic decomposition in the sump and hydrogen created from that. How much is that hydrogen? Because I don't think we've done a good job of answering that, and is that of -- how much risk is associated with that, and what's the concern with that. So all I wanted to try to do with this slide was to just try and show the Committee, you know, just to give some order of magnitude of the source terms and what we're going to look at, because what the PRA or NUREG-1150 finds is that the risk from these source terms doesn't even warrant really modeling in the -- for early containment failure or prior to vessel breach, it's not a threat to containment and was not modeled, and in late containment failure it was seen as a very, very small contributor. So once we accept that modeling or that level -- once we accept that modeling, then these conclusions fall out that igniters were not needed for large drys, and I think a subsequent fallout is that the recombiners are not a risk-significant system. So what I did was I went to NUREG-1150 to see how hydrogen was modeled at the Zion plant, which was the large dry that was looked at, which would be equivalent to San Onofre. And basically containment failure was looked at in four categories: early containment failure, which as I said is prior to vessel breach; late containment failure, which is after vessel breach; no containment failure; and bypass. And basically about 73 percent was no containment failure, 24 percent was late containment failure, about 1-1/2 percent is early containment failure, and the rest was bypass. Now for early containment failure the containment failure was composed of in-vessel steam explosion, overpressurization, which includes high-pressure melt ejection, and containment isolation failure. The expert elicitation showed that or had concluded -- in earlier drafts they modeled hydrogen but found that the contribution to failure was so small that it was not modeled in the final. It was modeled for late containment failure, and what they found was 99.5 percent of the late containment failure came from basemat meltthrough and only .5 percent comes from hydrogen. So you can see that hydrogen combustion was not seen as a big threat or risk-significant from 1150. DR. WALLIS: What happens in those .5 percent cases? What do you mean by containment failure due to hydrogen ignition in those cases? MR. SNODDERLY: Meaning that hydrogen combustion would result in containment failure due to hydrogen in containment due to the zirc metal-water reaction and additional hydrogen from core concrete interaction. DR. WALLIS: Containment failure, I mean, what happens? MR. HOLAHAN: I think we're talking about the kind of case that San Onofre was talking about, overpressure rupture of part of the containment. DR. WALLIS: It's a big hole? DR. POWERS: I think NUREG-1150 used a specific definition, and it was 10 times the design basis leak rate. DR. WALLIS: So essentially you've lost the containment then? MR. SNODDERLY: Yes. DR. POWERS: Well -- MR. HOLAHAN: I think it would be better characterized as a large leak rather than catastrophic failure. DR. POWERS: Well, one containment volume per day I think is what they used as a definition, so it has some ramifications in that it's not lost. But for the purposes of this discussion I think it's satisfactory to assume that you're getting very little mitigation from the containment. DR. KRESS: Whereas you would get some mitigation in the 99.5 percent basemat meltthrough. DR. POWERS: I think it's fair to say that we don't understand exactly what the release characteristics of a basemat meltthrough are, except that they surely must be much less than a badly leaking containment. DR. KRESS: Yes. So comparing the .5 and the 99.5 is really not a real good comparison. DR. POWERS: Well, I think what it tells you is that in a large, dry containment hydrogen combustion failures are not the dominant mechanism of containment failure. DR. WALLIS: Well, I guess I'm asking because I thought we'd just heard the hydrogen ignition does not fail the containment. Now you're telling us it can. MR. SNODDERLY: Let's put some risk numbers with that, then. So you're at -- excuse me. DR. KRESS: You wouldn't have any LERF at all. DR. WALLIS: It may not be relevant The message I got from the previous presentation was hydrogen ignition is not a problem. MR. SNODDERLY: It would be very, very small. DR. WALLIS: It can't fail the containment. MR. SNODDERLY: That .5 percent, Graham, would be something on the order of 10 to the minus 8th. DR. WALLIS: I don't know what you mean, because you don't say how many of the hydrogen ignitions resulted in containment failure. DR. KRESS: You can always -- you can postulate an accident. DR. WALLIS: You are talking about a different world. I mean they essentially were talking in a deterministic world. And I think the message was hydrogen ignition doesn't challenge the containment, the containment can take it. MR. SNODDERLY: Right. DR. WALLIS: Now put something up here which suggests that hydrogen ignition does fail the containment. I am just trying to reconcile the two statements. MR. HOLAHAN: Dr. Wallis, I don't think they are so inconsistent. I think the San Onofre presentation said, under, you know, rare circumstances you would show that the pressure exceeded the failure pressure of containment, but they had to search for rare cases. And I think this analysis is a similar conclusion. MR. SNODDERLY: This would be -- I would argue that 10 to the minus 8th number would be in that 5 percent of that 95 -- outside the 95 percent confidence limit. I mean I think the two are in agreement. It is a very -- it is a negligible contributor to risk. DR. KRESS: You will fail containment under some rare circumstances without hydrogen there. The contribution to the LERF due to hydrogen is very small is the point. You are going to have a LERF anyway. DR. WALLIS: I don't know, I am sort of puzzled. The whole -- the previous presentation made no reference to risk calculations whatsoever. MR. SNODDERLY: Yes. DR. WALLIS: But you did some. MR. SNODDERLY: I am just saying the staff in NUREG-1150 looked by hydrogen combustion as -- they did not -- I guess what I am saying, although the modeling is not exactly the same and they did not find any failures, we found a very, very small fraction and that I would not characterize that as rising to the level of something that needed to be looked at. DR. WALLIS: So they didn't make a real risk analysis, they did some qualitative stuff. You did some real calculations? DR. KRESS: Now, these are going back to some real calculations? DR. WALLIS: Or somebody made some real calculations. DR. KRESS: Yeah, these are real calculations that exist already and there are risks. DR. POWERS: The question, of course, comes as these are for a containment over a Westinghouse reactor of a particular design that does not have the same pressure capabilities of the combustion engineering containment, or the containment around the combustion engineering plants on the coast of California. The relevance is a bit missing, but interesting nonetheless. DR. KRESS: Yes. DR. WALLIS: I am tempted to say this doesn't matter, but I am not quite sure why it is being presented. I am trying to reconcile. DR. KRESS: Well, it is illustrates for large dry containments that the possibility of having much contribution from hydrogen to a LERF is pretty small. MR. SNODDERLY: As a reviewer, I would have to say, did we find any risk significant sequences from hydrogen combustion that I thought that we should look at or have San Onofre look at, from the standpoint of would the recombiners impact or have an effect on those sequences? And what I am trying to say is, from 1150, I didn't see that. I did see that they did not model late containment failure, meaning after vessel breach, in their PRA. From these numbers, it would not make me say, ACRS, or to my management, I think they need to go back and look at this, because it is such a small contributor. That is what I think -- that is what I did as a reviewer, Graham. I didn't see this as rising to the level to say -- DR. POWERS: I know I am going to jump ahead in your viewgraphs, but maybe it is time to bring it up. MR. SNODDERLY: Yes. Thank you. DR. POWERS: Suppose -- I know this cannot happen at the San Onofre unit, but suppose that there were some sort of an event, a perturbation, a change of mode and damaged some equipment, otherwise did not damage the core greatly and whatnot, put a little steam into the containment and whatnot. The NRC sent out an augmented inspection team as to the plant. A confirmatory letter shut the plants down for a year. And we have for then outrageous circumstances, a buildup of hydrogen taking place in the containment. The hydrogen recombiners could be used to eliminate that hydrogen? MR. SNODDERLY: Yes, they could. DR. POWERS: And without them, what would we have? MR. SNODDERLY: What we would have, I believe is circumstances where they would be needed, but that there would be a great deal of time, I would estimate on the order of 7 to 10 days, before it would build up to the lower flammability limit. I think in that time, we could come up with some type of a contingency action, either getting a recombiner from some off-site and brought onto the site and could be used. DR. WALLIS: But you know, this came up in the subcommittee meeting, it seems to be very speculative that someone is going to find something in a month, which they then move to the site and hitch it up and it does something -- unless it is already being planned for. So, assuming it is going to be found, somehow appear within a month. MR. SNODDERLY: Well, there are a number of utilities that already share -- 5044 allows you to share recombiners between sites. DR. WALLIS: If someone has actually, yes, has researched it and said, yes, there exist these things and, yes, they can arrive within two weeks, and, yes, they will work, that is fine. But if it based on the sort of an assumption they might be available, that doesn't -- MR. SNODDERLY: And I think if we were talking about reaching the lower flammability limit in 24 hours, Dr. Wallis, I would agree with you that we would need more pre-planning ahead of time. But I think when we look at the timeframes that we are talking about of in the 7 to 10 range, I think that that is -- it is justified to have that kind of -- DR. POWERS: I kind of think in the 7 to 10 day range, but we are not talking about that. We are talking about months for my scenario to approach 2 or 3 percent of hydrogen in there. You do run the risk of headlines, but regulation by headlines is something that I am not willing to do here. But, yeah, they can take of these small sources of radiolytic and corrosion based hydrogen in ways that are different than venting, which, quite frankly, in my scenario that I posed to you, if you didn't vent the thing, rather than hooking up some jerry-rigged thing, I think you are out of your mind. DR. KRESS: Well, they are still going to have the recombiners in there they said. DR. POWERS: Well, yeah, but maybe, but I still think the appropriate thing in my particular scenario is to vent it. DR. KRESS: Vent it. I think you are probably right, through the filter system. DR. POWERS: Go ahead, please. MR. SNODDERLY: Sure. So after looking at NUREG-1150, the other thing I wanted -- I did some value of the recombiners to address radiolysis from a severe accident. So the scenario I postulate is that you had the zirc metal-water reaction occurring over a two-hour time period. It overwhelms the recombiners, you have a burn inside containment. The containment survives but now you have a slow buildup of hydrogen. So what I did, working with the Office of Research, we went and used the COGAP code, which is what we used to model the current design basis hydrogen source term to come up with the type of curves -- to confirm the curves that San Onofre showed earlier. Basically, what we did was we tried to go in and take out the conservatism. We zeroed out the metal-water reaction contribution and we tried to -- as I shared earlier, I looked at NUREG-1465, looked at the amount of solid fission products and I estimated that to be around 8 percent, so I increased from 1 percent to 8 percent the solid fission productions in the sump. We modeled 75 percent of the iodine released to the sump and we looked at that radiolytic source term with a G value of .5 and then another run with a G value of .4. The theoretical maximum is .44 for pure water. And when we did those runs, we found that it was 7 percent when you used a G value of .5 and 5 percent when you used a G value of .4. DR. POWERS: When you say a theoretical maximum on the G value of hydrogen, what theory? MR. SNODDERLY: I was going by a paper done by Parczewski and Benaroya that was presented and looking at hydrogen source terms. I can give you the reference, Dr. Powers. DR. POWERS: Even the vintage would be enough. MR. SNODDERLY: I'm sorry. DR. POWERS: The vintage. MR. SNODDERLY: 1984, generation of hydrogen and oxygen by radiolytic decomposition of water in some BWRs, Parczewski and Benaroya. DR. POWERS: There's been just an excellent piece of work coming out of Canada that looked at all of these data, did some additional data, and came up with the g values for low LET and high LET radiation as a function of temperature and pressure by Elliot, 1993. MR. SNODDERLY: Thank you. DR. WALLIS: Well, is this okay, 5 to 7 percent? What do you learn from that? MR. SNODDERLY: I think that what we learned from that is that yes, the recombiners would prevent that subsequent burn, but that that burn would not -- that level would not be reached for 30 days -- the lower flammability limit of 4 percent will not be exceeded for 11 days, and in that time contingencies could be made to either bring a recombiner from offsite onsite or, if worse comes -- or venting, as Dr. Powers mentioned, or just to let it burn. That that burn would not be a threat to containment. DR. WALLIS: Because? Because the 7 percent is too low to -- MR. SNODDERLY: Right. The pressurization associated with 7 percent -- DR. POWERS: When you say it's not a threat to containment, I mean, that's a strength argument, but busting something in the containment can have impacts on other things. What kinds of other things do you worry about in containment? MR. SNODDERLY: Equipment survivability. DR. POWERS: That's right. I'm thinking of the TMI telephone -- MR. SNODDERLY: Right. DR. POWERS: Substantially scorched on one side. MR. SNODDERLY: As a result of the TMI event there's been a lot of work that was done at the Nevada Test Site and at Sandia, SCETCH facility, and the basic conclusion to determine if something needed to be done from an equipment survivability standpoint, the conclusion was that the equipment most likely would survive up to a single burn of a 75-percent metal-water reaction, 11-1/2 percent. There was -- when you talk about multiple burns in a single compartment, there was more of a concern with failure, but overall that conclusion is what I think -- DR. POWERS: Not much heat load, is what you're saying. MR. SNODDERLY: Right. And that was considered. So with those two data points I believe I came to the same conclusion that San Onofre did, and so then I think -- I don't want to spend a lot of time on this, because I think we're in agreement with San Onofre now about the importance or the need to -- or the part that the recombiners would play within the Severe Accident Management Guidelines, and the fact that they don't intend to rip the recombiners out at this time, and that if they did, they would notify the staff I think gives us a good assurance that because the plant -- because the recombiners exist in the plant that they would be part of the severe accident management guidelines as directed by NEI 9104, which really provides the guidance on how to develop a plant's severe accident management guidelines. DR. POWERS: The previous speaker brought up I thought an interesting philosophical discussion. He says if I've got a problem in this plant, I'm out to do two things -- three things I guess he said. Shut it down, cool it, clean it. And I want my operators to focus on that. You think about those issues that, you know, if I've got lots of levers available to the operators, I'm just creating more confusion, more complexity in the system, when in fact all I want him to do is shut it down and cool it and clean it. I mean, it's nice that they're going to leave these things in here and notify the staff if they take them out. On the other hand, maybe there is a big advantage in human performance space to take them out. MR. SNODDERLY: Yes. And I agree with you, Dr. Powers, and I think that in the short time frame the procedures that the operators are going to be concentrating on in their emergency operating procedures which the recombiners are going to be pulled out of and the hydrogen monitoring or the requirement to get it quickly aligned is -- we've recently granted a relaxation in that area. We briefed the Committee on -- remember the hydrogen monitoring request from 30 minutes to -- so to support your point, yes, those things should be taken out of those emergency operating procedures and put into Severe Accident Management Guidelines, which are going to be used more by the technical support center and those folks, the decision makers. So we're taking it out of the operators' hands and not burdening him but putting the technical support center, which I think is more appropriate for those folks to then support the operators. DR. POWERS: Yes. MR. SNODDERLY: That's really all -- I wanted to just let the Committee know what the staff was doing, what our perspective was from 1150 and some supporting analysis we had done. The remaining slides it just talked about the hydrogen monitoring which it appears that San Onofre has reconsidered and it's going to come in with PASS requests which you were briefed on I think within the last two weeks, and we would of course expect that submittal to be consistent with the position that was taken and in support of your -- and consistent with your letter, which we resolved. If there aren't any questions or if there are any questions -- DR. WALLIS: I guess I was looking for a more what I'd say is a crisper evaluation where you'd say in came this request and then we applied certain procedures and criteria philosophies, principles, whatever, and on the basis of those we checked that certain claims by the applicant were okay. It seems a much more sort of discursive, ad hoc presentation that you made. MR. SNODDERLY: Well, let me see if I can allay your concerns, Dr. Wallis. If you go to page 6, basically the basis for the exemption request is going to be in accordance with 50.12, and the criteria for exemptions, and I guess -- I'm sorry, wrong slide. DR. POWERS: Slide 5. MR. SNODDERLY: I can go from memorization -- Slide 5, 50.12, and what 50.12 -- thank you, Dr. Powers -- on Slide 5, application of the regulation in a particular circumstance would not serve the underlying purpose of the rule or is not necessary to achieve the underlying purpose of the rule. And what I would argue is that the purpose of the rule, the underlying purpose, is to show that an uncontrolled hydrogen-oxygen recombination would not take place or that the plant could withstand the consequences of uncontrolled hydrogen-oxygen recombination without loss of safety function. And I think what I've done in my review is confirmed that the plant could withstand the consequences of an uncontrolled hydrogen-oxygen recombination without loss of safety function as demonstrated or as modeled in NUREG-1150 as supported by the San Onofre's IPE and the additional analyses that we did. Now if that's not the case, then I don't think we have a basis for an exemption, but I think that is the case, and that's what I would argue, yes, the basis for the exemption. DR. WALLIS: I'm glad you got to Slide 5, because you made a bigger point of it, and I think it was appropriate at the Subcommittee, what was the basis for your decision. And then today we seemed to get a bit away from that to more fringe areas. MR. SNODDERLY: Well, thank you. I forgot about this slide. I'm trying to -- better make sure I -- DR. KRESS: If there are no more questions then, I'd like to thank Southern California Edison and the staff for their presentations, and I'll turn the floor back to you, Mr. Chairman. DR. POWERS: Thank you, Tom. I guess I still come back to the idea that this was a test case to see if we could make use of 1.174 to get rid of something, and I guess I turned to Gary and said was it a successful test case or just kind of a test case? MR. HOLAHAN: Well, like a lot of test cases, it probably has some pluses and minuses. I think a lot of people would say it's probably taken us too long to come to a conclusion. I think the conclusion's a very reasonable one. I think like a lot of these cases it's going to force us to deal with marginal issues. In this case in a sense we are lucky to have accident management guidelines as a place to put issues which I think otherwise we wouldn't be comfortable with walking away from completely. If you remember, in the South Texas graded QA issue in fact the licensee created sort of a middle territory of -- they didn't put everything as very important or unimportant, they created a middle territory where they targeted some QA activities. And to that extent I think it's a useful exercise, because in the future we're going to have to, you know, not only use risk information to decide what's really important, we're going to have to deal with other values and some of these middle grounds. And we're just going to have to, you know, find ways to deal with these issues. DR. POWERS: I think what you're saying is that our PRA tools are simply not adequate for addressing some of these what you call marginal issues, and so you can't get a cutting decision that way. MR. HOLAHAN: Sure. I agree. DR. POWERS: And I think we will learn to regret a tradition of finding halfway measures. MR. HOLAHAN: I don't feel that way. I'm not sure I would describe the kind of issues here as halfway measures. In fact, I think one of the reasons we have a risk-informed approach rather than a risk-based is that we do recognize there are other values besides core damage frequency and large early release, and we're trying to find a way to deal with some of those values. DR. POWERS: As are we. With that, I'll take a recess until 20 of the hour. [Recess.] DR. POWERS: We'll come back into session. We'll now turn to the status of the pilot application of the revised inspection and assessment program, and our cognizant member for this is usually John Barton, but we have pressed into service Dr. Shack, and so I will turn it over to you. DR. SHACK: Okay. I guess since the last time we have looked at this program a few things have happened. The Watch List has been eliminated. The recognition of superior performance is gone. We will be hearing today about some items from SECY-99-007A, which presumably will give us some more information on the pilot program, which has been one of the focuses of where this program is going, and I believe we will start off with Mr. Madison, who will begin the presentation for the Staff. MR. MADISON: Good morning. I am Alan Madison. I have to my left Morris Branch, from NRR also, and we also have Frank Gillespie, Phil Dean, Michael Johnson, and Gareth Parry available if there are specific questions. I will apologize in advance not necessarily having a slick presentation ready for you. With yesterday's extra day off, our preparation time got -- [Laughter.] MR. MADISON: -- got shrunk to this morning. The reason we came, wanted to come talk to you is to give you the Staff's update. It's been some time since we have been to talk to the ACRS about the improvement process, talk about where we are today, and we are ready to start the pilot, as Dr. Shack mentioned, also talk about the significance determination process and the feasibility review, which I know there was a great deal of interest at the ACRS, and talk a little bit about what is yet to be done. I am going to address the first and last item briefly, and hopefully leave sufficient time for Morris to go in depth or somewhat in depth in the significance determination process. DR. POWERS: I hope that we have some opportunity to discuss how we interpret results of a pilot. MR. MADISON: We had not prepared to talk about that at this time, but I would be interested in your comments. DR. POWERS: Maybe just philosophically discuss it. I mean my concern -- I look upon pilots as experiments and so when I think about experiments I think about experimental error, representativeness, and making use of the results. Perhaps you used the pilots to say are there any problems with this, and you can come away and say conclusively for this plant there are no problems or yes, there were problems. The question is how you extend it to the larger range of plants and the larger safety philosophy when you have done one plant for a very finite period of time, which may not be indicative of the problems you will have at other plants. MR. MADISON: I am going to talk briefly about some of that, Dr. Powers, but we can talk a little bit more, as you wish. Next slide, please. We have issued SECY-99-007A, and we briefed the Commission on March 26th on this document, talked about some of the changes that are additions to the process, which included adding a significance determination process, and dealing with some questions in the enforcement arena. The Performance Indicator Manual has been redrafted and we are Draft Number 2, which looks like this. We have done an awful lot of work in conjunction with NEI to get this document ready at least for the pilot program. There is still more work to be done on it. There's some improvements that we'll be looking at during the pilot and subsequent to the pilot, but it is ready to go to be published as a draft manual for pilot implementation. The baseline inspection procedures have been written. We have gone through concurrence, not necessarily concurrence but a round of comments with all the regions and some headquarters staff, and on May 20th this package, which includes all the baseline inspection procedures was signed out by Sam Collins and for comment. I don't know if the ACRS has received a copy of this. We also expect to put this package out for public comment in a Federal Register notice. We have done a significant amount of work in the assessment process development and looking at changing the action matrix, which we can talk about some of the details in that, if you wish. We have also developed our definition of what we are calling unacceptable performance, which is the last column on the action matrix. That's made part of this package. In the significance determination process arena, which Morris will go into more detail, we do have a process now that we have tested several times, one by doing a feasibility review, and I might mention that industry did kind of a parallel feasibility review with the same data that our folks were working with, and came out with consistent results, so we have got kind of a double-check there. We do have some work yet to be done in this area. We are still working right now on fire protection, in the fire protection arena, coming up. We have a process that is in draft form. We are working with industry and others to get that in better shape. We are looking at the shutdown area, to add some improvements to the STP process and in the containment area. In the enforcement arena, I mentioned that briefly, and I am not really prepared to discuss that today, but we have issued a draft enforcement policy. We do have an enforcement policy for the pilot program that ties very closely to the significance determination program. DR. POWERS: I guess you alluded to difficulties that you encounter with respect to shutdown and operating states and I am going to be interested in how you do significance on things like emergency diesel generators when you don't have that kind of risk information about shutdown states. MR. MADISON: As I said, we are still working in the shutdown area right now but what our direction would be is to empanel, and I am going to go into this a little bit, a group of SRAs and risk experts to help us in that arena until we do develop a significance determination process specifically focused on shutdown issues. DR. POWERS: Now your SRAs, I mean, are acquainted with the techniques for doing risk assessments during shutdown? MR. MADISON: Our intent is we would make sure that they are, the ones that we would use, yes, and we would be asking Gary Holahan and Rich Barrett to help us out in that area. Next slide, please. DR. POWERS: And Gary, I take it, is an expert in shutdown risk assessment? MR. MADISON: Gareth, you will have to help bail me out on this one. MR. PARRY: I think there are people in the PSA Branch of NRR and DSSA that have studies shutdown and shutdown risk assessments quite a lot, particularly for things like the shutdown rule, for example. DR. POWERS: Okay. MR. PARRY: Okay. DR. POWERS: So we will have a knowledgeable cadre of people working on this issue? MR. PARRY: Yes. Also, I am sure you are aware, that Research is developing a shutdown program, which you have heard about a couple of weeks ago, I think. DR. POWERS: No, we will hear about it in this meeting -- MR. PARRY: Oh, okay. DR. POWERS: -- and the substance of that is that they don't yet have a program. MR. PARRY: Right. They are planning it. DR. POWERS: And that they are relying on the literature survey to get some insights into what is and is not feasible here, and I'll telegraph to you that there are members of the committee that think that that is not a useful activity, in fact that we have to be much more groundbreaking in this area and not -- that the existing literature is not adequate to address the kinds of issues that are going to be encountered in this kind of activity. MR. PARRY: Okay. MR. MADISON: Some of the other things -- DR. POWERS: I could be wrong too. MR. MADISON: Some of the other things we have accomplished to date -- we held three workshops so far to train staff, industry and the public, or at least make them knowledgeable of the proposed processes in the pilot program. The first workshop was held in the Glen Ellen area on April 12th, the week of April 12th, to train the pilot program industry and NRC Staff primarily on the performance indicators and how to report and how to analyze the performance indicators. The second workshop was a closed workshop with NRC Staff and state personnel, state inspection personnel, to train our folks on the inspection process, the baseline inspection procedures, how to plan inspections, assessment of inspections and the enforcement policy. That was held the week of April 26th in the Atlanta area or in Atlanta. The third workshop was just recently held the week of May 17th near the Philadelphia Airport and that was a large public conference. We had approximately 350 to 400 people present and we talked about the entire process and directed primarily at the pilot program. We plan a second inspection training workshop. We were able to get over 120 inspection staff trained in the Atlanta workshop, but we found -- we feel that we need more inspectors trained to fully implement the pilot program, so we are going to go for an additional cadre in the week of June 21st and we will do that in the Region IV area, planned in Irving, Texas. We have accomplished, we think, a lot of communication activities. I say "we think" because you never know with communications whether you fully communicated or not, and a lot of that is based upon the feedback we get. We have gotten a lot of questions. We have responded to those questions. We continue to collect questions and respond to them. We have actually put some of our frequency-asked questions on our internal website. We are working towards building an external website and we will put those frequently asked questions on that website. We have held at least two briefings of various staff down to the inspector level and senior management level at all four regions. We have also held briefings here at headquarters for various inspections staff and review staff. We are also planning on doing public workshops at each of the pilot facilities. And in fact, I will be going out next week to do the first one the evening of the 8th, in the Quad Cities area. As I said earlier, we are ready to start the pilot program. We have developed some ground rules, upfront ground rules which we have published and expect to maintain during the pilot program. We have established some criteria in advance, actually guide a few criteria in advance. Some of them were, as I mentioned at the March 26th Commission briefing, some of them are numerical or quantitative in nature, but a lot of them are qualitative. Because of that, we have established a couple of different things that we are looking at during the pilot process, and this may address a little bit of your concern, Dr. Powers, of how you assess the success of a pilot program. The qualitative ones, we felt uncomfortable with just the NRC staff saying, yes, we have been successful. So we established what we are calling the Pilot Plant Evaluation Panel. It will be chaired by Frank Gillespie. We have members of industry, in fact, one member from each region that has a pilot plant. We have -- NEI is represented on there with Steve Floyd. We have one member from the NRC staff in each region. We also have a representative from the State of Illinois and Dave Lochbaum from UCS participating on the Pilot Evaluation Panel. DR. POWERS: I guess one of the things that continues to concern me a little bit about this whole thrust is how the inspectors themselves are reacting to these rather different activities that they are being called on. Are they going to have some input on the judgments of success and failure here? MR. MADISON: The inspection staff has had input from the beginning. DR. POWERS: Yeah, I understand that. MR. MADISON: Of course, you are aware that they helped write the procedures. We made sure we incorporated their comments during the writing of the procedures, during the comment period of the procedures and during the development of the program. We have kind of made it open season on the transition task force. We will collect any comments received and work towards responding to them and incorporating them into the program. During the pilot program, we have established a support group that includes a regional coordinator for each region to collect and collate and ensure that the comments are addressed during the pilot program. We also have, as I mentioned on the slide, an oversight group that will be chaired by Bill Dean on a biweekly basis, poll and actually solicit comment from the pilot plant branch chiefs and inspectors to how the plant -- how the pilot program, how the program itself is being implemented, how it is -- the problems that they are running into, where improvements can be made. And we will consider actually revising the process and the details of the process during the pilot program, on a limited basis. We are going to try to -- we want to minimize the editorial type revisions. We will kind of hold those to the end of the pilot program. But if there is something that is presenting an implementation problem, we will make the change during the pilot program. We also have -- it is noted up here as an SDP, but during the pilot program, Rich Barrett's branch, under Gary Holahan, is establishing a -- what are you calling it? MR. PARRY: It is an operational support group. MR. MADISON: Operational support group that will be reviewing the significant findings. They will review all Phase 2s from the SDP and they will participate in any Phase 3s during the pilot program to ensure that we maintain -- have a consistent application of the process. DR. SEALE: During the presentations we had earlier this year, a lot of us were very impressed, and I think we said so in some of our letters, with the input that this whole process had received from people from the regions and from the inspection people. Since that time, other things have happened. In particular, there has been a reorganization of NRR and in that process a lot of us were concerned with losing the independence and objectivity of the old AEOD functions that were integrated into -- partially into NRR and partly into Research. But another effect of that reorganization, at least it appeared, was to take the relationship between the inspection process and the staff people here who were working on this evaluation thing and sort of put them at the poles of the NRR activities rather than shoulder to shoulder, from an organizational point of view. I hope that all of these things you have listed here are an attempt to overcome the adverse effects of that kind of long communication line within the linear aspects of an organization chart in getting this process done. And I certainly will be sensitive to hearing about how well you are continuing to carry out the kind of cooperative creativity that the early efforts in this revision of the inspection process showed. Hopefully, you can keep that up. MR. MADISON: I echo that hope. That is definitely one of the major objectives of a lot of the processes that I have described, including one that I didn't mention was the -- we have assigned owners of the inspection procedures, and one of their functions is to solicit comments and collect comments and make sure that they are incorporated. We have tried to make sure that the staff, the inspection staff, feels fully engaged, fully involved in the process and that they are being listened to. One of the things we -- one of the nice things we got out of the Atlanta workshop was to see an actual change in attitude from inspectors from the first day of the workshop to the last day of the workshop. Once they had seen the details and actually had a chance to talk about the details, challenge them and talk to the people that wrote them, there was a lot more acceptance. There was a lot more of an attitude of, well, yeah, let's get on with it, let's try it now. And we are seeing that -- in fact, I got an e-mail from a Region I inspector this morning that said the same thing, you know, we are -- he provided a criticism, a comment, but he said, don't get me wrong, you know, we are anxious to get going. We are ready to get started on it. We just want to make sure you are still listening. And we want to make sure they know that we are still listening. There is still a lot of work to be done. We are not complete yet. Still that is why we are piloting. We are piloting with draft procedures to let -- again, to let folks know that we are still looking for comments. As I mentioned, we are still -- we are considering sending out and will send out a Federal Register notice with the baseline procedures. I want to kind of highlight that a little bit. This will be I think the first time the agency has ever sent out for public comment its inspection procedures, but we feel that is appropriate with this major change that we are making with the process and the program. The pilot program is going to run till the end of the year, so we have got a lot of work yet to do. We are just beginning, we are in day three of the pilot program -- day four. It started April -- or May 30th. We still, as I mentioned earlier, have some work to do on the Significance Determination Process in the areas of fire protection, shutdown, risk in the containment area. Fire protection, we are getting pretty close. We have held several public meetings. We have a draft process, we are working now at incorporating that process into the general Significance Determination Process, or at least incorporating their thinking. Industry has had a chance to look at it, comment, and we are going to do a tabletop exercise with industry to make sure we have the same understanding of how the process should work. We are looking at performance indicator changes not just revisions to the draft, but possible additions in the performance indicator area, either during or following the pilot program, in the areas of shutdown, fire protection and we need -- we think we need in the near future an unreliability indicator, and we are working very hard to get that. We are getting assistance from the Office of Research in those areas, and they may be talking about that more later this afternoon. I have training up here as a bullet. We feel that we are accomplishing some training as we go. We probably have over 200 -- about 50 percent of what we look at as the inspection staff, somewhat trained, at least trained adequately to implement the pilot program. But we need to train the rest of the staff. And we are developing right now in conjunction with the Technical Training Center plans to train the rest of the staff, to update the staff that has received training already in the fall and winter time periods, and we will try to accomplish that such that all four regions will receive adequate attention, as well as headquarters staff. We are also looking at holding a public workshop in all four regions, and we would do that in conjunction and cooperation with NEI, to inform the public and be able to respond directly to public questions as well as industry questions. And I already mentioned that we are going out and doing public meetings at each of the pilot facilities. Members of the technical -- or the Transition Task Force will also be going out to the pilot plants. Morris, Don Hickman in the performance indicator area will avail themselves at each of the pilot plants, to not just the NRC staff but to the industry staff. DR. SEALE: Will you keep us informed as to the location and time of those public workshops in the regions? MR. MADISON: Certainly, I will make sure you get that. One other thing that is not on the slide here, that didn't appear on the slide, it was meant to be on the slide -- it was talking about what I will call right now the above the baseline inspection activities. What happens when a performance indicator crosses the threshold? What happens when an inspection finding crosses the threshold? What do we respond with? We have just begun staffing to develop that. We have a concept that would include both that focused inspection to follow up, reactive inspections, a broad spectrum including incident investigations, AITs and other, and also generic. What do you do with generic safety issues, how do you continue to follow up on generic safety issues? As I mentioned, we have just begun to staff up developing that program and expect to have it in place prior to full implementation. That concludes my status and yet-to-do remarks. If there aren't any questions, I am going to let Morris get started on the Significance Determination Process. MR. BRANCH: Good morning, I am Morris Branch of NRR and I am currently assigned to the Transition Task Force. I am the task lead for the Significance Determination Process that I will describe, and, also, I was the task lead for the feasibility review that we did in February, that I will describe later. I am here today to discuss these two tasks that were key elements in the development of the new reactor oversight process. And both of these areas are described in SECY-99007A. Before I begin, I would first like to say that this effort involved a wide variety of agency assets. Our task group included members from Research, NRR, the Office of Enforcement, the Technical Training Center, and all four regional offices and their members that are here today as well. My background is inspection. I was an NRC field inspector for 16 years, I was a resident inspection and senior resident. Since I have been in headquarters, I have led several of the AE design reviews, and I brought the end user perspective to the project. I would now like to briefly describe our efforts to date in developing the process to assign a risk characterization, which we refer to as the Significance Determination Process, to an inspection finding. This -- you didn't talk too much about the process. Anyway, the reason that we needed to apply a risk characterization or significance to an inspection findings was so that the inspection finding could be dovetailed with the performance indicators during the assessment process. And we wanted to develop a process that would give a significance scale that was similar to that used in the PI thresholds. The process is intended to be used by field inspectors and their managers, and this presented a challenge to our task group in that -- to try to make a simple tool, but at the same time provide some fidelity to the process. And for the most part, wed use the old enforcement criteria as a basis for coloring the findings in the non-reactor cornerstone area, and in the reactor cornerstone areas, the process was developed using inputs derived from other agency products. We used Reg. Guide 1.174, NUREG-5499, which was an AEOD study providing the likelihood of initiating events. We also used NUREG-4674 and that describes the as-screening rules we used the as-screening as part of our process. And we used typical equipment and human performance reliability values generally consistent with those obtained from PRA models. From this slide you can see that we have developed several processes to address inspection findings in all the cornerstone areas. Our guidance to the inspectors will be contained in the NRC inspection manual and there will be attachments to describe the many processes. And I will briefly describe some of the non-reactor cornerstone areas. As Alan mentioned, we still are working on the shutdown Significance Determination Process. Containment, currently in our process we have the inspectors calling the risk support group in headquarters for help and the SRAs in the region. And as Alan mentioned also, for the fire protection, we are still working on development. DR. POWERS: This flow chart that you have here, are you intending to walk through this? MR. BRANCH: Yes, I was going to show you that basically what I was -- the purpose of this chart was to show that we do have different processes. If it is radiation safety safeguards or emergency preparedness, we are going to have an attachment appendix to the manual chapter that will cover those. And we used a different methodology, we didn't really use risk as well there, we used more of a process flow barrier type. And, also, the chart was to demonstrate that as you go through the process, if the issue is of low risk significance, then we send it to the licensee and let them correct it through their corrective action program. And also, after we go through the significance determination, the output of all of these processes is the input to the assessment and the enforcement process, and that was the purpose of the slide. DR. POWERS: Here is the question that comes to my mind. I started at the top on this and I said I have a finding or an issue, and I have a question here. It says, affects emergency preparedness, radiation safety or safeguards. Now, those are the areas where you don't use risk, you have some other criteria. And I said, now, can I imagine a finding or an issue that I wouldn't answer yes to that if I used my imagination a little bit? And I said, even if it was a finding that I would subsequently answer the questions lower down on this, I come yes and it takes me out of the loop. I mean there has to be some real constraint on me there because I have this radiation safety term in there. And with a little imagination I can say "affects radiation safety, yes, it affects radiation safety." Any finding that I come up with I can say, "Yes, it affects radiation safety." How do I avoid always going on the yes loop there? MR. BRANCH: The purpose of the slide was just really to show the different processes. It wasn't intended to show any real logic to it. You don't have to go through that gate first. I mean, we could have had this coming in at all ends. It was just to show that there are different processes that we'll have. We had to use a different method to develop for the nonreactor cornerstones. That was the purpose. DR. POWERS: Okay. So really the situation is that in reality, and the way it would be actually implemented is that I would first check those things where I could use risk, and if it didn't fall in those categories, then I would go to the nonrisk ones or something like that? MR. MADISON: Yes, that would be the right way to do it. MR. BRANCH: Okay. This process diagram describes the SDP for the power reactor cornerstones. You can see from the diagram the first step in the process is to clearly identify the concern. That's this part up in here. You have to -- doing the process development and doing our feasibility review it became clear that the inspector's concern and any assumptions has to be formulated prior to using this tool. And this part of the process is similar to performing an engineering calculation. You first state the problem, the assumptions you're making, and then you can use the process and expect repeatable results. This is an assumption-driven process. The next step we refer to as phase 1, and that's our screening process. And this screening will be accomplished by field inspectors. We believe that many items will be screened as nonrisk-significant in this step and will be passed to the licensee for resolution through their corrective action program. Our process forces the inspector to make reasonable but conservative assumptions. Therefore, inspectors will most likely pass more items than necessary to the phase 2 review. But that's okay, because false positives at the inspector level could be refined later during the phase 2 review we'll describe later, and also the phase 3 if it's necessary. After you go through the screening, and I'll go into the screening process in a little more detail after I finish going through the logic here, but after the screening you've determined that an item requires a phase 2 review, the inspector has to ask what initiating events are impacted by the finding. There may be more than one scenario that has to be reviewed. We have attempted to provide guidance to allow a field inspector to conduct the phase 2 review. However, until the inspectors become more familiar with the process, we anticipate additional risk analyst help will be needed. The next step in the phase 2 review involves determining the frequency of initiating events, and I'll go into the Table 1 and Table 2 after I finish going through this process. You determine the frequency of the initiating event and the duration of the graded condition, and then you determine the likelihood of the occurrence of the initiating event while the degraded condition exists. And then you consider the availability of mitigation equipment. Mitigation of the risk-significance of an issue is based on the equipment available to perform the high-level safety functions: heat removal and inventory control, et cetera. The general rule of thumb is that each line of mitigation available that will describe in the Table 2 represents an order of 10 change for the better in delta core damage frequency. After you finish the phase 2 review, you will have determined the final worst-case risk significance of an issue. This determination is represented by a color scheme similar to that used in the PI threshold values. We build into the process a phase 3 review if needed. This review will be performed by risk analysts and will allow refinement of the risk characterization and the significance of an issue prior to final action. I'd now like to briefly describe the screening process and the use of Table 1 and 2, and also, as Alan mentioned, NEI worked through several examples of issues -- DR. APOSTOLAKIS: Before you go there, in phase 1 screening, in step 1.2 -- MR. BRANCH: Yes, sir. DR. APOSTOLAKIS: Potential impact on risk means CDF and LERF, I suppose, not risk itself, CDF, core damage frequency. MR. MADISON: Yes. MR. BRANCH: Yes. DR. APOSTOLAKIS: Impact on risk. You mean potential change in CDF and LERF for that plant? Is that the impact we are looking for? MR. PARRY: Yes, it is. I mean -- and we're using the risk here and CDF and LERF as surrogates just like in Reg Guide 1.174. DR. APOSTOLAKIS: Now if I follow the approach for setting thresholds described in SECY-99-007, taking the 95th percentile, the plant-to-plant curve -- MR. PARRY: We need to discuss that. That's -- DR. APOSTOLAKIS: Yes. MR. PARRY: Not the right interpretation, I don't think. DR. APOSTOLAKIS: But let me just complete the argument here. MR. PARRY: Okay. DR. APOSTOLAKIS: Then of course four or five plants will be above the threshold the moment I select it. So now I take one of those plants, but there is no impact on risk, because that is already part of the PRA for that plant, right? So you will say the threshold tells me that the unavailability of this system is higher than the threshold, so I go and look for potential impact on risk, and that's zero. Because that risk value came from that number. That number is already part of the PRA. And that's the problem with using plant-to-plant curves to define these things -- one of the problems. DR. KRESS: I think you could interpret that differently, George. MR. PARRY: Yes. DR. KRESS: You could say that the value of the performance indicator for that plant is a measure of whether or not the system is matching the availability that you put into your PRA number, reliability number. So that that is an indicator telling you whether or not that was a good number in the first place, or whether you're matching it. DR. APOSTOLAKIS: But that's not the use of the thresholds right now. You are making it now plant-specific. DR. KRESS: Yes. That would be -- DR. APOSTOLAKIS: That's what I've argued all along. DR. KRESS: Yes, it would be. DR. APOSTOLAKIS: But if it's the way it is now as described in the document without any oral interpretations, automatically this plant is above the threshold, then I take the unavailability and I go to -- DR. KRESS: I think most of those unavailabilities are not plant-specific; they come out of a data base that's industrywide. DR. APOSTOLAKIS: No, but they have plant-specific numbers. MR. PARRY: This is -- the way that the setting of the thresholds was described in 99-007 doesn't give you the full story. What the threshold was set on was data from all the plants that we had in the data base, and it chose the worst value of that indicator over the period of observation, which I think was something like three years. So in a sense you should interpret that graph as meaning that over the period of observation that five or so of the plants at one time or other exceeded the threshold, but it didn't mean to say that they were stuck there. DR. APOSTOLAKIS: Oh. MR. PARRY: It actually fluctuates over the years. So it's a way of setting -- it's a way of trying to catch outlier behavior. Now if you remain at that level for a long time, then this calls for concern. But remember also that these indicators are only a license to dig a little deeper. They're not necessarily saying that the plant is unsafe at that point. DR. APOSTOLAKIS: Sure, but unless you make the indicators plant-specific, you will have logical problems all the time. Here is one. MR. PARRY: You will, in the sense that I think that the way the threshold's been set for the PRAs, they're set to the -- I don't want to call it the lowest common denominator, but it's the smallest increase of the plants that we looked at. So therefore for some plants that when we say they're going into the yellow, for example, if they were to do a plant-specific PRA, they would certainly not be. They could well be in the white, they could even be in the green, depending on the details of that plant. And that's one of the things you have to live with when you have a generic process as opposed to a plant-specific. DR. APOSTOLAKIS: Can we settle this thing once and for all, plant-specific or -- I mean, now you're giving me new information which was not in the document I reviewed. MR. PARRY: Yes, I think you're right. DR. APOSTOLAKIS: Which would certainly affect -- MR. MADISON: I don't think we've ever advertised them as plant-specific performance indicators. DR. APOSTOLAKIS: But they have to be plant-specific. It seems to me -- MR. PARRY: The measurement is plant-specific. MR. MADISON: But the thresholds and the idea of the performance indicators were on a band, provide a band of performance and to be in a generic manner, there's a couple of reasons for that. One is the time frame to develop plant-specific performance indicators we'd still be doing that, just that part of the work. Well, there's another issue. MR. PARRY: There's another issue there too, and that is that, as you know, not all the PRAs are done to the same standard. Therefore, if you want to develop plant-specific thresholds, then I think you have to effectively open up the whole question of whether the PRA is adequate. MR. MADISON: But I think you also need to look at what is the response, what's the NRC response to those crossing that white threshold, and as Gareth mentioned, that's a license or a permission or direction actually to the staff that you may need to look at this closer, you may need to consider followup inspection. They have -- there's a significant -- they deviated significantly from the norm, and NRC, you need to keep an eye -- it doesn't say that we're going to issue an order or that we're going to take any punitive action against the licensee for crossing that threshold other than we need to investigate it. DR. APOSTOLAKIS: But even in the investigation, there has to be a discussion of the issue. Let's take the other case. Let's say that somebody -- a plant has a low unavailability of a particular system, and they have maintained it for years, and the PRA uses that number, and that happens to be, you know, a factor of four below the threshold you choose, because the threshold is industrywide. Then the plant looks at that and says well, gee, you know, even if this goes up a little bit, I will never cross the threshold, so nobody's going to notice. Is that something that you would like to see? So the plant now has deviated from its practice before that time because now you've set the threshold higher, so now the numbers slowly shift up and nobody notices, because it's still below the threshold. So the excellent performer is allowed to become a little worse. You see, you really need a good section there describing exactly what you are trying to do. At this point either you have to tell me what's hidden behind the data that you show, and that was, you know, an important piece of information, or it will be left up to the people who will implement this later to make judgments like this. I think the issue of plant-specific versus industrywide performance is an extremely important one here. DR. BONACA: Yes. I would like to add from that respect also, because you have the generic curves addressing an individual system, for example, for different plants. But these plants are all different. DR. APOSTOLAKIS: Yes. DR. BONACA: If they were identical, I would say that is meaningful, but it's meaningless in my mind to call generic when you're looking at different plants for which an individual system has different importance. DR. APOSTOLAKIS: That's right, and they recognize that. They recognize it when it comes to the diesel unavailabilities. MR. GILLESPIE: Yes. I think we're kind of -- these indicators were picked as indicators, not as measures. They are not measures of risk. I think we were very up-front with that in the beginning. They are -- and there's multiple parameters involved. And they were set at basically a level that was a 95-percent level from '92 to '96 as an indicator, not as a measure. George, I don't disagree -- DR. APOSTOLAKIS: I understand. MR. GILLESPIE: That it wouldn't be nice to have them plant-specific, but then we're going to get into high-fidelity PRAs, staff review of PRAs, for every plant. Now maybe we get there in the future, and then -- this is not competing with it. If you break an indicator, it's only an indication. Then you get an inspection. The specific inspection results go through this process, and then you're plant-specific. So evaluating the risk isn't done by the indicator, it's done by the plant-specific inspection results. DR. APOSTOLAKIS: But it does trigger some action, though. MR. GILLESPIE: It triggers us looking at something, but if the unavailability -- South Texas, for example, told me they're within 28 hours on one particular system on unavailability, and they said they feel like they're riding the threshold very, very, very close just with their normal plant maintenance. DR. KRESS: I'd certainly -- MR. GILLESPIE: We know that. DR. KRESS: Yes. MR. GILLESPIE: If they break the threshold, we're going to know that. It's, you know, if -- I'm not sure what your -- DR. APOSTOLAKIS: Well -- Tom, do you want to -- DR. KRESS: I'd certainly disagree with the concept that these indicators ought to be on a plant-specific basis. I think that's just the wrong way to go. You guys are confusing comparing the plant-specific performance to the criteria that they ought to have met. You want to set a performance criterion of some kind, and the question is now how are we going to arrive at what these thresholds of performance criteria are. There's lots of ways you can set the baseline values you want. You can derive them -- you can pull them out of the air, for one way. You can derive them in sorts of ways. This was a way to set an industrywide threshold set of criteria that you want performance to meet. Then you compare on a plant-specific basis, those are plant-specific comparisons, see whether they meet these performance. It's a plant-specific comparison of an industrywide set of performance indicators. I think that's the way it ought to be. DR. BONACA: Let me just say my main concern is only one, that in one particular case a plant may have a system that sticks out, just because of the configuration of the plant. Now what will happen is the inspection program will recognize that. But I'll tell you, that will be like a stigma. It will never go away. It will never go away. It happens that way, because it is too high, the management will say why are we high, and back, and then you'll have people looking at precursors and that thing -- the plant will be for 40 years -- it might be 60 if we go to life extension -- and discussing why. That's not a big issue with them. DR. KRESS: What you're raising there is the issue of whether these are indeed generic performance indicators. Maybe you throw that one away if you have one like that and don't use it anymore. But if these are generic performance indicators, that should never happen. DR. APOSTOLAKIS: This is -- the fundamental question is the following, at least in my mind. First of all, I've heard many times that it's good to know who's doing better than you are and, you know, why they're doing better. The fundamental question is is this a job for the oversight process or is it a job for industry groups and organizations like INPO. Each plant should know whether they are doing worse in certain areas than other plants. But that's an industry issue, not an NRC oversight issue. These plants have been declared as safe by the Commission. That's why they're allowed to operate. The oversight process wants to make sure in my mind that what you thought or what you verified was happening yesterday will be happening six months from now, not to compare six months from now this plant with other plants. See, that's the fundamental disagreement. MR. MADISON: George, I read your comment on that, and that is definitely one way to do it. In fact, I would say that's the way we've been doing it for the last 20 years. That's the deterministic method to look at whether the licensee's complying with their license, their commitments, and the regulations. But nothing in this process tells the licensee that they don't do that anymore. But we chose to go a different way, a different direction, to assess the licensee's performance. And we have already explained our method, our madness in the way that we chose to go. It is different, it is not looking at the licensee's commitments. It is not looking at the regulations directly. And it is just a different way to look at the assessment process. DR. APOSTOLAKIS: I understand that, Alan, and I thought the difference was going to be the risk information, not to include other plants into the assessment process. MR. PARRY: I think it is -- we have got to back off this a little bit. The reason for choosing the population of plants is that what we were looking for was a level of acceptable performance across the industry. And you have already said that all of the plants -- DR. APOSTOLAKIS: Why is that the job of the oversight process? MR. PARRY: It is not a job of the oversight process, it is a job of trying to fix a criterion for the performance indicators. MR. MADISON: When it would be appropriate for the NRC to engage. MR. PARRY: What we have said -- MR. MADISON: And that helped us determine that level. MR. PARRY: What we said was, exactly as you pointed out, that all the plants have been reviewed and considered as being safe. So that given that, if we look at the performance of the individual systems in that population of plants, then we should be able to set a performance goal for those systems that is consistent with the plants being safe. And that is just setting a criterion, just one way of doing it. DR. APOSTOLAKIS: Well, no, I guess I disagree with that. I don't think there are many ways and this is one way. I think the fundamental issue is what are you trying to achieve. And in my mind -- MR. PARRY: All we are trying to achieve is indicators that point out when the performance in a particular system becomes significant. Now, we have chosen the green light threshold as saying if it is below that, then it is not particularly significant. DR. APOSTOLAKIS: Yeah, I understand. I guess the question is, what is significant? Significant with respect to what? And in my mind, the determination of what is significant should be with respect to the performance, anticipated performance of this particular plant. If San Onofre decides to spend $5 million to make an improvement someplace, I should not be penalized for that because I have been licensed, you have looked at my IPE, my IPEEE, I am doing fine. Now, if the industry wants to put pressure on me to perform better, that is the industry's job. MR. PARRY: Are you under the impression that these thresholds are going to change continuously? DR. APOSTOLAKIS: No. MR. MADISON: No. DR. APOSTOLAKIS: Even if they don't, you start -- there are two problems. The plants that are above the threshold automatically get a stigma that Mario mentioned. And the plants that are below now, I don't know, if their performance deteriorates, how you will catch it, because they will still be below the threshold. And there are other questions as well. That is why I am saying that you should really spend a little more time thinking about this. You talk about randomness and so on in the report. You just old us that the data we saw in the report were over three years. There is no discussion that I recall as to what the sample should be. When you calculate unavailability, what do you do? You do it every month and then you compare with the threshold? Or you do it over a year, over two years? MR. PARRY: Actually, I think it is a three year rolling average, I think. DR. APOSTOLAKIS: But that is the data you have. Now, when you start -- MR. PARRY: No, that is the way the indicators are going to be evaluated, I think. And if Don Hickman is -- DR. APOSTOLAKIS: But your pilot is for six months. So you are going to go back three years and get the data? MR. MADISON: Yes. We have already collected data from the licensees that will give us that backward look. DR. APOSTOLAKIS: Now, why three years? Why three years and not one? I mean -- MR. PARRY: Well, the longer the better, I guess, but in some ways. For smoothing out. DR. APOSTOLAKIS: I don't know. That is what I am saying, that I have not seen evidence that you have really thought about it very carefully and you decided to do this and not do that, and the implications or consequences of certain choices you have made. MR. PARRY: Actually, I think we have thought about it quite a lot. And one of the -- DR. APOSTOLAKIS: You just don't want to tell us about it. MR. PARRY: One of the problems, as you know, that the events we are dealing with are not very frequent and they are fairly random in their distribution. When you have got things like that, you have got to accept the fact that there is going to be some statistical uncertainty. We need some averaging. DR. APOSTOLAKIS: And it is not clear to me how you are handling that, that is what I am saying. It is not clear to me how you are handling that. Are you going to have a band? I mean that is what they do in quality control, they don't deal with rare events, so, you know, they can check. MR. PARRY: Well, the bands are the white and the yellow bands, if you like. DR. APOSTOLAKIS: No, but there may be a statistical variation that takes the number a little bit above the threshold, and I don't know whether that justifies going -- MR. PARRY: Well, to some extent, that has already been accounted for, at least in the green and white threshold. Now, I agree that when we set the white, yellow and the yellow, red, because we are dealing with average values, we are dealing with PRAs, which deal with means that don't capture these statistical fluctuations, there may be a problem there. But, actually, if you look at the thresholds, and I was looking at some of the data this morning, it is not -- it is very rare that they are actually going to cross into the yellow for sure. DR. APOSTOLAKIS: Well, clearly, then, the arguments you gave me, because I don't think we are going to close this today, did not convince you when it came to diesels. You felt that for two train plants, the threshold should be at .05, while for three or more trains, it should be .1. So there is an attempt to come closer to the design because you couldn't get away with it. And I will come back to my comment and Dr. Bonaca's comment -- why not make the whole thing plant-specific instead of playing with trains and this and that? I don't know that the high pressure injection system may have the same problem somewhere. I mean you are a PRA guy, you should know that. If there is one thing I have heard a million times that we have learned from the PRAs is that they plant-specific. MR. PARRY: I understand that. And, clearly, that would be a nice goal. But as you know -- DR. APOSTOLAKIS: It is unattainable. MR. PARRY: Well, certainly in the short-term it is unattainable. DR. APOSTOLAKIS: How the maintenance rule, didn't we ask the licensees to set some threshold values there? Why can't we do the same here and have them do the work if it is too much for us? We didn't say that the unattainability -- DR. KRESS: George, these are performance indicators. Would you give me a high level principle that allows me to work from to derive a value, an acceptance value on a performance indicator? DR. APOSTOLAKIS: The principle is this, the plant has already been declared safe. Even if the CDF is greater than the goal, as you know very well, we don't do anything. So all I want him to tell me is, I have achieved this number by this kind of performance, which means this train has this unavailability, that train has that unavailability, that is my threshold. If I exceed that, then I am deviating from my performance and the NRC ought to know about it. I think that is a perfectly logical way to do it. DR. KRESS: Would you, in a risk-informed world like 1.174, instead of performance indicators, if we had risk indicators like CDF or LERF -- let's ignore the fact that you are never going to see those right now, of course, and assume there was some level thing, would you like to see different CDFs and LERFs for different plants as acceptance criteria? DR. APOSTOLAKIS: For the oversight process, yes. DR. KRESS: You would like for each one of them? DR. APOSTOLAKIS: My fundamental problem is, what is the objective of the oversight process? I am not saying that is the objective of the whole agency. But when I say I am going to inspect, I go with very stringent boundary conditions. All I want to know is that six months from now, they will be doing what they are doing today, which I have approved. And if they deviate from that, I want to know about it. Now, whether I don't like the overall performance, this is not the program to tell me that. MR. MADISON: But, again, as I pointed out, we do not have the same objective that you are stating for this oversight process. We have -- our objective is to assure that the agency meets its mission through the safety goals, by meeting the objectives of the cornerstones, that is, of the process. MR. JOHNSON: George, this is Michael Johnson. I haven't spoken yet today, and I almost hesitate because we have got so much to cover to try to jump in, but I just want to remind us that what we are up to is something that is very simple and we really are about trying to figure out what is happening with respect to overall plant performance. Yes, we have seized upon some performance indicators, some better than others, and we hope to improve those performance indicators. Research is going to talk this afternoon about how we will make those improved and more risk-based PIs, but we really wanted to come up with a simple way to look for variations in performance that tell us that we need to engage, become increasing engaged before plants end up in a situation where we need to take some drastic action to ensure that our overall mission in protecting public health and safety is not being threatened. In essence, it is very simple, and I almost -- my medical analogy that I always throw out applies. If we were trying to figure out, if we seize on one indicator, one medical indicator, it would be your temperature, your body temperature, and your body temperature is different from my body temperature, which is different from Alan's body temperature, but we could simply agree, I think, that if we have let's say that exceeds 99 degrees or whatever it is, you tell me what it is, that may be an indicator that there is something going on that I need to do some additional investigation on -- and I could probably measure a delta body temperature -- DR. APOSTOLAKIS: Mike, nobody licensed you to live with a 99 degree temperature. MR. JOHNSON: Exactly. Now this threshold that we are talking about is a threshold that is overall performance. It is well -- it is below, it is at a level where we have time to take some additional interaction. We are not at the point where we are looking at even changes that are significant changes that could be measured in terms of delta CDF, as you point out, a the Green to White threshold, so we are looking for a threshold that enables us to go as the regulator and begin to engage, very low consequence for the action, begin to look at the licensee to say what is going on, is there something going on? There will be instances because of the crudeness of the PIs because we are not plant-specific where the answer will be is the something that is going on is the plant is a little bit different and so we don't need to take additional action. There will be other instances where there really is something going on with respect to performance that we need to take additional action, but again we are trying to come up with a single template that enables us to take a look across the spectrum of plants to be able to recognize performance declines before they get to be significant. That is what we are about. DR. APOSTOLAKIS: Well, I guess we cannot reach closure today, so the best way to approach this is to air the difference, especially make sure the Commissioners understand that there is a difference of opinion and then leave it at that. If they say go ahead, this is what we like, they are the ultimate authority. I am not even sure the industry understands this. Did the issue come up in your workshops? MR. MADISON: We have talked about having individual plant performance indicators. I don't know, Tom, if you want to comment on that. Tom Houghton from NEI. MR. HOUGHTON: We looked at that. We tried looking at that and what we decided was that it would be best to use a generic level of performance that would trigger the beginning of a look. In other word, the idea being that currently any deviation by the licensee becomes subject to inspection and the inspection doesn't have any boundary and we thought by setting a random error band in performance -- if you were above that threshold, then the NRC would have justification for saying, okay, your performance is a good band and we'll just look at what you are doing. On the issue of, if I could answer another question -- thank you -- that had to do with would the performance, you know, would the utility allow its performance to slip if it stayed within the green band, the indicators are going to be tracked graphically so that will be visible if the performance slipped. I don't think they are going to let it slip because I know that utilities are setting up administrative thresholds higher than the Green-White threshold so that they don't come in danger of exceeding the Green-White threshold. DR. APOSTOLAKIS: Well, one last comment and I am done. Let's say we follow this process and South Texas is doing well or San Onofre. They are above the threshold, so the first time around the NRC says, well, gee, you are above the threshold, so they look. They go through Phase 2 and possibly Phase 3, so the risk analysts get together and say, oh, yes, sure, we looked at the PRA and so on, and so this particular value is indeed above the threshold but for this plant is it okay. Now from then on, would the threshold be this new value for that system for that plant, given all this analysis, or we forget about it and we'll go back again to the generic threshold and we repeat the thing a year later? -- because if you say the Phase 3 risk refinement demonstrates that this value is acceptable for this plant and from now on this is our point of reference, then all you are doing is you are becoming plant specific but at a slower pace, which is fine with me. But if you say, well, I convinced myself this time. Now a year later we look again. It's higher. We go through the process again. Then it doesn't make sense. MR. MADISON: Well, philosophically -- MR. PARRY: We learn from the process. MR. MADISON: I think to answer your question, Mike was starting to, but we are going to look at the process. That is part of the reason for the pilot. It is a living, dynamic process. It is not going to be static. We will continue over the next couple years to look at the process and make those adjustments if they are deemed necessary. You are mixing apples and oranges in one respect though. We are not using the significance determination process to review the performance indicators. The significance determination process reviews inspection findings and so the risk associated with that is equal at any plant. The idea was to make it fairly and maintain that fairly generic at this point until we develop better confidence in specific findings. DR. APOSTOLAKIS: No, what I meant, Alan, is you say specific finding identified. It is above the threshold. You go all the way down -- MR. MADISON: But the performance indicator is not the finding. What we are talking about are inspection findings and actual issues and Morris is going to go into a little bit of detail, I think, of the types of issues that we are talking about. Exceeding a performance indicator threshold is not an issue or a finding that would go under the SDP. MR. PARRY: There is not an equivalent Phase 3 evaluation, if you like, currently, but you're right. You could probably for a specific plant go in, look what the value of that indicator does for you, and use that as a reason to show why it is not significant, although it has not been written into the programs. DR. APOSTOLAKIS: And that should be from that time until forever the new threshold -- MR. PARRY: Or at least the knowledge -- DR. APOSTOLAKIS: Because otherwise you repeat the process every time. MR. MADISON: Definitely. Yes. And there is no intent to do that. DR. APOSTOLAKIS: So you don't think that would be useful to put in the document? MR. PARRY: It might be. I mean it might be something we should think about. MR. MADISON: We think those are exceptions and we will deal with those as they come up. DR. APOSTOLAKIS: I am not sure they are exceptions, Alan. I am not sure. MR. MADISON: That is part of what the pilot program is going to tell us. DR. APOSTOLAKIS: You would have 104 exceptions at the end. I'm done. DR. SEALE: Stick a fork in him. [Laughter.] DR. SHACK: Do you remember where you were? MR. BRANCH: Earlier I described the process as a Phase 1, which is a screening process, the Phase 2 as our simple assessment of the risk, and then the Phase 3 we didn't talk about that in as much detail, but in Phase 3 will be a refinement, if necessary. In the Phase 2, the screening process, the first -- and we have tried different methods of teaching this to the inspectors. We have worksheets that we are working on to try to get the information out to them. This was just a simple diagram we tried to use doing the feasibility review, but basically what you do is once you have your finding, and let's say it is an MOV that is not operable that affects a train of equipment, you have to ask yourself, well, does it affect more than one cornerstone. If it could impact or initiate an event as well as mitigation, we go right straight into a Phase 2 review. If there is no expected impact, in other words if it is just qualification issue under 91-18 to still operable, we would screen that out and the licensee would fix it under the corrective action program. The next question we ask ourselves is does it represent the loss of a safety function of a system. If the answer is yes, it's a loss of function, we go into a phase 2. And then the next question we ask is does it represent a loss of safety function of a single train greater than the tech spec allowed outage time. Again, if the answer is yes, we go into a phase 2. If it's no, then we ask is the impact -- this is for the initiating event -- if the impact of the finding is no greater than increasing the likelihood of an uncomplicated reactor trip, if that's yes, we screen it out, and if we go through this process, we'd end up into the phase 2 review. In the phase 2 review I mention we look at once we've screened the process in and we're in the phase 2 review, what we developed is a table -- this is Table 1, and then Table 2 -- and Table 1 is basically -- and this was based on the AEOD's recent study as far as the generic probability of initiating events -- once we decide the piece of equipment that's inoperable, let's say it affects a steam generator tube rupture, aux feed equipment, you have to run through this process several times. But you enter the table with a probability of occurrence, and then you ask yourself how long did the condition exist. If it's less than three days, we're in this column over here. Three to 30, in the center column. And greater than 30 days, we remain at the initiating event frequency. We then go through the process, we select a color -- excuse me, a letter -- and then we go to Table 2. And Table 2, if we -- let's say if we picked a -- I don't know, what was the -- tube rupture was a C -- if we came in on this table here with a C, if we had no mitigation, we would color -- this finding would be colored red. If we, as we go over, we allow credit for recovery. This is one in ten, and this is 10 to the minus 2, 3, 4, 5, as we go over with more equipment. DR. APOSTOLAKIS: What does nonmitigation mean? MR. BRANCH: Well, for this event here, if there was no equipment to remove heat -- DR. APOSTOLAKIS: It seems to me that a fundamental requirement for all these -- now I guess it comes back again to what is a finding, Alan? What is a finding? MR. MADISON: It has to be -- it's associated with some inspection activity. And as Morris said, I thought I defined what we would rise to the level of finding. It has to be more than what would be considered today a minor violation. DR. APOSTOLAKIS: Okay. So you don't really expect that for any of those findings there will be no mitigation capability? MR. PARRY: No. DR. APOSTOLAKIS: Otherwise -- MR. PARRY: The finding may be related to the availability of a system. DR. APOSTOLAKIS: I thought Alan told me that's not the case. MR. MADISON: It won't be related to the performance indicator, but it may be related to this piece of equipment was unavailable for an extended period of time. DR. APOSTOLAKIS: But I thought the general guideline was the performance indicators and the various things we inspect are at a low enough level so that if there is a violation, we are not in trouble. So how can you ask the question is there mitigation capability? There must be a hell of a lot of mitigation capability. MR. MADISON: Well, that's not necessarily true, George. It's just like the same with the scram. The scram indicator counts scrams. But there are all kinds of scrams. There's a very simple, uncomplicated scram which would just be one tick on the performance indicator, it would be minimal followup by the NRC, but then there are other scrams that are more complicated that identify that there were pieces of equipment that were not ready to perform their function, that there were a lot more -- there was a lot more going on and the cause of the scram was a lot more complicated. Those types of scrams are going have to have more followup by the NRC, and they may fall into a finding, may generate a finding that goes into the significance determination process. MR. PARRY: Just to give you an example of a finding that might help your understanding of what that means. When we were doing the feasibility studies there was an example of a finding related to the automatic switchover to pump reset, that it would not have worked under certain circumstances at a plant. And in those cases for that particular scenario, which would be a small LOCA, or even a large LOCA, it would not have -- the amount of mitigating system would have been zero for that case, which is -- it's the finding that puts you in that column. MR. BRANCH: Right. One of the issues that we did do in the feasibility review for -- it was a D.C. Cook LER issue that dealt with the containment sump, and as we went through the process, again, you know, I indicated earlier you had to state your assumptions. And our assumptions were that the condition would not allow the water to get to the sump, therefore all ECCS was inoperable when you went into the recirc phase of operation. And with that we went in and credited no mitigation. Now I think on that one it was the medium-break LOCA turned out to be the one that drove it into the red area, because of the frequency of occurrence. MR. PARRY: Because that was the condition that was in existence since the plant was licensed, I guess. MR. BRANCH: Right. MR. PARRY: I mean, it turned out later that it wasn't that serious, but the initial LER had that finding in it, and that's what we analyzed. MR. MADISON: Also keep in mind that the numbers of issues that are going to actually rise to the red level are going to be very small. DR. APOSTOLAKIS: Yes. I hope so. MR. MADISON: On an annual basis. We hope so too. DR. APOSTOLAKIS: I was sort of arguing that it would be none. MR. MADISON: In the foreseeable future it may be possible that there are none, and that would say that the industry's performance has risen to that level. MR. BRANCH: The purpose of this slide is just to demonstrate that in the emergency preparedness area -- and we have a slide also for the radiation protection, and then one for safeguards -- and this was just to demonstrate how our process logic goes. And in this case here you have a violation identified -- I'm having a hard time seeing this -- and then in the area of emergency preparedness we use planning standards out of the 10 CFR as our basic -- like a barrier that failed, and as you go through the process again, these logic gates that you just -- if it's a failure to implement a planning standard during an actual event, you go to another chart. And in this case here the way the criterion was set up was that if you didn't implement the planning standard, and it's a risk-significant planning standard, during an actual unusual event, on alert, then as the issue as you process through here would color the finding. So that was the logic we used for emergency preparedness. This one here is the one we used for safeguards. We've had some question, and during our training we're working on developing better definitions for some of these. But basically as you go through the logic, and in this area here, the issues that typically would fall out here are the ones that are currently covered by enforcement policy as a level 4 violation. In the future they'll be NCVs. DR. POWERS: I don't know how I answered the first question. MR. BRANCH: It is subjective. We are in the process of working through this. MR. MADISON: We are developing guidance in that area. MR. BRANCH: We are developing guidance. We're working with -- DR. POWERS: How do you think about that? I mean, how do you go about developing guidance in that area? MR. BRANCH: Well, the first logic gate is if you just had a failure to follow procedure that dealt with something that was minor, the intent here was to go right straight into considering it a green finding, handing it -- passing it on to the licensee for a resolution through the corrective action program. DR. POWERS: Well, I'll admit that I can probably find items that any reasonable person would say well, that's kind of a trivial thing. But I'm not sure that I am in a position to take a collection of incidents and know which ones of those are not a trivial thing. MR. MADISON: And that is why we have asked -- I don't know if Dick Rosano, from the security group is here, safeguards group is here. We have asked them to develop guidance not just a definition of terms, but actually provide some examples to give the inspector guidance in how this would be implemented. MR. ROSANO: This is Dick Rosano, Chief of the Reactor Safeguards Section. That question does not ask what is the likelihood of a potential attack. That question is intended to ask, how much would that finding, that vulnerability assist the design basis threat in accomplishing sabotage? What we need to do is better define, and we do need to do that, we need to better define the risk that is associated with that finding and how much would it assist the design basis threat in accomplishing sabotage. It is not the probability of whether the attack would occur. DR. POWERS: Well, you haven't helped me very much because if I say that first one is how much does that assist, then I don't understand the purpose of the second question. MR. ROSANO: Okay. The second question goes to some issues that have a lot of history. For example, let me begin with an example so that you can test it out through the first block. If you found some administrative errors or minor vulnerabilities relating to barrier protection, but these -- let's say, for example, the intrusion detection system, be it microwave, refill, whatever the intrusion detection system is, its operability is essentially invisible to an attacker. A zone in a fence may be down. When I say the zone, I mean the intrusion detection system, not the fence itself. That would be a vulnerability, that would be a finding. And I am not going to assess right now whether it is minor or not because we need to define risk. But that would be a finding that you would put into the process. It would have perhaps a low probability of assisting an attack because it is invisible. DR. POWERS: Okay. So it is not easily exploitable and it is not predictable. MR. ROSANO: Well, no. No, but -- DR. POWERS: It seems as if you are answering the same -- 1 and 2 have gotten clouded with each other. MR. ROSANO: The two are very close, but if I can continue with that example, if there are several zones that are down, or if there is a single zone that is down for a significant period of time such that the staff on the site knows it, you have gone past the first block of whether there is some risk, because you are talking about several zones. You get into the second block of whether it is predictable or easily exploitable. That has to do with how many people know it, how long the vulnerability exists, and what extent or what percentage of the entire area is now degraded. DR. POWERS: It seems to me that the first question that is properly worded, it is not -- it seems to me, to take your example, you have got a zone down, you know that there is some risk. Now, you come to the second question, you say, yeah, but nobody can tell that it is down. It is not common knowledge on the site because it is only down for an hour. MR. ROSANO: If a single zone was down, it would probably fall out on the first question to green, if it was a single zone of a short duration. DR. POWERS: Well, okay. MR. ROSANO: But let's say, for example, that there was a single zone that was down over a period of a few shifts, so that it was known to several different shifts, no correction had been made. It would pass the first block and go to the next one. And then you would ask, was the noncomformance predictable or reasonably exploitable? And that would be an assessment of whether this information, or the fact that the zone was down or out of service was known and whether there was a threat, whether there was a heightened security condition at that time, whether there was some known probability of an attack being generated against nuclear power. A lot of other factors now become part of it to decide whether it could be seen and whether it could be exploited and not just whether -- the point of the first block was to eliminate small findings, things that, again, would be NCVs or 4s, or used to be 4s, to just drop them out so that they not be processed according to the other blocks having to do with exploitability. DR. POWERS: I think I understand now. The identification of a nonconformance, ipso facto, means that there is some risk. The first block really is a question on whether that is above some threshold of risk. MR. ROSANO: Yes. DR. POWERS: Okay. So now we have to presumably work out in the future what that threshold is. MR. ROSANO: That's true. And as Alan has pointed out, we have a problem with that definition right now and we need to work out some details. But I think that the point is that we don't want to enter this process with minor findings that wouldn't have gotten any attention in the past. We are trying to have an avenue to drop those out of the process before we start analyzing their significance in terms of exploitability. DR. POWERS: Okay You just want to find what threshold do you go to the second question, what takes you to the second question. MR. ROSANO: Yes. DR. POWERS: And you don't have that yet. MR. ROSANO: Not yet. DR. POWERS: And do I have any understanding on how you get that? I don't know what the equivalent of engineering judgment is in security space. I know that some people at Los Alamos failed in their security judgment. But -- MR. MADISON: Well, Dick is going to get a lot of help because they are also working with industry to come up with what that definition is. DR. POWERS: Well, hopefully not Los Alamos. MR. ROSANO: We will certainly get a lot of input. I am not sure whether we are going to get a lot of help. But I am being optimistic. DR. POWERS: Well put. Let's move on. DR. APOSTOLAKIS: Where do the numbers two events, three events, two or three events come from? MR. MADISON: That would be inspection, basically that number of inspection findings. MR. BRANCH: Yes, that was inspector judgment . DR. APOSTOLAKIS: Why 12 months and not three years that you will be using for other things? MR. BRANCH: Well, the indicators, some of the indicators are 12 month indicators. Some of the indicators are -- but this was basically an assessment cycle is 12 months. DR. APOSTOLAKIS: And there is a logical explanation of this somewhere? DR. POWERS: Why isn't it a cycle? MR. MADISON: I beg your pardon. DR. POWERS: Why isn't it a cycle? MR. MADISON: We are redefining the assessment cycle to be an annual cycle. DR. POWERS: Yeah, but why not a refueling cycle? MR. MADISON: It is a good question. We chose to have the assessment process fit with not only providing an annual assessment, which we feel is required of industry performance, but also fits in with our budget cycle for planning purposes. MR. FALLON: Yeah, and just almost arbitrary, I would guess, but the Commission preferred an annual report or an annual assessment of the performance of the plants. And so we said annual sounds reasonable to us. MR. BRANCH: Any more questions on this slide? DR. APOSTOLAKIS: An annual assessment using three year data, is that it? MR. MADISON: Some of the performance indicators have a three year -- they are three year rolling average. But the look is an annual assessment. MR. BRANCH: Again, this process is just to describe our logic dealing with public radiation safety. And as you go through the process, we have developed -- are in the process of developing transportation process logics as well, but, again, this is a way that we would go through to take a finding dealing with public radiation exposure and run it through our process and color it into a scheme that we could use for assessment. MR. MADISON: We have kind of used up our time. I don't know how much more you want to -- DR. POWERS: We have already handled that. We are letting this session go on till 12:15. MR. MADISON: Okay. DR. POWERS: You started 10 minutes late, so. MR. BRANCH: And we have done the same with occupational. Now, in this area here, I think earlier we talked about there wasn't a clear tie back to risk on some of these non-reactor cornerstones, but in this area here, when we were developing this, we did use some risk insight, and, also, I mentioned earlier, we used the enforcement policy. But in the case of exposure, we changed where the current enforcement policy on overexposure may be a Level 3, which would be the equivalent of a white finding, but in this case here, we ran it through the process and we -- DR. POWERS: I guess this one confuses me. Well, but it may be just because of the words on the slide and not the words in the text. If I have an extremity overexposure, that is one thing. If I have a whole body overexposure, that is quite a different thing. Is there a difference of the exposures that are going into this process? MR. MADISON: Roger. MR. BRANCH: Roger Pederson. MR. PEDERSON: Yeah, this is Roger Pederson, I am a Senior Radiation Protection Specialist in NRR and I helped develop this with NEI and our other stakeholders. No, we didn't look at the differences between the types of overexposures. We looked at the dose limits that are in Part 20 for the different types of exposures and an overexposure is an overexposure. DR. POWERS: Okay. It is just that the thresholds are different to finding what an overexposure is. MR. PEDERSON: Yeah, there are different dose limits for different types, for skin dose, for organ dose, for what used to be considered a whole body dose. Now it is 5 rem TEDE is a basic dose limit. But there are different dose limits for minors, different dose limits for fetal dose. So that overexposure is if it is any of those, if it is a dose in excess of any of the limits that are in Part 20. DR. SEALE: You chose not to argue with the ICRP. MR. PEDERSON: Excuse me? There is a little bit of a problem with skin dose and hot particles. But that unintended exposure block up at the top, we -- that is actually a definition that is in the PIs, the unintended exposure and that is 2 percent of the stochastic limits, 10 percent of the non-stochastic limits that are in Part 20, 20 percent of the minor and fetal dose limits, and 100 percent of the skin dose from a hot particle. So there is some adjustment there in terms of risk. MR. MADISON: Thank you, Roger. DR. POWERS: We are back to this marvelous question that I don't know how to answer. Substantial potential. I mean potential I understand, substantial potential, I don't know how to answer that question. MR. MADISON: This is not just -- the definition of terms is an issue in all of the non-reactor charts. MR. BRANCH: But on this one here, though, the term is actually in the enforcement policy today. MR. PEDERSON: Yeah, that is exactly right. In the enforcement policy today, in the supplements, an overexposure is the Severity Level 3 or a substantial potential for overexposure. That term "substantial potential," we defined in the enforcement manual about five years ago. It basically says if the situation, if the exposure event that happened, that did not exceed the dose limit, if a minor alteration to the circumstances as happened would have resulted in an overexposure, that is a substantial potential for overexposure. MR. MADISON: Thanks again. DR. POWERS: And so now all we have to do is understand what a minor alteration of events is. I mean you are always going to run into this problem. MR. MADISON: I should stop trying to sit down. MR. PEDERSON: In that definition, there is a number of examples to clarify that point. MR. BRANCH: When I started off I indicated I was going to basically describe our feasibility review in SECY 99-007, which outlined the new oversight program, it describes the Staff's plan to test the workability of the new reactor oversight process in early 1999. This was advertised as a limited review of a few plants using the available data to demonstrate the ability to assign a risk characterization of items typically contained in the plant's issue matrix, the PIMs, and the Staff also planned to conduct an exercise of the new process of the assessment matrix on limited data and to reach conclusions related to the actions to be taken using the new process. Because of schedule constraints, the feasibility review was performed at a time when many elements of the new program were still under development. We felt that was okay because the review was intended to identify improvements needed to support the pilot and the pilot is intended to identify and correct additional issues prior to full implementation in early 2000. The plants we looked at were D.C. Cook, Units 1 and 2, for the 1996-97 time period, Millstone Units 2 and 3 for 1994-95. To balance out plants between regions we picked St. Lucie out of Region II, Units 1 and 2, for 1997-98, and Waterford 3 for 97-98, and the reason we picked the time periods was in this process we had to align the inspection findings again with the performance indicators and these were the periods of time when the plants were actually operating, plus for Millstone and Cook this also encompassed the period where we had what was considered I guess at the time significant issues that we could run through the process and make sure that our Significance Determination Process would work for us. The participants for the one-week feasibility review consist of several inspectors or first-line supervisors from the four regions and we had several risk analysts from Headquarters. We had a member from the Office of Enforcement and also we had members from Technical Training Center. The first day we spent training, giving an overview of our new process that we had developed to the team. We broke into two groups, and during the second and third day we processed as many PIMs items as we could through the Risk Significance Determination Process. We could only effectively review about 20 or 30 issues per the groups in the two days allotted. However we did process items that we expected to be of risk significance, hardware items from LERs, and also this challenged the assessment processes I have described. DR. POWERS: Let me ask you a question. MR. BRANCH: Sure. DR. POWERS: When you went through the Significance Determination slides, there were lots of questions that I was unable to answer but presumably more educated and better trained people could answer, but that burden falls to do all those answers falls on somebody. In the steady state application of this, I presume it falls on the inspectors, is that true? MR. MADISON: That's true. DR. POWERS: So now we are going to enhance the amount of desk time of inspectors at the expense of plant time. Is this a good strategy? MR. MADISON: That was one outcome of doing the feasibility review, and one of the reasons why during the pilot study we are going to capture or collect the time spent on using the Significance Determination Process. One of the things though that we think there won't be such a great expenditure in this area is the screening process. The screening process allows the inspector to screen to Green a lot, a great many of the findings that they have without going through this process of the two tables that Morris showed, without going through the process of determining what are the mitigation factors that you can give credit for or how long it was an issue, so that again most of the findings will be screened through the screening process, Phase 1, which we think should be a relatively short period of time to accomplish. DR. POWERS: When you do an assessment of that, and it seems logical to me that a screening process helps you a lot on this, but I am thinking, gee, I am an inspector, I have been trained on something that I am inspecting, then I have to get trained on this other thing, and nobody is going to be happy with me being trained once. They are going to have me doing refreshers and things like that. All these things are conspiring to keep me out of the plant if I am an inspector. I mean there is a burden, even if you screening saves -- MR. MADISON: I understand your concern and that is why we are going to capture the time and monitor it, and we may need to make adjustments to the overall program, but I think there is a strength and we think there is a strength with this process in that it causes the inspector to ask the right questions. What is significant? What is important about this issue that I need to go find out as part of my inspection activities to determine the overall risk significance of my finding? So we think it will also help focus that inspection activity into the right areas. MR. GILLESPIE: Alan, in fact this may cause the inspector to be in the plant more, because you have got to consider what he is doing today and has been doing in the last several years, and the first Phase 1 of this is going to screen out likely a lot of issues that we spent a lot of time on the telephone about in morning calls. A lot of stuff that happens at the plant gets called in in morning calls and the inspectors are on morning call every morning, so we need to try this during the pilot. This could end up with the inspectors spending more time in the plants rather than less and part of the initial thinking on this whole thing as already kind of gotten into the current enforcement policy where people are screening things with a little more discipline already. DR. POWERS: I don't discourage you in your enthusiasm but I am willing to bet that burdens only go up on inspectors here, that they will still get morning calls about things and discuss things on those morning calls that would not pass the screening, and it will take a generation of inspectors. MR. GILLESPIE: Well, that's also true, because it's not just the inspectors. You have to get their supervisors trained to say this is a trivial scram, let's not spend hours discussing this. DR. POWERS: Yes, you are going to have a lot of people to train there, some of whom are less trainable. MR. MADISON: Yes, and that is part of the reason for the oversight group to monitor the implementation during the pilot program and provide that feedback and training to not just the inspectors but to their managers. DR. POWERS: Well, I hope when you -- I hope you will come and talk to us about your findings from this and I think it would be particularly interesting to see how the lives of the inspectors -- how badly for the worse the lives of the inspectors have changed. [Laughter.] DR. SEALE: Make a value judgment. DR. POWERS: I know which direction it is going. 'I just don't know the magnitude of that vector. MR. BRANCH: Like I say, we processed as many issues as we could through the Significance Determination Process in the two days. The fourth day we ended up generating PI data from these four plants. We could only get initiating scrams, transients. There were only about three or four PI that we could actually get generated -- MR. MADISON: Six. MR. BRANCH: There were six, I think, total. We processed, we aligned those up to the different cornerstone areas and on the fourth day what we did is -- the fifth day, excuse me, we then took the inspection finding as we colored them through the Significance Determination Process. We took the PI data and we aligned it at a cornerstone and we did a plant assessment, and we asked the regional representatives to bring in the information as to what was actually the actions that the agency took on those plants, and then what would we recommend with the new process. Basically, as the slide says, we determined that the process is feasible to pilot, that the Significance Determination Process was successful for the items that were within the scope, and again we took some "to dos" back. We are working on checklists for the inspectors, for the screening tool. We changed some of the logic in our process. We learned a lot from this feasibility review. Also the actions proposed by the new assessment process were similar to past actions but again we made our decisions based on the whats that occurred at the plant and not the whys. We took the equipment that was the Red findings and those were hardware type issues, and we passed that information back. We also determined that we did need to do a lot of training of our inspectors, and as Alan mentioned, we are in the process of doing that now. MR. MADISON: This concludes our remarks. MR. BRANCH: This concludes it. DR. APOSTOLAKIS: One question? DR. POWERS: Please. DR. APOSTOLAKIS: The White-Yellow threshold and the Yellow-Red are defined by looking at changes in CDF, 10 to the minus 5 and 10 to the minus 4. I don't remember exactly now how it is described but do you use plant PRA to define those thresholds for particular systems, or is it done also in a generic way? MR. PARRY: Are you talking now about the SDP or the PIs? DR. APOSTOLAKIS: The PIs. MR. PARRY: The PIs, all we did was we took a number of plant specific PRAs, as many as we could actually get, did the evaluation on a plant-specific basis, but then took the lowest of those values. DR. POWERS: When you took plant specific PRAs and what-not, were these ones that the NRC had done or -- MR. PARRY: No. These were ones that the NRC has but they were licensee PRAs that the NRC has Sapphire. DR. POWERS: And we have some confidence that these are useful in high quality PRAs that somehow will reflect the plant as it actually is -- I mean it is a notable defect in the IPE assessments, that we have found sometimes that the IPEs are of questionable fidelity to the plant. MR. PARRY: Some of them, yes, but a lot of them were actually fairly good too, and I am trying to remember which plants they were, but I think they were. DR. APOSTOLAKIS: These you had reviewed? MR. PARRY: They had been reviewed, yes. DR. POWERS: Well, I mean the IPE review itself was for conformance to a generic letter -- MR. PARRY: Right. Well, that's true. DR. POWERS: Conformance to utility in the inspection and assessment process. MR. PARRY: That's true, but I am going to look at that. I am not going to comment on the quality of these PRAs because I am not sure i have looked at all of them. MR. MADISON: That was about a lifetime ago in this process. It was at least six months ago that he looked at in detail. DR. POWERS: I marvel at the speed with which you have gone from a concept to now virtually implementation in this process. You're congratulated on what must have been a heroic effort. MR. MADISON: Thank you. There are a lot of people in the region and in the Headquarters that put a lot of time and a lot of good quality effort, I think, in developing this. DR. POWERS: The committee has documented it. In fact, it's been impressed at the breadth that you threw your net to drag people into this effort. MR. PARRY: Could I make a quick comment on the SDP with respect to plant specific issues here, just so that you get the picture right. The number that we use in Tables 1 and 2 are sort of generic, but the systems that you take credit for would be done on a plant-specific basis, so it's getting more towards Dr. Bonaca's concern, that at least the design features of the plant are going to be handled in that way. DR. APOSTOLAKIS: I second what Dr. Powers just said, by the way. I think this has been a heroic effort, and all I'm trying to do with my objections is improve the product. I do believe that these numbers eventually must become plant-specific, and if you follow the process that was described to us by the NEI representative and keep track of these assessments and so on, I don't see any reason why in maybe two, three inspections you won't end up with reasonably plant-specific numbers. MR. MADISON: What may happen, though, George, is -- and you'll hear this afternoon about Research's work on risk-based performance indicators -- may be a moot point in another year. We may be replacing these performance indicators with something new as developed by Research. DR. APOSTOLAKIS: And I will raise the same point this afternoon. MR. MADISON: Make sure you do that this afternoon. DR. APOSTOLAKIS: At least the same day. MR. MADISON: I'd also like to mention that OPA has just reissued NUREG-1649. We've revised it to make it more current with the processes. And I think we can also make available to you the slides that were used at the May 17 workshop that might provide some better -- we had several left over that I'll make sure that we make available to you. DR. POWERS: That would be useful. DR. APOSTOLAKIS: Are you expecting a letter from us? MR. MADISON: No. This was just primarily to provide you with status and try to address some questions. DR. POWERS: Yes. I didn't -- I mean, you've gotten sufficient feedback about the questions, and I think it's -- I think the only thing that I would caution is when you come back, clearly we need to schedule a little more time, because there are lots of issues that we're really interested in on the outcome of the pilot program. MR. MADISON: I understand. DR. POWERS: And what not. So with that I think -- DR. SHACK: You were going to make the inspection procedures public; right? MR. MADISON: Yes, we intend to make those publicly available. That hasn't occurred yet. We're working on the mechanics of how we do that. As I said, this is the first time I think the Agency's ever done it, but we feel it's the right thing to do. DR. POWERS: I think it will -- I think it has the potential of adding credibility to the Agency to make these things public. With that -- MR. GILLESPIE: Actually, all of our inspection procedures have been public for years. This time we're asking people to tell us what they think of them. That's different. [Laughter.] MR. MADISON: And actually listening to their response; yes. MR. GILLESPIE: And I would suggest that we need this pilot program to answer kind of George's question how many exceptions are we really dealing with. You know, is the South Texas where they know they're within 28 hours of a threshold, is that a problem for South Texas? Is it a problem for others? We need the six months of implementation time to see what comes out of it. Because one of our ingoing assumptions is that the thresholds that were set in the PIs, and industry concurred in this, are reasonable for the purpose they're being used relative to indicating whether we should look a little more. And they're certainly more reasonable than the process we've documented and used to date, which is really a very subjective process of do we want to go look at some more there. There's no structure to it. The value of this is we've got a decision structure we can now criticize and we can change. There's a skeleton there. There's a logic flow. We may not have everything defined, but we know we have to define it. And then we can argue about the definition. That doesn't exist today. So I think that's going to be the major contribution here is the structure it puts us in. DR. POWERS: I hope that you have steeled yourself for the possibility that this pilot, which is very valuable, is just too short, that you really need to go to a cycle. DR. SHACK: Let me get one more question in before you bang -- DR. POWERS: Certainly. DR. SHACK: Nobody actually addressed George's question of, you know, when you had a PI that you found for a specific plant was, you know, there was an explanation for it and, you know, would you really go back and do the whole process again, or would you simply adjust the value? MR. GILLESPIE: Oh, for that plant, we would -- yes, we would not react, because we know there's -- in fact, there's -- South Texas was one -- the Combustion Engineering plants have a problem with loss of normal decay heat removal if they don't have a bleed and feed, and newer ones don't. So where we know the specific explanation of what's happening, we wouldn't go out and re- and re- and reinspect every single time it happens. We document it once, and that's when we have to step back and reevaluate okay, if the Combustion plants, do we need to make an adjustment on X for the BWRs. DR. APOSTOLAKIS: So you are going to move toward a more plant-specific number -- MR. GILLESPIE: I think we're going to move there. I'm not going to say we're going to move rapidly, but as the exceptions start showing up, and since we're getting this data reported each quarter, that's actually a fairly frequent milestone. If there's the same explanation quarter to quarter to quarter to quarter, and it's programmatic yet it's acceptable, then we'd go in and say okay, what kind of adjustment do we need to make in this one. And it might be plant-specific, it might be design-specific. DR. SEALE: Well, similarly the difference -- we spent the morning, earlier part, lauding large drys, and we may find that there's a systematic difference between those and ice condensers. MR. GILLESPIE: Yes. [Laughter.] DR. SEALE: We might. DR. POWERS: I will recess until 1:30. [Whereupon, at 12:30 p.m., the meeting was recessed, to reconvene at 1:30 p.m., this same day.]. A F T E R N O O N S E S S I O N [1:30 p.m.] DR. POWERS: Let's come back to into session. We're going to turn now to an allied topic to our last presentation, in their performance-based regulatory initiatives and related matters. Our cognizant member on all things performance based and -- is, of course, Professor Apostolakis, and so I'll defer to him and expect an outstanding performance. DR. APOSTOLAKIS: This is a risk-based performance indicator program, right? [Laughter.] DR. POWERS: I'm sorry. DR. SHACK: They all look alike, George. DR. SEALE: A pretty low success criteria. DR. APOSTOLAKIS: Okay. Well, what we received before today's meeting was a copy of the viewgraphs that you guys plan to use. We did not get the report -- or the section that you're working on. Is that ready now? MR. BARANOWSKY: No. What -- is this working? DR. SEALE: Yes. MR. BARANOWSKY: We didn't get the report done. What we did was we put together these viewgraphs -- they're really sort of an annotated outline of the report. And we were going to write the report and we realized we weren't going to get it done in time to have it here before this meeting. So, we put extra information, if we could, into this so-called set of viewgraphs. We don't plan on covering every page and every word. So that we did the best we could, in terms of getting you information. But, eventually, we would like to have one; we just haven't had the time to do it. DR. APOSTOLAKIS: Now, we did get the copy of the paper that -- MR. BARANOWSKY: Yeah. DR. APOSTOLAKIS: -- will appear in PSA-99. MR. BARANOWSKY: Right. We did -- we did take this information and put together the paper that we're presenting at PSA-99. But, that's really a short version of explaining exactly what we're trying to accomplish over here. So, hopefully, that information we gave you is enough background for you to have formulated some questions and I'm sure they are -- DR. APOSTOLAKIS: Now, I'm a little bit confused about the various programs that the agency has, and there's this other one that we are to discuss this afternoon on performance-based regulatory initiatives. What you are doing is part of the oversight -- part of attempting to revise the oversight process? MR. BARANOWSKY: That's -- the plan is to see if we can come up with some improvements to the oversight process that brings in a little bit better performance indication on a risk basis, to be used in a risk-informed process. It's not meant to be a risk-based process. I want -- that's one point I want to make clear, when we give this presentation. But, yes, this is meant to address, and I hope we will address, some of the questions and concerns that were raised earlier this morning about plant specific nature performance and so forth. DR. APOSTOLAKIS: Okay. Now, is this also the appropriate -- were you here this morning? MR. BARANOWSKY: Yes, I was hiding back there. DR. APOSTOLAKIS: Okay. So, this is the project, then, that may look into this issue of plant specific versus generic thresholds and so on? MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: Because, you are the one thinking about the performance indicators. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: And I, of course, as you know, is heavily involved in the initial development of -- DR. APOSTOLAKIS: Sure. MR. BARANOWSKY: -- the framework and the current set of performance indicators, and I'm pretty well aware of what their limitations are. DR. APOSTOLAKIS: Yes. You are one of the usual suspects. MR. BARANOWSKY: Exactly. DR. APOSTOLAKIS: And you expect -- what do you expect from this meeting, a letter? MR. BARANOWSKY: No. What we're looking to do is to brief the ACRS and hear comments back, so that we don't come here late in the game with something that's pretty far along and developed and having spent a lot of time and effort on, and then get some fundamental questions on high level concepts and direction that we could have been debating here today and maybe made some changes in our program earlier. DR. APOSTOLAKIS: So, there is no need, then, for me to bring up again the questions I raised this morning. You were here and you -- MR. BARANOWSKY: Well, I was here, but you may want to bring them up at an appropriate point, and I'm sure you will. DR. APOSTOLAKIS: No, no, no, you took the steam away from me. Well, judging from the viewgraphs that we received, what you are attempting to do here is pretty much consistent with your presentation sometime ago -- MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: -- as part of the oversight -- the new oversight process. MR. BARANOWSKY: Right. And we, also, briefed the ACRS, it must have been over a year ago, in which we had some barely sketchy ideas that weren't quite as well thought out as where we are today. And so, we want to, as I said, try to get the ACRS aware of what we're thinking right now on this conceptual level, because, pretty soon, we're going to start diving into this and really moving forward with it. DR. APOSTOLAKIS: Okay. So, the members, I think, are familiar with the overall approach. MR. BARANOWSKY: Okay. DR. APOSTOLAKIS: So, why don't you start your presentation. MR. BARANOWSKY: Okay. Let's go do the second viewgraph. This shows the content of the rest of the presentation. And I'm going to cover information up through the bullet that says, "Programs supporting risk-based performance indicators" and then Steve Mays will pick up on the principles, attributes, and concepts, etc. I, also, want to mention that the very first item on here, the risk-informed reactor oversight process, was put in there, because when we first started crafting a white paper, if you will, we felt we had to put some background material in there. And I think we still need to lead off a little bit with that, because this work needs to be looked at in the context of that process. This is not something that's gone off on its own direction and then, hopefully, you're going to try and sway the whole process into some new way of doing things. I think you'll see that our objective is to show that this can be an improvement to and consistent with the current process. Our basic objective is progress, not perfection, and so there will be things that we can't do and we'll tell you about that may take a longer time to do. But the things that can be done to fix some of the deficiencies that have been identified, we'll try to address them as soon as we can. DR. APOSTOLAKIS: When do you plan to finish this work? MR. BARANOWSKY: We have a schedule and it's sort of a multi-phase schedule. I think the first operational set of risk-based performance indicators would be January, 2001, and then others could come after that. And you'll see the kinds of things that we're having some trouble with. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: The other thing I want to mention is that we want to cover the viewgraphs in somewhat of a limited way; in other words, not go over every single word, because they were fairly detailed. And so, we'll talk about some of the figures and then we'll just have some discussion. DR. APOSTOLAKIS: That's fine. MR. BARANOWSKY: So, let's go to the first figure on page three. And I'm not going to go over this, because, obviously, you know what the new oversight process is. But the key point is that there are both performance indicators and there are risk-informed inspection findings, and that those two sets of information provide a complementary indication of performance at the plant. I won't call the risk-informed inspections performance indicators. But, if you think about it, they are measures of performance. And so, we've taken that into consideration. That's sort of a new thought, I think, that wasn't in some of the earlier concepts that we had been working on, and that's going to be factored into some of the risk-based performance indicator development activity. DR. POWERS: I think understand pretty well what human performance means in this slide and I may understand problem identification and resolution, but the safety conscious work environment is new terminology for me. Would that be at all related to something I know of as safety culture? MR. BARANOWSKY: Probably, and as you might have heard us say in previous discussions, we don't currently see a way to measure, or whether we even should measure safety culture directly; but that if we select the right types of performance indicators, the impact of that so-called culture would be seen through the performance indicators, themselves. DR. APOSTOLAKIS: We'll come back to that -- MR. BARANOWSKY: Yeah. DR. APOSTOLAKIS: -- when you talk about the limitations. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: But, actually, it's not probably. In the report, you very clearly state that some people call this safety culture. MR. BARANOWSKY: Right. And -- DR. POWERS: Well, I think it's interesting that on a viewgraph of an outline nature, boxed diagram nature that only highlights the most important things, that something that we shy away with -- from so carefully actually shows up. MR. BARANOWSKY: The other thing is, of course, this is meant to be representational of the current framework that was put together for the new reactor oversight process. And I think those three cross cutting issues, in particular, were discussed, because there was a lot of interest in them. And those were the kinds of things that were focused on, in one way or another, in some of the performance assessment activities that were in the old process, let's say. And we recognize that it's pretty difficult to measure those kinds of things, but that people still have an interest in them and they fit somewhere. And that's what we're trying to show here. So, if there are any other question? Let's put up the figure two, Tom. This chart is the equivalent to a framework for us, on this project, much like the cornerstone framework that was presented for the reactor oversight process. We tried to devolve risk, if you will, into some constituent elements. And what I want to do is talk a little about what are risk-based performance indicators. Simply, they are indicators that measure either changes in risk or the constituent elements of risk associated with a particular plant, or they could be measurements for the total industry. And there are some role, in this whole business, for a total industry risk indication. The risk-based performance indicators use parameters that you are familiar with in risk assessment, such as frequency, reliability, availability, and probability. And they use them to calculate these -- either calculate or, as part of the measurement of the various indicators that are associated with risk, through an absent sequence logic. And lastly, the risk-based performance indicators, they have thresholds that are objective indication points for judging the significance of performance changes associated with corresponding changes and public risk. DR. UHRIG: Is this a bottom-up type approach here? MR. BARANOWSKY: Yes. DR. UHRIG: That is to be read -- MR. BARANOWSKY: What we did was we said, you can measure or calculate risk at various levels. And those levels go from the highest level, which is basically what is the public risk associated with operating a plant in a certain mode, versus what are the risk implications of either equipment or people performance, as they affect trains and components and so forth. And what you are going to see is that we're going to talk about developing performance indicators at different levels, some of which would be primary indicators, much like the current set of performance indicators that we have; and others would be more secondary, that allow you to understand what are some of the causative elements associated with changes in risk-based performance indicators. So, one could have really a large set of performance indicators. And some industries -- I've been looking into what other either licensees or other industries are doing. They will have sort of a pyramid shape, if you will, to their own performance indicators. The top one might be corporate performance, and then you get down to various lower levels, which feed those different levels that get up to corporate performance. And we're taking a similar approach here to looking at performance indicators. So, it's not just measuring risk only at the highest levels, such as public risk; it could be reliability or some other parameter, which ultimately feeds into risk. But, it feeds into risk through a logical framework. It's not just sitting down and saying, I think reliability is important, so let's go measure some reliability. We have a notion as to how that reliability fits into the risk picture, so that we can eventually set performance thresholds that might be related to plant specific differences in design. The other thing I want to point out with this is we're measuring performance in risk significant areas. We're not trying to do a plant specific risk assessment. We're measuring performance, in order to determine if there are performance changes that are risk significant, that warrant NRC actions. Now, that's a subtle difference, but our intent is not to come up with a plant-risk indication, because for one thing, we're going to have some holes in our model. DR. APOSTOLAKIS: In your, I guess the previous slide, which you didn't show, what are risk-based performance indicators, I think you're addressing -- MR. BARANOWSKY: Yeah. DR. APOSTOLAKIS: -- now. It would be nice to have a slide with a maybe slightly different title, listing requirements, perhaps, for developing the performance indicators and so on. One thing that comes to mind, for example, is: what is the level of redundant information that you would like to receive from the performance indicators? In other words, if I have a performance indicator -- and I trust it, should I say, I will do nothing on the basic events now, because the train level is good enough; or the other extreme is, of course, is to have a train level indicator on every single indicator you can think of at the event level; or some sort of an optimal in between. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: I don't -- I haven't thought about it, myself, but something like another -- another requirement or question. Should we try to develop -- to establish performance indicators at the highest possible level in that diagram or not? MR. BARANOWSKY: I think that's a principle that we believe is true, but we, also, recognize that when you do that, you lose a lot of information -- DR. APOSTOLAKIS: That's correct. MR. BARANOWSKY: -- and it gives you the ability to take timely actions. And I think Steve is going to cover some of this. DR. APOSTOLAKIS: So, there are objective -- MR. BARANOWSKY: Yeah. DR. APOSTOLAKIS: -- I mean, competing -- MR. BARANOWSKY: Right. DR. APOSTOLAKIS: -- requirements. But, I suppose what I'm saying is that it would be nice to see a separate discussion of these, you know, setting the stage, so to speak, of what drives the performance indicator -- MR. BARANOWSKY: Right. DR. APOSTOLAKIS: -- definition to follow later. MR. BARANOWSKY: Let's see if Steve's -- DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: -- first couple of charts cover that and then we can add to it. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: That's the kind of -- we're looking for that kind of comment. DR. UHRIG: But, let me go back to system level indicator. You've got one in there called capability. Why is that an indication of performance? Or do I misinterpret what you're driving at. MR. BARANOWSKY: Oh, those are -- yeah, that's -- DR. UHRIG: I understand -- MR. BARANOWSKY: No, that's a good point. DR. UHRIG: -- reliability and availability, but not capability. MR. BARANOWSKY: That's a good point. One of the objectives for certain sets of systems is that they be reliable, available, and capable. And if the system isn't capable, meaning that its design or, for other reasons, it's not able to perform the intended function, even though it, in theory, is operable for technical specifications, or thought to be operable, then that's an indication of performance that's unacceptable for that particular piece of equipment. DR. UHRIG: But -- MR. BARANOWSKY: I mean, in theory, you can pull that into reliability, if you want to. And we had some debates on this, when we did the reactor oversight framework cornerstones. But, we thought that, in light of the way we do inspections and what this means to inspectors and licensees, that we would pull capability out from reliability and availability, wherein availability would mean more or less is the system in or out of service, as planned or unplanned; reliability meaning is it able to perform on demand; and capability meaning if it's in service and it starts up on demand, is it able to provide the necessary flow for the particular accident that we're talking about, especially since some of the tests don't necessarily capture full capability. DR. UHRIG: But that's a function of design of the plant. MR. BARANOWSKY: Right; correct. DR. UHRIG: And that's the history -- MR. BARANOWSKY: Yes. DR. UHRIG: -- unless you go back and revamp the plant. And it strikes me that that's not fair game for an inspection demerits, so to speak. DR. BONACA: But, you have the degradation taking place in components. I mean, there are some systems that are limited by design, although they were effective enough, and then, because of aging, they get limited in performance. MR. MADISON: This is Alan Madison. There's also on inspection, and I've been involved in this and several diagnostic evaluations, you run into not just degradation, but you, also, find out new information. The qualification of the equipment may impact it. Capability, as Pat was talking about, was, is the system function capable of performing its design capability, its design function. DR. UHRIG: That has to do with aging, then. MR. BARANOWSKY: It may have to do with aging; it may have to do with other things you learn about. MR. GILLESPIE: The net positive suction head problems we found in the AE inspections -- MR. BARANOWSKY: Okay. Some of the -- SPEAKER: Is it inspected for 20 years? MR. BARANOWSKY: Some of the logical system -- distribution system type problems that we've identified -- DR. APOSTOLAKIS: So, there is a fundamental unspoken condition here that you are -- the performance indicator is related to quantity that may change with time. Because, I think, Dr. Uhrig is right, I mean, the original design must have made sure that the components were capable, at that time; and because of aging or other reasons, you may have a change. That does not apply, for example, to separation criteria. I can't imagine that something that -- if two components are, say, five feet apart, three years later they will be only three feet apart. MR. BARANOWSKY: Right. [Laughter.] DR. APOSTOLAKIS: There's nothing that -- MR. BARANOWSKY: No, I think you would -- DR. APOSTOLAKIS: So, that's not part of a performance indicator. MR. BARANOWSKY: Right; correct. If the -- let's say a new generic issue came up, in which we determined that we need five feet of separation instead of three. We wouldn't call that a performance issue. DR. APOSTOLAKIS: No, that's right. MR. BARANOWSKY: But if the licensee was supposed to install new wiring with five foot of separation and installed it with three, and we observed that through an inspection, then that is a performance finding, that their design and installation performance is not up to what we expected. So, it's got to be treated somewhat differently. I'm not saying to put them all in the same pot, but you have to account for this kind of thing in your performance assessment. And it's not the same kind of performance characteristic as one gets from the reliability of equipment failing, say on demand; but still under certain accident conditions, it doesn't meet the specifications for performance. DR. BONACA: That would include also issues of code compliance, where a system is functional, but is not operable, because it doesn't meet the code? MR. BARANOWSKY: I think that what we're not going to do is account for things that don't affect the performance necessary for the risk significant acts and sequences. That might be true for some design bases envelope. And compliance will be treated -- as Al said this morning, we find compliance issues that are in risk insignificant areas. I think they showed that in the STP process. They go in to inform the licensee and they will take care of them. In the risk significant process, that's another story. Keep an eye on that. DR. BONACA: Okay. Let me understand this, because I'm confused: so this really would be risk-based -- MR. BARANOWSKY: Yes. DR. BONACA: So, therefore, a component that is capable functionally would be considered operable, by this standard -- MR. BARANOWSKY: Correct. DR. BONACA: -- even if it doesn't meet a specific code. MR. BARANOWSKY: Right. DR. BONACA: But, it's capable of performance. MR. BARANOWSKY: Right. Now, we wouldn't even have that information probably get factored into the indicators here, because we're going to have an approach that only collects the information necessary to get the risk significant, say, design or capability defects. The other parts of the inspection program might pick that up, but that's not necessarily what we're trying to do here. DR. BONACA: That's a significant distinction. MR. BARANOWSKY: Yes. DR. BONACA: I mean, it's a huge departure. MR. BARANOWSKY: It is. DR. SEALE: At some point, it does dovetail with the new way of handling category four violations. MR. BARANOWSKY: Correct. That's what we're doing. DR. APOSTOLAKIS: I think what we need, Bob, is that report or -- MR. BARANOWSKY: Yeah; okay. DR. APOSTOLAKIS: -- definition of performance, and bring up the issue of time. MR. BARANOWSKY: Good. One -- bring this questions up and we can fold them in. I think that's good. That helps us out. DR. APOSTOLAKIS: I think we need to define what we mean by performance. MR. BARANOWSKY: Okay. DR. APOSTOLAKIS: And then we have performance indicators. But, it seems to me there is a fundamental element of time. Something can change, in other words; otherwise, you don't need them. Then, it's a design issue. So, all the example you gave us really indicate that something may go wrong over the lifetime of the plant and you need some indicator to monitor that. MR. BARANOWSKY: The design and capability part of this whole thing is really one more difficult ones to describe, I think. I'm not so sure it's hard to treat, but I think it's hard to describe. And I think I agree, we need to put something up. DR. APOSTOLAKIS: Let's go on now, because we're spending -- MR. BARANOWSKY: Okay. DR. APOSTOLAKIS: -- too much time. MR. BARANOWSKY: That really does finish that viewgraph. Let's go to item seven, Tom. I am just going to briefly run over this, why we're having risk-based performance indicators. I think the first three bullets are the usual types of things, you know, because of the policy on use of risk. I guess I will say the second bullet, the framework, it's very important. I've had the opportunity to work on that new oversight process framework, and it's hard to figure out how to pull things together in a logical framework, where you can put things on an even playing field that's not associated with a risk model of some sort. And so, risk -- the risk framework provides that ability for capturing and analyzing and comparing and integrating information. Integration is important, too. So, I think that's a pretty important one. We expect that the risk-based performance indicators will obviously be an improvement to the current set of indicators. You talked about some of those this morning. And the next viewgraph, I think, probably identifies some of the limitations a little bit more. And then Steve will be talking about some of the holes that are filled in by the risk-based performance indicators. They're going to more closely link performance to risk and to risk thresholds. And, in particular, they're going to provide the ability to look at things on a more plant specific basis, so that some of these concerns that we have about using the same threshold with indicators that are basically designed, say, on a train basis, giving some misleading indication of how risk significant that performance is, can be handled through this process. I don't know exactly what that would do, in terms of changing the thresholds, but we would have the ability to do that. So, let's go to the next viewgraph, number eight. What this viewgraph does is makes a list of the kinds of areas of improvements, or limitations, as we call them here, that the current set of performance indicators has; and the first one being that there's no reliability indicator, the next one being that there's limited indication of availability for certain systems. We don't have indicators for barriers or for shutdown or external events. We have very limited indication capable of capturing things like human performance and design and procedural quality. There's no integrated indicator. DR. APOSTOLAKIS: Pat? MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: It so happens that the Swedes are doing a lot of work on this. And I don't know if you're aware of it, there was a report from SKI, published in February, '99. Have you seen it, Steve? MR. MAYS: Yeah, we have it. DR. APOSTOLAKIS: Okay, you have it. And they are actually using expert opinions. One of the experts is sitting to my left here. It's interesting, by the way, that I counted that they used 21 experts, more than 50 percent were from the industry. They developed a list of performance indicators. I mean, you can question what are the things they do, but there's all the work that went into it. And they -- the experts identified the top five. Now, this deals with culture, human performance and so on. So, number one is the annual rate of safety significant errors by plant personnel, contractors, and others. The fifth one is annual rate of plant changes that are not incorporated into design basis documents, by the time of the next outage following the change. Now, these are things that you can see, I mean. So, perhaps we were too quick to say that we can't find performance indicators for these cross cutting issues. There are -- these experts, at least, were able to identify -- they started with a list of 11, I believe, and they ended up with five most important. DR. BONACA: Actually, 11 were cut down further manipulation. There were over 100. DR. APOSTOLAKIS: Over 100. DR. BONACA: And then went down to about 11, and then down to about five. DR. APOSTOLAKIS: About five. So, I suggest that you read this report carefully and see whether there's anything there that you want to borrow from them. MR. BARANOWSKY: Yeas, we looked at it. That's -- we're looking for anything that has been done in the last couple of years, and I know that was just done. DR. APOSTOLAKIS: Yeah. MR. BARANOWSKY: In fact, we contributed an expert to that panel, too. DR. APOSTOLAKIS: I know. MR. BARANOWSKY: Okay. DR. POWERS: Similarly, we've come to the -- that no indicators -- or things like fire, they have been FPA performed, according to creating a performance-based fire protection rule. And they clearly have slots and a rule for performance indicators. Are there indicators not defined or don't meet your criteria? MR. BARANOWSKY: I'm not sure I know their performance indicators. MR. MAYS: I don't know what their performance indicators are and I haven't seen anything right now that tells us what they're looking at. DR. POWERS: But, I mean, you have a member of the NRC staff that is on that panel, trying to create those. MR. MAYS: That's correct. MR. BARANOWSKY: We can look into it, I guess is what we can do. DR. APOSTOLAKIS: Since you have the Swedish report, you don't seem to change, though, your position that you can't get indication of cross cutting issues. I mean, you've said that to us several times. MR. BARANOWSKY: All is aid was we had limited indication. I expect us to have some indication on cross cutting issues. I think we are planning on that. DR. APOSTOLAKIS: Oh, okay. MR. BARANOWSKY: Now, what we are trying to do a little bit different than the Swedes, and we could fall back on what they're doing, what we're trying to do is make a more direct link between the indicators that we're looking at and risk, that doesn't rely on expert opinion so heavily. Now, if it's not possible -- DR. APOSTOLAKIS: If you could, yeah, that would be great. MR. BARANOWSKY: Yes, of course. But, if it turns out that that's the only alternative open to us -- we use that a lot on quantification of different events and so forth, and we wouldn't ignore it. DR. APOSTOLAKIS: So, you don't feel constrained by the current climate of not getting into organizational issues? MR. BARANOWSKY: Well, I'm not sure I'd call that organizational issue. DR. APOSTOLAKIS: See, that's the thing. I think a lot of it has to do with communication. MR. BARANOWSKY: Yeah. If I was to say I'm concerned about some of these parameters that you mention, because they reflect poor organizational performance, then I'd be off on a different tangent. But, I think what we're going to do is look at it and say, what, even on a judgmental basis, does that information tell us about the risk significance of these kinds of things. And if we can make an argument that connects it to risk, even if it's connected indirectly or implicitly, instead of explicitly, okay. But, if it's connected through, I just think it's important, because I think organizational factors are important, unless directed to, I don't think we see that in this program. MR. MAYS: I think Pat talked about earlier, the key thing was, is that we're putting together a framework related to risk, as opposed to a framework related to anything I can think of that relates to the performance. I mean, if you just say performance -- and you raised the question, what do you mean by performance. I think the key thing here is that we're looking at the performance, as it relates to risk, as opposed to the performance, as it relates to management effectiveness or public confidence, or any other metric you might want to choose there. And I think if we don't have a way to take any information and relate it back to risk through some model process, we're probably going to be reluctant to include it as a performance indicator. DR. APOSTOLAKIS: But, these top five that they identified, are the top five that affect the failure rates of components or the frequency of initiating events. So, they are making the connection to risk. The question is whether that connection is quantitative. But, again, you can have a qualitative connection. Everybody agrees that this is important. Now, the fact that you cannot quantify the precise impact perhaps should not deter you from having some indication of how well the plant is performing, with respect to that -- MR. BARANOWSKY: I think we're open-minded on that. DR. APOSTOLAKIS: Yeah. Because, I think you are getting into an area, where quantitative relationships are a little difficult to abide. MR. BARANOWSKY: Right. But, this stuff will have to stand up to a much wider and more critical audience than whatever the Swedish report went through. DR. APOSTOLAKIS: No, I understand that. And, again, remember that the audience may be critical if you don't address the issue at all. MR. BARANOWSKY: Agreed. DR. APOSTOLAKIS: You would be criticized anyway, but -- [Laughter.] MR. BARANOWSKY: Okay. We've got the message. Let's see, where was I. And we know that there are some problems with trying to come up with indicators in the areas of emergency preparedness, radiation safety and security. And also, lastly, I want to mention that there -- the current indicators, at least some of them, are heavily count-based, and you run into some statistical problems on count-based indicators. DR. APOSTOLAKIS: The maintenance rule, they don't have any reliability base? MR. BARANOWSKY: Well, we aren't using the maintenance rule. I would think they have -- they can do the maintenance rule anyway they want, as long as they can relate it back to whatever their performance objectives are and, in some way, back to their PRA. My understanding is it's done quite differently at each plant and it's going to be hard for us to use that, as a case in point, to make a judgment on this. But, I'm talking about indicators of numbers of reactor trips per, what, day or year or whatever we're talking about; or numbers of failure counted. If you use those kinds of approaches, you run into some problems with statistics. And I think the ACRS has gone over this over the years. The last point, we have to be realistic about what can be done, in light of the lack of certain things, in particular risk models in areas. It will be awfully hard for us to develop risk-based performance indicators, when we don't have a good model framework to relate the basic event information, that is the essence of the indicator to the risk. We would be back into the business of using judgment, as to how important things are, and we can do that probably outside the framework of this particular project. DR. BONACA: One comment I'd like to make is that, you know, one difficulty that we, also, have in the Swedish study was the effort to try to connect human factors or cultural indicators to risk and to plant operation. And that's a very difficult task, because dependancies are not so clear. So, you go to an elicitation process and so on and so forth. From a practical standpoint, however, you could -- you could select groups of performance indicators from cultural areas, and they're being done -- utilities out there. I mean, there are certain specific groups, which are focused on that -- for which you don't have a direct correlation to risk. Right now, it doesn't exist, although we suspect there is, of course. And then, you could monitor them separately, the group of risk indicators, which are top, and then some of the cultural indicators, okay. And, again, those, you know, I can't just make -- I have the list, actually, I'll present to you and the rest of the committee the day after tomorrow. But, safety conscious work environment, there are that that is being monitored of the units: operator challenges, corrective action program, self assessment, quality of maintenance. I mean, there are really literally eight or nine of those boxes that utilities track pretty religiously right now, because they realize that all the past experience, in so far as failures of system and availability and so on and so forth, is tied to those kinds of general areas of performance. And, I mean, it seems to me that you don't have necessarily to identify an analytical dependency. That's a challenge always to do so, to try to sort out some cultural indicators. DR. APOSTOLAKIS: So, when you say risk-based, are you excluding these kinds of indicators? MR. BARANOWSKY: No. I think what we're saying is we need to at least have a logical connection, which can be through expert judgment, but that what we're not going to do is have a pure human performance indicator that takes the 10,000 records that we have in the human performance database and say, here's the plant norms and here's the 95 percent thing, when only 20 or 15 or 10 percent of those are related to risk. We want to be able to deal with the ones that are most risk significant. That's what I mean. DR. APOSTOLAKIS: Even though the connection may be qualitative? MR. BARANOWSKY: Even though the connection might be qualitative. DR. APOSTOLAKIS: And just to make my position clear, I do not necessarily agree with everything that's in the Swedish report. All I'm saying is look at it and see what you can use. MR. BARANOWSKY: I agree. DR. APOSTOLAKIS: I'm not saying this is something you have to use. MR. MAYS: And, as well -- and, again, as Pat indicated earlier, this is an oversight process of which indicators is a part. If you talk to Alan and the people, who have been putting together the risk informed base line inspections, one of the things they tried to do was to go perform inspection activities in areas like the corrective action program and self-identification of problems and other things, to get information about whether plant performance was heading in the right direction or wrong direction, independent of a direct indicator. So, it's not a necessity that the indicators cover everything you would ever want to know about plant performance, because we, also, have a baseline inspection program, which is supposed to go out and examine other areas of performance that are not readily amenable to indicators. DR. BONACA: Well -- but, that's what concerns me a little bit. I agree that you're looking at those things. And I appreciate that fact that you have that layer there of, you know, safety conscious work environment and so on and so forth. It is the statement that you cannot build indicators that I guess I have some problem with. DR. APOSTOLAKIS: But, they didn't really mean it, though. DR. BONACA: Okay. But that's what I heard. MR. BARANOWSKY: Did we say we couldn't? I'm just looking to see if we said that. But, I think we said it would be difficult. DR. BONACA: Yeah, okay. MR. MADISON: It depends upon how you use the indicators. If you use the indicators for enforcement or for sanctions against the licensee, I don't think you're going to be getting them through judgment. I think there are going to have to be some qualitative guidance associated with those performance indicators. If we're using performance indicators, on the other hand, to perhaps make some minor adjustments in our resources, I think you can be a little more judgmental. MR. BARANOWSKY: That's a good point. MR. MADISON: It really depends upon what the use of the performance indicators is going to be. DR. BONACA: But the way -- the way in which you are going to either use those indicators to just a portion of your resources. It's good to give a very clear signal to the -- the utilities that are going to follow some -- I think you're providing an example -- MR. MADISON: You're absolutely right. And that's why we, again, have to be very careful with -- for example, a good manager is going to have a lot of different performance indicators. At the sites now, they have a stack of performance indicators probably about that thick, that they get a report on, on a monthly basis, that they manage their plant with, and that's an appropriate use of that level of performance indicators. As a regulator, we don't want to get to that depth. DR. BONACA: Yes. MR. BARANOWSKY: I mean, as an example, I think to what Alan is talking about, I can have a motor-operated valve performance indicator. That may not be one that I want to use to judge a licensee's performance in the assessment process, but it might be one I want to use to plant risk-informed inspections. I have some latitude in performing those inspections. So, that might be used to point us in the right direction. DR. APOSTOLAKIS: Well, I think what you -- the focus of the discussion has shifted a little bit. When we say subjective, or when we use that word subjective here, we didn't mean -- I don't think that Dr. Bonaca and I meant subjective indicators. The indicators are quantitative. I mean, these guys are telling you to count the number of times that they did not recommend something. But maybe subjective is the connection of that indicator to risk. MR. MADISON: Right. DR. APOSTOLAKIS: But the indicator, itself, is quantitative. Now, that, according to Bob, is something that he will consider. MR. MADISON: But, also, the thresholds for action, that's where the rubber meets the rubber. What are you going to do, based upon those performance indicators? DR. APOSTOLAKIS: I agree. MR. MADISON: The thresholds for action are going to have to be more than subjective, depending upon the action you take. DR. APOSTOLAKIS: And they will have to be plant specific, of course. MR. BARANOWSKY: Well, to some extent. But, I mean, we are definitely not going to have 150 new performance indicators, even though there might be 150 possible ways of measuring performance. DR. APOSTOLAKIS: I think we've reached the point of agreeing. MR. BARANOWSKY: Okay. DR. BONACA: Just one last comment I have. DR. APOSTOLAKIS: I'm sorry, yeah, sure. DR. BONACA: Just one example, okay. When I look at a power plant and I see under operator challenges -- let me just, you know, I have an indicator, I'll call it that way, that there are a very large number of operator work-arounds. There are a lot of temporary modifications in place, bypass jumpers, and all kind of stuff like that, and, for example, you know, control room enunciators lit. That is a big problem. It is truly challenging the operator on a daily basis with his ability to perform effectively. Now, you know, I had, personally, put together an indicator that is being used somewhere that embodies those kinds of elements. It's just a number counting, and it has to be a point at which you say this is too much. You cannot allow that. That's also a terrific indicator of management philosophy, because those things don't happen -- management doesn't accept that kind of operations. It's really similar to you driving your car with, you know, lights lit in your dash, because, well, you know, that's okay and these breaks are okay. But, that's not the way to operate. That's an example of an indicator, which I believe it will give you a lot of insight in cultural. And, you know, I don't see why we shouldn't pay some attention to it. DR. APOSTOLAKIS: I agree. MR. BARANOWSKY: -- in terms of the ones that are connected more qualitatively to risk. DR. BONACA: I just wanted to give an example -- DR. APOSTOLAKIS: That's fine. No, I think that's very useful. MR. BARANOWSKY: Okay. DR. APOSTOLAKIS: Now, which slide will you jump to, now? MR. BARANOWSKY: Now, we'll jump to number nine, if you don't mind. DR. APOSTOLAKIS: Ours are not numbered. So, it's an interesting exercise to figure out what nine means. DR. SHACK: Get today's package. MR. BARANOWSKY: You don't have today's package? They're not numbered? DR. APOSTOLAKIS: Well, I marked up the one that was mailed to me, so I don't -- but today's packages are numbered? MR. BARANOWSKY: They're different. We made some changes in the viewgraphs, too. MR. MAYS: It was trick to see if you were paying attention. [Laughter.] DR. APOSTOLAKIS: Well, what do you know, they're numbered. MR. BARANOWSKY: Okay. We've actually shown this before, when we talked about the programs that we had, when we were back in EOD. And what you see here are a number of projects that we have ongoing from data, to industry-wide analyses, to plant specific analyses, all of provide some of the building blocks for the development of risk-based performance indicators. In fact, what we're talking about doing is trying to have sort of an off-the-shelf approach to building these indicators, wherein we take models, data, and methodology that we developed for these other activities and put them together, as building blocks, to come up with the risk-based performance indicators, so we can try them out in a pretty short period of time. Now, not everything is going to come off the shelf. Some things are going to take a little bit of developmental work. But, our objective is not to try and advance the state of the art in risk analysis technology, but to take risk analysis technology that we found that we could use successfully in the past, and then just use it to build these indicators. And the elements that you see, starting from the bottom, you have the operational data. We have a number of raw data systems, some of which are fed by the nuclear industry. For instance, the Equipment Performance Information Exchange System, EPIX, some through NPO, and others that are associated with data collected more directly by NRC, such as LARS. They are provide the input to the models. In industry-wide analyses, we had developed the framework for analyzing systems and components and for pulling that information together in an integrated fashion. We presented both the methodology and the results at the ACRS before and that information has been used, for instance, to help define some of the risk-based inspection areas that should be looked into. And we continue to do trending of primarily industry-wide performance using that analytic information and data. And, also, we are able to get plant specific results. The problem is that the data density that we were working with prior to what we're trying to do over here was fairly sparse and it took probably five years or so to see a change in performance that was statistically meaningful. But, now, we're talking about making a few changes in the data and the way we're handling it, to enrich the statistics, if you will, and we think that we'll be able to see some changes a lot more quickly. DR. APOSTOLAKIS: So, how should I read this diagram, now? MR. BARANOWSKY: What you should read this diagram is that we came up with a number of programs that existed that have -- DR. APOSTOLAKIS: Come closer to the microphone again, Pat. MR. BARANOWSKY: We have a number of programs that start with data, that gets fed up into models, that we perform analyses that are covering the kinds of areas that we're talking about having risk-based performance indicators in. And then we have plant specific, what I call integrated models, in which we can look at the more complete risk picture associated with these piece parts to the risk. DR. APOSTOLAKIS: But what do you call industry-wide? MR. BARANOWSKY: Industry-wide -- DR. APOSTOLAKIS: Includes plant specific analysis. But, when you're doing initiating events, you look at each plant; right? MR. BARANOWSKY: Okay. What I haven't explained is that we have taken into account enough of the design and operational differences in our models to account for plant specific differences. What we haven't done is modeled every single thing in the plant. We've only modeled those things that PRAs and the operational data has shown us to be important, in terms of discriminating performance amongst plants. DR. APOSTOLAKIS: So, when you do an initiating event study, you have -- MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: -- to look at each plant? MR. BARANOWSKY: Yes. And we come up with an industry-wide analysis, which has both plant specific elements summarized over the industry. And that's what I mean. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: And that's true for everyone of these elements that we have in here. It's all plant specific information, which is done for the full industry. So, we have essentially simplified models for initiating events or components or systems, trains, and so forth, a number of which we've presented at the ACRS. They're either on a plant class or individual plant basis. And the model only goes to the depth that we understand is significant from past PRAs and from the operating experience analyses that were done, for instance, in the Accident Sequence Precursor Program. And the Accident Sequence Precursor Program and the models we have there are undergoing some work, to make them of sufficient quality, that we can use them to integrate all the indications, whether they come from the performance indicators -- what I haven't shown here is the risk informed inspection results, which could also be folded in through the Accident Sequence Precursor models, in order to come up with an integrated assessment. So that you would then have your indicators at different levels, systems, trains, and so forth, as I showed in that prior viewgraph, or at the higher levels, if you want. But, that's -- the point is all these are the building blocks to get to that capability. And they're pretty much in existence. Some of them need a little refinement. If they needed to be completely developed from scratch, we'd be talking five or six years down the road to do this project. Instead, we're talking maybe two. The other, I think, good thing about this is a number of the analyses that are done on an industry-wide basis are used to feedback into the regulatory process, generic lessons learned. DR. APOSTOLAKIS: The only box that's not here, to me, the operator error probability studies. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: Actually, why? MR. BARANOWSKY: Why it's not what? DR. APOSTOLAKIS: Well, I don't -- what do you mean there? I mean, what -- MR. BARANOWSKY: We have -- DR. APOSTOLAKIS: What do you mean? MR. BARANOWSKY: What we mean is we know, for instance -- like, say, common caused failure is a special element of reliability and availability that needs to be taken into consideration, in doing an analyses that aren't apparent from just looking at, say, component failures, when you're trying to do that. And the same thing we think is true for operator error. You have to factor that in somehow into your model, to reflect the fact that there are going to be some plant specific differences that aren't always captured within the data that makes up the components or system performance that relates to how well the operators will perform during an accident, okay. DR. APOSTOLAKIS: So, this would include, then, the annual rate of plant changes that are not incorporated in the design basis document? Is that an operator error? MR. BARANOWSKY: No. DR. APOSTOLAKIS: What do you mean operator error? MR. BARANOWSKY: What I mean by operator error is I mean operator fails to initiative, actuate, control, and do whatever they have to do, to systems and components and so forth, during an accident. DR. APOSTOLAKIS: Would the Wolf Creek event be here? MR. BARANOWSKY: Yes. The one where he -- DR. APOSTOLAKIS: No accident; they started the accident. MR. BARANOWSKY: Right. You could -- it could be an initiator. We didn't say it couldn't. DR. APOSTOLAKIS: I don't understand what you would do with that. That was a fluke, right? Or was it systemic? How do you decide what's systemic -- MR. BARANOWSKY: Don't now. I don't have the answer for that over here, but I am telling you that we need to try to factor that into the risk-based performance indicators. Certainly, we would either account for that through the inspection type finding, where we take the significance of the incident and roll it up through some sort of an integrated model; or we could capture several events and if they seem to be through a model outside the norms, then that would also give indication that there's a problem. I think there are really two ways of getting at these things. One is if you have an event and it's important on its own, you can deal with it. The other thing is, if you have small things occurring over time, such as equipment malfunctions or personnel errors, that individually aren't important, but over time, when you have them in a model, indicate that there's a problem, you have to deal with them in that way. DR. APOSTOLAKIS: But, it seems to me, Pat, that you have to, at some point, look at the root causes of things. I mean, to say that at Wolf Creek, they created the path because they opened the valves inadvertently, doesn't help you much, until you dig deeper to understand why that happened. For example, there may be a history there of postponing work and doing it at another time, without checking what other things are scheduled for that -- MR. BARANOWSKY: Yes, I think we agree with that. What we're saying is -- DR. APOSTOLAKIS: So, is that what you're doing there? MR. BARANOWSKY: Yes. What we're saying is either through the significance of the incident, itself, or through the accumulation of information on lesser incidents, you will determine you need to know the root cause. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: This isn't a root cause program. DR. APOSTOLAKIS: I understand that. So when are you going to come here and talk to us about that program? MR. BARANOWSKY: The root cause program? DR. APOSTOLAKIS: Yeah. MR. BARANOWSKY: That's be Alan Madison talking to you about that. DR. APOSTOLAKIS: Alan Madison again? [Laughter.] MR. BARANOWSKY: No, that's part of our -- that's part of our follow-up inspection. When a performance threshold is tripped, whether it's on an individual incident or inspection finding, or if it's associated with performance indicators that collect a number of pieces of information, if you go over a threshold, part of that investigation by the inspection folks is to look at the root cause and corrective action. MR. MADISON: It may be part of that. And it will be the supplemental inspection activities that I talked -- that we're in the process of developing now. We're staffing up to do that. MR. BARANOWSKY: So, we're not trying to provide indicators that take the place of that supplemental inspection. DR. APOSTOLAKIS: I'm sure Alan wants to come back have the pleasure of another meeting. MR. MADISON: Several more times, I'm sure. MR. MAYS: I think we're also, George, missing a key point that's -- that Pat talked about earlier. DR. APOSTOLAKIS: Is this the first time, Steve? MR. MAYS: Let's see, last -- what was the time frame? I think the key point is the -- in putting this program together, as Pat mentioned, we're trying to make it related to risk and consistent with the regulatory oversight process. And one of the fundamental principles in regulatory oversight process is that we're trying to get away from trying to figure out all the possible causes of things going wrong and trying to control them at the Nuclear Regulatory Commission, to a process where -- that's the licensees job. And our job is to monitor performance and to monitor it a level, where we can interface with them, when that process isn't working. And our interface will become greater and greater, as the risk significance of that process not working becomes more and more evident. So, I think that's the basic reason why we're not as interested in this program, in trying to come up with causal indicators, as -- we're not saying that they're impossible, that they're not useful; we're just saying that they're not exactly the same fit with the regulatory philosophy that's embodied in this new process. DR. APOSTOLAKIS: Well, we can discuss this. You have 26 minutes, Pat. MR. BARANOWSKY: Right. And now I'd like to turn it over to Steve, to go to the next chart, which is number 12. MR. MAYS: And as a matter of fact, that was a perfect lead in. There are two fundamental principles that we're trying to maintain in this particular project. And the first one was, we want to have the risk-based performance indicators measure performance that's related to the cornerstones in the reactive oversight process. So, we're not trying to put together some new process that's brand new complete and conclusive just to replace it. We're saying, what we find out from a risk-based perspective that will make this process better. And we have a list of several attributes that are listed there. DR. APOSTOLAKIS: So, here, is where you will define performance next time? MR. MAYS: That's correct. DR. APOSTOLAKIS: Okay. MR. MAYS: And the second principle is, they have to be related to risk, and we spent a lot of time already talking about that. And that figure two that we had up earlier gives you our concept of what we think that means. And, again, to reiterate something we talked about before, we're looking at progress versus perfection here. And our purpose is also not to go out and measure total plant risk and then compare that against some standard. That's not what we're about here. What we're trying to do is take the performance as it is and assess its risk significance, wherever we find it, and then be able to take regulatory action in accordance with that risk significance. DR. APOSTOLAKIS: Wait a minute. Take the performance as it is -- MR. MAYS: That's correct. DR. APOSTOLAKIS: -- and compare it to what's expected from that plant. MR. MAYS: That's correct. DR. APOSTOLAKIS: Otherwise, what's significant; right? MR. MAYS: Well, we had -- see, we have a whole set of regulations out there that the plants are expected to meet. DR. APOSTOLAKIS: You're right. MR. MAYS: And what we're dealing with is, when the plants are out there and we have a situation where they're not meeting our expectations or they're doing something differently than what our regulations do, the amount and the level and the nature of the involvement of the regulator we're basically saying should be a function of the risk significance of that performance deviation. And that's what we're really about. DR. APOSTOLAKIS: Right. MR. MAYS: So, that, in essence, strains a lot of things that might be in the general universe of things about performance indicators -- DR. APOSTOLAKIS: Right. MR. MAYS: -- or causes that limit what we're trying to do, in this particular thing. DR. APOSTOLAKIS: And what a plant at another part of the country does is irrelevant. MR. MAYS: So, the next thing I wanted to talk about on that -- DR. APOSTOLAKIS: Now, one other thing that we commented on earlier. Desired attributes, would you like them to be as high as possible in that top down thing? MR. MAYS: You're getting down into the second -- two slides down from me. MR. BARANOWSKY: Do you mean in terms of measure -- we can measure or calculate? MR. MAYS: Yes, you're reading the script ahead of time. If you put up slide 13 here, Tom. Some of the technical characteristics make things related to risk, and we've talked about most of these already. But one of the points that came up this morning, and I want to make sure we talk about here, is we want to have definitions for these risk-based indicators. The definitions need to be consistent, so the industry knows what it is we're looking for and how we're going to collect data and count it. But, there also has to be capable accounting for differences in plant design. And you can do that a number of different ways, and I'll talk about that and a little bit difference. The second thing is: accounting for the sparseness of data, and I've used this example before and I'll put it out again, if we have 10- plants and they test a piece of equipment 10 times a year, and there's one failure last year -- that would have been one failure in a thousand test, and that would be roughly 10 to minus three probability failure on demand. But if you go to that one specific plant and look at the data alone, only in that plant specific context, you get one and 10. That's two orders of magnitude worse than the industry average, if you take that kind of a look. Well, that would be inappropriate. And so -- MR. BARANOWSKY: That's the count-based problem. MR. MAYS: That's the count-based problem. This is the problem that served as the basis for the many discussions, letters, and other stuff on trigger values -- DR. KRESS: Trigger values. MR. MAYS: -- that we had with diesel generators years ago. And what we're trying to do is avoid that same scenario coming again. DR. APOSTOLAKIS: But, this issue was addressed many years ago, when methods for developing plant to plant variability curves were developed. The fundamental question is whether the evidence that you just described for 1,000 -- not a 1,000 -- from 100 plants can -- whether the components of those plants can be treated as being identical. MR. MAYS: Or similar enough. DR. APOSTOLAKIS: Similar enough, yeah. But, if you can answer that question in the affirmative, then you can do this; otherwise, you can't. It's as simple as that. MR. MAYS: Agreed. DR. APOSTOLAKIS: So, I'm not saying that generic information should not be used. What should be use is plant specific information. Now, if you decide that something that happened at another plant could happen in yours, that's plant specific, too. MR. MAYS: Agreed. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: No, I think we agree with what you're saying, George, because, I mean, we would be using that in some sort of updated procedure. DR. APOSTOLAKIS: Sure; sure; fine. MR. BARANOWSKY: Okay. MR. MAYS: Okay. And the other thing is, we wanted to have thresholds consistent with the regulatory oversight process, that would identify a couple of things for us: one, when deviations were occurring outside of what the normal expected variation and the performance was and when that continued to degrade to the point where it became risk significant at different levels. DR. APOSTOLAKIS: Now, by the way, I know you don't mean that, but the statistical fluctuations should not be compared with the epistemic distributions, right? I mean, the epistemic distributions shows the variability in the mean value or the parameter value. The fact that you had three occurrences instead of one is something else. It's in the aleatory part. MR. MAYS: Correct. DR. APOSTOLAKIS: It's not clear in the 007 report now, these things. I'm sure the authors understand these differences, but it's not there. MR. MAYS: Okay. So on the -- DR. APOSTOLAKIS: This is the first time probably we have to deal with those underlying aleatory models. In the PRA, you never do. You deal only with the epistemic, the variability, the mean, and the medium, and that kind of stuff. But because you are dealing with the real world, you have to -- the aleatory model dels with the real world, not the epistemic. MR. MAYS: So the next thing I want to talk about in some of the PI -- risk-based PI concepts we want to look at was that we wanted to have the ability at the highest level possible providing risk indication. If we had it and could do it at the plant level for all the things associated with public risk, that would be great and we'd want to do that. I don't think we have the data and I don't think we have the models to do it, at that level. So, we're going to go down to the next level where we came. And sometimes, you'll get that directly through data and sometimes you have to do it through data and models to build up to it. And that's why we have a program here that will have information in different levels. The accident sequence logic from risk assessment is the tool we're going to use for establishing the order and priority of how we try to do this stuff. So we're trying to come up with indicators that would have data from industry and peer groups, as well as plant-specific. When I say "generic data", I mean data from similar plants and similar circumstances. I don't mean just a number somebody came up with, and say that generically applies. So, we're looking to gather data in similar groups and similar structures that relates to that performance of that group, and then apply probabilistic principles to determine what the plant-specific performance is, related to that. As we talked about the other indicators, one of the interesting things is, we're not only going to be getting indicators as it would indicate from a risk model. For example, you can look at a BWR HPI system an d get information about turbo-driven pump performance and valve performance, and put it all together in that train reliability. But again, the time between tests and demands and that is pretty small, compared to what the expected probabilities are, so you're not going to get very much on plant-specific. And even when you use that industry-combined data, you're going to have some limitations on your data. So, one of the ways we're going to look at this is to also look at motor-operated valves, air-operated valves, motor-driven pumps and other things across the plant to be able to see where the greater density you may be able to see some of the implications of some of those cross-cutting issues that people seem to be concerned about. That's why we have a common-cause failure database. That's why we have component studies underway, to be able to come up and say, how would those kinds of causal factors manifest themselves into something that we can get counts on that is related to risk. So, that's the purpose of doing that effort. The other thing that was raised this morning is about thresholds. Here's an example on this side about thresholds. We may have the exact same indicator definition for reliability of emergency diesel generators. But a plant that has two diesel generators and needs at least one of them in the event of a loss of off-site power to safely deal with that particular initiator might have a different threshold than a plant that has four emergency diesel generators of a similar nature but only needs one of them to be able to cope with the loss of off-site power. So you should be able to get credit in your performance thresholds for the amount of redundancy and capability you have in the plants. That's one of the ways you can deal with plant-specific issues. Another way is a situation which we've used, the example here of the AFW system. There are so many different designs in the AFW systems, we came up with 11 plant classes when we did the AFW study. And even within those classes, there was variability among them. What we would do in that case is try to come up with indicators that are the piece parts of that, use the models to come up with the reliability of those systems, then the threshold might be the same. If you go above an x-number increase in the reliability of AFW system, that transforms you into a particular white, green, yellow, or red zone. But the particular model would have less capability in it because you have had less trains. So we would devolve it in that direction. So there's a couple ways we can deal with that. We haven't made up our mind yet on any particular ones, which one we're going to use, but we intend to look at that to see which ones make the most sense. DR. APOSTOLAKIS: But my argument this morning was not with the ways dealing with it. MR. MAYS: It was whether we would do that. And the answer is, we are going to. MR. BARANOWSKY: Well, let me say, we're going to look at it. DR. APOSTOLAKIS: That's a definitive answer, I guess MR. BARANOWSKY: Before we say we're going we're going to look at it and see what the implications are. DR. APOSTOLAKIS: You're going to think about it. MR. MAYS: In fact, a lot of the work that we're going to do is run a number of test cases to see what the implications are. How much detail do we need, and does it make sense to put all this stuff in there? DR. APOSTOLAKIS: When I say "plant-specific", I don't necessarily mean, you know, that you have to take into account every little detail pertaining to that plant. If you decide that for the purposes of this program, knowing the numbers of trains, or redundant systems that you have, is good enough, that's good enough for me too. You have looked at it from the perspective of making it plant-specific. Now, in some cases, in some systems, they may have cross-ties, for example. And that may not be important, so you will have a third category. But it doesn't have to be, boy, for each plant, for each component, we're going to have one number that's unique to that plant. I didn't mean that. MR. MAYS: That's right. And there's a spectrum there between having completely generic and being totally plant-specific. DR. APOSTOLAKIS: Exactly. MR. MAYS: And what we're trying to do is find some sort of a balance in there that's practical to do, that maintains progress in this oversight process, as opposed to perfection, and moving on. I just wanted to make sure you understood, that you knew, that we understand that. We've looked at enough data to tell us we think we can do that. And we're going to go try and see where we can. DR. APOSTOLAKIS: Great. MR. MAYS: The other thing about the structure that we had here is that, when we talked about the highest levels to the subsequent levels, one of the leading issues we keep getting talked about is, what about leading indicators? We want to know about things before they happen. Without getting into the causal discussion again, if you think about that picture, what you see is that in order for you to get to public risk, you have to have several layers of things go wrong. So you can look at that chart as, the bottom level as the leading indicators of the next level, as the leading indicator of the next level, so when you have to talk about leading indicators, their first question is, leading of what? So, what we have here is a structure that will allow us to have indicators on multiple levels, where you should be able to see and reflect what the implications are. For example, suppose you had diesel generator performance going bad, and reliability was decreasing. Well, what if, at the same time, the loss of off-site power frequency was getting lower and the reliability of the other mitigating systems was going up? Well, right now you have no mechanism for dealing with that. But the kind of thing we're talking about in building an integrated model, you'll be able to put those together in a logical sequence, based on accident sequence logic, that would say, is that a big deal or is it not? Similarly, one of the concerns that was raised in the original framework was, what if you have a lot of little things that don't seem to be very important? Well, supposed of the reliability of diesels only went up a little bit, but at the same time, for the BWR HPCI unavailability went up a little bit, and the RPCI reliability went up a little bit, and off-site power frequency went up a little bit? Together, those might be important. DR. APOSTOLAKIS: It's like a million here, a million there. Pretty soon -- MR. MAYS: Right. Pretty soon it adds up to real money, as Senator Dirksen once said. And so that's the purpose of having indicators at multiple levels is to enable you to deal with some of these issues that were raised, as we were putting together this framework back in the workshop in September of last year. It's not going to be perfect; it's not going to capture everything. But it's going to be progress. And that's why we're looking at. DR. APOSTOLAKIS: So what you're saying then is that a performance, a set of performance indicators that are not risk-informed cannot be leading? MR. MAYS: No. We don't know what they're leading to. That's the problem. DR. APOSTOLAKIS: Well -- MR. MAYS: They could be leading, but we don't know what. DR. APOSTOLAKIS: Well, we had a very interesting presentation by Scientech at the subcommittee meeting, where they did make the point that, you know, something that was leading, that is significant with respect to one criteria may not be with respect to another. For example, if you had a cladding temperature as an indicators you know, how far you are from 2,200 degrees that does not necessarily give you, that is not necessarily a leading indicator with respect to core damage frequency. MR. MAYS: Right. I agree. DR. APOSTOLAKIS: So that's a fundamental weakness, then, of a non-risk informed, performance-based system. MR. BARANOWSKY: That's our point. We agree with that completely. You can have a lot of indicators that don't tell you very much useful information. But they just get people confused about where to spend their resources. MR. MAYS: Some of the limitations about risk based PIs -- again, they're not going to be perfect and do everything. If you have, for example, in this regulatory framework, the idea is you have indicators of some variance. You get worse; we go look at you a little more. But the simple fact of the matter is, we don't have enough data to calculate the plant-specific steam generator tube rupture frequencies. They don't occur that often. So it would be senseless to put together a steam generator tube rupture indication because, you know what? The next one that occurs, we're going. We'll be there. And we'll have inspection people out there and we'll be trying to find out what happened and what the cause was and how to fix it. So, there's no point in making a risk-based performance indicators on everything that's associated with risk because some of them we're not going to have data for. And one occurrence is going to be enough to trigger our increased regulatory attention. So we're not going to try to put performance indicators into that area. We talked earlier about yeah. The risk-informed inspections is where we're going to rely on people to go and look at the programs and the processes to try to keep those particular infrequent events, keep them infrequent. DR. APOSTOLAKIS: That's why you need -- MR. MAYS: And there's a lot of money being spent right now between the industry on how to deal with steam generator cracks, and all that other stuff, as mechanisms for preventing steam generator tube ruptures, from both financial and risk considerations. There's a lot of work going on out there and we're not going to try to do that. DR. POWERS: Let me ask, this class of things that you need to have indicators on because they are so very risk-significant that they're going to attract attention the first time you have one. I mean, that's like, a steam generator tube rupture, as you're saying. The steam generator tube people generator tube. People are going to be interested in that. Is this a complete set that you've listed up here, or is there a broader set of these? MR. MAYS: That's not necessarily the complete set. We're trying to convey the concept that as we're going through and determining which risk-based performance indicators we can do and are practical to do, one of the things we're going to do is take the principle that if it has a long period of occurrence, such as data's sparse, and it's such a situation that regulatory attention would be focused on it regardless, then that's a candidate for not putting together one. That's the principle and what we wanted to get some agreement on. I don't think that's necessarily a complete list in this slide, but it's exemplary. MR. BARANOWSKY: But the other important part is that there would be risk-informed inspection tied around that because the consequences are potentially high. DR. POWERS: How about risk-significant scrams, unplanned? MR. BARANOWSKY: Well, they're in between, probably. MR. MADISON: Again, I think you're getting into the supplemental inspection activities. I described the reactive inspection is going to be more kind of a broad scheme that's going to go all the way from the risk-significant scrams up to an IIT or an ARG, up to the TMI event, is the scope of the development of that process. DR. POWERS: Really, what I'm understanding is, you've got a principle here and I think you're inviting us to comment on things that don't need an indicator because they're so significant that they're going to attract attention. I'm just trying to understand the bounds on what your thinking is right now. How about inadvertent control rod movements? MR. MAYS: I don't know if we have an answer for every one of them, to be honest with you. MR. BARANOWSKY: I don't know. DR. APOSTOLAKIS: That's one that, those are items that have caused substantial amounts of regulatory attention at a couple of plants here in the recent years, that maybe, maybe that wasn't necessary. So I'm just wondering where that stands. MR. BARANOWSKY: Personally, I wouldn't put it in there, but we really haven't gone through all these types of things the way you're talking. DR. BONACA: But activity management is a major issue in the industry. I mean, that would fall right into that category. For example, INPO has paid a lot of attention to the issue, and driven it in the -- so, it's significant. MR. BARANOWSKY. It could go to into an operator performance indicator. DR. POWERS: And I think, you know I mean, that particular one, it seems to me what you're wrestling with is the kind of thinking that INPO would have, which is, that's really a bad thing to do. I mean it's just bad conduct of operations to be moving, jerking around your control rods in ways that you don't know. On the other hand, from a regulator's point of view, because nothing happens in there and it is difficult to see how control rod movements result in core damage with any kind of probability at all, that the regulator may well want to take that one and say, okay, that's an indicator that's a mark against his inventory of bad operator performance. And if he gets enough of those, then he has a decreasing trend in operator performance, and that probably raises questions about his training program and things like that, that I need to go and inspect. DR. BONACA: That's exactly right. What I mean is, you went to the training now. The bigger issue, in fact, is not necessarily an individual event, but the fact that if you have a significant sequence of minor events -- one may be rod movement; then you have a slight boron dilution, or something is happening out there; or you have a number of events that you classify as activity management. And then you look at your program and you're saying, this guy does not understand some of the basics. So that's really where you come back, and now you have a higher indicator that you want to look at because that indicator there is telling you something very substantial. It means that maybe your training program is not adequate. Maybe there isn't sufficient respect for the core in the power plant. So -- DR. POWERS: It's understanding when you use an event, a happenstance, as the springboard for giving close attention in and of itself versus using it as just one more checkmark in a demerit list, that if it gets too big, then it prompts you. MR. BARANOWSKY: I think that it's the latter that we would be thinking about, not a particular event but just some modest error. DR. POWERS: But you know what I'm talking about. MR. BARANOWSKY: Yeah, I know what you're talking about. I've seen them. I mean, it provoked a huge amount of response. DR. POWERS: Which under this system I don't know. It may or may not. But it would seem like it would not. MR. BARANOWSKY: Alone it might now, but taken in the context of a number of other things, it might be quite important. It might be just the straw that broke the camel's back at that point. DR. APOSTOLAKIS: Did you say insufficient respect for the core, Mario? DR. BONACA: Yeah. DR. APOSTOLAKIS: I like that. DR. POWERS: I mean, that's the concern. DR. BONACA: Because there is -- what you do, you tend to have people who have run the plant for a long time, they become complacent. They don't understand the implications of this. DR. POWERS: And it happens so easily. DR. APOSTOLAKIS: It's just the expression that I like. Not the content the content I understand. The expression. DR. BONACA: We always hate to talk about how you know, what kind of power there is inside that. DR. APOSTOLAKIS: So do people go and pay their respects to the core in some plants? DR. POWERS: It's when you're working with high energy processes, who work well, through those periods of times, you get this tendency toward the cavalier. DR. BONACA: Yes. DR. APOSTOLAKIS: Yes. DR. POWERS: And you say, I can do anything to this and it's okay because I know this system. And those are the ones that bite you. DR. BONACA: We had an example this morning with regard to working with nitric acid in the lab. DR. POWERS: Oh, yeah. Would you have respect for that? DR. BONACA: Oh yeah, one had better. MR. MAYS: That's all we need, by definition all we'll take. Okay, this table here is designed to give you it's probably what you'd call a target-rich environment. But what we've done is indicated, in the two phases that we're looking at working on risk-based performance indicators and we've tied these things to the corner stones, as well as to operating modes of the reactors. We've given you an indication of, visually of how much more depth and breadth we're trying to cover in terms of making progress in this process of risk informing the oversight process. You can see where there are several indicators that we're looking at right now that we've at least gone and done a preliminary look and say, is this ones we think are worthy of going after? So if you don't see one in here, it means that so far we haven't found it worthy of going after, but we'll be happy to take into consideration any of the ones that you think we ought to. DR. APOSTOLAKIS: Do you want to identify those as you go along? MR. MAYS: Yes. DR. APOSTOLAKIS: Because fire frequency doesn't mean anything to me. MR. MAYS: Well, the fire frequency was one we were looking at to see what it is we could come up with related to add power as an indicator of performance. Everybody will agree that the frequency of fires can be an important contributor to risk, but it's not the complete picture. DR. APOSTOLAKIS: Well, what fires? See, that's important in my opinion. MR. MAYS: That's correct. But we have to go we have a fire report that we put out and you guys have seen. We classified different levels of significance of fires. And we're looking to see what we can use with that to determine what we can put together as a performance indicator. The key thing for you to look at here is that issues associated with the barrier cornerstone are primarily going to be for phase II, which means we're not going to be looking at them right now. Our primary focus on phase I is on the initiating event and the mitigating system cornerstone, and putting together an integrated indicator for which there is no integrated cornerstone, but that would take into account those two cornerstones and put them together in a risk structure so we'd be able to come up with a little better picture of what the risk significance of things were as we go through this process. MR. BARANOWSKY: And also, not only just mitigating systems and so forth, but shutdown conditions versus power. I think we have a pretty significant deficit there that we're looking to try and fill as soon as we can. DR. APOSTOLAKIS: Good. Anything else? DR. POWERS: Well, I guess the target-rich environment that you advertised there, it really comes to maybe the words that are associated with Table 1. The basic premise is, utilize readily available data and methods and models that are already developed. That's an enormous constraint, it seems to me. Why? Why can't you make use of data that are not so readily available, or why is it not possible to develop models? MR. MAYS: It's not that it's not possible. It's not that it's not possible to get additional data that we either don't already have or are expecting to get. As Pat mentioned earlier, we're expecting to get some significant amounts of new data and new data density through the voluntary reliability and availability reporting that we're getting through EPICs, which was in response to our proposed rule. We got the first batch of that data in a month or so ago, and we're loading it up into that rad system, which is the developed database that you see on that program slide. So we are taking advantage of stuff not only that's currently off the shelf, but that is going to be on the shelf very shortly, in putting these together. The biggest issue with respect to doing brand new projects was, if we wait until we get to develop new stuff and get new data, we're going to be three, four, five years down the road before we make any improvements or any changes to the current reactor oversight process. And in light of some of the current concerns about limitations that have been already expressed, we thought it was a more prudent course to take stuff that we already had available that might answer some of those concerns. That might broaden the risk coverage of the indicators and the risk focus of the inspection activities to make a more robust system now while we can and then build this as an iterative process, rather than try to wait until we got a really, really good one five years from now and then find out that maybe it really took six or seven years to be a really, really good one and in the meantime had nothing in between. So that was the concern we had. We wanted to make some consistent, real progress in a fairly short time period. There's nothing fancier than that as the reason for wanting to do that. DR. POWERS: I mean, that sounds very laudable get something that's better than what we have right now. It's not contributing to what I would call regulatory stability by any means. MR. MAYS: I think the -- DR. POWERS: Continuously generating the PIs. Every time I turn around, it gets very difficult to know how to run a plant. MR. BARANOWSKY: Well, I think we're not trying to just generate PIs for the sake of generating them. MR. MAYS: We're going out and talking to people about what indicators we have now, and what we're going to have to rely on through inspection, and how to combine those two together in a coherent process, that's more coherent than we have now. And we've from people inside the agency and from people outside the agency that they have concerns about some of the capabilities and limitations that we have now. So what we're saying is, let's go ahead and try to address those issues that we can. I recognize that that will be new PIs and new data, but we're trying to get data and information that's either currently reported in EPICs, or currently reported in to INPO's SSPI indicators, or current information to minimize that kind of impact on people. And we're trying to put the ones together that we can get general agreement will have an impact on the way we do business in a positive way that makes it more logical and coherent and predictable to the industry. We're not just going to do any old indicator because we can. We're going to sue indicators that we can agreement that adds pretty much value to the process, otherwise we wouldn't be going. DR. POWERS: If I look at this table, it seems to me that in comparing current and proposed refining, expanding and refining your performance indicators in an area that you already have for many of them. Some sort of an indicator. What I don't see is a table that says, here are those areas that currently I don't have very good performance indicators. They're the security safeguards, radiation protection -- sort of security and safeguard sort of things. And I don't see a corresponding slide that says, here's how I'm going to attack those problems. MR. BARANOWSKY: I guess our corresponding slide is, we're not going to attack security and safeguards, or radiation safety cornerstones. They aren't within this risk framework. MR. MAYS: 17. MR. BARANOWSKY: Now, if someone said, please do something with those and come up with better indicators, we would give it a shot. But that's not the kind of area that we necessarily have the expertise in doing. I think we know how to develop indicators, but on the first set of cornerstones is the areas that we do have the expertise to do that in. And by the way, that's the bulk of what we're going to be looking at as a regulatory agency anyhow. We're only covering a small fraction of the risk with the current indicators. DR. POWERS: Well, at least from what we head this morning, we were presented with some flowcharts that posed challenges to me to try to answer the questions in the other areas. And so I guess it surprises me that a research organization doesn't want to attack those juicy targets. Not many footprints in that sand. MR. BARANOWSKY: I'm not saying that we don't as a research organization. I'm saying this project is not doing it; it could happen elsewhere. DR. APOSTOLAKIS: So you're final thoughts. MR. BARANOWSKY: Just the schedule, I guess. MR. MAYS: The schedule slide indicating that was asked earlier. We're looking to complete our definition of the candidate set of risk-based performance indicators and start apply the data we're getting from EPICs and other things to see which ones we can put together and use, and which ones make sense. We plan on coming back in the November timeframe to talk to ACRS about which ones we were able to put together, what they showed us, what they didn't show us, why we decided to do or not do things. Then we would go to the Commission, and then go out to the public, with the proposed risk-based performance indicators and see what kind of comments we got about those. And then after responding to those comments, getting more data on an analysis under our belt, we would complete the final set of risk-based performance indicators and come back in November of 2000 to ACRS and the Commission, and with their blessing would start implementing them in January 2001. So, the current PI report that you've been used to seeing that was done in AEOD and is still being done in this branch when we removed research the Year 2000 would be the last one of those you would see. And in 2001 it would be replaced with a new set of indicators, which would be these risk-based indicators. DR. APOSTOLAKIS: So the indicators that you use, Alan, in your pilots, preliminaries -- MR. MADISON: Those are, that's the set we're going to go with as far as implementation of the program. That's the ones that we've gotten. Remember the performance indicator program for the oversight process is a voluntary reporting program by the licensees. And we have agreement in that program with what we have. We'll have to go through in addition to the process that Steve's described, we're going to have to go through a process of review in the industry to get industry buy-in so that they continue the voluntary reporting program. Otherwise, we're going to be in the process of writing a regulation. MR. BARANOWSKY: Well actually, all the information we're talking about here, I believe, is currently being reported voluntarily or planned to be reported voluntarily. So it's not new information. I mean, we're mindful of that situation. MR. MADISON: Yeah, it will go through I don't think that process is going to be, going to slow up the implementation that Steve described. But we'll have to do that in conjunction with that. DR. APOSTOLAKIS: Okay. Any other -- you're not requesting a letter? This is just information. Any comments from the members? Thank you very much, gentlemen. It was very useful, as usually. Back to you, Mr. Chairman. DR. POWERS: And I will recess until 3:15. [Recess.] DR. POWERS: Let's come back into session for the third in our trilogy. And that is a discussion of the performance-based regulatory initiatives, which I assume are different than those that we've already heard about. And once again, we call on Professor Apostolakis. DR. APOSTOLAKIS: Thank you. We wrote a letter back in April of '98 after we had a presentation by the Staff on what we're planning to do on performance-based regulation, which we really performance-based regulation without the benefit of risk information, as I recall. We made some recommendations, which were not really used by the Staff when they met with public stakeholders. The Subcommittee on Probabilistic Risk Assessment and the Subcommittee on Regulatory Policies had a joint meeting last April 21st, when we had a presentation on NUREG CR-5392, Elements of an Approach to Performance-Based Regulatory Oversight. There was a number of issues that were raised at the time. There was an interesting discussion on leading indicators. The contractor showed, that, as I said earlier, what is a leading indicator with respect to one particular criteria may not be leading enough with respect to another. But they also presented what they called the diamond tree, which at a very high level tries to show how management actions and practices affect programs, and then hardware, and so on. And some members that one got mixed reviews. As I recall, some members liked the idea of a high-level diagram that showed how all things came together. Other members, including me, felt that it was too high-level and not of much practical use. And it was another name really for influence diagrams. An accommodation was made at the time or at least the Subcommittees felt that it would be useful to have an approach to this issue, similar to the one the Staff followed when they developed the Regulatory Guide 1.174. Namely, state a set of principles and expectations before you start developing the performance measures. I don't know whether the Staff plans to do this. One other development after that meeting is that the Nuclear Energy Institute sent a letter to the Chairman, I believe, where they no, to Mr. Rossi where they argued that this research program is not needed because we already have two performance-based initiatives. The first one is the one we discussed today on the oversight process, and the second, the efforts to make 10 C.F.R. Part 50 risk-informed. I must say, I myself am a little bit perplexed. Why? What is the objective of this program, and why is it needed? I'm sure we'll hear something about it today. So, with that, I will turn the floor to the Staff, and Mr. Kadambi, it's yours. MR. KADAMBI: Thank you MR. ROSSI: Mr. Prasad Kadambi's going to give the presentation for the Staff. Some of it is the same as what you saw the last time, and it reflects everything that we have learned and discussed since the last time we met with you. So with that, I'll let Prasad go ahead. I would like to point out that this effort is now in the Regulatory Effectiveness Section of the Regulatory Effectiveness and Human Factors Branch in the new research organization. DR. APOSTOLAKIS: Can you finish by 4:15, do you think? MR. KADAMBI: I believe we can. DR. APOSTOLAKIS: Okay. MR. KADAMBI: Thank you very much. Good afternoon, Mr. Chairman and members of the ACRS. Basically, what I'll present this afternoon will be an abbreviated version of the presentation on April 21st, augmented with some insights that we've had from additional input. This is just an outline of the presentation I expect to make. We did meet with the Subcommittee, as you mentioned. At the time, we thought that we would not have the time to meet with the full Committee, but we did get an extension on the Commission paper that was due at the end of May and is now due at the end of June. And so I guess we have this opportunity for additional input. We've used the inputs received so far to define more clearly, I hope, what we intend to achieve within the available resources. In this presentation, what I'd like to emphasize is what I feel is the statement of the problem that we are trying to address, as articulated in a series of questions, which I will get to toward the end of the program. This is briefly an overview of the presentation. I believe the Committee has copy, a draft copy of the Commission paper that we had sent around for concurrence review with the offices. As a result of the review process, we have learned about many performance based activities that are currently going on, and what we seek to do is to integrate and coordinate all the other activities. Some of these are, as you mentioned, a revision to the reactor regulatory oversight process, the risk informing of 10 C.F.R. Part 50. There is also the ongoing effort with regard to the maintenance rule and the implementation of Appendix G, the performance-based part of it. Also, there are many activities within NMSS that really do fall in the category of improving the performance-based approaches, application of performance-based approaches. Basically what we want to do is leverage our modest resources to learn about the benefits and drawbacks of performance-based approaches. What we will focus on, our observation of what is going on. Based on the observations, you know, we would hope to develop a technical basis for addressing questions on performance-based approaches. And these questions are those, right now, that I'll address later on in the presentation, based on the inputs that we've received so far from various people, including the ACRS Subcommittee. DR. APOSTOLAKIS: Now, are the last two bullets, something that the speakers of an hour ago really are addressing as well? MR. KADAMBI: Certainly, and that's the point I'm trying to make, is that this is really part of a lot of other things that are going on in the agency. And we want to really integrate it with the ongoing efforts, and at the same time provide DR. APOSTOLAKIS: Can you remind us what the SRM said? MR. KADAMBI: I'll have a separate slide on that. DR. APOSTOLAKIS: Okay. MR. KADAMBI: Just to set the stage, historical background this is rather text-intensive here. But the point of the slide is to show that the Commission has provided a lot of guidance on performance-based approaches. It goes back to 1996. You know, we dealt with direction setting issue 12. And the strategic plan also has specific language on it. In 1998, there were two SECY papers, one on performance-based approaches, and the other, the white paper on risk-informed performance-based regulation. The white paper was finalized on March 1, 1999, and that provides quite specific guidance in preparing the plans. And the SRM from the SECY 98-132 on February 11th very clearly lays out the specific elements that the Commission is looking for. The white paper, this is basically an extraction from the language of the white paper dealing with performance-based regulatory approaches. These are the attributes that are associated with it. The first is to establish performance and results as the primary basis for regulatory decision making. I believe that this is tied to the Government Performance and Results Act, GPRA, as we call them. It requires us to do this. There are four attributes that have been identified. These are words directly out of the white paper, I believe. To some extent I have summarized. But, the first one is that "parameters either measurable or calculable should exist to monitor the facility and licensee performance." The second is, "objective criteria to assess performance are established based on risk insights, deterministic analysis, or performance history." So what this says, I believe, is that risk insights are not necessarily the only source of the criteria, but, you know, I guess it depends on the particular situation where performance-based approach is being applied. The third highlights, I think, one of the expected benefits of performance-based approach, which is flexibility. "The licensees, if the performance-based approach is applied correctly, should have the flexibility to determine how the performance criteria are met, and hopefully it will encourage and reward improved outcomes." DR. APOSTOLAKIS: I'm sorry, I don't remember. I thought there was a period after established performance criteria. This is a new thing now, in ways that will encourage and reward outcomes? MR. KADAMBI: That's in the -- DR. APOSTOLAKIS: The white paper? MR. KADAMBI: The white paper. DR. KRESS: What does the "in ways" refer to? MR. KADAMBI: I guess that's the implementation of the -- DR. KRESS: Implement the rule in such a way. MR. KADAMBI: Right. MR. ROSSI: Well, I think the idea is that if you have a performance-based rule that provides licensee a moderate amount of flexibility, that will encourage them to find better ways to meet the performance criteria and even exceed them in some cases, in a cost-effective way. I think that's the intent of this. DR. KRESS: That flexibility will automatically give this. MR. ROSSI: That flexibility, right. DR. KRESS: Okay. I understand that. MR. KADAMBI: The fourth attribute, I believe, is equivalent to providing a safety margin expressed in performance terms. That is, "Failure to meet performance criterion will not, in and of itself, constitute or result in an immediate safety concern." Now, what the white paper also says is that these things could be done either using the regulation, license condition, or regulatory guidance that is adopted by the licensee. DR. UHRER: Just a quick question "performance-based approach" here implies that these are almost deterministic results. That is, you either performed satisfactorily or not. Is there an implication that this would be automated? That there would be a as an attempt to eliminate judgement? MR. KADAMBI: Not at the level we are talking about right now, which is I don't think you can ever eliminate judgment as a practical matter. But in terms of how the criteria are established and how the decision is made, whether the criteria are met or not, the second attribute speaks to that by saying that the assessment of performance that are established based on risk insights or deterministic analysis or performance history. DR. UHRIG: But I was looking at the words "objective criteria" in that second bullet. That implied to me that it might be an attempt to eliminate judgment. Okay. You've answered the question. MR. ROSSI: I think that one of the examples that I think you've very recently heard about, if not today, is the Reactor Inspection and Oversight Program improvements, where they have a green band, a white band, etc. DR. UHRIG: Yeah. MR. ROSSI: And that's probably a very good example of performance-based approach to the reactor oversight process. There, they use mitigating events and that kind of thing. And there was judgment used in setting the criteria, but I do not believe that that could be automated. That is an example where you use risk techniques to establish the parameters you're going to measure and the criteria that you're going to use. DR. UHRIG. Okay. Thank you. MR. KADAMBI: This is where, Dr. Apostolakis, we will address the specific elements of the SRM to SECY 98-132. The first one says that all the programs offices should be involved in identifying candidates for performance-based activities. It has become clear that there is a lot going on in the program offices, and our Commission paper should reflect that. We expect that it will. The second is that the Staff should take advantage of pilots to improve understanding and maturity. Again, there are many pilots going on in NRR as well as NMSS, and we will certainly learn from these pilots. The solicitation of input from industry is a continuing effort, industry and other stakeholders. We've done quite a bit of it in the past, and there's a lot going on there also. And again what we intend to do is take advantage of what's going on. DR. POWERS: Let me bring you back to your question on pilots, or the issue on pilots. We formulate these pilot programs, we get some plant to volunteer to participate in the pilot, and somebody says okay, this pilot's going to go on for, to pick a number, six months. How do we decide that six months is an appropriate time for a pilot? I guess what I'm asking you overall, is there any technology, any insights on how you design pilots? MR. KADAMBI: Not that I know of, but there may be others who know better. DR. POWERS: Why isn't there? MR. ROSSI: Well, I think that the criterion is a qualitative one that's used, and that is that you want to have the pilot go on for a long enough time to collect some meaningful information to allow you to know how well it's working, but you also don't want it to go on for so long that you cannot bring the pilot to an end and make a decision on whether you should implement a permanent change by rulemaking or something of that sort. DR. POWERS: But how do I decide -- MR. ROSSI: But I know of certainly no systematic approach that's been used other than what I just described to decide how long a pilot ought to go on. DR. POWERS: But we're making -- I mean, we seem to have pilots going on for everything I can think of. Pilots to the left of us, pilots to the right of us. If they are really going to volley and thunder, it seems to me that there ought to be some sort of understanding on what constitutes a good pilot and what constitutes a so-so pilot. We don't have any bad pilots. MR. ROSSI: Well, I think when you do a pilot, you want to look and see, you know, how do the licensees feel about that new approach to whatever it is that you're having a pilot on compared with something that's purely deterministic, whether they think that it does give them flexibility, whether they believe it reduces their cost or increases it, whether they believe that it creates a lot of additional confusion in dealing with the NRC on whether they have met the regulations or not. Those are the kind of things that you look at. And then the time period -- DR. POWERS: So really the pilot is more for the licensee and less for the NRC? MR. ROSSI: No, I think it's also for the NRC, because I describe what the licensee would be looking for. What the NRC would be looking for is the degree to which they believe that they can judge whether the licensee is or is not meeting the rule that is used to have a performance-based approach and whether what is being done is maintaining the proper level of safety, whether we can make objective decisions on whether the licensees are or are not going the proper kinds of things. So I think it's both ways, both the NRC and the licensees try to learn those kinds of things from the pilot. DR. POWERS: It just seems -- always seems curious to me when a pilot for inspection and assessment doesn't go through a full fuel cycle. MR. ROSSI: Well, I think -- I don't know how long they're going to go through the oversight process, but I think that's -- DR. POWERS: Six months? MR. ROSSI: I don't know whether it's going for a full -- is it going six months? DR. POWERS: It's six months. I mean, it's not even -- DR. SEALE: Well, it's gotten started a little earlier than that. DR. POWERS: But, I mean, you don't see everything unless you go through -- you may not see everything going through a full fuel cycle, but at least you have a chance of seeing most things. Let's continue. MR. KADAMBI: Okay. MR. ROSSI: Well, at six months you see a variety of things. You may not -- depending on the number of plants that you have pilots at, you may see the kind of things that extend over a whole fuel cycle in total. But you also don't want the pilot to become a permanent thing that goes on without coming to a decision and making something permanent. MR. KADAMBI: In addition to pilots, the SRM suggested that we should incorporate the experience from the maintenance rule and Appendix G. It also said that a plan should be submitted before expending additional resources. Right now we're planning on about less than one FTE effort, and incorporate the existing activities, comments from the Commission, and by comments from the Commission I believe is meant the various SRMs that have been issues in this area, the DSI-12 paper, and the strategic plan, et cetera, the ACRS and other stakeholders, and we've I believe done these. It also says to consider providing guidelines and simplifying the screening and review process for selecting activities. Lastly, it directs really that we should look for opportunities in new rulemakings, and the suggestion is made that when it goes out for public comment, when a rule goes out for public comment, the question should be asked whether some requirement is overly prescriptive. But in terms of what's going on, the process that the staff follows, this is a place where the rulemaking activity plan, which is a compilation of all the rulemaking activities going on, including NRR and NMSS, would be very useful. So these are the sorts of ongoing activities that we can take advantage of. Now, on April 14th we had a stakeholder meeting here and we solicited proposals based on a set of questions. I would like to just go to the questions and then come back to the slide. The questions posed to the stakeholders recognize that risk information is hard to apply in some areas such as quality assurance, security, fitness for duty, et cetera, and we asked for input on where performance-based approaches may provide a more cost effective way to address these regulations. We asked for suggestions that cover NMSS types of activities. The methods, tools and data, these are all I think reasonable obvious sorts of questions and, you know, we also brought up the ongoing activities such as the reactor regulatory oversight process and the revisions to Part 50. DR. APOSTOLAKIS: Doesn't your first bullet, though, send a message that you view performance-based approaches as competing with risk-informed approaches? MR. ROSSI: No, I don't think it does. As a matter of fact, I wanted to address that first question up there, because I believe one of the major intents of looking separately at performance-based approaches was to complement the use of risk-informed techniques in performance-based approaches and look to see whether performance-based approaches could be used in areas that were not readily amenable to reliance on PRAs. And as Prasad indicated, we had the stakeholder meeting in April. We also had the same subject on the agenda at a stakeholders meeting with the industry on September 1st in Chicago, and we made a concerted effort to solicit input from the stakeholders that attended these meetings on specific examples or specific areas where would could indeed look at applying performance-based approaches to rules, regulations, other areas of the regulatory process. And we basically got no specific new ideas beyond the things that had developed in the regulatory oversight and inspection improvement process. And I think the NEI letter that came in in response to our stakeholders meeting in April did not have any additional areas that they felt ought to be taken on at this point in time. So we have looked very hard at finding areas that are not amenable to the use of PRA for applying performance-based approaches. And I believe these areas exist, but I think what the situation is right now with the stakeholder is that they are very interested in the reactor oversight and inspection process and what is going on there. In NMSS, I believe both NMSS and their licensees are very interested in pursing the things that are underway now and there is a reluctance to take on other areas beyond those at this particular point in time. And so a lot of what we are deciding to do is based on that kind of stakeholder input and the fact that we do have a lot of performance-based approaches underway now that we feel we can learn from. Now, I recognize that the ones that are underway now to a large extent, at least in the reactor area, depend on the use of PRA, but, nonetheless, I think we will be able to learn from them. And so I think the real key is that eventually what we want to do is to address the questions that are in Bullets Number 1, 4 and 5 in areas that are not so amenable to PRA. But it may very well be that at this point -- that this is just not the point in time to be aggressive in that area, which is the reason that we are suggesting doing what we are suggesting here today. MR. KADAMBI: Going back to the stakeholder input, these are some of the issues that were raised, and I think the first one addresses the question you just raised. And I believe that the relationship is a complementary one and they are not competing with each other. People wanted to know the relationship of the performance-based activities to other NRC programs, and as of right now, we don't really have a road map to offer on this, and some of the things that we are proposing to do will hopefully help in that. Going down, I believe what the two -- the written comments that we received afterwards constituted quite significant, I think, inputs as part of the stakeholder input. There was one attendee who proposed a concept of a composite index, collateral or composite index, and NEI comments suggested that the oversight process revisions and the Part 50 rulemaking revisions were sufficient performance-based activities. I believe ACRS has received -- DR. APOSTOLAKIS: Now, let me understand that. I thought the staff was trying to risk-inform Part 50, not to make it performance-based. Is that correct? Or both? MR. KADAMBI: Well, one of the objectives stated in 98-300 is to make it as performance-based as practicable, I believe. So the stakeholders also suggested that, you know, the rules -- at least the general feeling was that the rules themselves may not be overly prescriptive, but, you know, some other aspects of how they are implemented may create the kinds of problems that are observed, and that it is important to learn the lessons of some of these ongoing activities such as the maintenance rule, et cetera. DR. APOSTOLAKIS: What is the bullet skepticism relative to accuracy of cost benefit analysis? MR. KADAMBI: Well, I think that was -- I figured that would come up. One of the questions we asked was, will work in the quest for performance-based regulations be cost beneficial? And I think, you know, this was a reaction to that. That is Question Number 3, one aspect of it, so. DR. APOSTOLAKIS: Do people really understand what the performance-based system is? MR. ROSSI: Well, I think that we discussed the characteristics that Prasad has on the one slide here, and they are pretty clear. I mean there are four attributes of performance-based approaches, and I believe they are quite clear. So I think people did indeed understand what is meant. And I believe generally people that that is what being done in the current effort on the reactor oversight and inspection program, is to do exactly that. And now that the challenge that we had hoped we could address was to find other areas that weren't related to the PRA, to pursue performance-based approaches on. And, again, that is complementary to risk-informed regulation, it is not in competition. It is certainly complementing. And I think you can use, obviously, use the risk-informed techniques along with performance-based approaches. I think that in many cases they go together. DR. APOSTOLAKIS: Maybe the skepticism is related to the costs that will be incurred to switch to a performance-based system. In other words, I am not looking at a steady system where you have a system, you see we have a system already. DR. KRESS: I think there is a longstanding skepticism in the industry about how NRC does it cost benefit analysis, particularly with respect to the costs. This may be just a carryover. MR. ROSSI: Well, I think there are also concerns on the cost to the industry of the maintenance rule, which was a performance -- basically, a performance-based rule, which may have driven some of this also. DR. APOSTOLAKIS: So are you saying then, Prasad, that the stakeholders present were not too enthusiastic about all this, is that correct? MR. KADAMBI: The bottom line is they did not offer very specific suggestions when invited to by the list of questions, and there was -- it wasn't very clear that anybody said, you know, the staff should be giving this very high priority and doing much more than we already are in it. MR. ROSSI: And you do have the NEI letter that states what Prasad just said I think pretty clearly. But I believe that that is meant to mean at this particular point in time. I don't think it is for all time in the future. But I am reading that into their letter. DR. APOSTOLAKIS: I am reading it the same way. DR. KRESS: Do we have any other information on the composite index concept? MR. KADAMBI: The letter from Mr. Jones is all that I have. I think it offered some interesting ideas which right now I can't really say very much other than, you know, having read it and tried to understand what was said. MR. ROSSI: I assume that letter has been available to the -- MR. KADAMBI: Yes. DR. APOSTOLAKIS: Yes. Making it available and making sure we understand are two different things. We certainly need to have somebody explain more. Do you understand it, Prasad? MR. KADAMBI: I believe I do. DR. WALLIS: Can I ask something naive here? If I were to try to choose between doing A and doing B, I would want to have a measure of whether A is better than B. And that seems to me you are always faced here with making choices. Should we make regulations this way or that way? Should we choose this path or that path? And here we have got some kind of sort of incantations or something, risk-informed, performance-based. I mean they are quasi religious or virtuous concepts, and, yes, you should do more of it. But when you are choosing between doing A and doing B, I like something like cost benefit where you can actually add things up and say A is better than B. I think what you see in the case of risk-informed and performance-based, how I get measures of whether doing A is better than doing B. MR. ROSSI: Well, the first thing I think you would do when you have an A and a B, generally, you have an A, and generally A is the way you have done it in the past or the way you are doing it today, and you are looking for a B that is a better way. And so the first thing you do is you compare B with A in both a qualitative and quantitative way and, to whatever degree you can, a cost benefit way, and decide that, yeah, B really looks better than A. And then what we are currently doing pretty much within the agency is we are doing these pilots which gives a better indication of whether B is really better than A to, as I indicated before, both to the licensees and to the NRC staff. And I believe, as I indicated before, that that is the major purpose of the pilots before you, ingraining this thing in concrete and offered up to -- or required of everybody. DR. WALLIS: But this business of, say, applying a principle of risk-informed, it sounds good, it makes sense, but has anyone sort of shown that if you do it in a risk-informed way, then B will be definitely much better than A? Is there some measure of how much it is better? MR. ROSSI: Well, the reactor plant oversight process is probably one of the very large efforts underway within the agency, and we are working with the industry there, and we are going to have pilots. So I think we are going to determine whether the risk-informed approach is or is not better. DR. WALLIS: Okay. So what is it you measure about the pilots that tells you it is better? Is it that people are happier? And costs are reduced? MR. ROSSI: No. I think you get an indication of the costs and efforts involved and the degree to which both the NRC and the licensees can come to agreement on what should be done. DR. WALLIS: Well, that's what bothers me, is that if it's just that two sides come to an agreement, that doesn't give me any objective measure of whether it's a good one. MR. ROSSI: Well, it's certainly that the industry is going to look heavily at the cost, and the NRC is going to look very hard at whether they believe -- and it may be qualitative belief -- whether they believe safety is being maintained as well or better with B than with A. And it may be a qualitative thing, although we do have a number of measures, very quantitative measures, of risk now, and I think you just heard that this afternoon in the risk-based performance indicators. They are a very good -- should be a very good measure of how safety is changing. DR. WALLIS: Well, let me assert that if you do your cost-benefit analysis right, then the payoff ought to show up there. MR. ROSSI: I would agree with you. No question about it. MR. KADAMBI: Okay. Moving on to the ACRS Subcommittee input, the main messages that I carried away from the April 21 meeting were that the staff should continue to develop answers to the question that was addressed last year as the result of the SECY paper that we were working on then, how and by whom performance parameters are to be determined and deemed acceptable, and how to place acceptance limits on them. And that is one of the things I believe we'll be able to address through the plans or at least the element of the plans that we are proposing. I believe the Subcommittee would also, I think as Dr. Apostolakis mentioned, the staff should seriously consider developing a guidance document analogous to Regulatory Guide 1.174, and also I believe there was a suggestion that some kind of a pilot application of the currently available methods should be considered. These were the three main points that -- DR. APOSTOLAKIS: I'm not sure it was stated that way, to tell the truth. MR. KADAMBI: If it isn't, perhaps, you know, I can correct -- DR. APOSTOLAKIS: I don't remember, though, the exact formulation, but we certainly did not urge you to jump into a pilot application before, for example, addressing the first two bullets. That was my impression, unless another Member recalls differently. Mike, do you remember, perhaps? MR. MARKLEY: I don't think it was anything that strong. It was discussed. DR. APOSTOLAKIS: It was discussed. MR. MARKLEY: I don't believe individuals made a recommendation to that effect. DR. APOSTOLAKIS: Yes. MR. KADAMBI: Well, that's fine. It was just something -- DR. APOSTOLAKIS: So the third one is a little bit fuzzy. MR. KADAMBI: Okay. Thank you. DR. APOSTOLAKIS: But the first two I think are definitely recommendations from the Subcommittee. MR. KADAMBI: Okay. By way of a summary of all of the inputs that we have received, the ongoing programs in the other offices represent quite significant efforts that we intend to take advantage of. There were no specific suggestions received from stakeholders that that would drive us towards any specific, proposing any specific projects. The NEI letter of May 13th suggests that the two big projects going on should be considered sufficient for the performance based area. DR. WALLIS: I'm surprised about Number 2. If it is a good idea, someone ought to say, gee whiz, if you really used it for this problem we would be so much better off. There ought to be some enthusiasm. MR. ROSSI: Well, again I think that there are areas that eventually we are going to get suggestions on. I think the situation is that we are pursuing some major efforts now with the industry and their belief is that we probably have enough to do in those areas in the area of performance based approaches to not warrant trying to expand it aggressively into other areas. DR. WALLIS: With the hope that you will then as a result of what you are doing now find some real successes to report and then there will be real embracing of the concept? MR. ROSSI: I think that is the crux of what we are suggesting on our very last slide, as a matter of fact, is exactly that, is to make sure we know about all of the performance based efforts that are underway within the NRC and to follow those closely and learn from them and then try to expand out over time from that, but let me not do the rest of Prasad's -- MR. KADAMBI: Thank you. Let me just go ahead and get to that point. The next two slides are basically, as I mentioned earlier, a series of questions which involved I believe an elementary planning process in which the statement of the problem should be made and in this case what I am trying to do is use a series of questions to do that. I am basically using the input from the Commission, the ACRS, and everybody else says, you know, what are you doing about these things? For example, how are performance and results providing the basis for regulatory decision-making? DR. APOSTOLAKIS: Do you mean that to end there? Do you mean how are performance results providing the basis? MR. KADAMBI: Well, I am taking the words out of the white paper. It says "Regulatory decision-making should be based on performance and results." I mean I think implied in it is the fact that there is a measurement of performance and that leads to results which are perhaps the outcomes, what we call outcomes these days, so this is a combination of the performance indicators as well as the specific outcomes. And the parameters that are used, what are they? Are they directly measured or are they based on a calculation or an analytical model, and how and by whom were they determined? This is the sort of question one could ask about an ongoing project. To what extent -- DR. WALLIS: Is performance-based unique to NRC or isn't there some history of using performance-based approach in some other application, where it was shown to work and learn from how it was done? MR. KADAMBI: I believe that there is plenty of evidence that it is being applied elsewhere, but I am not aware of anyplace where something like the list of attributes that we have tried to incorporate into a performance-based approach has been articulated in this way, but the NUREG CR-5392 that the subcommittee heard about, you know, does have a literature search in which there were examples given of other agencies that have tried to employ performance-based concepts. And when we are talking about specific projects, we can ask whether it's being implemented through a regulation, license condition, or a regulatory guidance, and these are the sorts of lessons that we can expect to learn. Now in terms of -- DR. WALLIS: It's strange, I'm sorry, following up, but I mean in something like car safety, you have seat belts and air bags. Now the real criterion should be that your car doesn't kill as many people as it did before. These are just -- and performance-based should be based on fatalities rather than on acquiring a certain kind of -- DR. SEALE: Appliances. DR. WALLIS: Appliance. That would be performance-based regulation, wouldn't it? MR. ROSSI: Well, I think the maintenance rule is probably an example where they look for maintenance-preventable failures, and that's their criterion as to whether they exist or not, and that's the criterion by which you adjust the maintenance program. So that's kind of analogous to what you're talking about with seat belts, I think. But obviously we have to start well before we have any kind of fatalities. I mean -- DR. WALLIS: That's a case where we have measures, though, we have fatalities, we have tens of thousands or something, and yet it doesn't seem to be a performance-based regulation that's used. MR. ROSSI: Do you mean for the automobiles? DR. WALLIS: We have a real -- all kinds of measures of how well we're doing in a way that we don't have so obviously in nuclear. DR. APOSTOLAKIS: But I think the State of Montana has decided to reduce the speed limit again back to 55 or 65? MR. ROSSI: It's 75. It's 75. DR. APOSTOLAKIS: Just last week. MR. ROSSI: That was 75, I'm quite sure, 75. DR. APOSTOLAKIS: Because last year -- DR. SEALE: It's 75 now. DR. POWERS: I'm not sure that seat belts and Montana are part of the discussions here. DR. APOSTOLAKIS: No, but it was performance-based. Last year the number of deaths went up, and they attributed it to the lack of a speed limit. DR. POWERS: Again, I'm not sure it's part of 10 CFR Part 50. DR. APOSTOLAKIS: I'm sure it isn't. DR. POWERS: I think we should stay on the subject I guess is my point. MR. KADAMBI: Thank you, Mr. Chairman. DR. WALLIS: I just mentioned it because it all seems so vague to me. I'm trying to tie it down to something I understand. If I could see an example in the nuclear industry where -- MR. ROSSI: Well, I think the oversight process I think has a number of things. I mean, they're tying things like initiating events that occur at a plant, safety system reliability and availability, they're tying those things back to the risk of a core damage accident, they're setting criteria on them, and they're looking, you know, they're looking at those things that are directly tied to risk. And that's a very quantitative approach. So I think that's a very good example, the risk-based performance indicators, of doing that. And that will allow performance-based decision making. But it's tied to the PRAs. DR. WALLIS: Well, again, the payoff is not in using something, it's in some other measure of success. That's what I guess I'm getting at. You don't get rewarded for using performance-based, you get rewarded because some other measure of something has improved as a result of using this method. MR. ROSSI: Yes, I would assume that in the case of the risk-based performance indicators you can plot the risk of the plants as a function of time and see if -- DR. WALLIS: Show some success. MR. ROSSI: Right. I think you can do that. MR. KADAMBI: And if we used the attributes of performance-based approaches, what it says is that there should be additional flexibility. So, you know, one of the things that we could ask is how is the flexibility being accorded to the licensees. And, you know, is the flexibility being used for improved outcomes. And improved outcomes may be reduction of unnecessary burden or, you know, improving the safety margins if that's what would be indicated by the risk analysis. For example, if the safety margin is decreased inordinately when the performance criterion is crossed, then it would indicate that the safety margin is not sufficient and should be increased. Whereas on the other hand if there is no change in safety margin even if the performance criterion is crossed, it may mean that the safety margin is much more than is needed. Those are the types of questions that I believe -- DR. APOSTOLAKIS: Actually I think your third bullet from the bottom should be rephrased, because, I mean, it's related to the question we just discussed. I don't think you have sufficient experience to make this judgment. MR. KADAMBI: Right. DR. APOSTOLAKIS: But it seems to me the question really should be can we design a system for determining performance goals that would translate to a predictable safety margin as providing a basis for consideration of changes. In other words, that's the heart of this: Can you come up with an approach to determining these performance goals? I would not call them criteria. Goals. That would give you a confidence I guess, would contribute to your confidence that the plant is operating safely. Because you're not going to get this message from experience. I mean, we hardly have performance measures operating. MR. KADAMBI: Well, you know, the idea is at least for right now to look at what's going on across the Agency in these multitudes of projects. DR. APOSTOLAKIS: I understand that, but, I mean, you say planning for performance-based approaches. MR. KADAMBI: Um-hum. DR. APOSTOLAKIS: Can you define an approach or devise an approach that will achieve this? That you will be confident if all the performance goals are met then that the plant is safe, or meets a certain safety goal. MR. KADAMBI: I guess I don't know the answer to the question. DR. APOSTOLAKIS: I know you don't. I don't either. MR. ROSSI: I think what you're suggesting is that we need to rephrase that -- DR. APOSTOLAKIS: Yes. MR. ROSSI: As a question. We ought to ask ourselves to make judgments on the effectiveness of the performance-based approaches. That's what you're saying. So it's a matter of not answering the question now but rephrasing the question and asking that question as we look at all the performance-based -- DR. APOSTOLAKIS: In other words, can you remove experience from there. I mean, can you logically define something like that. Okay, you have minus one minute. MR. ROSSI: Okay. I would like to point out that I think many of what Prasad has done here is he's tried to collect the things that we ought to look at, both qualitatively and quantitatively, in judging the performance-based approaches, and many of these I think were covered in questions that you asked, like cost-benefits and a number of the others are the things that we talked about even before we got to these slides in the meeting. DR. APOSTOLAKIS: Why shouldn't we agree with NEI? MR. ROSSI: I beg your pardon? Why should we what? DR. APOSTOLAKIS: Why shouldn't this committee agree with NEI that the agency is doing enough in this area and should stop this effort? MR. ROSSI: Well, first of all, the effort -- why don't we talk about that with the last viewgraph? DR. APOSTOLAKIS: Fine. MR. ROSSI: Because the last viewgraph may be -- that may answer your question. DR. APOSTOLAKIS: Good. Go to your last viewgraph. MR. ROSSI: So we are very clearly not going to take an aggressive approach to this thing, so we are close to doing what you suggested. We are going to expend modest resources in this area and do the things that are up there. DR. POWERS: When I look at that list I am hard-pressed to find a product coming out of this work. I see observe -- develop if appropriate -- enhance. I don't see "and by doing this work we will be able to do something tomorrow that we cannot do today." MR. ROSSI: We are not guaranteeing that we are going to -- the product up there, by the way, would be develop guidelines based on the efforts to better identify and assess issues and candidate the performance-based activities, and we say "if appropriate" so the product would be developed if our observations indicate that that is an appropriate thing to do. Basically what is being suggested is that we -- and we have already done a lot of this -- we clearly collect the list of all the things that are under way in the agency that are performance-based activities by comparing the activities with the characteristics of performance-based approaches, so we know what is going on and we feel that there's going to be a variety of these sorts of things that we can then learn for and we will look and see if we can learn things from them to apply them in other areas. You will notice that the last bullet up there says that we, based on what we do we will either institutionalize or terminate this activity, depending on what we observe. But we still believe that there must be areas where performance-based approaches can be used to address rules that are not amenable to addressing through the PRA approach. That may not be correct. I mean we may find that everything that people think of to apply performance-based approaches is dependent upon the use of PRA. I doubt if that is going to be the answer but it may be. If it is, then we jump to the last bullet up there I guess. DR. POWERS: Well, I certainly listened this morning to people struggling with how they would do performance-based approaches in connection with radiation protection and safeguards and security and things like that. I just don't see how this helps. You can't do those with PRAs but I don't see how this effort helps. MR. ROSSI: Well, I think what this effort is going to do is follow what is done. I mean this is an agency-wide program. It is not a separate program. What the tie with this effort is is that we will follow what they are doing in the areas of radiation and security and see how they are going about it, see what the answers are to the various issues raised on the previous two slides and then look to see if that can be carried into other areas. This is not going to address what is done in the radiation area or the security area. It will learn from those. MR. KADAMBI: What I would like to point out is a lot of these activities are going on right now. I mean we don't necessarily know what there is to learn from it until we really do try to pose the kinds of questions that the Commission has implied, I believe, in articulating the attributes of a performance-based approach. The other point is that there is an SRM over there that the Commission says, you know, the Staff should give us answers or respond, address these items and really we are proposing based on what we see happening in the Staff and with the limited resources available what we can do right now, and hopefully if there is something, if there is a gold mine out there, you know, this will lead us to it. DR. WALLIS: Well, the gold mine could be that the regulations are far too conservative and that if you really redesign them based on all these principles that there will be a tremendous pay-off in terms of productivity of the industry. That could be. I'm just speculating. I am not saying that that is necessarily the case at all, but if there's a gold mine it's got to be something like that. MR. KADAMBI: And that could well come out of even a modest effort like this, that there is a lot of margin over there. MR. ROSSI: I suspect that that particular gold mine that you described would be found by several of the different things that are going on now complementing one another to help lead that way. MR. KADAMBI: Mr. Chairman, that's my presentation. Thank you. DR. APOSTOLAKIS: Any other questions, comments? Thank you, Prasad. MR. KADAMBI: Thank you very much. DR. APOSTOLAKIS: Back to you. DR. POWERS: I think we can go off the transcription at this point. [Whereupon, at 4:23 p.m., the meeting was recessed, to reconvene at 8:30 a.m., Thursday, June 3, 1999.]
Page Last Reviewed/Updated Tuesday, July 12, 2016
Page Last Reviewed/Updated Tuesday, July 12, 2016