Plant Operations - May 9, 2001
Official Transcript of Proceedings NUCLEAR REGULATORY COMMISSION Title: Advisory Committee on Reactor Safeguards Plant Operations Subcommittee Docket Number: (not applicable) Location: Rockville, Maryland Date: Wednesday, May 9, 2001 Work Order No.: NRC-203 Pages 1-257 NEAL R. GROSS AND CO., INC. Court Reporters and Transcribers 1323 Rhode Island Avenue, N.W. Washington, D.C. 20005 (202) 234-4433 UNITED STATES OF AMERICA NUCLEAR REGULATORY COMMISSION + + + + + ADVISORY COMMITTEE ON REACTOR SAFEGUARDS PLANT OPERATIONS SUBCOMMITTEE + + + + + WEDNESDAY, MAY 9, 2001 + + + + + ROCKVILLE, MARYLAND + + + + + The Subcommittee met at the Nuclear Regulatory Commission, Two White Flint North, Room T233, 11545 Rockville Pike, at 8:30 a.m., John D. Sieber, Chairman, presiding. COMMITTEE MEMBERS JOHN D. SIEBER, CHAIRMAN GEORGE E. APOSTOLAKIS, MEMBER MARIO V. BONACA, MEMBER THOMAS S. KRESS, MEMBER GRAHAM M. LEITCH, MEMBER WILLIAM J. SHACK, MEMBER ROBERT E. UHRIG, MEMBER GRAHAM M. WALLIS, MEMBER MAGGALEAN W. WESTON, STAFF ENGINEER STAFF PRESENT: DAVID ALLSOPP, IIPB/DIPM TOM BOYCE, NRR/DIPM/IIPB EUGENE COBEY, NRR/DIPM DOUG COE, NRR A. EL-BANIONI, NRR/DIPM/SPSB JOHN HANNON, SPLB/DSSA DON HICKMAN, IIPB/DIPM J.S. HYSLOP, NRR/SPSB JEFF JACKSON MICHAEL JOHNSON, IIPB/DIPM PETER KOLTAY, NRC DIPM ALAN MADISON, NRR/IIPB GARETT PARRY, NRR/DSSA PHIL QUALLS, NRR/DSSA/SPLB MARK SALLEY, NRR/DSSA/SPLB MARK STORIUM, NRR/DIPM STEVEN STEIN, NRR/DIPM JOHN THOMPSON, NRR/DIPM LEON WHITNEY, NRR/DIPM/IIPB PETER WILSON SPSB/DSSA SEE-MENG WONG, NRR/ADIP A-G-E-N-D-A Agenda Item Page Introductory Remarks . . . . . . . . . . . . . . . 4 NRC Staff Presentation Introduction . . . . . . . . . . . . . . . . 5 Significance Determination . . . . . . . . .28 Lunch NRC Staff Presentation (cont.) Performance Indicators . . . . . . . . . . 203 Cross-cutting Issues . . . . . . . . . . . 234 General Discussion . . . . . . . . . . . . . . . 244 Adjournment. . . . . . . . . . . . . . . . . . . 257 P-R-O-C-E-E-D-I-N-G-S (8:30 a.m.) CHAIRMAN SIEBER: The meeting will now come to order. This is a meeting of the ACRS Subcommittee on Plant Operations. I am John Sieber, Chairman of the Subcommittee. ACRS members in attendance are Dr. George Apostolakis, Dr. Mario Bonaca, Dr. Peter Ford, Dr. Thomas Kress, Mr. Graham Leitch, Dr. William Shack, and Dr. Robert Uhrig. The purpose of this meeting is to discuss the reactor oversight process, which today will include the significance determination process and performance indicators. The action matrix will be discussed at our next meeting in July. We had our last subcommittee meeting with the staff on oversight processes on December 6th of last year. Maggalean W. Weston is the cognizant ACRS Staff Engineer for this meeting. The rules for participation in today's meeting have been announced as part of the notice of this meeting published in The Federal Register on April 16th, 2001. A transcript of the meeting is being kept and will be made available as stated in The Federal Register notice. It is requested that speakers first identify themselves and speak with sufficient clarity and volume so that they can be readily heard. I also request that all speakers please use the microphones to aid the court reporter. We have received no written comments from members of the public regarding today's meeting. I think we should now proceed with the meeting. Mr. Mike Johnson of NRR will introduce the topic and the presenters. Mike? MR. JOHNSON: Good morning. Thank you. I am -- my name is Michael Johnson from the Inspection Program Branch. I'm joined at the table by Doug Coe, who is also from the Inspection Program Branch. He is the Chief of the Inspection Program Section. And as was indicated, we have a variety of topics to talk about this afternoon -- I'm sorry, this morning, and spilling over into this afternoon. And there will be a bunch of additional participants, including representatives from the Plant Systems Branch. I've got John Hannon, who is the Branch Chief of the Plant Systems Branch; J.S. Hyslop and Mark Salley, who will be talking about specific issues of interest to the ACRS and the significance determination process, and other participants. So, you'll see participants cycle in and out for efficiency purposes throughout the presentation this morning and, again, into this afternoon. As was indicated, today's briefing really does focus on the SDP and the performance indicators. This is, again, a continuation in a series of presentations that we've had, the last one being in December where we specifically talked about issues relating to the SDP and performance indicators. We appreciate the opportunity to talk to -- talk to ACRS on an ongoing basis on these and other issues. We have, in the past, benefited from these exchanged. And, in fact, in preparing for today's presentation, I read over the transcript from our last meeting, and we talked about many of the issues, I think, that are on ACRS's mind with respect to the ROP. And we'll continue dialogue on those very issues today. So, it's been a fruitful -- a fruitful exchange for us. And we know that this fits into your schedule -- this meeting today fits into your schedule, along with a presentation, I guess, in July -- Mag, is that correct -- MR. JOHNSON: Yes. MR. JOHNSON: -- to talk about the action matrix and getting ready for September's session with the committee in preparation for your letter to the Commission on the ROP. And so, we're happy, again, to be in front of the ACRS to talk about these various issues. In preparing for today's presentation, we provided background materials. One of the primary background materials are SECY-99-007, or excerpts from SECY-99-007, that provide a lot of the basic information for the concept of the ROP. And we've talked about many of those issues many times with the ACRS. In addition, Inspection Manual Chapter 0609, which is our manual chapter that talks about the significance determination process, was provided. And we'll spend, again, a good portion of what we do today talking about and responding to questions on the significance determination process. We were able to work closely with Mag, I think, to understand what the issues were that you wanted us to cover. Hopefully, we've been able to factor those into our presentation. And I know to the extent we haven't been able to, you won't be shy in getting us to address the issues that you care about. Today -- next slide, Doug -- we're going to really focus on four, specific things. First of all, I want to just say a few words about the initial implementation status, and that is, that overall result of the ROP to date, to bring you up to speed with respect to where we are. Then, after that, we will get directly, again, into the significance determination process. We've got a series of examples that we want to go through with you to help you better understand the significance determination process and how it is being implemented. And in addition to some examples in the reactor safety area that Doug is going to talk about, we specifically will cover some fire protection -- the area of fire protection, the fire protection SDP, and an example in that area, again to help the ACRS better understand how we're implementing the significance determination process. Following that, we have a topic on performance indicators, again to respond to your questions on performance indicators and the issues, again to continue the dialogue on issues that we've talked about with respect to performance indicators and respond to your questions. And finally, we want to wrap up with some -- we call them selective issues, but they really are overall topics, if you will, that don't relate necessarily to the individual topics that we would have hit in getting there; so, again, an agenda, I think, that responds to the questions that we know you're interested in. Next slide. Let me just say a couple of words about the overall results. We are -- we have wrapped up the first year -- or, I should say, are wrapping up the first year of implementation of the ROP. Last week, as a matter of fact, Regions 2 and Regions 3 -- Region 3 conducted their end-of-cycle reviews. The end-of-cycle is the review that happens at the end of the assessment year in which the regions look at what has gone on in that year with respect to the performance indicator results, the trip thresholds, and the inspection findings of the trip thresholds in terms of looking at, again, what actions the Agency took in accordance with the action matrix and getting ready for issuance of the annual assessment letter that provides to the licensees and to other external stakeholders the results of the oversight process for that particular year. Following the end-of-cycle, there will be an Agency action review meeting. And just to put this -- the Agency action review meeting in context, think of the Agency action review meeting as a revamped senior management meeting. Again, it is the meeting of senior managers somewhat different from the previous process in that this meeting really is an opportunity for senior managers to review and provide an affirmation, if you will, of the results of the ROP with respect to plants that ended up with significant performance problems, and that means, for us, plants that ended up in the action matrix in the multiple repetitive degradative cornerstone. So, these are plants that have really had some performance issues. But secondly, in the Agency action review meeting, we talk about industry trends, and what have industry -- what has industry trends told us about whether we've been able to maintain safety. And finally, we look at self-assessment results -- the results of self-assessment for the first year of implementation, and what are the lessons that we've learned, and what is the feedback that we've gotten from stakeholders, and what changes do we need to make to the process based on those? So, that's where we are in the process. MR. LEITCH: Could you say again what you call that meeting? MR. JOHNSON: That is called the Agency action review meeting, the AARM, the Agency action review meeting. MR. LEITCH: Thank you. MR. JOHNSON: We believe that we've substantially exercised the ROP during the first year of implementation. If you were to look at the action matrix in terms of where plants fall in the various columns of the action matrix, we had a number of plants, a majority of plants, in the licensing response band. We had plants that ended up in in the regulatory response band. We ended up having plants that were at degradative cornerstones. That is a further degradation of performance. And in fact, we had a plant -- a plant IPS, Indian Point 2, that ended up in the multiple repetitive degradative cornerstone. That has enabled us, because of those cross thresholds, be able to exercise all of the supplemental inspection procedures. We've been able to do all of -- to do our event follow-up procedures. I almost said "all of our event follow-up procedures," but I stopped myself, Doug, because we didn't have an IIT, thank goodness. But we've got -- we've had a wide range of performance, and therefore, a number of opportunities to exercise many aspects of the ROP. And we think that's been a good thing. We've made several significant changes based on lessons learned to date where we found what we believe were flaws that needed to be corrected that we couldn't wait on. But our intent in going into the first year of initial implementation was to try to maintain the process stable, if you will. And so, we held off making wholesale changes until the end of the year where we could do a more considered self-assessment on what changes we needed to make. And you'll see those changes talked about -- being talked about in a Commission paper at the end of the year. And again, this will be talked about at the Agency action review meeting, and we'll brief the Commission on those results in July. l DR. APOSTOLAKIS: But you'll talk about them today as well? MR. JOHNSON: I would suggest that we talk about them maybe in the meeting in July. We'll be closer -- we'll have a better opportunity to have done the roll-up of self-assessment activities. We'll be closer to the Commission briefing, and we can give you a better idea of what we'll be telling the Commission. Finally, we believe that -- and we'll -- I'll just remind you that we've talked all along about establishing some objectives for the ROP. And you're well aware of those because you helped us form those. We wanted this process to be more objective, for example, to be more understandable and predictable. And we think that the process has, in fact -- is more objective, and more understandable, and more predictable, and the other attributes that we're measuring with respect to the fundamental objectives of the process. And we base that on some of the data that we've collected with respect to the matrix. We base that on the feedback that we've gotten from internal stakeholders, and the feedback that we've gotten from external stakeholders. We do continue, again, to collect data on the ROP. We have a set of matrix, if you will, with criteria associated with those matrix in some cases to enable us to draw some objective conclusions with respect to how well the ROP is meeting its intended goals. And we'll continue to collect that data and make decisions based on the effectiveness of the ROP and to indicate -- implement changes based on what that tells us as we go forward. So, those are the overall results of the first year of implementation. And again, we think we've made a fair amount of progress with respect to implementing the ROP. DR. APOSTOLAKIS: Now, it says there on the fourth bullet, "successful demonstration." I wonder what the measures of success were. I mean, what -- what could have happened that would have you made you declare it unsuccessful? MR. JOHNSON: We have -- that's a good question. We have -- but it's not one, George, I think you want me to answer today because that would take us -- and I think the question goes to the self- assessment process and the matrix that we've established, and measure the goals of the ROP. We have established those matrix. For example, with respect to the process being predictable, we measure things like did we -- did we implement the procedures in accordance with the criteria established for them? So, there are various criteria we've established, various matrix to identify each of the various goals. And what I would suggest, again, is that in that briefing that we do in the next meeting in July, that we come back and talk to you a little bit about what those self-assessment measures have told us about the various -- DR. APOSTOLAKIS: But the ultimate goal of this is to make decisions. So, without getting into details, have you made any decisions using this process that would have been different if the old one had been followed? MR. JOHNSON: Have we -- DR. APOSTOLAKIS: I mean, is the new process leading to more rational decisions, or better decisions, or decisions that make the stakeholders better -- I mean, happier? MR. JOHNSON: Yeah, we -- DR. APOSTOLAKIS: Isn't that the ultimate criteria? MR. JOHNSON: Yeah. In general, we have a -- we have a good sense of comfort with respect to the ROP in its overall ability to achieve the objectives that we set out for it. So now, all I'm suggesting is I can't -- I can't show you the matrix that enabled us to get there. In fact, we're still evaluating those matrix because, again, the year just ended. But yeah, we believe that the process is -- has been more objective, is more understandable. We've gotten specific feedback that says that the process is more understandable. The external stakeholders tell us the process is more understandable. The internal stakeholders tell us the process is more understandable. So, yeah, we believe that the process, again, at a high level, achieves its objectives. Now, I've got to caveat that -- and that's why I want to have this conversation again in July -- with several things. First of all, it is early. We're still analyzing the data. Secondly, for some of the matrix, because the matrix are new, it's hard to make a call on things like -- one of the things that we're going to measure, for example, with respect to measuring whether the program meets the NRC's performance objectives, is does the program increase or enhance public confidence? Well, that's a tough measure. We've gotten some ways that we're going to try to measure that. We've gotten some early bench-marking results, if you will. But it will take a year, or a couple of years maybe, before we can have some strong conclusions with respect to whether it does that. So, again, what I'd like to do is to come back to you and talk about the self-assessment process a little bit and the results. CHAIRMAN SIEBER: I think one of the aspects that licensees look at, which I think is important and you ought to evaluated, is whether the licensees perceive the process as being fair. You know, there were some times, years ago, that perhaps some enforcement action was interpreted as not as fair as it could have been. And it seems to me, with the structure that you've developed here, that the chances and the opportunities to be fair are much enhanced over what they have been in the past. But I think you ought to look at that process. And I guess another question that I have, which is really a follow-on to George's question, is are you making more decisions or less decisions, given the state of the industry, with the new process, as opposed to what you would have done under the old process? MR. JOHNSON: Okay. CHAIRMAN SIEBER: Go ahead. MR. JOHNSON: We might -- we might actually get into -- give you a better sense as to whether we're making more decisions or taking more actions as we go through -- as you see the SDP exercise, for example. We'll tell you how we come out on issues. And we'll try to -- we'll try to give you a sense for what -- how the -- how the old program might have dealt with those issues, to the extent we're able to. But in general, we've established, in this ROP, what is a -- what is called a licensee response band. And that means that we've come to the recognition and the realization that there is a level of performance and there is a level of performance degradation at a very, very low level that really falls within the responsibility of the licensee to correct. So, and that's different from the old process. In the old process, we would have engaged, perhaps, on issues that fell within that level. CHAIRMAN SIEBER: That's right. MR. JOHNSON: Under this process, we set aside those that are in the licensee response band. So, intuitively, the answer is that we make -- there are fewer interactions. DR. FORD: Mike, I have an even more basic question. I'm new to this, and I'm trying to learn. Can you tell me, in two sentences or three sentences, basically what this is all about? Are you trying to be proactive? Are you trying to reduce bureaucracy? What are you trying to do? MR. JOHNSON: Certainly, I'll try in two sentences, and, Doug, kick me if I get much beyond two minutes. The revised reactor oversight process, the reactive oversight process, grew out of an effort that we took on really early 1998, late 1997, out of some concerns that the Commission had really with respect to how we were assessing the performance of plants and deciding what actions we were going to take. And at that time, we had a number of processes, a number of different processes, in place. The Commission was concerned about subjective they were. The Commission had a very strong sense that -- that subjectivity shouldn't be central to our assessment process; that we ought to be objective as possible. For example, the Commission was concerned about the fact that we had -- we could be -- we could sense conflicting and sometimes overlapping messages through our various assessment processes. And so, the Commission directed the Staff, or we got permission from the Commission, to do an integrated review of our overall assessment processes and to develop a replacement. Around the mid-1998 time frame, we were talking to ACRS. We were talking to external stakeholders about that process. And we got feedback on that activity. And the nature of that feedback was still very critical, not just on where we were, but with where we were trying to go with respect that particular initiative. That caused us to step back, to take a fresh look, and that fresh look became what is the reactor oversight process. And in essence, what this reactor oversight process is, is it's a -- it's a process that starts with -- it's a hierarchical process that starts with the notion that there's a mission. It identifies strategic performance areas that have to be satisfied in order for the Agency to achieve its mission. And then, we went and identified individual cornerstones within those strategic performance areas, the cornerstones being the key, essential information that if we're able to satisfy ourselves with respect to the performance plans, we can have confidence that our overall mission is being achieved. And so, that's what the reactor -- how the reactor oversight process is structured. Now, within each of the cornerstones, we have performance indicators that -- that is, objective things that we can measure about the performance of the plant, that give us information about the performance of the plant. We also have inspections because we recognize that performance indicators don't -- cannot possibly tell us everything that we need to know with respect to the individual cornerstones. And we take those inputs from the performance indicators and from the inspections and we apply for thresholds to decide whether we ought to take, as the regulators, some increased regulatory action in accordance with an action matrix, a structured matrix that enables us to meter out, if you will, what our response ought to be based on the performance of the plants. And we take actions based on, again, the performance of the plant. So, that -- that's what we -- that's what we're about with respect to the ROP. Now, today, we're going to talk about performance indicators, so you'll get a better sense of what that -- how they work. We're also going to talk about the significance determination process. It turns out, when you do inspections, you've got to have a way, in this objective process, to be able to look at the findings from inspections to decide whether they're significant and warrant us taking some increased action, if you will, or whether they're minor, minor in nature. And that's what the significance determination process does. So, you'll get a sense for how that works also today. MR. LEITCH: Mike, I would just say, my perception is, too, that -- and just to amplify what you said, is that this was an effort to make the regulatory process more predictable, and to give licensees an early warning of regulatory issues. I think, in the -- in the late 90's, my perception was that it seemed to be -- the regulatory process seemed to be very brittle in the sense that a plant would be going along, apparently in good condition from a regulatory viewpoint. And then, all of a sudden, a situation would occur, either an operating event or some inspection would discover some particular flaw. And then, once that opened up, it seemed like it rapidly spread to the plant being effectively in a regulatory shut-down sometimes initiated by the licensee, but, in effect, a regulatory shut-down. So, I think the effort here -- correct me if I'm wrong, Mike -- but my perception is the effort here is to try to -- is to temper those actions, make them more predictable, and anticipate declining regulatory performance and take action before it gets all the way to "The sky is falling; we've got to shut this plant down." DR. FORD: Since the utilities are stakeholders in this, are they part of the team? MR. JOHNSON: We have had a number of opportunities -- provide routine opportunities, as a matter of fact, for stakeholders, external stakeholders, to interact with us. And that began back in -- back in 1998, as a matter of fact. It was sort of the watershed workshop that cast the structure for this. The framework of ROP was an external meeting where we had stakeholders; we had industry; we had the Union of Concerned Scientists; we had -- we had everyone that would show up involved in helping us develop and get alignment on how that process out to be laid out. And that continues today. DR. FORD: But they're not part of this -- these results? They weren't part of the team that came up with these results so far? MR. JOHNSON: They -- well, we -- how can I explain this? I'm trying to be very brief, and not -- and not take a lot of Doug's time. We have -- we've had a number -- as we implement the process, there are a number of opportunities for external stakeholders to remain involved. For example, when Don Hickman talks about performance indicators a little bit later on, we're going to talk about, for example, the fact that some of the performance indicators caused -- that is a reporting criteria that caused licensees to raise questions that require some interpretation. Well, in resolving those questions, those scruply-asked questions we call them, we have a monthly -- about a monthly meeting with the NRC and the industry, attended by NEI. And it's a public meeting where we take on those individual issues and work to agreement on the decisions with respect to who we should interpret the criteria or whether, in fact, we ought to change those reporting criteria to address a question. That's an example of sort of the ongoing interchange -- exchange that we have with external stakeholders in implementing the process. And so, they are -- the industry is. I mean, when we talk about the results of the process, we're going to tell you -- in July, we're going to tell you how we've implemented the process from an internal perspective. But we're also going to tell you how we think that process has impacted the performance of the industry. So, it's hard to separate the two. DR. BONACA: Yeah, just one comment I would like to make; it was simply there is an impression almost that there was no significant determination prior to this system. There was, and the significance was based on the degree of compliance. And today, the significance is an elimination process based on risk. That's really the big shift there, okay? So, compliance, alone, is not anymore material. I mean, typically -- I mean, if you had a finding, nobody very much looked at, you know, is it significant from a safety standpoint? It was, you know, how far are you from compliance within the acceptable regulation? And that really was the basis for determination of significance. DR. APOSTOLAKIS: This Agency has been accused of, in some past instances -- it's overreactive. Would this process help us not to overreact in the future? MR. JOHNSON: Doug, do you want to take that? In fact -- DR. APOSTOLAKIS: What did you say, Doug? MR. COE: Yes, you're exactly right. It helps us not to overreact, and it helps us not to under-react. We want, as Mike indicated earlier, a consistent and more predictable process. And I think that your points were right on. I think that the prior process, although there was an attempt to be thoughtful and to be consistent, it was more subjective. And over time, there were differences that arose as to how we reacted to various things, either under-react or overreact. And so, this was the essence of the concern that ended up where we are today. DR. APOSTOLAKIS: Okay. CHAIRMAN SIEBER: And in fact, that gets back to the statement that I made earlier about the perceived fairness of it all and -- which, to me, is a very important aspect of what it is you're doing here. MR. COE: And people say fairness is predictability and understandability -- CHAIRMAN SIEBER: Predictability and -- MR. COE: -- transparency -- CHAIRMAN SIEBER: -- consistency. MR. COE: -- and consistency, yes. CHAIRMAN SIEBER: Right. MR. JOHNSON: Now, you might -- you might remember Chairman Jackson's -- one of Chairman Jackson's favorite words was "scrutability". And we think this process goes a long ways towards helping us be very clear about what the issues are, what are determination of those -- the significance of those issues is, and how we got to where we end up with respect to what actions we ought to take. So, we think the process -- and again, that goes back to one of the key things that we're measuring about the process, what we think the process should measure. Okay, that's what I was going to talk about under "overall results". Now, Doug is going to start the SDP discussion. MR. COE: Thank you. Just building on what we just have talked about, the SDP is necessary to characterize the significance of inspection findings as one of two inputs to the action matrix; the other being the performance indicators. And the scale that was -- we tried to achieve with the SDP is intended to be consistent with the scale, the threshold scale, that is used for the performance indicators and when -- and when we take certain responses, based on those performance indicators. It started with the application of risk insight and risk thinking from a reactor safety standpoint. But as you'll note, we have seven cornerstones, some of which, in the safeguards area or the occupational or public radiation health area, of the emergency preparedness area, may not have a direct link to a core damage frequency risk matrix. And in those cases, we still have an SDP because we still need an SDP to characterize inspection finding significance so that it can feed the assessment process. And we try, in those cases, to make a consistent parallel with the risk matrix of the reactor safety SDP in order that the response that the Agency would give to particular inspection findings is consistent across cornerstones. That's a more subjective judgement, and it's one that we're -- you know, as we get more experience, we continue to refine. Today, we're going to talk about the reactor safety SDP because we understood that this was your primary interest. And so, what I'm going to show you are -- actually, I've got four examples, two of no-color findings -- and I'll explain what conditions arise, or what circumstances arise, that we would not colorize a finding -- one green finding, and one non- green finding. These are all real examples. In fact, for three out of the four, I basically cruised our website and plucked those three examples, the first three that you'll see, right out of our website. And I've referenced the inspection report numbers if you care to look further. The fourth one, the non-green finding, is also a real example, but it hasn't been published yet. So, I've sanitized in terms of its -- the description of what -- what the plant was and so forth. But it was, in fact, a real example. So, we'll get on with the first example. The no-color finding category are findings which don't affect the cornerstone, or which have extenuating circumstances. These are the two primary categories of no-color findings. The decisions on whether to colorize the finding or not is made prior to entry into the SDP. It's -- the guidance that governs that is Manual Chapter 0610*, and there's a series of questions that are asked. I'm going to try -- I'm going to show you, kind of at a high level, how those questions result in a no-color finding. The first example that I've got here was an inspection procedure that asked the inspectors to look at licensee LERs. And the finding that was reported in an LER was the missed surveillance test for the control room oxygen detector. Now, the guidance in 0610* is that if it -- if a finding does not affect the cornerstone and, therefore, cannot be process by an SDP, then it is documented as a no-color finding. In the example that we have here, the cornerstones in the reactor safety area are initiating events, mitigating systems, barriers, and emergency preparedness. And that's under the reactor safety cornerstone. So, looking at that particular finding, the lack of a -- or the failure to do a surveillance test for a control room oxygen monitor, when looked at from the standpoint of does it -- does it actually affect the cornerstone, and would it -- would it be possible to evaluate that finding through the SDP process using delta-core damage frequency, or delta- large early release frequency as the matrix. The answer would be no, and that's what came out of the 0610* lodging. DR. KRESS: Doug, what's the purpose of that oxygen monitor? MR. COE: The purpose of the oxygen monitor, I would presume -- and I can't say for sure, but based on my general understanding -- is that oxygen monitors are there in case the control room is enclosed, becomes enclosed, sealed, through, you know, control room isolation functions. And therefore, then there's a -- DR. KRESS: A chance of depleting -- MR. COE: -- monitoring process -- a monitoring instrument that, then, would tell the operators that they were getting dangerously low oxygen levels. DR. APOSTOLAKIS: But then, it seems to me that it would be under the broad category of reactor safety, would it not? DR. KRESS: That's what I was wondering. MR. COE: You could say it could be under the broad category of reactor safety. The next question you could ask is how would you characterize it was significant? If you're looking at delta core damage frequency or delta large early release frequency, there's really no -- there's no connection there. DR. APOSTOLAKIS: Well, I would say that its contribution to the facility is really negligent. I mean, that's probably a more accurate statement. And one does not need to do an analysis to see that. MR. COE: I wouldn't necessarily disagree. But what we're trying to do is come up with guidance that does produce the right results. And in fact, in very early stages of the development of this process, the criteria that we're discussing here on how to color -- how to choose not to color an inspection finding didn't exist. And there was a -- there was a thought being given at that time that there were -- if you couldn't meet the threshold for greater significance, it would be a green finding. And we would basically have a lot of green findings, okay? Now, somewhere along the way, in the development of this process, it was decided that having a no-color category would be useful, I think initially because of the extenuating circumstances that we're going to talk about in a minute. DR. APOSTOLAKIS: But in this particular case, if we would go back to the previous -- slide five -- no, this is six, yeah -- MR. COE: Yes. DR. APOSTOLAKIS: -- the definition says, "findings which do not affect the cornerstone or which have extenuating circumstances." It seems to me saying that the finding does not affect the cornerstone is too strong. I mean, has a negligible impact; I think that's more accurate. Maybe that's what you mean by "does not affect." And when you are elaborating on it, you actually -- that's what you said, that, you know, calculating that was really a waste of resources for this particular case. MR. COE: Yes. DR. APOSTOLAKIS: And the end result is known in advance. It's going to be very, very small. But it does fall under the cornerstone of reactor safety. DR. SHACK: But what is the question the man asked to make that decision? Does he say, "Does this affect -- is this going to affect the initiating events?" What are the actual questions he asks himself so he comes up with that answer? MR. COE: Well, those are questions that are articulated in 0610*, Appendix B. And for reactor safety cornerstones, not including emergency planning, they include the following questions: Could the issue cause or increase the frequency of an initiating event? That's the first question. The second question is, could the issue credibly affect the operability, availability, reliability, or function of a system or train in a mitigating function? There's four questions. The third question is, could the issue affect the integrity of fuel cladding, the reactor coolant system, reactor containment, or control room envelope, the integrity of those things? And four, does the performance of the issue involve degraded conditions that could concurrently influence any mitigation equipment and an initiating event? In other words, could you -- could you affect the likelihood of an initiating event at the very same time with the -- with the same issue that you would degrade a mitigating function? So, those are the questions that are asked. And I don't disagree that -- DR. APOSTOLAKIS: Doug, let me -- I think communication and using the right words are very important whenever you do things like this. I mean, we found that out and PRAs and so on. Instead of saying that we will not do -- I mean, we screened things out in the PRA repeatedly, and nobody objects, okay? And nobody has come back and said, "Gee, you know, you really missed it, after 25 years of experience." Instead of saying we're not going to do it, maybe a better way of saying it is that a crude evaluation shows, or a conservative evaluation shows, that the CDF is negligent. That sends the message that you have thought about it; you have evaluated it. You have not made the decision in advance not to evaluate it; which I think you are evaluating in some sense, you just don't want to spend too much time on it because, you know, professional judgement evidence shows it's not going to make any difference. So, I think sending the message in a different way is probably better. MR. COE: We can take that comment because it gets at a discussion, a dialogue, that has occurred since this guidance was formulated. And there are persons on the staff who feel much the same way you do. I would say that there are other examples, and perhaps this isn't the best example, but, for instance, a finding which involves a missed surveillance test, which then, the surveillance test is subsequently performed and found to be acceptable. Now, was there an impact on the cornerstone? Was the cornerstone functioned -- were any of the characteristics or attributes in the cornerstone for mitigating systems affected? Well, the answer would be no, not at all in that case. So, maybe that's a better example of a finding which doesn't affect a cornerstone. And to try to define that threshold, does it or doesn't it, is somewhat subjective, I would have to say. DR. APOSTOLAKIS: I understand. And I think -- you know, I think you understand the spirit of my comment. But I have another question. If the finding does not affect the cornerstone, and you guys have declared that these are the things you care about, why bother? MR. JOHNSON: I'm sorry, why -- DR. APOSTOLAKIS: Why bother to look at it at all? MR. COE: Typically -- DR. APOSTOLAKIS: You know that -- MR. COE: Typically they are violations, that -- for example -- oh, I don't know; we've got violations of, like I say, missing surveillance tests or of other administrative regulatory requirements that can't really be processed through the SDP. And in fact, one of the reasons why the no-color finding category came into being in the first place was to assess whether or not these findings that could not be processed through the SDP warrants a significant determination process of their own. And this question has come up in a number of areas, such as -- and most particularly, I think, in the cross-cutting areas, human performance issues, where mistakes are made, or errors are made, in the cross-cutting areas of performance -- problem identification and resolution. So, you know, it's a broad category of things that we find cannot really -- don't really -- don't really comport with an SDP that's been created. And we really can't make a link to core damage frequency changes or delta alert changes. So, we're left with this set of findings that may be regulatory issues, may be regulatory violations, that we're not sure what to do with. So, we put them in this category. DR. APOSTOLAKIS: I thought we were trying to get away from that. MR. JOHNSON: We are, George. Let me give you an example -- let me give you another example that perhaps adds to the examples that Doug has given that were very good. One of the things that you'll recognize that we ought to care about, as a regulator, that may not have a direct impact on the cornerstone, as we've been able to measure in terms in terms of the results of an inspection finding, is something, for example, that would impact the regulatory process or our ability to -- to effectively regulate the performance of the licensee. For example, let's suppose -- and this is a scenario that we've come -- we've had some concerns with respect to performance indicators. And that would be, for example, a situation where a licensee inaccurately reported a performance indicator, you know, or let's -- the example where there were some willfulness, a violation that was willful in nature. And maybe that would have a -- maybe there would be an element of that that would have an impact on the plant, that you could run through an SDP that would have an impact on the cornerstone. But the willful nature, or the inaccurate reporting, or you know, those kinds of issues are also issues that, again, when you look at the questions and the things in 0610*, the excerpt that we just handed you, are not things that you necessarily get through an SDP on, but things that we ought to care about as the regulator. Those are other examples. DR. APOSTOLAKIS: But you care about them because they're supporting things that need to be done -- MR. JOHNSON: Absolutely. DR. APOSTOLAKIS: -- because they affect the cornerstone. MR. JOHNSON: Or they eventually -- DR. APOSTOLAKIS: Could affect -- could affect. MR. JOHNSON: Could affect the cornerstone. DR. APOSTOLAKIS: Good, good. DR. SHACK: Let me come back -- I didn't like the answer to the surveillance test one because this one -- you know, this one, if I answer these four questions, if the thing failed the surveillance test, I think I would have answered the four questions in the same way. You know, when it comes to you, a surveillance test, and you say, "Okay, it was an important component; I missed the surveillance test. But when I did test it, it was okay, and it had no impact," that's an answer I don't think I like because that tells me I got lucky. You know, if I can't -- you know, it seems to me these things should be hypothesized. You know, if I missed a surveillance and if the surveillance test was negative, then I could still answer these as no significance. If I missed a surveillance test and it happened -- you know, the thing that has always bothered me about these things is everything is going to be green until something really happens. You know, yeah, it's no problem if you miss a surveillance test as long the thing is working well. You know, I either find out the thing is not working when I need it or in a surveillance test. MR. JOHNSON: Well, and I know Doug has got a perfect answer for this, but let me just cut in with my less than perfect answer, and then he can correct me. You know, when we say -- when things make it through the findings threshold; that is, they are more than minor, we're not -- we're talking about, in every case, something that we want the licensee to do something with. No-color findings are not -- for example, a missed surveillance test or, you know, anything that we documented in the inspection report as a finding or as a green finding, you know, a finding on very low risk significance, are all issues that the licensee needs to correct. It's not that we're setting them aside, that they can -- that they can have the option of doing nothing with. They've got to put them in their corrective action program, and we look at their corrective action program as part of our periodic PI&R -- PI&R, problem identification and resolution inspection procedure, to make sure that they're doing something with those issues. So, we're not -- we're not setting them aside, but they are clearly less significant than in a situation where you would have had, say, a missed surveillance test found -- that a surveillance test was found that there was a problem with that component. And that component, when you go back and you look and see, there was a condition that was brought on by some issue that happened a long time ago. And so, you can really look at how long that particular situation existed. MR. LEITCH: So, why do you take -- DR. APOSTOLAKIS: So, why -- MR. LEITCH: Excuse me. I was just going to say, could you take me through the line of reasoning that would apply? Here's the licensee that missed one surveillance test, and this is the only one he has missed in a year, versus another licensee that has missed ten surveillance tests in a year. And every one of those goes through the analysis, and every one is no-color. Is there some kind of a trending? How do you -- how do you deal with that, or would you like to deal with that? MR. JOHNSON: The way we do that is -- and Steve Stein, is he wondering around the audience? Make sure your ears perk up for this. The way we do that is, if we have a finding, and that finding is more than minor -- I'm sorry, a finding is more than minor, and we're documenting it in the inspection report. If there is some cross-cutting element of that finding, we document that in the inspection report. And when I say "cross-cutting," I mean things like -- the cross-cutting issues are things that are -- have impact on whether the licensee has a good PI&R system, problem identification and resolution system. If they're human performance in nature, if they're human performance things that are going on, that are, again, cross-cutting, and if there are safety conscience work environment issues that are going on -- different from safety culture -- safety conscience work environment -- by that, we mean is there something that is indicative of there being a chilling effect, if you will, a hesitancy on the part of the plant staff to raise issues. Those are cross- cutting issues. Well, if we have a finding in an inspection report, and there is this cross-cutting nature to it -- performance, problem identification, safety conscience work environment, those get documented in the inspection report. And as a part of our problems identification and resolution inspection, we -- today, on an annual basis -- and we're changing the period as to that a little bit, and making some other changes that we think improve that inspection. But we look at those issues, the collection of those kinds of issues, to see if that tells us that the licensee has what we call a substantial -- a trend, an average trend with respect to substantial problems in this cross-cutting area. And we document those in the assessment letter. We talk about those with licensees to get licensees to get them resolved. CHAIRMAN SIEBER: Does that mean that you're actually doing a bean count as you go through the period of missed surveillance and other kinds of things that licensees do that cause a non-cited violation because you can determine whether a cross- cutting issue is there or not? MR. JOHNSON: I wouldn't say -- I wouldn't use the word "bean count". In fact, the Commission was very careful with us to give us -- the Commission told us to be very careful with respect to how we -- how we treat issues that are green. The Commission was concerned that we would take a collection of -- we would count greens, things that have a very low risk significance -- CHAIRMAN SIEBER: Right. MR. JOHNSON: -- and we would somehow amalgamate them, if you will -- CHAIRMAN SIEBER: Right. MR. JOHNSON: -- and roll them up into something and make a big splash. CHAIRMAN SIEBER: Now, that's the old system. MR. JOHNSON: Right, that was the old system. But we think it's very -- CHAIRMAN SIEBER: And you can always write a finding against your QA program. MR. JOHNSON: Exactly. DR. BONACA: Let me ask you a question more specific. Go to the next slide, if you could. Look at disposition of finding, "confirmed entry into the licensee corrective action program." I mean, we come back to this, as we came back before. Here is what -- are you abandoning the issue once it's in a corrective action program, or are you looking for how timely they're going to address the issue, and whether or not this is a repeat issue? I mean, these are two fundamental elements of the corrective action program. And that answers a lot of questions. If you say, yeah, we're going to count it, and we keep an eye on that, then I am comfortable with this. MR. JOHNSON: That's exactly what we're saying. DR. BONACA: Okay. MR. JOHNSON: We're saying that we -- DR. BONACA: So -- MR. JOHNSON: In this PI&R inspection, that's exactly what we do; we go look at what is in the corrective action system. We ask ourselves, you know, is the licensee dealing with issues? Are there issues there that are significant that the licensee hasn't dealt with; you know, are there -- are there patterns? Steve, do you want to -- now is a good time for you to jump in. MR. STEIN: Steve Stein, Inspection Program Branch; I just wanted to clarify one point on that previous example. What made it a no-color finding was that -- was the equipment, was the control room oxygen monitor, not the fact that it was a missed surveillance. The missed surveillance on a mitigating system, on an injection valve, or pump, or on the EDG, would fall within a cornerstone and would go through the SDP. And all the conditions associated with that could make that more than of very low significance. So, that's the point I wanted to make, that what made that no-color was the equipment, not the fact that a surveillance was missed. MR. JOHNSON: All right -- DR. APOSTOLAKIS: Which seems to me supports what I said earlier, that you really need a conservative analysis as to be in your mind, and dismiss it, which I think is perfectly all right. I mean, that's how we do all these things. MR. JOHNSON: Yes, because -- DR. APOSTOLAKIS: If we have a problem somewhere, and we analyze it, and we find out the delta CDF is delta less -- or smaller than something, then that's a green. So, green is good. MR. JOHNSON: No, no. Green is not good. Green issues are still issues that we think the licensee needs to do something with. DR. APOSTOLAKIS: So, if the number of scrams is smaller than the number you specified, which is good, you give the guy a green, don't you? MR. JOHNSON: With respect to -- okay, we were talking about inspection issues. Now, green with respect to the performance indicators means that that performance is in the expected range, sort of this nominal range, of licensee performance. DR. APOSTOLAKIS: Right. MR. JOHNSON: So, in that case with respect to scrams, yes, scrams -- DR. APOSTOLAKIS: But green cannot mean -- I mean, is one of them light green and the other is dark green? MR. JOHNSON: They're green. DR. SHACK: Well, but still, no color does, in fact, highlight the fact that it's green. I mean, there's a difference between no color and green. DR. APOSTOLAKIS: Exactly, that was my question. Why bother? Why not declare this a green? MR. JOHNSON: We -- DR. SHACK: What's the difference? MR. COE: That's a dialogue that has occurred on an ongoing basis within the Staff. MR. JOHNSON: And in fact, this was an issue that we talked about. We recently had an external lessons learned workshop to roll up the results of the first year of implementation and to talk with the industry and other external stakeholders. And this issue of no-color findings was one that we talked about, and the kinds of concerns when something like -- you know, you've got a system that uses colors. We can understand the significance of colors. But here are these findings that you don't assign a color to because you say they're outside of the cornerstones. And what's the significance of those? There seem like there are a lot of those, perhaps. We should really do something with no-color findings. And in fact, we went into that workshop with a proposal that we were going to turn those no- color findings into greens. And what do you think the industry's response was? The industry said, "Don't make those things all greens. We care about greens just like we care about everything else." And so -- and so, the dialogue continues with respect to how to treat these issues. DR. APOSTOLAKIS: Okay, so let's put in a different way then. Why don't we go to the performance indicator for scrams; and if you are below the appropriate threshold, that is a no-color finding instead of green? MR. COE: That's not a finding. DR. APOSTOLAKIS: It's the same rationale. MR. COE: You see, that's not a finding, George. That's -- you've got performance indicators, which are just data collection, and then you've got findings which are actual -- some kind of deficiency occurred. The licensee's performance was deficient in some respect, and that was the source of the finding. DR. APOSTOLAKIS: So, "finding," you're using it in the regulatory sense? MR. COE: That's right. DR. APOSTOLAKIS: Again -- I don't know, this no-color business is not very comforting. MR. JOHNSON: We agree, we agree. But again, having said that, there are issues -- you can get yourself to the point where you can find issues that when you try to treat them through the SDP, you cannot. But when you ask yourself the group three questions in that excerpt that we handed you, if they still are things that we ought to be concerned about as an Agency, then those are things that are no-color findings. DR. BONACA: Actually, I think, you know, in part is the issue that -- if you look at the specific of this, you know, for it to be significant, you would have to have a significant event: a release, a problem with the control room, the need for oxygen in it, and then find that you don't have it because you didn't monitor it right. So, the risk, in itself, is minute. And yet, there are thousands of activities where compliance is important because without compliance, you don't have the assurance that in case yo have that kind of residual event that happens, you can deal with it. And I think that somehow we have to still deal with this thousands and thousands of compliance issues. And so, I'm trying to understand how -- I'm not disagreeing at all with you, George. I'm only saying that they still are significant individually because if you don't maintain some level of significance applied to them, well, you would have literally collapse of this commitments; I mean, particularly -- DR. APOSTOLAKIS: Well, I think there are two issues. DR. BONACA: -- what people are going to say, "Well, likelihood for it to happen is so remote, why should I" -- you know? DR. APOSTOLAKIS: It seems there are two issues, Mario, that have been raised here. One is the consistency of the approach, the self-consistency. DR. BONACA: Sure. DR. APOSTOLAKIS: In other words, when we do something with the PIs or the significance determination process and we declare something to be no-color or color of this, we have to be self- consistent. The second is I appreciate that, you know, we want to have a higher degree of confidence by doing -- having these additional requirements. But I thought over the last four or five years, we were trying to move towards a risk-performance based system, which is different in spirit. So, I don't know how -- I mean, it seems that we are still worried about things that are -- admittedly, the risk is insignificant. DR. SHACK: But I think their four screening questions are very good. You know, they seem to me to, you know, have answers that are scrutable and, you know, do a first cut at coming up with those questions in a way that an inspector, I think, has a chance of dealing with them. DR. BONACA: Absolutely, and look at the disposition of this; it goes into the corrective action program. I mean, it simply says "Do it." And the only activity is they want to monitoring if the corrective action program works. So -- DR. APOSTOLAKIS: If I look at the action matrix -- which we're not going to discuss today -- but if I look at it, I think if I have a green finding someplace, do I ask them to do something? MR. JOHNSON: You're not -- DR. APOSTOLAKIS: No, no, it has to be a licensee corrective action. When do I involve the corrective action program and ask them to do something? It has to be white? MR. COE: The action -- well, no. MR. JOHNSON: The problem identification and resolution inspection that I talked about happens as part of the baseline, happens for every plant, regardless. When the action matrix gets invoked is when thresholds are crossed. So, if you had a white issue, then you'll see -- you'll see that you change columns. DR. APOSTOLAKIS: Yes, so, nothing green; you don't do anything when it's green? MR. JOHNSON: Right, but we'll talk -- we'll go through the action matrix. MR. COE: Remember, a licensee that has findings and performance indicators in the green range is considered in a licensee response band. That's the characterization we've given it from an assessment point of view. So, what is it, 80 percent of the plants, or whatever it is, are in the licensee response band, and we expect them to deal with the lower level issues, the ones that we don't feel a need to engage them on. And so, their corrective action program is expected to correct those lower level issues before they become manifested in larger issues. And their motivation to do that, of course, is to -- is to continue to be treated in the licensee response band; that is, not get extra NRC attention, and inspection effort, and activity. DR. APOSTOLAKIS: Now -- MR. JOHNSON: Before we leave, can I also just add to you -- DR. APOSTOLAKIS: We're not going to leave. MR. JOHNSON: -- add that when we talked about this issue, you know, we really thought the stakeholders who would be most concerned with no-color findings would be members of the public. But in fact, Dave Lochbaum, who was -- who was there, didn't really share our view. He didn't -- he wasn't all that concerned about no-color findings, to be all that honest. And maybe it's because when we started off the year of initial implementation, we had -- we had a number of no-color findings. But that has gradually decreased as we were able to get out guidance with respect to these screening questions. And so, the numbers really are -- and I don't want to leave you with the impression that there are a lot of these things going around. There truly are not. And I think this may be a concern that we were more worried about than either the industry or others, like stakeholders. DR. APOSTOLAKIS: You see, Mike, one of the -- I am concerned on self-consistency. You have an example later where an inspection finding led to green, correct? MR. JOHNSON: We have one, yes. DR. APOSTOLAKIS: Okay. Now, inspection finding, by definition, is -- Doug just told us is some sort of violation somewhere. You forgot something; you did something incorrectly. MR. COE: It could be a violation or it could also be some kind of deficient performance that was not a violation. That's fundamentally it, but which contributed to an increase in risk, for example. DR. APOSTOLAKIS: And you declare that as a green. On the other hand, when it comes to performance indicators, green means expected performance, you just told us. Isn't there an inconsistency there? MR. COE: Yes. Actually, in that respect, there is. Both the -- well, the performance indicators include performance that we expect to occur as well as, in some cases, that which we don't expect. For example, unavailability performance indicators in the reactor safety cornerstone have a component of unavailable that occurs due to normal, routine maintenance, which is acceptable, as long as it's performed under the maintenance rule guidance. And then, there might be additional time, exposure time, of unavailability of equipment that's due to some kind of deficiency that is, then, added to that performance indicator. So, that's a particular performance indicator where you've got a combination of poor performance contributions to that indicator, as well as acceptable performance. But fundamentally, you know, you're right. An inspection finding is always associated with some performance issue; a PI may not be. DR. APOSTOLAKIS: But ultimately, the inspection findings feed into the action matrix too. MR. COE: Yes. DR. APOSTOLAKIS: So, now, it seems that the green means something different for those two, and we have different questions for the whites and the yellows -- MR. COE: Well -- DR. APOSTOLAKIS: -- which presumably will mean something different too. MR. JOHNSON: But George, with respect to the action that we take as an Agency, there's really no difference. If you have -- if you have a collection of only these kinds of findings that we've been talking about, they end up in the licensee's corrective action system. The licensee takes actions to address them. We periodically go out and look, and that's it. The licensee -- and the licensee -- the performance is in the licensee response band, and so our actions are to do the baseline inspection. If a plant has a scram or, let's say, two scrams in 7,000 critical hours, and the threshold is there scrams for 7,000 critical hours, once again, from a regulatory perspective, we're not doing anything. The licensee is in the licensee response band. Now, if a licensee doesn't get increasingly concerned as they get close to that threshold, we think that's a problem. But again, we're not going to engage because the licensee -- the plant is in the licensee response band. It's only when they trip that threshold that that is that deviation from nominal performance to the white for those performance indicators, or we have an SDP result that is white in any of the -- in any of the SDPs. It's that -- it's that that gets us to increased action, based on the action matrix. So, I -- now, they -- DR. APOSTOLAKIS: I guess -- MR. JOHNSON: -- they come together in a way that is consistent. DR. APOSTOLAKIS: Well, we're going to have another subcommittee meeting to discuss the action matrix, so maybe a lot of these questions will come back when we do. DR. SHACK: We're beating this to death. You just cut my -- DR. APOSTOLAKIS: It's already comatose. DR. SHACK: I would have thought you'd had zillions of no-color findings. But what's really happening is the inspector is not going out and zinging them for things that -- I mean, he's providing another level of screening before he even asks the four questions, rather than playing gotcha. Because I can't believe an inspector couldn't go through a plant and just keep writing up everything under the sun if he had a quota of citations that he had to fill. MR. COE: That's right. And the NRC has enforcement guidance on what constitutes minor violations. And we've actually tried to incorporate on that and expand on it a little bit in this process. So, the type of thinking that you're -- that you're thinking about now, that the inspector does, is actually -- we've tried to formalize this in this guidance. I'm just not discussing it right now. But it's -- it is there. DR. SHACK: So there really is even another level -- MR. COE: Yes. MR. JOHNSON: Yeah. In fact, they are -- we haven't talked about them, but they're the group one questions. If you look at the hand-out, there are some group one questions that really help the inspector try to distinguish what is truly minor. And I guess we don't -- I don't want to take Doug off and have him go through those. But yes, there is screening even before that. MR. COE: But the overall objective is to get inspectors to be sensitive to the things that are potentially risk-significant, the things that are potentially significant in the other cornerstones. And we've given them the -- the yard stick of the SDP is to help to define that in a much better, clearer way than we have in the past and, you know, towards the goals of objectivity and consistency. So, let me pursue this now because the next example I've got on no-color findings is actually in the other category, which is, perhaps, maybe, a little bit more clear, less subject to, you know, dialogue and debate. That is that that kind of a finding, which as extenuating circumstances -- and we define "extenuating circumstances," in our guidance. But principally, it's issues that may involve willfulness or issues in which the regulatory process is impeded because certain information which was required to come to us did not. Okay, and in this particular inspection finding, the licensee submitted an application for operator license -- I believe that -- yes, an operator license application was submitted. And it was -- it incorrectly stated that certain training had been completed. So, we were about to act on a license application to give an exam to an operator, and the information that we had was incorrect. The operator had not received the training that the license application stated that he had. Okay, per our guidance, this is a finding that potentially impacts our ability to perform our function, since we have been given information that's incorrect, okay? And in that case, if the impact of that can -- you know, does not affect the cornerstone; and in this case, it clearly did not because we caught this before the license -- before the operator was examined and put on-shift, then it's a no-color finding. So, again, it's exactly as before. We confirm that the licensee entered that into their corrective action program, and then we treat it as a non-cited violation. DR. UHRIG: Was this just an accident, or was it an error, or was this deliberate? MR. COE: Well, I can tell -- I can certainly say that it -- our assessment was that it was not deliberate, okay? Because if it was, it would have been captured in a different -- in a different way. In fact, willfulness, in many cases I think what you would expect to see, not just as a non- cited violation. We would probably examine it for enforcement action above the non-cited level, as a severity four or a three, or higher. So -- DR. UHRIG: Usually, most of these things are errors somewhere along the line. MR. COE: Yes, yes. But when we -- when they -- when we find them, we have to have a process to deal with them -- DR. UHRIG: Yes. MR. COE: -- and to deal with them in an appropriate way. And again, I think that if the -- if the process is set up to disposition lower significance items, or findings, it allows the inspectors to do more -- to spend more effort, on areas that are potentially of greater significance. And that's the intent. MR. LEITCH: Does it enter into your decision at all whether the item has been already entered into the licensee's corrective action program. Like say, for example, this issue here, say in -- before it comes to your attention, say the licensee has reviewed the matter and said, "Oops, we found a glitch in our training program. This fellow didn't really get this training," and they put it into their corrective action program, would that -- would that, in any way, affect this? Might this then -- MR. COE: Yes, actually -- MR. LEITCH: -- drop off the -- drop off the -- and not even be considered a finding? MR. COE: Our guidance does not provide for inspectors to, what we call, mind the licensee's corrective action programs, except in one specific case, and that's our periodic problem identification and resolution inspection procedure. MR. LEITCH: Right, right, yes. MR. COE: And there, we send in a team of inspectors on a periodic basis to do just that. And we look at the corrective action program, the items that are in there, in a -- you know, we try to look at them in a risk-informed way or, you know, looking for the items of greatest significance and looking for trends and patterns and that sort of thing. But the findings that come out of that have to be linked to the SDP in terms of their significance. In other words, we -- again like Mike said earlier, we're not allowed to go in there and aggregate things and then make them -- if they're all green issues, or would be green issues if we put them through our SDP, we could not make a bigger deal out of that than the most significant of those findings individually. DR. UHRIG: Suppose that this had happened before, had been put in the corrective action program, and it failed, and now it's showing up again. How do you -- what is the impact of this having happened before? MR. COE: Well, if this is a repeat -- DR. UHRIG: Yes MR. COE: -- that you're talking about, a repeat kind of condition, the philosophy that we're operating under is that licensees should be correcting these things at a lower level; and that if they don't, if this continues to repeat, and if the source of the continuation of this repeating problem is a more fundamental issue associated with their management controls or what-not, that we would expect ultimately that we would have inspection findings and/or performance indicators that would cross the threshold from green -- from the licensee response band -- into white in which we would engage. DR. UHRIG: But it would be in the corrective action program, not here? MR. COE: Yes, it would -- if their corrective action program was not functioning, we would expect, over time, to see these kinds of issues manifested as higher significance issues. If the licensee was doing a good job, and maybe, you know -- they're managing at a lower level, clearly. They're trying to keep the -- you know, the problems at a low level. And the real question is, for us as an Agency is, has the threshold -- I mean, we're going to allow -- and I think somebody said earlier, we're just waiting for things to happen. Well, we're not really because what we're doing is we're trying to define a threshold so that when things happen of a certain significance, we engage the licensee. And the intent is, is that we engage the licensee at a level before a significant impact to public health and safety occurs. So, when a licensee issue comes up that's greater than green, it goes into the white region for instance, then we engage at a certain level. We expect that that was -- is still not a significant impact on public health and safety. And we are going to engage at that level. We consider that an early engagement as the problems have now departed from the licensee response band, and now they're in the regulatory response band. So, we're going to get involved. MR. JOHNSON: Good, good. I just wanted to add one thing to make suer that we leave you with the right impression. If, for example, an inspector comes across an issue, they do not not make an issue an issue because the licensee already found it. If it's an issue, we set aside whether the licensee found. We look at that issue and treat that issue in our process. Now, when we go out and we do our supplemental inspection, in the case where issues have crossed thresholds, then is when we would recognize what the licensee has done with respect to finding the issue, and correcting the issue, and so on, and so forth. But there's no -- there's no provision -- well, inspectors do not -- you won't see one of the questions in the screening questions that is, has the licensee already found it, or is it already in the corrective action program? That's not the -- we don't want licensees -- we don't want inspectors thinking about those kinds of things. But again, what Doug has said is true; we don't want inspectors also -- we also don't want inspectors living in the licensee's corrective action program where they're simply, Doug's words, minding the corrective action program, looking through them for issues that we can bring up and document as our own inspection reports. MR. LEITCH: But the fact that you have few and declining numbers of non-white inspection findings would seem to indicate that that's happening anyway, right? The licensee must have 10 or 15 of these a day, issues entering the corrective action program. CHAIRMAN SIEBER: That's about right. MR. LEITCH: And many of those could be -- could somehow be a non-white inspection finding. So, there must be a de facto going on, a kind of -- these many low level issues are just not even surfacing as non-white -- non-color issues. MR. JOHNSON: Right, right. Yes, and that's what we mean when we say we don't want inspectors to mind the corrective action program. We think it's healthy for licensees to find their own issues, to put them in their corrective action programs. And we don't -- we don't want a program that discourages that by raising those -- pulling those issues out, raising them, documenting them, you know, just for the sake of getting greens on the docket. CHAIRMAN SIEBER: On the other hand, even though an issue may be licensee-identified, if it is truly risk-important, you would still have enforcement action regarding that. For example, the failure of all emergency diesel generators to start or load, even though the licensee may have discovered that and corrected it, it still is a matter for enforcement -- MR. JOHNSON: It still matters -- CHAIRMAN SIEBER: -- is that not correct? MR. JOHNSON: Exactly, it still matters in the reactor oversight process. It's still something that we would take action on if they cross thresholds, including -- including, perhaps, enforcement. CHAIRMAN SIEBER: Right, okay, thank you. MR. COE: Right. At the moment, there's -- it doesn't matter who finds it or whether it's self-revealed. We'll assess its significance, and we'll utilize the action matrix accordingly. CHAIRMAN SIEBER: Right. MR. COE: Okay, that's the last example on no-color. The next example is a green inspection finding. Again, this is under the reactor safety or the mitigation cornerstone -- mitigation systems cornerstone. In this case, during the conduct of an inspection procedure that was looking at surveillance testing, inspectors identified that an RHR system bypass valve had been temporarily modified to be in its full-open position. However, the licensee hadn't done any evaluation following that modification as to assure that the technical specification flow requirements were being satisfied. However, the subsequent evaluations that the licensee performed showed that the system flow did meet its surveillance test requirements. Okay, now, this differs somewhat because -- from the previous examples because there was a definite impact, is considered to be a definite impact, on the cornerstone; in other words, a safety function or a function of a component was affected. The flow was reduced. There was an impact, a physical change, a difference. And it was an adverse difference. It was the flow is less, okay? When screened through 0610* questions that we discussed briefly before, this conclusion is drawn: that it did affect the mitigating systems cornerstone and, therefore, its disposition would be, again, just as before, to confirm that the issue had been entered in the licensee's corrective action program. In addition, since it did affect the cornerstone, we would proceed to do a phase one SDP analysis. And the question that the phase one SDP asks in this particular instance is whether or not the system function had been affected, but whether -- not only had -- did the system function -- was the system function affected, but was operability and design function maintained? That's one of the questions. In fact, that's the first question that is asked of an issue in the mitigating systems cornerstone and when it enters the -- this SDP. And in fact, if that answer is yes, that operability and function were maintained, the issue screens, at that point, as green. And the licensee is expected to correct that, or the conditions, the underlying conditions, which caused that. But we would not engage further with any further inspection of its root causes or, you know, we would leave that up to the licensee. MR. LEITCH: Would a notice of violation have been issued in this case? MR. COE: That's a good question. I don't know the answer to that. I'd have to go back to the inspection report. I didn't -- that didn't jump out at me. MR. LEITCH: Just another similar question: is a non-color synonymous with a non-cited violation? CHAIRMAN SIEBER: No. MR. LEITCH: Are those two categories -- MR. COE: No. A green -- a violation which is given green significance is going to normally, in almost all cases, be given a non-cited violation. MR. JOHNSON: Yes, that's true. MR. LEITCH: Yes, okay. But a no-color is always a non-cited? MR. COE: Well, if a no-color finding arises because of willfulness or an issue which significantly impedes regulatory process such that we would consider it -- you know, we may consider for a severity level enforcement action up to, and including, civil penalties. MR. LEITCH: But it would still be a no- color? MR. COE: But it would still be a no- color. MR. LEITCH: I see, okay. MR. JOHNSON: Yes, we're sort of a little squeamish on answering the questions, your specific questions, only because -- and I'm looking at Steve. Steve is not even looking at me now because he knows that we're having an ongoing dialogue with the Office of Enforcement on this issue. And in fact, one of the things I intend to do when we come in July is bring the Office of Enforcement along with us because I think we ought to be able to talk about -- to address your questions about what enforcement also comes out. But to be quite honest, with respect to what is a no-color finding, Doug is exactly right. There are issues -- there are issues that receive traditional enforcement. And the Office of Enforcement doesn't consider those to be no-color findings. And I don't want to make this -- I don't want to make this overly convoluted, but let me just say that we have to -- we're still working with how we -- with this whole topic of no-color findings and how we eventually end up with respect to what is a no-color finding. When we set up the action matrix early on, we intended that there would be two kinds of things. There would be things that you could run through the SDP that would receive a color, and that color was synonymous with where they would fit, that we could run through the action matrix and come up with a result. It was also the recognition that severity levels would still apply for things that received traditional enforcement that were -- that is, things that were outside of -- things that could impede the regulatory process. For example, those would receive -- willful violations, those would receive traditional enforcement. So, you could have a situation where you have a finding that -- you'd have a collection of things: things that receive colors, things that received a severity level, right? And so, it's not as -- and so, when you ask a question, do all no-color findings -- are all no-color findings NCVs, well, that really depends on how you define a no-color finding with respect to how you treat these traditional enforcement items under that definition of no-color findings. In July, we'll have our act together because we will have -- we will have closed the loop on the dialogue with respect to no-color findings, and we'll be able to answer that. So, if you can hold whatever questions you have for that -- MR. COE: All right, the next example is the white inspection finding. In this case, an oil leak was identified by an inspector on an emergency feedwater pump. This is in a pressurized water reactor which had, over a period of time, forced the operators to make daily oil additions in order to even maintain visibility in the oiler. And when the residents had questioned this, the ultimate answer, the licensee determined that there were some bolts on the bearing housing that had been loose. The oil had been leaking into a drain sump, so it hadn't been puddling on the floor. So, it was basically, you know, not gathering the kind of attention perhaps that it should, other than the fact that the operators were adding oil every day. Ultimately, it was found that if the pump had been called upon to operate, it would have only run for a few hours. And then, the bearing oil would have gone away, and the pump bearing would have -- would have become irreparably damaged, and that that condition occurred for 39 days. Once again, the 0610* documentation threshold was met because it did affect a mitigating system cornerstone. The pump was actually inoperable, unavailable, for that 39-day period of time. Now, again, it would have operated for a few hours, but from a -- from the standpoint of a significance determination, the going in assumption was that it would not meet its mission time. It would not have satisfied its safety function in the long- term. Phase one SDP, another one of the questions that the phase one SDP asks is whether or not an actual loss of function of a single train has occurred. And the threshold that's put on that is if it's greater than its tech spec allowed outage time. And that's not necessary a risk-informed threshold, but it is a threshold that has -- we have historically used, even in the Accident Sequence Precursor Program. And we have borrowed it and continued its use in this program. DR. APOSTOLAKIS: Are these questions in this Appendix B? MR. COE: No, these particular questions -- in phase one, now we have left 0610*, which is essentially the documentation threshold and the discussion of no-color findings. And now, we've entered 0610, Appendix A, which addresses the significance determination for reactor safety findings, okay? And in that document, you would find a -- in fact, you're going to see it in just a moment -- a worksheet that lists these questions for phase one -- for the phase one SDP process. And this question -- my only point here is that's the question which kicks this issue into a phase two. In other words, if there was an actual loss of safety function and it was greater than tech spec allowed outage time, and therefore, it is deserving of further attention, further analysis; not that it wouldn't come out potentially green upon further analysis, but we can't say that for sure right now. And it needs to be looked at further. Again, in an overall sense, we have a graduated approach here. Phase one is a relatively broad screening process that allows an inspector to assess that something is not -- does not need further analysis or review from a risk standpoint, that they can call it green. They don't have to accept that, by the way. The guidance that we have put out is that if an inspector wants to exercise a phase two process, they're more than welcome to do that. And in fact, we encourage it, even if they think it's a green, because that starts them and continues them into the process of gaining risk insights into that particular issue and, in fact, into that plant in general. But in this particular example, that's the screening question that's relevant, and it forces the next phase of SDP analysis. Now, the next phase of SDP analysis is phase two. And I've got a more detailed set of documents here that we can look at. But the high- level picture of the phase two analysis is the following. The worksheets are utilized that are plant-specific. We've created inspection notebooks, risk- informed inspection notebooks, which contain a series of worksheets. And the purpose is to identify what this impact has had on the dominant accident sequences. DR. APOSTOLAKIS: Now, why dominant? MR. COE: Because that's what drives the risk. That's what drives the significance. In other words, the question that we're asking is which -- for this particular degradation, this particular deficiency, what effect has that had on the sequences, the accident sequences? Which sequences were affected by that particular degradation, and how much remaining mitigation capability was left to mitigate those accident sequences? DR. APOSTOLAKIS: But if you have a problem somewhere that affects a system that does not appear, say, in the top five sequences, but affects, you know, the following seven, what happens then? I mean -- MR. COE: Okay, it's a good question. Your question sort of infers that we're only listing the dominant accident sequences for review. In fact, the sequences are written at a very high level. Any one sequence is essentially a functional sequence. The sequence that's represented here, which was the one that came up the highest, when you look at it as far as what's changed, is a transient with a loss of power conversion system, essentially loss of the main condenser and the turbine, followed by a loss of all the other emergency feedwater components. And therefore, you have a loss of function of emergency feedwater, followed by a loss of primary feed and bleed. Now, that's the sequence -- DR. APOSTOLAKIS: How can you lose feed and bleed? MR. COE: Well, you didn't in this case. In this case, the only thing that was affecting that sequence was EFW, the EFW pump. DR. APOSTOLAKIS: You mean you lose the capability to feed and bleed? MR. COE: This is simply a functional accident sequence. The end result of this sequence occurring is core damage, okay? DR. APOSTOLAKIS: Right. MR. COE: So, what we're doing is we've got a whole bunch of sequences listed in the worksheets. And the idea is which of those sequences was affected by a -- in other words, what changed? Which sequence baseline value for that core damage frequency risk contribution has changed? Well, this one changed because this one used to have two motor-driven pumps and a turbine- driven pump. It now, for a period of 39 days -- DR. APOSTOLAKIS: Has one. MR. COE: -- there is only one motor- driven pump and one turbine-driven pump. So, that's the change that has occurred, okay? DR. APOSTOLAKIS: Right. MR. COE: So, this element has changed; the other two have not. And they retain their original baseline assumptions on frequency of this event occurring and likelihood or probability of this loss of function occurring as well. This essentially is providing defense in- depth. And the remaining mitigation capability that remains here is providing us a defense in-depth to sustain reactor safety -- DR. APOSTOLAKIS: So, do you have -- MR. COE: -- even when you had this problem. DR. APOSTOLAKIS: Do you have those sheets here and -- MR. COE: Yes. DR. APOSTOLAKIS: -- the information that the inspector has? MR. COE: Yes. MR. JOHNSON: Yeah, I was going to tell you that Doug has -- you are actually going to go through those sheets, aren't you? MR. COE: We'll go through them in as much detail as you wish. CHAIRMAN SIEBER: That's going to take a long time. MR. COE: This is just the high level treatment. I'm giving you the answer. And then, as you have interest -- DR. APOSTOLAKIS: Wait a minute -- MR. COE: -- we'll go through the details of how we get there. DR. APOSTOLAKIS: So, all these, pages 16, 17, 18 -- MR. COE: Yes. DR. APOSTOLAKIS: -- 19 -- do you want to get to that right now or -- CHAIRMAN SIEBER: No, I think that what we -- what our best bet is to continue and finish with the overall explanation as to what went on. And then, we can take a break and come back and -- DR. APOSTOLAKIS: Okay, good, good. CHAIRMAN SIEBER: -- dive into the details. MR. COE: Okay, the -- what comes out of the analysis essentially, for this particular sequence, is that you have to define what the likelihood of the initiating event is. And actually, it's -- we have a table, and I'll show it to you. And we've characterized bands of likelihood, which are essentially probabilities, with the letter characters that represent those bands. In this case, this is the highest frequency. In other words, this -- what this represents is, is that the initiating event frequency is greater than one in ten years, and the exposure time is greater than 30 days. So, the likelihood of this event occurring within the 39-day period of time is characterized as "A". We'll get into what that -- a little more detail of what that means later. In addition, the mitigating system credit that I was talking about earlier, there were two remaining motor-driven -- I'm sorry, there was one remaining motor-driven emergency feedwater pump left. So, the mitigation credit for that function here is two, which is a representation of 10-2 likelihood that that remaining motor-driven pump would not be available, plus one, which is representing the turbine-driven emergency feedwater pump availability of -- in this case "1" represents 10-1 likelihood that it would not be available. So, those are added, two plus one here. And then loss of feed and bleed, normally we give feed and bleed about a 10-2 credit for unavailability. And in this case, that's represented by this "2". So, you add these up and you get "5". You don't assume that you can recover that damaged pump should it have been called upon to function. And so, you are left with a credit -- a mitigation system credit of five. When combined with the likelihood rating of "A," you enter another table and you end up with a significance result of white, which is a representation of a change in core damage frequency because that pump, that one pump, is degraded for those 39 days. The change is between E-6 and E-5 per year, okay? DR. APOSTOLAKIS: So, the credits are really the exponents? MR. COE: Yes, that's essentially the negative log rhythm of the unavailability figure -- DR. APOSTOLAKIS: Good. MR. COE: -- the baseline unavailability figure. CHAIRMAN SIEBER: And I guess that's on page 23, right, how you get -- you know, the -- MR. COE: Page 23? MR. JOHNSON: Yeah, we'll get -- we'll get there. MR. COE: Yes, page 23 is the table which actually produces the final result -- CHAIRMAN SIEBER: Right. MR. COE: -- for that sequence, okay? Now, phase three -- you know, it's acknowledged, and it was acknowledged right at the very start, that phase two is a crude process. Its value is that it's in the hands of the inspector, who is the closest person to the actual plant design, plant operation, and can be the best person suited to identify if any of the assumptions that are being used in this level of analysis are incorrect. And that's -- DR. APOSTOLAKIS: What is -- I'm sorry. What is the CDF of this particular plant? Do you know the baseline CDF? MR. COE: I think it's about 3E-5 or 4E-5 per year based on their IPE. And I don't know if that's been updated. DR. APOSTOLAKIS: So, this is one-tenth of a -- MR. COE: Pardon? DR. APOSTOLAKIS: The change was one-tenth of that, right? MR. COE: It's in that range, yes. It's in that range of 10-6 to 10-5 per year. DR. APOSTOLAKIS: Right. MR. COE: Right. The phase three process was an acknowledgement -- the need for it was an acknowledgement that there would be occasions -- yes? DR. APOSTOLAKIS: Is yellow or white worse? Which one is worse? MR. COE: Yellow is worse. DR. APOSTOLAKIS: Yellow is worse? MR. COE: Yes, by an order of magnitude. MR. JOHNSON: And red is worse even. DR. APOSTOLAKIS: Yeah, I knew that. (Laughter.) MR. COE: So, phase three was essentially an acknowledgement that risk analysts would have to probably get involved at some point to either confirm that a phase two analysis was correct, that the assumptions were appropriate, and that the analysis was producing an answer that was defensible, or it may be that the SDP, itself, the phase two worksheets -- there are certain cases where the SDP worksheets will not work. For example, if a component is not considered totally unavailable, but a deficiency has reduced its reliability, the phase two worksheets won't work. The only way to assess that is through the use of adjusting the -- you know, making some judgement about the change in reliability, and then processing that through a computer-based model. That's the only way to do it. So, there are occasions when phase three, we anticipate, would be needed and, for the most part, because the worksheets have been so delayed in coming out. For the first year of operation -- of implementation for ROP, we've done a lot of phase three analyses. And we're hoping to relieve the burden somewhat on the risk analysts that have been doing these analyses by the implementation of these worksheets. CHAIRMAN SIEBER: I presume the region analyst is doing the phase three evaluation and the licensee is doing it with their own PRA at the same time. And the chance of the answers being exactly the same are probably nothing. MR. COE: There's always differences. CHAIRMAN SIEBER: So, how do you resolve that? MR. COE: Well, we talk about them and understand what the differences are. In many cases, the licensee is making assumptions that, you know, at least initially, we are not necessarily willing to accept. So, the process allows for us to come forward with a preliminary analysis and put it on the table and then, in fact, offer the licensee the opportunity to come back in a formal way, through a formal, docketed process, a public meeting, to come back and give us further information, upon which then we can make a final decision. We don't have to come to consensus with a licensee in order to produce an SDP result. But we do offer the licensee the opportunity to provide us with as much information as they feel is pertinent so that we can make a well-informed decision. CHAIRMAN SIEBER: So, this would be a matter for an enforcement conference if one were to occur? MR. COE: We call it a regulatory conference, yes. CHAIRMAN SIEBER: Right, okay. MR. COE: And the purpose is primarily to gather the information that we need to make our final assessment. CHAIRMAN SIEBER: Okay, thank you. MR. COE: Right. I would point out that, in many cases -- and this is an ongoing issue for us. And how we do this better is to account for the external initiating event contributions. As you're aware, the level of detail and sophistication and, in fact, the complexity -- because of the complexity of external initiating events is -- we don't have as much information as we have for the internal events. And so, we do the best we can with the information we have. But in some cases, it has to be a fairly qualitative judgement as to whether or not there's a contribution that would affect the analysis results. In this particular case, an enforcement action was issued as a notice of violation, or will be. Again, this hasn't come to completion yet. But the expectation that this kind of a violation would be a cited violation, okay? Earlier, we were talking about the non- cited violations at the lower-significance levels. When an issue of -- when a finding reaches the white or above significance level, then NCVs are not an option under our program. We cite the violation. We require the licensee to respond in writing on the docket. CHAIRMAN SIEBER: Now, when you issue a cited violation -- and let's not use this as the example, but just in general -- you still have the levels under -- like you had under the old enforcement policy? How do you determine what level you're in? MR. JOHNSON: Actually, when -- for an issue like this that you've been able to run through the SDP, we've set aside those levels. CHAIRMAN SIEBER: Okay. MR. JOHNSON: The significance is given by the color of the finding. And because it is an issue that's associated with a violation, it is a cited violation. So, with respect to the ROP and things you can run through the SDP, in general, we don't have severity levels. In general, you don't have civil penalties. And I'm saying "in general," because if you had a situation where there was some actual consequence, it is possible to have a color, to also -- but then again, run it through and assign a severity level and, in fact, issue a civil penalty. But in general, for most violations, we're talking about a color and a citation if it's a violation. CHAIRMAN SIEBER: That's enough, right? MR. JOHNSON: Yes. CHAIRMAN SIEBER: Okay, thank you. MR. JOHNSON: Okay, that's actually -- that's the high level treatment of SDP for the reactor safety. I would also point out that you're going to hear a little more detailed treatment of a fire protection example a little bit later. The fire protection example has an entire, separate appendix that they work through. But then, the outcome of that is an input to these kinds of worksheets, these SDP worksheets that we've been seeing here. So that there's some initial work that has to be done in order to process a degraded fire barrier issue, that type of issue. And we'll get to that a little bit later. Now, I don't know if you wanted to take a break now, because what I would be prepared to do is to go into some more detail on this particular white finding issue -- example. CHAIRMAN SIEBER: I think right now would be an appropriate time to take a break, at least for me. And let's reconvene at 10:30. (Whereupon, the proceedings went off the record at 10:10 a.m. and resumed at 10:30 a.m.) CHAIRMAN SIEBER: I think we can resume at this time. Mike? MR. COE: Okay, I'd just like to walk through some of the details of the white inspection finding example that I showed you a moment ago. We'll go into whatever amount of detail that you care to. The first part of an SDP in the reactor safety arena is a clear documentation of the condition. Factually and for the purposes of establishing exactly what the impact was on plant risk, we have to not include hypothetical situations, such as single-failure criteria. So, we basically ask the inspectors to document factually what the condition is. We also ask them to think about the systems that were affected, the trains that were affected, the licensing basis or the design basis of the system. That's sometimes not necessarily the whole story because it might have risk-significant functions that go beyond the design basis of the system. But at least as a matter of completeness, we wanted to ensure that that function was articulated. And that's done here on this sheet. Maintenance rule category is important when we ask some of the questions in the next -- in the next part of the phase one process. In the time that the identified existed, or is assumed to have existed, is important, again, from the standpoint of the final risk determinations because it's one of the influential inputs. So, this is just a more complete description of the identified finding that I illustrated earlier at a high level. So upon documentation that you saw there, the inspector is given a worksheet that looks like this to help them identify if, in fact -- or which cornerstone was affected. At this point, the decision had been made already that a cornerstone was affected. And based on the earlier questions, it's anticipated that it would be the mitigation systems cornerstone. But this worksheet lays out all three cornerstones in the -- or three of the cornerstones in the reactor safety strategic area, and asks the inspector to clearly identify what, exactly, is the cornerstone of concern. In some cases, it might be a combination of cornerstones. It might be -- a single issue might affect both an initiating event frequency, as well as a mitigation system cornerstone. But this is just to lay out the assumptions. And I would point out that the documentation expectations for the SDPs in this area, and in other areas across the board, across all cornerstones, are expected to be clear and reproducible, such that an individual member, a knowledgeable member, of the public could take the inspection report and the description of the significance basis and take our guidance documents, such as Manual Chapter 0609, the SDP process, and arrive at the same conclusion, so that it's reconstructible -- the basis is reconstructible. Okay, so that's -- here again, the mitigation cornerstone is the one that we're -- it should be this -- this next sheet here should be the next one in your package. I hope it is. CHAIRMAN SIEBER: No. MR. COE: It's not? Well, go to the next page, then. We might have reversed these two. CHAIRMAN SIEBER: Page 18. MR. COE: Okay, yes, page 18 is this sheet here. And it's actually the next one after the one I just showed. Here, the inspector is given a series of questions to determine whether or not the issue can be screened as green at this point and processed accordingly, or whether it needs a further treatment which may, or may not, result in a higher level of significance, but at least warrants the further treatment in phase two. In the case of mitigating systems -- mitigation systems, the inspector asks these series of questions. If the issue had impacted the initiating event cornerstone or the containment barrier, part of the barrier cornerstone, then you would ask -- he would ask these questions. In this particular example, the feedwater -- the emergency feedwater pump example, the first question was, is the finding a design qualification deficiency that does not affect operability? The answer is no, it did affect operability. And so, you go to the next question: is the finding an actual loss of safety function for a system? Well, that was defined in the first page, what the system was. The system, in this case that was affected, was the emergency feedwater, and the whole system had not been lost. So, the answer to that question is no. The third question is, does the finding represent actual loss of the safety function of a single train of a system for greater than a no-tech spec AOT? And this gets to the criteria that I indicated earlier, causes this to be answered yes. And therefore, we go to a phase two analysis. In other words, there is a need to look at this issue in further detail to assess its significance. The other page that's labeled 17 is not -- is not used at this point in the process because it deals with external initiating events. And it's a set of screening questions that would be used at the end of this process. If this process established a level of significance above green, then we would come back and we would look at these questions, and determine whether or not, either qualitatively or quantitatively, depending on the information we had available, whether or not external initiating events were a contributor to this significance. But we'll -- we can come back to that after we finish the internal event treatment. Now, each plant -- basically, now we have established the assumptions that we're going to be using in utilizing the various worksheets in the risk- informed, plant-specific inspection notebooks, which include these phase two worksheets. This table here represents, for this particular plant, the initiating event frequencies for all of the initiating events that would be subject to consideration by this SDP. And this table basically requires the inspector to -- and there's another table that will help in just a moment. But this table allows the inspector to determine what the initiating event likelihood is for the period of time, the exposure time, for the degraded condition. So, in this case, 39 days is this column here, this first column. And as we will find out -- I have already -- like I said, I've already given you the answer. But as we will find out, the initiating event that prompts the sequence of greatest interest here is one that has a -- starts with a reactor trip with a loss of power conversion. So, that's this initiating event here. The assumption -- the going-in assumption of its frequency is in this range. And since it's a 30 -- greater than 30-day period, that's why it resulted in this estimated likelihood rating of "A," which represents, once again the likelihood that that event will occur within that period of time. Okay, and we can come back to that as need be. Now, how does the inspector know which sequences he needs to examine in order to assess which have been affected by this particular problem? Well, the affected system; that is, the system that has the problem, is emergency feedwater. And this table, which again is specific to each plant, each plant's design, and is incorporated into that plant-specific notebook -- this would indicate that the EFW system, which is composed of two motor-driven pumps and one turbine-driven pump, appears as a mitigation function in all its initiating event sequences, with the exception of medium LOCA, large LOCA and loss of instrument air. So, in other words, the inspector's next task is to pull all of the worksheets out -- and we'll be going through a couple of those in a moment -- that -- except for these, okay? Because in all of those other worksheets, EFW appears as a mitigation function. And so, it has now been affected. So, therefore, the likelihood of those sequences occurring has been affected. And then, ultimately, we're going to seek to answer the question, how much have they been affected, okay? The next -- the next two pages are just a continuation of that table. CHAIRMAN SIEBER: Now, these are plant- specific; is that not correct? MR. COE: What I've gone through so far is all in the plant-specific, risk inform notebook. CHAIRMAN SIEBER: Okay. MR. COE: One is generated. There's about 70 or so such notebooks that are either developed or in the final, the very final, stages of development. And they cover all 103 operating reactors. CHAIRMAN SIEBER: All right. This table here is a generic table. It appears in our overall guidance document for the reactor safety SDP in Manual Chapter 0609, Appendix A. And what will happen here is that for those sequences that the inspector has identified as having been impacted, each sequence will be given an initiating event likelihood based on the particular initiating event for that sequence. And then, each sequence will be judged in terms of the remaining mitigation capability that remains given that this one aspect of its mitigation -- of the mitigation function is unavailable. And depending on how much mitigation function capability remains -- again, this is remaining mitigation capability -- and these are just examples. They're not a complete set of examples; they're just examples -- then, for that sequence, we'll generate a color. And the color will reflect the delta core damage frequency change associated with that sequence on an order of magnitude basis. Okay, so we'll come back to this table as well. In fact, I would suggest you maybe put a marker in that because we'll want to refer to that as we go through the analysis. This is the first worksheet that -- and, in fact, is the one that had the white sequence in it that I identified as dominating this particular analysis. The worksheet is -- the way it's laid out is the first row up here, the first line, carries forward information that has already been determined from the previous tables. This comes from that first table, and it indicates that this particular initiating event frequency is found in row one of that table, and that the exposure time assumption was that it was -- that -- from that table was greater than 30 days, and that that result was "A," as we saw, okay? Now, this next section of the table defines what these mitigation functions are that appear in these accident sequences. Again, these are high-level functional accident sequences. So, for each function, such as EFW, EFW will be described in terms of the plant components that are available to mitigate that -- to provide that function in order to mitigate that initiating event. And this is describing the full, creditable mitigation capability for each of these safety functions. So, this is as much credit as you can take for that safety function for that particular plant. In the case, in the specific case, of emergency feedwater, again emergency feedwater, in this plant, can be achieved -- the safety function can be achieved -- with any one of two motor-driven EFW trains and that the two of them together, therefore, comprise a one, multi-train system amount of credit. And there's a numerical value associated with that credit, and we'll talk about that in just a second -- in a minute. In addition, the turbine-driven EFW train, there's one. And it is a full, 100 percent capable train. So, one of one of that train is also capable of providing the full function. Now, in addition, there's -- there needs to be steam relief on the secondary side through either one out of two ADVs -- that's atmospheric dump valves -- or one out of 20 safety valves, okay? And that's -- that's a necessary -- that's there basically to -- for completeness. It doesn't really factor into the credit because you have so many options there. What you're really limited by is the ways of putting feedwater into those steam generators. Okay, now, looking down here at these three sequences that are listed, we see that EFW appears in all three sequences. Therefore, all three sequences have be affected by this degradation, by this problem. So, treated individually, we say how much mitigation capability remains for each affected sequence? In the case of EFW, we have one motor- driven EFW pump remains, and one train, one electro- mechanical train, is given a credit of two, that it represents 10-2 unavailability. Okay, a turbine-driven emergency feed pump, because it's a -- what we call an automatic, steam-driven train, we have only given it a credit of one, 10-1, unavailability. So, there's one in ten chances that that turbine-driven EFW pump would not function upon demand. But there's one in 100 chances that the motor-driven EFW pump would not function upon demand. And that's based on our generic insights in terms of the differences between electro-mechanical train reliability and steam-driven train reliability. So, that's the amount of credit we're willing to give in this particular, rough analysis. The other function that's associated that, if it were to fail along with these other failures or events, would lead to core damage is high pressure recirculation. The high pressure recirculation function is achieved through -- you know, for -- and I said "high pressure," so it's the high pressure function -- is achieved through one out of two high-pressure injection trains, which, in this case, there's a note here indicating that there are actually three pumps for two trains -- taking suction from one out of two low pressure injection trains. And all of this requires operator action. Now, in order to assess the value or the credit that's given to that function -- and remember, that function has not been impacted by this deficiency. So, we're going to give full credit. You'll notice that, for the EFW, we only gave the credit that was remaining. The fact that the other motor-driven pump was degraded or unavailable is reflected by the fact that it does not appear as credit for mitigation capability. In the case of high pressure recirc, there has been no impact on that function. And therefore, full capability is creditable. How much credit is that? In this case, operator action essentially is the most restrictive feature because if operator action doesn't occur, the function will not be met. And there is guidance in our 0609 document that describes how that should be treated. In this case, we give credit of three, which represents 10-3 likelihood that operators will not successfully implement high pressure recirculation. If you sum those numbers up -- oh, and this column right here, failure -- recovery of the failed train, in each of these cases of an identified, actual impact, the question often arises, can the issue or can the degradation be recovered by an operator action? For example, if this had not been a bearing oil problem; if it had been a switch left in the wrong position and an operator in the control room, based on indications, could identify that and recover from that deficiency, then credit for that recovery might be warranted. We give credit under -- only under certain meeting -- if the action would meet certain specific criteria, which are actually listed on this worksheet in the next -- in the next slide. But in this particular case, we are not giving any operator recovery credit because our going- in assumption is that the bearing will fail, and it will fail catastrophically. And therefore, there's not time available to recover from that. So, the function is completely lost. Now, the sum of these, if our math is correct, should be six. And that value of six is the mitigation credit for that sequence. And if you go back to the table, page 23, that we had up there just a moment ago, that sequence had an initiating event likelihood and a remaining mitigation capability of six, which is this column. So, that puts it in green. But notice that green is right next to a white, so we're high green, okay? I'm not going to use dark green or light green. (Laughter.) But clearly, we're up there -- we're less -- we're about an order of magnitude away from a white. And that may be significant later, so -- and we'll talk about that in a minute. The second sequence is treated the same way. In this case, early inventory high-pressure injection is satisfied by a multi-train system up here indicating one out of two high-pressure injection trains -- again, there's three pumps, but there's two trains. And that injects from a borated water storage tank. One multi-train system is given a credit of three, and that represents the combination of the individual components, the individual train reliabilities, plus the added factor of a possible common-cause problem mode. So, at a high level -- in other words, if it was -- if it was two independent trains, totally independent, each train would have a credit of two. And you could add those together if they were two independent and diverse trains to get four for our total mitigation credit. But if the two trains are identical and they're part of a multi-train system, then you can't -- you can't just add those up without accounting for the potential for common-cause failures. And when you add that in, that drops you back to about a 10-3 in a rough sense. So, that's the basis for that. So, that also, then, is given a credit of three, which differs in basis from what we gave up here, because up here, it was based on operator action. Down here, it was just simply based on the electro-mechanical train unavailabilities. Again, it produces a remaining mitigation capability rating of three -- of six, excuse me. And we're still dealing with the same initiating events, so we're still dealing with "A". And if you go back to the table, it's identical to what we had there before; again, green, and I've added here next to white. "NTW" stands for next to white. Now, the third sequence is the one of interest, and this is the one that I represented earlier on the high-level slide. The EFW credit is the same as before. In this case, the feed and bleed credit -- if we look at the feed and bleed function up here -- or in this case, it's defined as primary bleed because it really is based upon the availability -- not only operator action, but the availability of bleed sources. Similar to what I described up here for high-pressure recirc, the things that drives this credit here is operator action. And for feed and bleed, primary feed and bleed, we allow a credit of about 10-2 likelihood of not succeeding. And so, that credit of two is what is represented here; and when summed, gives five total, which, again, if you go back to the table that I showed earlier, would get you into a white range. DR. APOSTOLAKIS: So, if the operator credit is one, what would happen to this finding? It would be what color? MR. COE: Well, if you -- if you assume that for feed and bleed, that operators were only -- you were only comfortable, for whatever reason, giving operators credit of 10-1 likelihood of -- CHAIRMAN SIEBER: It would be red. MR. COE: -- of failure, you would be in the next -- you would go to the next color up, right? Because this would be one; this total would be four. DR. APOSTOLAKIS: So, it would be yellow. MR. COE: White would go to yellow. DR. APOSTOLAKIS: Now, that -- you know, the operator actions of this kind, you know, it is very difficult to quantify the probability because it's not the probability of failure. It's the probability of failure to decide to do it, hesitation, all that. So, I mean, what would that mean now? This is independent of the -- of the violation or the finding, right? MR. COE: This is independent of the violation, but it's assumption that is used as part of the significance determination. DR. APOSTOLAKIS: But your action, though, from the enforcement matrix, would be different; I mean, white versus yellow, right? And the full plant will be penalized depending on assumptions that have been somewhere else. MR. COE: That's right. DR. BONACA: For example, in this case, you do have -- in the simulator training, they are begin trained to enter into the -- DR. APOSTOLAKIS: It's not a matter of doing it correctly. It's a matter of deciding to -- DR. BONACA: Absolutely, and that is always absurd -- that is always absurd how fast they get into it. I mean, that sequence -- so, there is information available on-site. DR. APOSTOLAKIS: No, but you remember the Davis Bessie thing where -- DR. BONACA: I understand that, but -- DR. APOSTOLAKIS: -- economic consequences were huge. DR. BONACA: Oh, yeah. DR. APOSTOLAKIS: So, I don't know that the 10-2 is on solid ground. I mean -- DR. BONACA: No, I'm with you. I mean, there is information -- at least in later years, I know that there is a of emphasis on are they doing it or not. MR. COE: And I would point out that, on the next page, which is a continuation of this table, there's a note that we've added that indicates that based on the license -- this particular licensee's IPE, the human error probability for this function that they used, in their own analysis, was 3.4 E-2, all right? So, we're not far off. But the main point that I want to make here, I think, is that all of -- you know, your point is exactly right. The objective here is to come up with what was the impact on risk? And we use a core damage frequency risk matrix as the means of getting to that answer. And it's all based on a lot of assumptions. And what we're trying to do here is to bring these assumptions out into the open so that they can be examined. And again, the inspectors, very often, are the persons who are in the best position to point to an assumption and say, "I know that's not true. I know that's not right." We want the analysis to represent the truth, as best we know it. And in order to get those assumptions out on the table, we're using this kind of process. And we're asking the inspectors to go through the same kind of thinking process that would prompt them to think well, have I seen any problems in the simulator with feed and bleed? Is there any evidence that there's a problem in this area that I can't -- that this assumption should really be one instead of two, and that maybe the color should be yellow instead of white. DR. FORD: I'm a material scientist, and therefore, used to making a hypothesis and examining it with fact. This is a very logical approach, but is there any way of going aback and double-checking it against experience, actual, factual experience? MR. COE: In terms of do we have a database -- DR. FORD: Yes. MR. COE: -- of operator performance? DR. FORD: Yes. I mean -- DR. APOSTOLAKIS: I think you are talking about the whole approach, and not just operations, are you? DR. FORD: Exactly. I mean -- DR. APOSTOLAKIS: The whole SDP? DR. FORD: For instance, your whole -- this whole table is based on input in item one. You're putting a "1" in that top, left estimated frequency. MR. COE: Yes. DR. FORD: What happens if it was a different frequency basically you've got a time- dependent degradation or whatever it might be? MR. COE: Sure. DR. FORD: How would that -- is there anything that you can double-check these estimates, reasonable although they may be, against observations? MR. COE: Well, actually, yes. The basis for the -- for instance, the initiating event frequencies that we're using comes from a study that was completed a couple of years ago by -- it started out by AEOD, and then became research. But these assumptions -- you know, the order of magnitude that we chose to use for these various initiating events actually came out of an initiating event study, which is published in a new reg. It represents the best insight that this Agency has, based on operating experience that has been gathered over the years as to what we expect the generic frequency of those events to occur at. Now, it's important that the inspector understands that these are assumptions. And when they're applied to their specific plant, that specific plant's experience may differ. But the assumption of where -- where we're starting out assuming that that initiating -- that frequency is, or what the mitigating system reliability is, is starting out with a generic value. And those assumptions are exposed through this process and thereby, allow the inspector to challenge them if, based on their own knowledge and understanding of that plant, they feel that they aren't true. DR. FORD: So, when you say you have 70 of these documents going out, which cover all 103 operating stations, they may well change, depending on the history of that particular plant? MR. COE: Well, we -- DR. FORD: Bad water chemistry control or whatever it might be? MR. COE: Well, I think that -- yes, well, what you're saying is that the plants will change over time. They may modify the plant. The reliability of certain equipment may change over time based on issues or problems. What we've tried to do here is establish these starting values at a more or less conservative level. DR. FORD: Okay. MR. COE: We think we've got a more or less conservative set of assumptions here for most things. And we're continuing to monitor the process to identify, you know, areas where something might come up and we might identify that our assumptions may not be as conservative as we expected. And so -- but in general, we think that if this process renders -- and I don't mean that -- whenever you do risk analysis, you really shouldn't be using conservative assumptions, right? You should be trying for the best, most reasonable assumptions possible because conservative assumptions can often, you know, cause the results to skew and may obscure other things that you're interested in. And so, we're not trying to be over- conservative in our assumptions, but the numbers we're using are based on systems studies, for example, such as that are generated by research now: ox-feed water, diesel generator systems. They've gathered information from LERs and other databases, and they've done statistical analyses. And the numbers that we're using, such as a credit of 10-2 for a single, electro-mechanical train, a credit of 10-1 for a automatic steam-driven train, come from -- or at least are checked against -- the results of those studies. DR. FORD: Okay. MR. LEITCH: Now, could you go back to slide 24 for just a second? MR. COE: Sure. MR. LEITCH: I have a question about how you get the "2" in the column there that's labeled "remaining mitigation capability" -- MR. COE: Yes. MR. LEITCH: -- "2 motor-driven emergency feed pump". Where does that "2" come from? I guess my question is, is this all pre-printed on this sheet, or is this the result of this specific -- MR. COE: No. Actually, it's a good question. There is a -- and I apologize; there is a separate table that's in 0609 that defines the credit given to a multi-train system. In fact, it defines a multi-train system. MR. LEITCH: Okay. MR. COE: And the credit that's given to a single train -- or in cases where operator action comes into play, the credit is actually given right here because operator action credit will change from sequence to sequence, you know? So, we don't -- we don't try to define that in a table. We put it right up here. MR. LEITCH: Okay. DR. APOSTOLAKIS: So, Doug, you said earlier -- but let's confirm it once again -- all of these tables are plant-specific? MR. COE: Yes, the tables that I'm representing here are plant-specific. DR. APOSTOLAKIS: And the numbers? MR. COE: Yes. DR. APOSTOLAKIS: The credits? MR. COE: Well, the numbers are -- start out to be generic, such as a credit of two for a single train, and one for an automatic steam-driven train. And the frequency of initiating events started out to be what was represented in the new reg study that research provided. As we've gone through the process of asking licensees for comment, they may have provided us with some additional information upon which we can make a decision that we should alter that initiating event frequency, or that we should alter that mitigating system function, or that we should alter that operator reliability, HEP value. DR. APOSTOLAKIS: Have many licensees actually asked you to make these more plant-specific by submitting such requests? MR. COE: Licensees typically gave us a lot of information that they felt was more accurate for their plant. You know, I think in almost every case, every licensee gave us feedback that they felt was better reflective of their plant. Now, we didn't accept that carte-blanche, obviously. And in fact, there is some advantage to sort of staying with some more generic assumptions as a start. Now remember, I said there was a phase three process too. If our phase two tool is a little bit over-conservative, we're willing to accept that because it's expected that if the phase two results are challenged by the licensee because they have a better analysis, and typically they will, then we'll get into a more detailed level of analysis that would -- would, then, have to account for some of the more specific differences that the licensee was using relative to our assumptions. DR. APOSTOLAKIS: But the determination of the color is not on a generic basis, correct? MR. COE: The determination of the color comes directly from this analysis and these worksheets, based on the plant-specific assumptions. DR. APOSTOLAKIS: No, but I mean you have a matrix somewhere that tells you that a five, right, is a white? That was -- CHAIRMAN SIEBER: Page 23. DR. APOSTOLAKIS: Yeah, yeah. MR. COE: Well, relative to that particular initiating event likelihood. DR. APOSTOLAKIS: Right, but it's -- this is not plant specific. MR. COE: This table right here is not plant-specific, that's correct. DR. APOSTOLAKIS: It is not plant- specific? MR. COE: Yes, that's correct. DR. APOSTOLAKIS: It appears to me it should be; I mean, because a plant-specific nature is already in the -- is it possible that the same number at one plant should be a green and another should be yellow? Does that make sense? MR. COE: It could make sense if the plants' designs for the green plant had more mitigation capability than the one that had the yellow plant -- or, I mean, the yellow finding. DR. APOSTOLAKIS: No, but I said the same number. If you had more mitigation ability, the number would not be the same. MR. COE: Well, if the number. DR. APOSTOLAKIS: You wouldn't get the same number. MR. COE: If the number was the same, the color would be the same. The color is representing the band, an order of magnitude wide, and that doesn't change. That is a -- that is a threshold, a set of thresholds, that is consistent with the PIs, and is essentially fixed. DR. APOSTOLAKIS: But -- DR. SHACK: But this is really sort of giving you an exponent on CDF. So, I mean, it really goes back to 1174. And so, it is the same for all plants. DR. APOSTOLAKIS: No, but 1174 uses a fundamental input, the baseline -- so, no, I'm not saying that it should be. It just occurred to me that the decision on the color is generic, but the input into the matrix is plant-specific. And I'm wondering whether this is consistent -- self-consistent. I mean, but I hadn't thought about it. MR. COE: Well, you raised the point about baseline CDF. And our metric here, remember, is the change in CDF. We're not referencing these colors to any baseline, any particular baseline CDF. They are referenced only to the change -- DR. APOSTOLAKIS: Right. MR. COE: -- delta core damage frequency and delta LERF. DR. APOSTOLAKIS: But even in 1174, when the baseline CDF is greater than 10-4, we drop the delta -- MR. COE: Yes. DR. APOSTOLAKIS: -- by another magnitude. MR. COE: Right. For permanent changes to the plant -- DR. APOSTOLAKIS: Permanent changes. MR. COE: -- that may be appropriate. DR. APOSTOLAKIS: Yes. MR. COE: These are performance deficiencies that have resulted in temporary degradations. DR. APOSTOLAKIS: Now, do you see a time in the future where all this will be computerized? MR. COE: Good question. Maybe. Initially -- DR. APOSTOLAKIS: Maybe you see a time, or maybe there will be? (Laughter.) MR. COE: It is possible. My thoughts are, it is possible that this is an intermediate step along the way to the use of -- the employment of more sophisticated risk tools by field inspectors. The challenge that we face today, and we have faced over the past few years when we've tried to risk-inform our processes, even before ROP, is that inspectors -- we were not able to give inspectors sufficient training to allow them to utilize the computer-based tools effectively that had been developed, all right? We established the SRA Program, the Senior Reactor Analyst Program, in 1995 in order to begin to get at that need. And it took almost two years of training before the SRAs were really, fully qualified. This is a way of accommodating the needs of ROP while, at the same time, in a very complimentary way, giving inspectors a better risk- informed perspective of their particular plant, and of the risk -- of the probabalistic framework that is, in many cases, not something that they had to deal with day-to-day in the past. They deal with a very deterministic framework of compliance-oriented and the -- the decisions as to what was important and what was not were based on their own experience and the various pressures that were brought to bear by their own management, but the licensee, and by the public or outside stakeholders. So, what we've tried to do here, and as we've said repeatedly, is to come up with a more predictable and objective tool. And this is -- the risk metric is the way we've chosen to do that. But the inspectors have a challenge of understanding better the assumptions and the limitations of the risk tools that we employ. And so, this is -- this is the way of accomplishing that. MR. JOHNSON: And I would only add that my -- the way I respond to your question, George, is to say that we think -- we think there is something that is valuable with having inspectors, at this stage, work through these sheets. And in the future, for efficiency purposes or for accuracy purposes, it might make sense to computerize it. But today, we think this -- we get -- we get maximum benefit in terms of enabling inspectors to understand not just what the significance is, but working through why it's significant. DR. APOSTOLAKIS: I fully agree with you that we can view this as a training period where people really understand what PRA is all about. But at the same time, as you know, the Office of Research is putting all the IPEs into a -- so far, they're calling is SPAR? MR. COE: Yes. DR. APOSTOLAKIS: So, after we have a SPAR model for each unit, maybe that would -- and that will not happen tomorrow, so -- MR. JOHNSON: Right. MR. COE: No. In fact, you know, we're struggling -- I know Research is struggling with the level of effort and the amount of resources that they can devote to completing the work on the more sophisticated -- the next revision to those computer- based models. But even once they're completed -- you know, even once they're written, an important aspect of that is to go out and check them against licensee analysis results -- DR. APOSTOLAKIS: Sure. MR. COE: -- and to make visits to the site to ensure that the assumptions that those models have in them are accurate. And then, there's the question of ongoing maintenance of those models, and how much effort we're willing to put into that. And then, there's a whole argument that says, well, maybe the licensees ought to just provide their own models for our use. And there's ongoing discussions at high levels regarding that. So, how it all plays out in the end, I don't know. I hold out that there's a possibility that inspectors will become risk-informed enough to be able to use the tools if they exist, the computer- based tools. But right now, I think the agents -- not only the inspectors, but the management, the decision- makers in the Agency, need to have a process that forces the revelation of these assumptions as they make these decisions so that we can legitimately claim that we have a risk-informed process. Because if all the assumptions are buried into computer models somewhere, and we're making the decisions based on the results coming out of a computer relative to some standard or some threshold, I'm not sure that I can call that risk-informed, okay? MR. LEITCH: I think I may be getting mixed up a little bit between core damage frequency and change in core damage frequency. I guess my question basically on the next slide, 24 I guess it is. On that last example, could there be a scenario where normal operations gets a green? DR. APOSTOLAKIS: Gets what? MR. LEITCH: Gives a green. In other words, you're running along normally with three -- with two -- you just had it there a minute ago. MR. COE: Yes, slide 24, right? MR. LEITCH: Slide 24, yeah. It's unnumbered. DR. APOSTOLAKIS: Slide 24 is up there. MR. COE: Oh, thank you. (Laughter.) MR. LEITCH: Say you had both motor-driven pumps and a turbine-driven pump, and you assume, say, one for feed and bleed. Does that give you a green indicator in normal operations? MR. COE: Well, first of all, you only look at these if they've been changed. So, if a baseline contribution of a particular sequence -- the baseline contribution of all full mitigation capability is potentially white, okay? I don't think that happens very often, but it's possible, right? MR. LEITCH: Yes. MR. COE: Because white represents a single, functional sequence that contributes anywhere from 10-6 to 10-5. You know, and most core -- most plant baseline CDFs are between 10-5 to 10-4. But the point is, is that you only look at this if there has been a change. MR. LEITCH: Okay. MR. COE: Now, the theory -- MR. LEITCH: But if you just -- MR. COE: Philosophically, what happens is, if you're only looking at the sequences that have changed, if we were to look at the core damage frequency with the change, we would add all the baseline sequences, the ones that didn't -- weren't affected. MR. LEITCH: Okay. MR. COE: And then, when we subtract off the baseline, all of those go away. All those sequences go -- the contribution to all those sequences goes away. So, all we're left with is the one that changed. MR. LEITCH: That changed, yes, yes. Okay. MR. COE: So, that's -- theoretically, that's how we can call this delta CDF. DR. APOSTOLAKIS: So, this is really CDF- oriented not LEFT? MR. COE: Well, we haven't talked about LERF yet. But the LERF -- the significance standard for LEFT is essentially one order of magnitude more sensitive than for delta CDF. DR. APOSTOLAKIS: But you do have the tables and everything? MR. COE: Right. We just -- this issue didn't impact that, so we're not talking about that today. DR. APOSTOLAKIS: Yes. MR. COE: Okay, so this is only one of several worksheets. Now, I mentioned that, essentially, the guidance for this example was that all the worksheets had to be looked at, with the exception of a few LOCA worksheets. DR. APOSTOLAKIS: I wonder if you have -- I mean, one of your cornerstones is emergency planning -- MR. COE: Yes. DR. APOSTOLAKIS: -- which is beyond LERF? MR. COE: Yes. DR. APOSTOLAKIS: So, how would you go back? MR. COE: We have a separate significance determination process specifically designed to address findings coming out of emergency preparedness inspections. DR. APOSTOLAKIS: So, you're using level there results? MR. COE: No, it's more -- the logic of that SDP is more related to the nature of the deficiency that caused the problem. DR. APOSTOLAKIS: It's not risk -- CHAIRMAN SIEBER: It's determined risk. MR. COE: Correct. It's really -- you can't really claim it to be risk-informed, although what we've tried to achieve with all of these other cornerstones that you can't link directly to core damage frequency or delta LERF metrics, is a consistent response. The Agency's response is commensurate with the type of deficiency that has occurred. And it is, I think in the formulation of those SDPs, somewhat more subjective. But what we're trying to achieve is the same goal, the same level of consistency. CHAIRMAN SIEBER: Okay. MR. COE: Okay? CHAIRMAN SIEBER: And you have the same situation in physical security, right? MR. COE: Yes, right. In fact, you know, as you're probably aware, we made an attempt early-on to incorporate risk-informed processes -- a risk- informed SDP process into the physical security SDP, particularly -- specifically for assessing the significance of force-on-force exercises. And that proved to be unworkable. And I -- you know, I was involved in trying to make it work, and I, and others, were just simply not successful. You know, there's too many differences from a -- you know, when you're talking about risk in terms of sabotage events and the level of intent that -- you know, and all of the variations that can occur in terms of recoverability of things under fairly stressful conditions. It just wasn't workable, so we're redefining that now. CHAIRMAN SIEBER: Okay. MR. COE: Okay. The other sequences that I thought I would show you -- I haven't gone through all of them here. But the other ones that came out, not white, but rather there was another sequence that came out green, you know, right next to white, was a loss of off-site power sequence. In this case, I just wanted to point out that the loss of off-site power initiating that frequency is in a different row in table one. It's in row two. Exposure time, of course, is the same. It's greater than 30 days. But the result now, if you look on that table, is "B," which represents a less frequent initiating event. Now, that means that you don't have to have quite as many mitigating -- quite as much mitigation capability on the average for that initiating event as you would for the one in the higher frequency category. And that -- that affects, you know, in a probabalistic way, what the outcome of the significance is. So, in this case, we look at EFW again. These have been affected, and the choices that are made here, in terms of remaining mitigation capability, are exactly what we've described before. In fact, even for this particular sequence, the feed and bleed is also the same. And in fact, that sequence, other than the initiating event, is exactly the same as the one that got us the white. Okay, in this case, because the LOOP frequency is less than that transient without loss of PCS frequency, this -- this value of five, instead of getting us a white because we drop down one on the -- on this table here -- we're in the -- the LOOP is a "B" likelihood here. And we come over here to five, and we're green next to white, okay? Now, you probably realize already that a real PRA sums all of these contributions up. And what we're dealing with here is sequence-by-sequence. And we're saying the most dominant sequence, you know, is the one that drives the color. But in fact, we acknowledge and recognize that an accumulation of sequences at lower levels may sum up to something greater than the threshold that we have set for green and white. And the way we accommodate that in a phase two level in the courser treatment that we give a phase two, is to establish a summing rule. And the summing rule says that if you have more than two of these sequences that are green next to white, then you should call that a white, okay? Now, that is a somewhat arbitrary choice, but we thought it was a reasonable one, at least to start with. And that's not to say that if you even had two greens next to white that that wouldn't, or shouldn't, prompt maybe a more thorough analysis, which is, you know, often easy to do with either our own tools or utilizing the licensee's analysis, okay? So, all I'm saying is that we recognize that that's a limitation of this particular phase two level of detail, and we've tried to account for that. And that, if nothing else, gives inspectors -- you know, reminds inspectors that that's really what's going on here, that there's a potential for aggregating, or summing I should -- you know, summing these lower level issues to something that was of greater significance. Okay, that actually completes the documentation that I had to show you and the example, unless there's any other questions about what we've done. MR. JOHNSON: Now, we're at a point in the presentation where we'd like to get into the fire protection SDP if that's possible. Does that fit? CHAIRMAN SIEBER: That fits with me. MR. JOHNSON: Okay. J.S., Mark, guys, would you come up and join us? MR. HYSLOP: Hey, Doug, I think you've got my transparencies. MR. COE: They're up here, yes. MR. JOHNSON: Would you like me to flip for you or -- MR. HYSLOP: Hi, my name J.S. Hyslop, and I was the co-developer of the fire protection SDP, which was developed over a year ago. And Pat Madden and I -- I'm in PRA -- Pat Madden, a fire protection engineer, also developed this. We did it together. And Pat has since moved on. And now, Mark Salley, beside me, is now responsible for the fire portion of the fire SDP. This first slide indicates that I'm going to give an overview in the presentation. Basically, it's just the general remarks that are going to be overview. From that point on, we're going to get into an example with a specific application of the fire SDP on a set of fire protection findings we had. And so, in that -- don't move on yet, Mark, I'm going to talk about the identification of the findings and -- clear identification of the findings. We're going to talk about the fire scenario, and there -- a realistic fire scenario where we, of course, have to take into consideration the figuration of the room as well as the findings themselves. And then, we're going to apply the SDP to estimate a color and talk about the basis for the degradation, as well as the failure probabilities used. Go ahead. I want to make some general remarks to just give you some insight into what we're doing with the process, as well as some information about it. We're using techniques and data generally accepted by the fire risk community. What do I mean by the "techniques"? Well, the technique involved consideration of a fire ignition frequency, the defense in-depth elements, barrier suppression, etcetera, and mitigating systems. We put all those together, using the appropriate probabilities, to get a change in core damage frequency. The date -- DR. APOSTOLAKIS: J.S., most utilities, the way I understand it, use the screening bounding methods like five in their IPEEEs. Would that -- wouldn't that make it very difficult to calculate delta CDF in this context? MR. HYSLOP: Well, what we're doing is we're trying to look at realistic scenarios. So, we're actually using the emission frequency associated with the scenarios. And we have tools to evaluate the damage done by the fire, quantitative tools that we're developing now. And you know, we're trying to estimate the damage -- as a result, we try to estimate the damage as reasonable as possible. DR. APOSTOLAKIS: So, what you're saying is that you are going to be use the IPEEE to some extent? MR. HYSLOP: We're using a lot of information out of the IPEEE, but we reserve the right to develop scenarios ourselves to disagree with those in the IPEEE because, you know, that's what our inspectors do. They go out in the field. They look at the fire sources, and they make independent judgements themselves about the damage done. And I'm going to be talking -- my next point is that it's an evolving process. And I'm going to give you some information that -- about some of the things we're working on to improve that right now. So, it's an evolving process. We just released the second version of the SDP. It's my understanding that that was distributed to you by IPB. And there, we've got some clarifications and just -- on identifying and evaluating realistic fire scenarios, as well as guidance to assist the inspectors to determine the degradation level associated with the weakness of our inspection finding. We also have -- it's an evolving process, so we have future plans. First of all, we have a tool development to assist the fire protection inspectors to evaluate the effectiveness of manual actions specific to fire. You know, you may have manual actions specific to fire because of evacuation of the control room, or you may have manual actions specific to fire because of heavy smoke being in the vicinity. This project was -- we got a lot of help from Research on this. Nathan Siu was using the data available to him through the Fire Research Plan. So, they provided the foundation. NOR has since looked at that and made some modifications to it. You know, the next step, of course, is to document this, and to go around with industry, get their feedback, as well as the other stakeholders, because that's the way the reactor oversight process works. Another tool that we have under development, really by Mark Salley, is the development of a quantitative tool to estimate the fire damage as the result of a fire ignition source. And Plant Systems is working on that now, and developing templates and a guide to use for the inspectors. And so, the next step is, what are we doing with our stakeholders? Are we telling them about this? And yes, we are. There was a fire protection information forum held, I don't know, a while ago I attended. And I talked to them about these -- about these plans that we have for the fire protection SDP. And then, there was the reactor oversight workshop, which was held a month or two ago, or so. And there, we had a fire protection break-out session where Mark and I attended, as well as some other fire protection people. Some SRAs came and, you know, industry came. And we talked about, again, what we're doing. And we've been talking to industry the whole time, in response to your comment earlier. Throughout this development process, early on, before we even used it, we had many meeting with industry: with NEI, small meetings with industry where we had a couple hundred people there -- here. And we ran -- DR. APOSTOLAKIS: What do you call a large meeting then? MR. HYSLOP: Well, I guess I would say -- DR. APOSTOLAKIS: More than a thousand? MR. HYSLOP: No, no, no. There were a hundred people there. There were a hundred people. That's a large meeting for me. I come from a small town. So, CHAIRMAN SIEBER: A thousand would be medium. DR. APOSTOLAKIS: It's a medium. MR. HYSLOP: Anyhow, the last point is, we have a state-of-the-art research plan going on ten floors -- being managed ten floors up, and Nathan Siu. And he's doing work on suppression. He's doing work on fire barriers. And I've told him that I'm interested in the insights that he gains from his program because this is an evolving process. And likewise, I'm interested in your comments. Again, it's an evolving process. Next. So, we're going to get into the example right off the bat. You know, as I said, this is based on fact. We had an inspection, and the inspection identified several findings. The first finding is the suppression system, was a CO2 system. I'm just going to tell you about these briefly. Mark Salley, later in the presentation, is going to get into these in more detail and tell you his basis for our choosing a level of degradation associated with these findings, okay? So, just briefly, the fixed suppression system wouldn't maintain the minimum concentration for the fire hazard. There's a minimum concentration required, and it was lower than that. Also, there was a barrier problem. The electrical raceway fire barrier system protecting redundant trains didn't meet the one-hour rating. It was substantially less. DR. APOSTOLAKIS: So, this was the same plant? MR. HYSLOP: This is the same plant. This is the same room. DR. APOSTOLAKIS: Oh, okay. MR. HYSLOP: Okay? This is one hour. So, you've got -- and I'm going to talk about the configuration, but you've got one room; you've got fire barriers in that room that are degraded. And of course, you know the regulations: when you've got a one-hour barrier, you've got a fixed suppression system also in tandem. And that fixed suppression system responsible for protecting that barrier was also degraded, okay? Now -- CHAIRMAN SIEBER: Let me ask a question. MR. HYSLOP: Yes. CHAIRMAN SIEBER: Could you imagine a case where the lack of functionality of the suppression system would cause the degradation of the fire barrier, and therefore, you get basically two issues out of one defect? MR. HYSLOP: Well, we look at these -- I'll let Mark Salley answer that more fully. But we look at these synergistically in the analysis. We say that these two compound the problem in the analysis. And you'll see later in the slide how we do that. Do you want to respond, Mark? MR. SALLEY: Yes, if I understand your question properly, are you saying the suppression system would degrade the barrier? CHAIRMAN SIEBER: Right, or the lack of functionality of the suppression system. For example, here, was the fact that the fire barrier did not meet the one-hour rating independent of the fact that the suppression didn't maintain the concentration? MR. SALLEY: That's an interesting question, and you take it all the way back to the barrier qualification, in and of itself. If you remember the whole thermal lag and the fire barrier issues, another one that came down the road was kao wool -- CHAIRMAN SIEBER: Right. MR. SALLEY: -- which was a ceramic, fiber-type material. CHAIRMAN SIEBER: Right. MR. SALLEY: And there, the hose drain at the end of the fire exposure would be very important to have got its qualification that the hose stream wouldn't remove it. So, that should have been looked at, at a lower level in designing the system. CHAIRMAN SIEBER: Okay. So, what you're saying is you do look at things in a synergistic basis? MR. HYSLOP: Yes, that's one of the strengths of this method. CHAIRMAN SIEBER: All right. MR. HYSLOP: And then, the last -- the last thing we have to consider is the time of this degradation. If you remember in Doug's presentation, the time affects the change in CDF. A lesser time -- since it's an annualized change in CDF, a lesser time has a less effect than a long time, okay? And we find that these findings existed greater than 30 days, and that's the largest range. There, you have no reduction in CDF for the time, and they existed simultaneously. And that was determined during the inspection. Go ahead. CHAIRMAN SIEBER: Another question: when you talk about the fire barrier, it could they used deficient material, or it could be the fire barrier is defective, like there's a hole in it. In this case, which was it? And in general, do you treat them the same way, either deficient material versus a breach in the system? MR. SALLEY: When you get into the actual evaluation, they would start falling in the same matrix of the degradation -- CHAIRMAN SIEBER: Okay. MR. SALLEY: -- as to how degraded they are. CHAIRMAN SIEBER: All right. DR. APOSTOLAKIS: J.S., I didn't understand this argument about the 30 days. You say that was greater than the maximum, therefore -- MR. HYSLOP: Yes, there are three time ranges in the SDP: zero to three days, three to 30, and greater than 30. And the greater than 30, essentially, you assume you've got 300 or some days' degradation. It's a factor of one that's used in there. So, you don't get a reduction in your core damage frequency if you're greater than 30 days, where you would get a reduction of ten if you're three to 30, and a reduction of 100 if you're zero to three. DR. APOSTOLAKIS: Okay. MR. HYSLOP: Next slide. Now, I was planning to jump right into the phase two, but I'll talk about the phase one a little bit, although I really don't want to spend much time on it because it's not as important for this application. Essentially, we recognize we have significant degradations in defense in-depth. We haven't talked about them, but you'll see that. And this fire barrier and automatic suppression protect essential equipment, equipment that's on those sequences, and loss of would have a big effect. That, alone, will put you into the phase two process, so now I'm going to talk about the phase two. The phase two, one of the earliest things we have to do is ask the following question: can we have a realistic fire scenario? You know, we've got a -- we've got a degradation defense in depth. Do we have a fire scenario that's going to challenge that? And so, when you do that, you know, you have a knowledge of the degradations, and you, of course, need to have an idea of the configuration of the room. And I'm going to talk a little bit about that now. This room was a 4160-vote essential switch gear room, so you had your safety-related switch gear. It was divided into three sections by partial-height marinite walls. These walls went nearly to the ceiling, but not all the way. And so, you've got three sections, okay? So, in each one of those sections, you had an electrical train -- electrical bus of switch gear where you needed two buses to support one mechanical train. That's the way the plant was set up. Now, if you had a fire in one of the far regions, then we still had too much of a train. So, you had a mechanical train of equipment. You really got into trouble if you had a fire in the center one because in the center, you had cables crossing over from each of those electrical trains, over the end of the center switch gear, okay? CHAIRMAN SIEBER: And they went over the wall? MR. HYSLOP: And they went -- yeah -- CHAIRMAN SIEBER: Okay. MR. HYSLOP: -- over the wall; over the end, right. CHAIRMAN SIEBER: Where the plume would be? MR. HYSLOP: Right, right. So, you know, you've got a -- you've got an ignition source over the end. A fire starts there, develops a plume, potentially does damage. Mark is going to talk more about this. So, do you want to go ahead, Mark? MR. SALLEY: Sure, this is a good time to pick it up. I'm Mark Salley from the Plant Systems Branch. Pat Madden originally had started this. I helped him a little bit. And Pat moved on, so I've been picking up a lot of the fire protection with J.S. from here on out. George made an important comment earlier about how this comes together. If you look back, the IPEEEs, Generic Letter 8820, Supplement 4, there is a starting point, especially for the people who used the five method. And they said, "Hey, look, we've done a lot of work with Appendix R. So, from that Appendix R starting point, we'll take this snapshot in time, and we'll do this IPEEE." From that IPEEE, the next progression is where J.S. and I are pretty much going. So, I think you can see, as we're moving along, that one bit of information is building on the previous one. To just giver a little summary here of what J.S. is talking about, we have our three vital switch gear, 4160, the vital buses, the three fire barriers -- DR. APOSTOLAKIS: Is there a reason why we don't have this? MR. SALLEY: Oh, I'm sorry. This was just an extra. I thought I'd give you -- MR. HYSLOP: We just made this one. MR. SALLEY: -- a real quick -- a little more clarity. DR. APOSTOLAKIS: How about a picture being worth a thousand words and all that? CHAIRMAN SIEBER: We'll pick up a copy. We'll get you a copy. He can make it available to all of you. MR. SALLEY: The fire barrier separator of the marinite walls that J.S. spoke about, the area of concern is where the cables from the three merged over the center unit here, okay? Now, in the Appendix R-type strategy for compliance, the requirement would say, okay, there's a number of ways to do this. This licensee chose to put one-hour fire-wrap, fire barriers, on those cables. And the room is -- has a full, automatic suppression system; in this case, a manual CO2 system. So, that was his method of compliance. As the inspectors looked at it -- CHAIRMAN SIEBER: I'm not sure how an automatic suppression system is a manual CO2 system. MR. HYSLOP: We're getting -- CHAIRMAN SIEBER: It sounds like it's manual. MR. HYSLOP: We're going to get into that on the next slide. MR. SALLEY: Yes, this licensee did have a manual here -- MR. HYSLOP: Sorry about that. MR. SALLEY: -- yeah, with this; you're correct. CHAIRMAN SIEBER: But these are original design problems with the construction of this room, right? MR. SALLEY: Right. This is unique to this licensee and -- CHAIRMAN SIEBER: So, this has existed forever? MR. SALLEY: Yes. CHAIRMAN SIEBER: Okay. MR. SALLEY: When the inspectors were looking during their inspection, they found -- they inspected the hardware in the plant. They, first off, review the fire barriers. In reviewing the fire barriers, what they determined was that they really weren't one-hour rated barriers as they -- CHAIRMAN SIEBER: Well, the walls weren't because they weren't full height. MR. SALLEY: Well, the -- CHAIRMAN SIEBER: And the wrap probably had some other defect. MR. SALLEY: The wrap is the concern here. DR. APOSTOLAKIS: Wait a minute; what wrap are we talking about? CHAIRMAN SIEBER: The way the cables are wrapped. DR. APOSTOLAKIS: Oh, the cables. So, the wall -- the barrier goes, what, not all the way to the ceiling, right? MR. SALLEY: Right. MR. HYSLOP: Not quite. DR. APOSTOLAKIS: So, what, a couple of feet or -- CHAIRMAN SIEBER: So, it's really not a barrier. DR. APOSTOLAKIS: What? CHAIRMAN SIEBER: It's really not a barrier, the way that drawing shows it. DR. APOSTOLAKIS: Well, I mean, for some events, it is. MR. HYSLOP: My understanding was it was quite higher than the switch gear, and it was a couple of feet from the ceiling. MR. SALLEY: This, of course, was probably back-fit to Appendix R, and it was a unique consideration where they put the non-combustible marinite in to try to get some compartmentation between the three pieces of equipment from their original design. DR. APOSTOLAKIS: So, what did you draw there now? MR. SALLEY: The area of concern is where the cables -- CHAIRMAN SIEBER: Is the middle. DR. APOSTOLAKIS: Right. MR. SALLEY: -- from the three units came together at a common point. Now, the licensee's strategy for compliance would be, okay, from where we've passed into this area, we need to install one- hour fire wrap on those cables, so they can survive a fire in this center unit. The inspector is looking -- MS. WESTON: I think we need to say this was probably -- was this an exemption -- DR. APOSTOLAKIS: No, you have to come to the microphone. MS. WESTON: I think -- DR. APOSTOLAKIS: Speak with sufficient clarity and volume. MS. WESTON: Your name? MR. WHITNEY: Speaking of clarity, was let's explain whether or not this was an approved exemption or not, that this doesn't meet the letter of Appendix R or does. Can you explain that, please? MR. HYSLOP: That was Leon Whitney. You've got to use yoru name when you -- MR. WHITNEY: Leon Whitney, Inspection -- MR. SALLEY: This is an actual configuration, as I said earlier, for a plant to do their Appendix R compliance come up with a strategy, or this is an exemption for the barriers. For example, in Generic Letter 8610, we provide a guidance which the licensee would use here. The requirement was still to have this one-hour fire wrap in this area, which the licensee claimed they had. As the inspectors looked into the detailed testing of the fire barriers, they determined that it really wasn't one-hour. In reality, it was probably 10 or 15 minutes of fire endurance from this barrier. So, that would get them -- to enter the SDP, there's a design requirement. They don't meet that design requirement, and that would be the start of this. In addition to that, they looked at the CO2 system -- DR. APOSTOLAKIS: Which plant is this? CHAIRMAN SIEBER: If it hasn't been issued yet, we shouldn't -- MR. HYSLOP: It was over -- it's over a year ago. I guess it is; I don't know. CHAIRMAN SIEBER: Let me point out that if the inspection report hasn't been issued, then we should not use the name here on the record, okay? MR. HYSLOP: It should be. I don't know. I don't keep up with that. CHAIRMAN SIEBER So, if you don't know for sure, don't tell us. MR. HYSLOP: We don't know. MR. SALLEY: We don't know for sure, but it's a real plant and this is a real case. MR. HYSLOP: It's old. CHAIRMAN SIEBER: Let's move on. DR. APOSTOLAKIS: Was this identified as a critical area? MR. HYSLOP: I can't remember. You know, I can't remember, to answer your question. DR. APOSTOLAKIS: Well, that's a good question, I think, to investigate because that would be a good test -- CHAIRMAN SIEBER: Well -- DR. APOSTOLAKIS: -- of the IPEEE. CHAIRMAN SIEBER: Yes. On the other hand, it depends on what the deficiencies are in the wrap. Was the material bad? Was the installation bad, or was there not enough of it? DR. APOSTOLAKIS: No, but the IPEEE, we don't look at deficiencies. I mean, this is a critical area because all the cables come together. CHAIRMAN SIEBER: On the other hand, if you met the regulations, then five would give you an answer, okay? DR. APOSTOLAKIS: Well, then, I'm curious to know what five is. CHAIRMAN SIEBER: Okay, right, and you're degraded from five's answer at this point. MR. SALLEY: Right, that's an important point, George, because five had screening tools with it. And one of the criteria, for example, was if you have a suppression system and you meet the NFPA standard, then you take credit for it. As the inspection here looks deeper into it and reviews that suppression system design, they say, "Hey, wait a minute; for a licensing basis, you don't meet your suppression system requirements." CHAIRMAN SIEBER: Right. MR. SALLEY: So, that could actually feed back into the five analysis. But the five was a snapshot in time. John? MR. HANNON: Mark, this is John Hannon. I would think it would be also important to recognize that part of the inspection program itself has us looking into the areas of most risk significance. And we would draw from the five analysis if that was what had supported the IPEEE. So, that would have been an initiating cause to get us to look at this room in the first place. MR. HYSLOP: That's a good point, you know? That's one of the things that they do, yeah. MR. SALLEY: Getting back to our scenario, we have the deficiency in the fire barrier, the cable wrap. Looking further into the suppression system, for the mechanics of the CO2 system to extinguish a fire -- now, the hazard here would be the cables. The cables would be a deep-seeded fire. If we took the minimum NFPA 12 for the CO2 system design, it would tell us that you need a 50 percent concentration, and you need to hold that for 20 minutes with a deep-seeded fire -- basically, by suffocation, removing the oxygen leg of the fire triangle. As they looked into the testing of the system, what they found was the -- I'm jumping ahead here -- what they found was the concentration was a little less. If everybody has a visual, I'd like to get back to the -- that was an extra that I shouldn't have brought, George. MS. WESTON: Now, let me copy it. CHAIRMAN SIEBER: Well, somebody -- you need to be here. MR. SALLEY: I just thought if we couldn't get a good visual -- DR. APOSTOLAKIS: You've been trained well in -- CHAIRMAN SIEBER: Yes, I have. DR. APOSTOLAKIS: -- these proceedings. (Laughter.) MR. SALLEY: Okay, there's one error in the slides. We have one double-printed. This example phase two will come later. So, please pass over that. MR. HYSLOP: So, skip the one slide and move to this one, please. DR. APOSTOLAKIS: Degradations. MR. SALLEY: Degradations. Now, the first degradation is the suppression system. Now, the comment was made about this area does have a deviation. And yes, for a manual actuation, it wouldn't be equivalent to an automatic actuation. So, you would begin the degradation right there that well, hey, this is going to require some human to find the release box, release the system in the event that they do have a fire. Looking at the system further, the CO2 concentration -- like I said, the minimum would have been for 50 percent, held 20 minutes, to extinguish the design basis fire, which would be a deep-seeded fire. They didn't meet that. They had a 46 percent concentration. And the third thing we discussed was the degradation to the fire barrier. How bad are we degraded? This is a good question that the inspectors get into routinely. For example, if I had a 49 percent concentration for 20 minutes, and the Code said the minimum was 50, you know, we start splitting hairs for the one percent of CO2. Then, you can get into things like well, gee, where are the cable trays? Are they in the top of the room where the CO2 is heavier and it's going to be lower? And we can get into a lot of technical arguments through the SDP as we move on. But the fire barrier, being approximately ten minutes, where we originally had required an hour, is a pretty good degradation. That's definitely a moderate to high degradation for the fire barrier. CHAIRMAN SIEBER: Was that due to damage or design? MR. SALLEY: I believe design in this case. CHAIRMAN SIEBER: All right. MR. SALLEY: And you can see the test that they had indicated the barrier's rating was 10 to 15 minutes. DR. APOSTOLAKIS: I guess I don't understand it. What tests are these? I mean, tests that had been done in the past? MR. SALLEY: Yes. The original tests -- DR. APOSTOLAKIS: And the licensee had access to them, and they misinterpreted them or what? MR. SALLEY: This all ties back to the whole '90/'92 era of Thermalag as to just what is a rated electrical raceway fire barrier system. This isn't Thermalag; this is a different vendor. So, we're seeing that same experience with different vendors going back and looking at the original qualification testing. DR. APOSTOLAKIS: Was the licensee aware of this fact, that, you know, based on the tests, the rating was about 15 minutes? MR. SALLEY: I believe the inspection, in this case -- DR. APOSTOLAKIS: Does that come back to what Doug was saying earlier about willful -- MR. SALLEY: I wouldn't say it's willful. I would say, how hard do you -- do you look? I mean, we operated a lot of years under the Thermalag where we were under the impression that it was good until we started really looking. You know, just what did you test this to? And just what was your configuration like in the test compared to the plan? We got into all the details and the rigor of the engineering -- CHAIRMAN SIEBER: But I think this one is different than that. I think the Thermalag was difficult to interpret the test results. And in fact, I think there was a finding that some of those test results were not accurate. On the other hand, when somebody designs a barrier system, you use the test results from a test of the material and then calculate how much of it you need based on the conditions you have in the room. So, it, more than likely, is an error in the application of the specific material to the configuration, as opposed to misinterpreting the test for a false statement, so to speak. MR. SALLEY: We've -- we've -- CHAIRMAN SIEBER: That's the way I would interpret this. MR. SALLEY: Yes, we've got great understandings into just how electrical raceway fire barriers work. And that's a whole discussion -- CHAIRMAN SIEBER: That's right. MR. SALLEY: -- that we've had numerous times about the physics behind the barriers. But this is all of that. The -- DR. SHACK: This is a latent design error, is our best guess? MR. SALLEY: That's an excellent way -- CHAIRMAN SIEBER: That's a good way to put it. MR. SALLEY: -- excellent way to capture it. The third thing, and getting back to the defense in-depth of this, is the fire brigade. On this site, they had a very good fire brigade. So, we would expect the fire brigade to perform well within their means. DR. SHACK: Just, again, what are the questions the inspector asks himself to decide that this is moderate degradation for the auto suppression system and moderate to high degradation? How does he pick those values? MR. SALLEY: That's a very good question. In the guidance that we provide in the Appendix, there's numerous suppression systems. Not all suppression systems are going to be created equal. Let me just give you some examples here. If you're dealing with yoru gaseous systems, your C2 and your halon systems, those are suppression systems. That means that when they go off, they will put the fire out. They are a pass/fail type of thing. It's the little-bit-pregnant argument. I mean, the system works or it doesn't. A suppression system, a sprinkler system, by its design, its original intent was to control the fire. You know, it could limit it into an area until manual suppression could come in and extinguish it. So, you have those two schools of thought in the fire suppression system design. Now, you get into degradation. Let's take a sprinkler system. Say a head had to be 12 inches from the ceiling, and, for some reason, they installed them 15, 18 inches below. Are they Code compliant? No. Is it a degradation? Well, yes. Why is it a degradation? Because the fire is going to have to get a little bigger and a little hotter for the heat layer to build on to actuate that sprinkler system. Will the system go off? Well, it will eventually go off. You may have a little more fire damage, but it should be creditable. With a gaseous system, it's not quite that easy. In this case, the numbers are very close, so we'd want to call that a moderate degradation. Let's take that same CO2 system and say the inspector found a check valve that was installed backwards. Now, the system will get called upon, and no agent would come on. So now, you definitely have a high degradation. Say he looked at the CO2 refrigeration system and the tanks were empty; it's clearly a high degradation. So, there is judgement calls. There is engineering experience by the inspector as to what category to pick. And usually, there's discussions about that. I'll give you an example of one I had -- a halon system in the past where it didn't make concentration. The original design was for a surface fire like you'd find in a flammable liquid. You know, the argument the licensee put up was, "Well, hey, we designed for a surface fire. We really didn't anticipate a deep-seeded fire." The only problem was the fire hazard was the cable spreading room where all the fires were deep-seeded and there was no surface fire to think of. So, that's the kind of dialogue you'll exchange with the licensee to get your degradations. CHAIRMAN SIEBER: Now, generally speaking, during the construction of the plant, or sometimes during hot functionals, all of these systems are tested, and the gaseous systems are tested, for concentration. Is that not the fact? MR. SALLEY: That's an interesting point. Yes, they are tested. And sometimes, we are finding systems, when we go back, and the inspectors are very -- doing a very rigorous, thorough look at is just how did your concentrations look and pulling the old strip charges from the original design. And we're fine, and some of it just quite didn't make it, and maybe someone justified off, and that's under question now. CHAIRMAN SIEBER: Okay MR. SALLEY: For example, if you missed by a little bit, they said, "Oh, my problem was from leaks over here, and I sealed those leaks. And I know that" -- CHAIRMAN SIEBER: Or my calibration was bad, and it deserves a correction. MR. SALLEY: Right, and so those are debatable things that still occur today, and they happen routinely. CHAIRMAN SIEBER: But the reverse check valve would have been found there, the fact that you may not have enough suppressing agent in a tank that would cause your system concentration not to be appropriate. MR. SALLEY: Right. CHAIRMAN SIEBER: Okay, thank you. MR. SALLEY: There has also bee some work -- in 1986, we had a big study with Sandia on just how much agent does it take to extinguish a fire? You know, the National Fire Codes look at a broad band. And deep-seeded to them is a cable fire in a nuclear power plant; it's also a bale of cotton in some other applications. They're all deep-seeded fires by their definition. We tried to refine it more to our hazards, which were the cables. So, yes, we know the numbers can be a little lower, and we have that guidance available. CHAIRMAN SIEBER: Okay, thank you. MR. SALLEY: So, that's the three key points here of the defense in-depth in this specific scenario. Knowing that and having assigned a rating factor with that, we get back to the analysis portion, which is where J.S. will pick it up, to how this all comes together now to define some level of risk. J.S.? MR. HYSLOP: Yes. I just wanted to say the documentation on those degradation levels are in the public domain now. So, you can access that and look at them for more explanation. As I said before, in a fire risk analysis, you're looking at the frequency of the fire, your defense in-depth elements, and your mitigating systems. This first term, FMF, fire mitigation frequency, really just deals with the frequency of the fire and the defense in-depth that's left. Of course, a fire which -- where the suppression system fail, where your barriers should fail if challenged, you know, these are fires that we're really worried about. So, that's why we developed the FMF. Now, the ignition frequency of the 4160 vital switch gear cabinet, we said that was the cabinet in the center bay. So, it was an ignition frequency associated with that cabinet that we're concerned with for this analysis. And the IPEEE had provided that. I'll give you numbers on the next slide; I just want to talk generally right now. The next terms, the automatic suppression and manual suppression -- really, we had a manual fixed suppression system here. So, "AS" was really a manual suppression. We just didn't think about that when we were writing the guidance. But we take into account that it's manual in the degradation rating, as Mark said. Manual suppression, that's typically the fire brigade and any type of early response that people -- that operators would have to put it out. DR. APOSTOLAKIS: But how -- I mean, these things are not really modeled in the fire PRA. So, I don't know how you can -- MR. HYSLOP: I'd like to get to that on my next slide. DR. APOSTOLAKIS: Okay. MR. HYSLOP: I'm going to talk about that. Let me just talk about it generally, George, and then we -- DR. APOSTOLAKIS: Okay, no, no, that's fine. MR. HYSLOP: -- can get into the details. DR. APOSTOLAKIS: That's fine. MR. HYSLOP: And so, for the suppression system, the manual suppression -- the manually operated, fixed suppression system, which is "AS" and the fire barrier, we had degradations. And we're going to use those numbers in this equation. Now, the fire brigade, everything was -- everything was great there. And so, we didn't have any degradations, so we'll use a lesser failure probability associated with it. And we have this term, "CC". It's really kind of like a common cause dependency term. There, we recognize that, for some cases, if you have a sprinkler system and you have a fire brigade, those common delivery systems can introduce common cause failures; fire -- your fire water pumps, you know, it's a pressure for each one of those. So, we recognize that there is an additional failure mode in there, and we've taken it into account. For this particular case, it was a gaseous system, so it wasn't an issue. Now, I've got the numbers on this page, and then I'll explain them to you on the next page. All I want to say is these are the numbers that we attribute for the various degradations. And the fire mitigation frequency essentially says we have a factor, a 10-5, leading into the mitigating systems. So, you know, we don't have to have a lot of mitigating systems to derive us to a green here. If we have none, then we're in white territory, okay? Let's move to the next slide. And what I want to say is, these numbers really are coined as exponents of ten. Remember Doug had the 1, 2, 3, or 4, or whatever; well, you know, these are exponents of 10. So, "-3" is 10-3. DR. APOSTOLAKIS: You have not included transient fuels, have you? This is just -- MR. HYSLOP: There weren't -- to my knowledge, there weren't any transients found during the inspection. DR. APOSTOLAKIS: But if you want to have a frequency of fire -- MR. HYSLOP: You're talking about having a probability of transient -- DR. APOSTOLAKIS: Yeah. MR. HYSLOP: -- fuels, even though -- we haven't included that That's something that we're going to include in the next version. We're going to be providing, in the next version -- this is another thing of the evolution -- a whole set of ignition frequencies for inspectors to use when the plant doesn't have them because some IPEEEs didn't go to this level of detail. They said, "We've got a room. We've got suppression, and we've got some severity factors." So, they never got into this. DR. APOSTOLAKIS: Right. MR. HYSLOP: So, that's going to be in the next stage of this tool. DR. APOSTOLAKIS: Okay. DR. FORD: So, where do these numbers come from? MR. HYSLOP: Turn to the next slide, and I'll tell you. DR. FORD: Oh, okay, you will tell us now, okay. MR. HYSLOP: okay, this table provides the origin of these numbers. The top column -- the top row of this table identifies the defense in-depth elements. And I checked that we just had a one-hour barrier, an automatic suppression -- or really, a manually initiated one -- and a fire brigade for this analysis. DR. APOSTOLAKIS: So, this is not just for this incident? MR. HYSLOP: No. DR. APOSTOLAKIS: This is something from a document? MR. HYSLOP: This is -- this is generic. And I'll talk you about, you know, the source of these, George, and how -- but let me get there. And so, the first column talks about the level of degradation, and we have three levels of degradation in this technique -- you know, you might say two levels, the moderate and the high. The normal operating status when we -- when rate something, we find it meets Code typically. But we still have some sort of failure probability associated with those. So, if we start talking about these, you know the first question is, where did these numbers come from? Is there any reference for these numbers? Did I have to develop them? You know, what's the answer? And if we start for the three-hour barrier for the normal operating state, we had -- new Reg 1150 developed these sandia during their preparation for the study, I guess. And in this particular study, these said that a wall, three-hour barrier, fire-rated three-hour barrier, had a one in a 1,000 chance of failing. That was the base probability associated with it. Now, if you had additional -- if you had dampers or doors in that wall, they collected data to support the unavailability of the door -- the doors in that wall. And that's what would drive the failure probability for the normal operating state. CHAIRMAN SIEBER: Does that mean the door is sometimes left open, or blocked, or does it mean the door is really a three-hour barrier. MR. HYSLOP: It means that if the door is left open or blocked. CHAIRMAN SIEBER: All right. MR. HYSLOP: That was -- that's my understanding. Let me tell you that there wasn't a lot of documentation in the new regs on these. And I told you that we we're working with the Office of Research. We've asked the Office of Research if they have any more insight to give us on these failure probabilities in this table, we're very interested. It's evolving and it's a state-of-the-art process. CHAIRMAN SIEBER: Now, the plants usually keep track of missing fire barriers and blocked doors and things like that as part of their fire protection monitoring system. So, there's a source of plant- specific data for that that could be used, I guess, if the licensee wanted to contest what you were doing? MR. HYSLOP: Yes, the -- as Doug said, you know, we have a phase three process. I'm talking about the phase two. The licensee in any -- in any and all of this study has the opportunity to present additional information to refine the results. Here, we've tried to provide generic data so that we can get through the process. CHAIRMAN SIEBER: Okay. MR. HYSLOP: Now, if you go to a three- hour barrier, one that has a high degradation, they are -- the zero means that we're not giving any credit for it. And the plant system documentation would support minimal credit for this particular high-level degradation of a three-hour barrier. DR. APOSTOLAKIS: So, the inspector is -- I think Bill asked that question -- is provided with information or guidance, how to declare something as moderate or high? MR. HYSLOP: Yes. MR. SALLEY: Yes, if I could jump in, J.S.? We have another example of a case we're working right now. In an area, it was required, for their original Appendix R compliance, to have a three-hour box built around a number of cables that had penetrated into an area. They didn't need -- they needed to rely on this A-train, we'll call it, inside the box for a fire in the B-train area. So, by regulation, it was always required to have a three-hour enclosure around it. And the licensee was moving along, thinking it was pretty good. The inspector went out; the inspector was looking at it and said, "That box up there," -- they said "Yeah, three-hour barrier for Appendix R." He said, "Great, can I see the test reports for it and the design basis?" The said, "Sure." So, they started digging through it. When they got into it deep, they really didn't have a test. It sounded like a good idea to take these non-combustible boards and assemble them here. And they've existed that way since the mid-80's and have taken the three-hour credit for it. Now, we got into a discussion with them, and what kind of a credit could we assign to this? Well, we have no testing. We just know that it was a box -- CHAIRMAN SIEBER: It's zero. MR. SALLEY: Right. Will the bolts fail, and the box fall off, even if it's non-combustible? They've covered it with phlomastic, which is a limited combustible. So, to enter into this, it entered in as a high degradation, zero. The licensee then, because we got the zero, started working through it, built a mock-up of this at a laboratory and tested it, and found that it got approximately one hour. So, we, in the analysis, further refining it, went from the high degradation to, here, a moderate because they did have some creditability to that box after having tested it. Once again, it was a good inspection finding to go and look at that. CHAIRMAN SIEBER: So, they had to conduct a special test to even come up with that? MR. SALLEY: Yes. CHAIRMAN SIEBER: Okay. Now, let me ask another question. And again, referring to "door" on there, do I interpret that to mean that any three-hour door in the plant that's expected to be open for 30 days a year? MR. SALLEY: The door thing, I just want to -- I understood a little different J.S., if I could expound upon that. We give you three levels there. We give you a -2, a -2.5 and a -3. CHAIRMAN SIEBER: Okay. MR. SALLEY: Not all three-hour fire barriers are the same. For example, if I wanted a perfect fire wall, I'd have 12 inches of concrete poured, solid pour, no penetrations, no doors, no nothing. I'd have a lot of confidence, and history has proven that, that that's a pretty good three-hour fire wall. However, in a power plant, if I introduce a door, well the door doesn't test the same as a fire wall. The door criteria is much more lax, just by the nature of the door. I mean, you have gaps. If you have some flaming remote on the other side, it won't perform as good as the wall, but it's still a recent -- CHAIRMAN SIEBER: It's still a three-hour barrier. MR. SALLEY: -- it's still a three-hour barrier. You need to have that in there. If this wall had numerous penetrations, I'd need to look at those penetrations? And do I have tests? Do I have designs? Do I have a comfortable feeling with all the penetrations? So, that -2 to -3 gives the inspector some room to customize it for his application in determining the -- CHAIRMAN SIEBER: It sounds to me a little subjective. MR. SALLEY: Engineering judgement. (Laughter.) CHAIRMAN SIEBER: That's another way to phrase it. Let's move on. MR. HYSLOP: Okay, so basically, the moderate degradation is a twist of a value between the high and the normal operating state. And we all -- to get to the values, you know, we used, we looked at the one-hour barrier, the fixed suppression system and fire brigade. The one-hour barrier in the normal operating state was taken to be approximately equal to a moderate degradation of a three-hour, in that, you know, a moderate degradation of a three-hour is somewhere between two and one hours. And so, that's what we expect for a normal operating state for a one- hour. And then, the logic is similar for the moderate and high degradations of the one-hour, as was for the three-hour, the basis for the choices. Now, if we talk about the fixed suppression system, there, the normal operating state of that is also taken from many studies. I know it's in five, and I know it's in the PAR Implementation Guide, the basis for this number. So, there, that's judged as a normal operating state. And again, for an automatic suppression with -- where we have some degradation which drives us to conclude that there's minimal credit, we give it zero. Now, the last one we talk about is the fire brigade. And really, it's a manual suppression. That's really what that is, is a manual suppression there. Let's be quite frank, because if you look at the fire brigade, you notice, for a high degradation, we give credit. And that's because there are fire watches, there are operators going around a plant, and there's data found in the PAR Implementation Guide that supports that these people do put out some fires before they get bad. So, we have some credit there. The -1 there is -- it's often scenario- important. But for cases where the IPEEEs looked at lots of fire sources creating severe fires, .1 was typically used in those analyses to support that. And that was really the origin of the number here. Let me see if I have any other comments. So essentially, you know, I guess to sum up, some of these normal operating states are supported by industry, or NRC, or both, and guides. And the other values were kind of deduced from common sense, good judgement. Go ahead. So, now, I'm going to move into the reactor safety portion of this because we've identified the fire mitigation frequency. This is the -- these are the fires where -- which have the opportunity to get big. Our suppression system hasn't worked, and so we have some elements of our defense in-depth that are going to fail to control this fire. So, what I'm going to -- let's move to the next slide. DR. FORD: Excuse me. The only data in this whole page 37 is this Sandia new Reg 1150, the previous one. The only hard data that you have -- MR. HYSLOP: Well, we have -- DR. FORD: -- 10-3 is -- MR. HYSLOP: No. Well, that was adopted by industry also. So, that's a number generally accepted in the PAR community. I think that was derived in the 1150 studies. I don't know if industry did any additional work before accepting that. This is one of the things I've identified to Nathan Siu, of the fire research plan, that we're interested in having additional information on because, you know, we recognize that this is one of the -- one of the area that -- you know, there's a limited information on fire on which to make our judgement. DR. FORD: But that's my point; the only referenceable data is that 10-3? MR. HYSLOP: Well, there -- no, there -- no, that's referenceable also in either the five or the PAR Implementation Guide, both of which are industry documents. DR. FORD: Okay. CHAIRMAN SIEBER: I would like to suggest that we're going through the basic principles right now of how this worked, but you do have a specific example. MR. HYSLOP: Okay. CHAIRMAN SIEBER: And maybe we can do that after lunch so that we can get through with the general explanation and let us know. MR. HYSLOP: Actually, we're doing the example, but we're almost finished with it. So, I think -- CHAIRMAN SIEBER: Well, it looks like a lot of sheets. MR. HYSLOP: Well, that's okay. I can do those in five -- in five minutes, and that's what I intend to do. I'd like to -- if you don't -- whatever you want to do. (Laughter.) CHAIRMAN SIEBER: I think this is a good place, then, to stop before we get into all this detail here. MR. HYSLOP: Okay. CHAIRMAN SIEBER: And even though it's extremely interesting -- MR. HYSLOP: Okay. CHAIRMAN SIEBER: And why don't we recess for lunch and come back at one o'clock? And we'll finish this up then. (Whereupon, the proceedings went off the record at 12:08 p.m. and resumed at 1:02 p.m.) A-F-T-E-R-N-O-O-N S-E-S-S-I-O-N (1:02 p.m.) CHAIRMAN SIEBER: We'll come back to order and continue with Fire Protection SDP. MR. HYSLOP: What we've done, just to remind you, was to calculate a fire mitigation frequency which was the frequency of the fires for concern were those that aren't extinguished or controlled by suppression and those which challenge our barrier. What I'm going to do is move on to an evaluation which involves the reactor safety worksheets. Doug Coe, in his presentation, talked to you about an application of those sheets and I'm going to talk to you about a different one. Let's move to the next slide. [Slide change.] MR. HYSLOP: The next slide is the worksheet for a small LOCA and the reason we're using the small LOCA is because it's fire induced. As you recall, we lost all the electrical trains. Losing those electrical safety trains means that we lose our component cooling water and our charging system and losing both of those induces a small LOCA or RCP-Seal LOCA. And this is consistent with the assumptions used in the reactor safety process. So now we have to say how significant is that fire induced small LOCA? And if you look at the sequences which are here on the left most column, you'll see that first of all one sequence which leads to core damage is a small LOCA and the loss of all high pressure injection. In this scenario, we're looking at the charging pumps and I think the SI trains injected a slightly lower pressure, but you could have some depressurization, therefore some mitigating capability if they were available. But upon losing all these electrical trains and mechanical trains you, in essence, lose your high pressure capability. So the reason I said we could get through this quickly is because we really give no credit for the mitigating capability in this particular scenario. So our fire mitigation frequency which had no reduction because of the length of time that these degradations existed, essentially serve to characterize the increase in core damage frequency fully. And for this example, we get a white and you would go through the same tables as Doug did. I've just short-cutted it. [Slide change.] MR. HYSLOP: So if you go back to that earlier slide it says that the resulting evaluation is white. Now what would happen if we had felt that that fixed suppression system wasn't worthy of any credit at all? If you remember, it was immoderate, based on the observations that the inspectors made. Then it would be in a yellow territory. As you've talked before, the yellow provides a different response than the action matrix as does the white. So we would have geared up a little more for this one. If we repaired the fire barrier, for instance, then we would have been pretty much at a green/white threshold in that case and depending on exactly where we were, we would have gone -- we may have gone with a white for that because this is a conservative approach and then allowed the licensee to come in with a refine analysis to support his work. So that's it for my presentation and Mark's presentation. MR. COE: If I may add one thing, it's important to make the point that the SDP process has not removed the requirement for the staff to make judgment and as you've seen here with the fire protection as well as the earlier presentation that I did, the judgments are now more rigorous, more disciplined by this framework that we've chosen to use, but in essence, there are still judgments and they occur at the assumption level or at the basic input level for these SDPs and the logic that then processes those assumptions to a final result is clear and is apparent to all of our stakeholders and is then the subject of dialogue and discussion. So I do want to make the point that we have not extracted judgment from this process. MR. SHACK: What's the feedback you get from the inspection people about whether they feel they can make these judgments? MR. SALLEY: Can I take that one? In the fire sense, let me pick that up and explain one other thing that's kind of important if we go back to our example. Now remember, this process is new. It's evolving. J.S. told you that. We're getting better. We're refining, we're doing, we're learning. If you think back to your question in this case here with judgment and such, there's one question that we just kind of glossed over and it was done this way in the actual example because it was so early on and it was looked at and that is what's the fire potential, okay? If I could argue from a licensee's standpoint and say well, okay, yeah, it's a 10 minute fire barrier, but when I go through the dynamics of combustion here, I get a 6 minute fire, worse case. So that gives us room to argue and move around within the evaluation. One of the things that we're currently moving on and this is the way we're seeing them now is what's the realistic fire threat. As the newer SDPs are coming in the findings, that seems to be one of the up front questions. Could I have had, do I have the chemistry there to give me the credible fire to challenge these degraded barriers' suppression systems. And that's where we're moving with the effort now. With the inspectors, one of the things that is -- if I for example say a gallon of combustible liquid, in each one of our minds we picture the fire that could be. It could be in a kerosene lamp and you've got a hurricane lamp or you could spill it all at once and get a big burn. How do we make those judgments and that's what we're working on. We have a quarterly workshop with the inspectors to review the cases and J.S. goes through the cases that we've been through in the last quarter and we're starting to introduce some of these new tools and methods on how to do the fire scenario development. That's what the process is currently today. That's what we're working on and going through the development methods is where we start all getting the same judgment and the same experiences, learning from the different ones. CHAIRMAN SIEBER: Now that may change the significance of a given set of circumstances. It doesn't change the fact that you would still be in violation of Appendix R which is deterministic to the violation whether it's cited or noncited or whatever color it is, it still exists. MR. HYSLOP: And as we've said, any and all of those still go into the corrective action program. They need to be fixed. CHAIRMAN SIEBER: On the other hand, it seems to me that the development of risk-based fire analysis is not too far along. If I look at NFPA 805, it discusses that to a great extent, but it seems to me that that is in addition to the deterministic requirements of Appendix R or branch technical position 9.5.1 or the guidelines or whatever class of plant that you're into. And until such time as you risk-inform Appendix R, if you ever do it, where it tells you, you don't need a one hour, three hour fire barrier, you need a 20 minute or a 60 minute or a 90 minute fire barrier based on fire scenarios and risk probabilities, it sort of puts us into an enforcement juxtaposition into what the regulations tell us to do, it seems to me. Can you comment on that at all? MR. SALLEY: The risk-informed performance based approach, I see the SDP portion of this is moving in the right direction and being fairly valuable. For example, in the past, if you were to just find the CO2 system, didn't meet its design concentrations and the fire barriers didn't either -- CHAIRMAN SIEBER: It would be a Level 4. MR. SALLEY: Right. At some point in there someone would say well, how bad was it and some engineer would walk out there and say well, you know, we've got this switch gear and from what I've seen a switch gear fire -- and it would be an opinion. A pure opinion. It's going to be real bad or it's not. Here, we at least are starting to framework and say okay, from a risk standpoint. How bad would it have been? What would the possible outcomes be and we have a nice structured framework to make a better determination. So I see it as a real improvement there. CHAIRMAN SIEBER: Now the example you describe here is a Phase 2 analysis under the SDP. What circumstances would cause you to do a more rigorous analysis and if so, how would you do it? There's a step beyond this, right, as far as the degree or rigor? MR. SALLEY: Right. And I guess we haven't seen a whole bunch of Phase 3s, but one of the areas that I've seen them go to is we go into the fire dynamics. CHAIRMAN SIEBER: Using what tools? MR. SALLEY: That depends. You know, C-FAST is a common piece of software put out that people like to use and make approximations with. So you would start seeing the fire modeling come into more -- but also you would see in a Phase 3 from my experience J.S., and please correct me, but the issue of fire frequency, okay, people wouldn't want to say what's the fire frequency of the room or what's the fire frequency of that specific piece of -- CHAIRMAN SIEBER: Equipment. MR. SALLEY: Equipment. And you'll see that things -- the fire frequency can change orders of magnitude, you can change colors. I like the colors like -- you guys want to keep them green. I want lime green and dark British racing green. CHAIRMAN SIEBER: If you change by a factor of 10, you change colors all together. You go from a green to a white to a yellow to a red, right? MR. SALLEY: You'd see more rigorous fire dynamics development. You'd see more rigorous on the fire frequency of the specific component rather than average or an area and I think between a Phase 2 and a 3 you would see things like the licensee taking it serious and going to perform a test to see what grading does that barrier really have. The NRC has given us zero and we can't argue with their zero. So they'd go out and try to get some hard number for it. MR. HANNON: This is John Hannon. I'd also add that there's an effort, we have underway now at the NRR staff to look at the fire events database to update that and that might provide more current information. Think of using the SDP as far as fire event frequencies, initiation frequencies. CHAIRMAN SIEBER: Yes. Now is the methodology you would use to do a Phase 3 analysis in a fire protection area proceduralized or documented or is this whatever you decide you want to do kind of thing? MR. HYSLOP: I haven't done any Phase 3s associated with this. I've looked at a couple of utility ones. We currently need better Phase 3 guidance and that's one of the things that we've asked the Office of Research to provide us as a result of this program. In general, your technique is the same, given the things that you're considering: frequency, defense in depth and mitigating systems. I suppose if someone could develop distributions they could think of something other than the mean and incorporate that. I don't know of anyone who has done that. So I really don't have a very good answer to your questions. MR. COE: When you're talking about Phase 3 guidance, you're talking really about what kind of standards exist in the general field or practice of probabilistic risk assessment. CHAIRMAN SIEBER: That's true. MR. COE: And you may be aware that ASME is working on some standards that the NRC is participating on that committee with and they should be coming out with a set pretty soon. CHAIRMAN SIEBER: Well, they actually have published a standard, but that's for regular PRAs, you know, the very comprehensive ones. And it doesn't seem to me, as I recall that standards that it tells you specifically what models to use, what assumptions to make, where you get your data from, how you derive all these quantities that go in there. The fact that it doesn't even describe initiation frequency, defense in depth, mitigating systems or any of that. It's sort of in the eye of the beholder at this point, right? MR. COE: Exactly, and I think that the process that we've devised here is one that helps the decision makers of this Agency that are about to make a risk informed decision, better understand the assumptions that went into it. And I don't know that that would change necessarily whether you're doing a Phase 2 analysis or a Phase 3. A decision made on the basis of a Phase 3 analysis should be just as understood in terms of the influential assumptions that were used as a Phase 2. CHAIRMAN SIEBER: I would think one reason why you would go to a Phase 3 is because your Phase 2 analysis was challenged and that being the case, then why not challenge the Phase 3 analysis? MR. COE: In any case, what this does is foster better discussion and a more focused discussion between the staff and the licensee, typically. I've seen this play out because anytime an issue is characterized as greater than green, it comes to a panel at headquarters. And the panel, subject to the panel, is whether or not we are applying the SDP process consistently and inevitably the discussion gets down to the level of confidence that the staff has and the assumptions that are most influential to the result. And then when we discuss this with a licensee, again, it focuses our discuss on those assumptions which are most influential to the result. And I think it's a more efficient way of processing, of communicating with, both internal to the staff as well as external. MR. JOHNSON: But we do hear your question and it's a good question. CHAIRMAN SIEBER: Yes. I guess the other thing that I'm thinking of is there really aren't a lot of fires in power plants if we ignore waste basket fires in some outbuilding some place. On the other hand, there is talking about mining for noncompliance, there's a lot of opportunities just due to the complexity of the regulations to find design deficiencies and testing deficiencies and so forth. I mean you could really make a living doing that. So I see the potential for enforcement actions, noncompliances, noncited violations, what have you, being always there. MR. COE: We hope our inspectors are sensitive to and looking for the most significant of those because I think anybody could agree that as large and complex a facility as these are, there will definitely be some level of deficiencies that exist all the time and the licensee should be identifying and correcting those and our interest would be in identifying those that are of greatest significance to the public health and safety. CHAIRMAN SIEBER: And that's what this process is intended to do. MR. COE: Is to focus our efforts as regulators, yes. CHAIRMAN SIEBER: Okay. I think that clarifies that for me. Why don't we move on. MR. JOHNSON: Okay. All right, Don will you come up? We're continuing through the presentation. If you look in your packages, we're going to shift gears now and talk about performance indicators and Don is here and we hope to be joined by Garrett Perry shortly to talk about a number of issues with respect to performance indicators. The first topic that I wanted to cover was to talk about thresholds in a very general sense, just to refresh your memory with respect to what we intended to do with thresholds, not just performance indicator thresholds, but thresholds in the ROP. We're then going to talk about the process for developing thresholds and I think there was some interest in having us look at mitigating system, an example of how we set those thresholds, so we're going to do that, right, Don? MR. HICKMAN: Yes. MR. JOHNSON: And then last, but not least, we're going to talk about PI reporting so you understand a little bit of the mechanics of how we get this PI data to the Agency. Just by way of providing some explanation or some reminder, if you will, about what we were trying to achieve with thresholds in the ROP, again and I made this point earlier, when we set out to do the ROP we had the notion, in fact, industry very much wanted us to recognize that there needed to be some licensee response band. We weren't going to be able to achieve zero defect. That was an unreasonable expectation. There, in fact, needed to be some area with which the licensees could operate their plants and have problems, but that wouldn't warrant necessarily an increase response on the part of the regulator beyond what we do with respect to doing sort of a baseline level of inspection at all plants to make sure we have the necessary information along with performance indicator information to begin to get an indication about the performance of plants. So there was this notion of a licensee response band. Well, in order to make that work we set up a series of thresholds and those thresholds really serve as trigger points, if you will, for us to take increased regulatory response. Again, the greater the degradation, the more thresholds, the more significant the threshold trip, the greater the regulatory response and we'll talk about the regulatory response when we talk about the action matrix in July. I do want to make the point that the thresholds aren't intended to be predictive. And in fact, we don't even like to use words like leading. And in earlier presentations for the ACRS and in multiple presentations, earlier presentations throughout the development of the ROP, we have typically gotten the question, are the thresholds leading, are performance indicators leading and every time we try to come back with a response that goes very much like we don't guarantee, we don't believe that it's appropriate for us to say that we can predict or present an occurrence of an event. We can't predict necessarily that at Plant A, whose performance is at X level today is going to be at Y level in a year from now. That's not what we set out to do when we set the thresholds. What we set out to do when we set the thresholds was be able to trigger ourselves early enough in a way that would enable us to take timely action because what we don't want to have happen, we don't have to have plants go into that unacceptable column of th action matrix. We're talking about that far right column of the action matrix where we've lost confidence in their ability to maintain the design of the plant. And we've got some words, some high levels words that were taken from the order, from things like -- like words we wrote in the Millstone order, for example, where the Agency has lost confidence in the ability of the plant to -- the licensee to operate that plant safely. And so the thresholds are intended to allow us to trigger, to respond in time to interact before a plant would go into that column. So we talk about timely, we talk about thresholds as enabling us to take timely action where we see these performance declines happening. And thus, that's what we were trying to do with respect to the thresholds. I guess I just wanted to pause for a second and let us talk about thresholds before we go further because I know there was and has been, continue to be questions about what we were intended to do with respect to the thresholds. Wonderful. MR. BONACA: I don't want to belabor it, but it's hard to believe that you can take timely action if you don't have some leading indications that you can work on. That's my comment. You're saying on the one hand you don't intend to have leading indicators. I can accept that. Then on the other hand you say you want to be able to have indicators that will give you the opportunity to have timely action which means take action before things happen. So that in and of itself implies you expect them to be leading. So I don't know where you're going with the two statements. MR. JOHNSON: And it's sort of timely -- that's a fair comment. It's sort of -- is it timely or is it leading to what and this is kind of the discussion that we have. One of the difficulties with the current thresholds in some people's minds is that with respect to the low level issues that they see at a plant, you can get into -- some people firmly believe that they can in terms of thing that you begin to see indications, low level indications of human performance, low level indications with respect to the way licensees find problems or treat those problems, that those provide an early indication, if you will and if the licensee doesn't fix those, they're going to end up with a problem. And I guess I'm trying for a shift in mindset. The old process used to have us look at those issues and react to those issues. We often drew conclusions based on a predominance of those kinds of things, extrapolated them to say hey, if you don't fix these things, licensee, you're going to end up on the last list and the problem with that is that we predicted about twice the number of plants that actually ended up on the last list based on an approach like that because what actually happens is that at a very low level, unless you actually see thresholds, unless you actually get to a point where thresholds are being crossed, much of what you see to cause you alarm because you never know whether what you're seeing is a tip of the iceberg or it is, in fact, what it is and there's not much beyond it. And so again, the rigor of the thresholds is to try to say if there are performance problems, we want to have the threshold set low enough so that we can trigger response as those performance problems begin to occur, but again, if you have problems that don't even reach that threshold, we're going to -- those fall in the licensee response band. That's the balance I'm trying to strike when I draw the line between what is timely. The notion of being predictive, I mean we've had, you'll remember maybe a couple years ago or three years ago or so in response to a direction that we got from the Commission, then I think the EOD looked at financial indicators and the notion at that time was that financial indicators would be an example of something, a type of indicator that would be predictive. And that the ACRS, at the strong urging of the ACRS, among other stakeholders, we backed away from that approach because again, what you would seize upon in terms of being predictive could give you bad results, you could end up seizing on something and thinking that you were getting a valid prediction and in fact, you weren't getting a valid prediction at all. So again, the emphasis on the thresholds was to allow us to recognize performance problems and begin to interact the action matrix providers with greater responses early on because again what we don't want to happen is we don't want to have a plant where tomorrow we decide for ourselves that that plant is unsafe. We want to have had an opportunity to engage and we think that engagement has to happen though through results, performance issues that reflect themselves and especially as they cross thresholds to the SDP or performance indicator issues that cross thresholds that we've set up. MR. HICKMAN: If I could add to that just a bit. The old AEOD performance indicators were sometimes criticized and we ourselves also wanted to try to make them predictive, leading. I know, criticized for the fact that they were. That is very difficult to do because you have to look at programs that will ultimately reflect in performance at the plant. What those programs operate through people and you never can predict how people will react to programmatic weaknesses. Instead of trying to make them predictive, what we always said we were trying to do was to try to make them as responsive as possible, as quick reacting to changes in performance at the plant so that we could identify that as early as possible. In fact, we did some comparisons of the trends of PIs against Agency actions, senior management, meeting actions and things like that. Putting on the watch list and those kinds of things. And that was kind of rather informative. But we want to be as reactive as possible, particularly for this program because one of the premises of the program is that if there's a risk significant problem at a plant it will eventually turn up in performance at the plant. If it doesn't do that, then we say it's not particularly risk important, if it doesn't reflect in some kind of a performance at the plant. So we're looking for those kind of performance problems to show up and we want to identify them as soon as possible so we can step in after crossing the first threshold into the white band and try to take some action to prevent them going further. That's the whole premise of the program. MR. KRESS: Since George is not here, I'll try to articulate a couple of questions that I anticipate he might have asked about this slide. One of them would be looking at the second bullet and the third, delta CDF due to some change in these performance indicators are likely to be plant specific. How do you know that these are the numbers that would be generic? How do you arrive at a generic number for what is like to be plant specific? That's one question. The other question is what's the rationale for choosing the 95 percentile for the first threshold? Why is that a good number to use? MR. JOHNSON: Okay, I'm sorry, was there a third question? MR. KRESS: Those two right now. MR. JOHNSON: We actually were going to get to those. Don was going to talk through the actual process for developing thresholds and when we get joined by Garrett Perry and I know Don's been anxiously watching the door for Garrett to come in, Garrett was involved in the original setting of the thresholds. We'll talk about those issues. MR. HICKMAN: Yes, we'll get into both of those. We'll start with the first bullet. The green-white threshold, the concept was to identify plants with performance as an outlier to the industry. We didn't go into this development with the concept in mind of 95 percent or two standard deviations or anything like that. When I show you this slide, I think you'll see that it's very obvious where the thresholds should be set and maybe I guess we should go into that one right now. [Slide change.] MR. HICKMAN: This is an example of what we did. This happens to be the safety system unavailability of the aux. feedwater system. Remember, now that we did this in the fall of 1998 and so we took the most current full three years of data that we had, that was 95 to 97 and this was -- we did all this in concurrence with -- in agreement with the industry, represented by NEI. We said that we would take those three years and make them our baseline. So we collected this data over that period for the best data we could get for each of these PIs. In this case, for our safety system unavailability indicator we used the same definitions that WANO had been using for many years. So they had been collecting this data on a quarter by quarter basis, taking 12 quarters and summing them up, calculating the indicator. We had three years worth of that. We had 12 quarters worth of that data. We took every plant, in this case all of the PWRs, there's 71 on here. We took the highest value during those three years and we plotted it and that's what you see. It's the worse case value, the highest unavailability of that system for each of those plants. MR. KRESS: Now, if I were going to draw a line as a threshold through that, I would have dropped down to the next level, instead of the one you have because there's, to me, it looks like two modes, two mode distribution. I would make the line right in between the two modes. I don't see a real rationale for the line you have up there. MR. HICKMAN: That's set at 2 percent which is the current threshold. If we had dropped it down to the next line, that's 1.5 percent. I guess you could argue about that. We looked also at the number of plants, two things we looked at. One was that there was a clean break. You didn't want to have a plant, two plants slightly, very small difference apart, but on opposite sides of the threshold. So we looked for a gap. And as you point out, it could have gone either place. MR. KRESS: Yes. MR. HICKMAN: We then also looked at the number of plants. And this is not a hard and fast rule. It wasn't like 95 percent was a hard and fast number. It was of that order. And so we captured five plants setting it at 2 percent, out of 71 in a three-year time period. If we had dropped it down we would have gotten 13 plants. MR. KRESS: I don't understand why you didn't, frankly. MR. HICKMAN: This one is probably a little more controversial than some of the others. Most of them were very clear where the threshold ought to be. This one we could argue about whether it's 1.5 percent or 2 percent. You're right. We felt that 5 plants was better perhaps than the 13. MR. UHRIG: You say this is just PWRs? There's a hundred and some odd plants there, unless I'm not reading it -- MR. HICKMAN: Well, the numbering system is kind of strange. These are the graphs that we got from NEI. They provided this data. And the numbering isn't quite right. But if you count the bars, it's actually 71. (Laughter.) MR. UHRIG: Okay. MR. HICKMAN: If all the plants are there, then there would be that number, but they're not all there. It's confusing. MR. BONACA: Again, this is not plant specific at all. What I mean is that it doesn't recognize the -- MR. KRESS: That was the other thing -- MR. BONACA: -- importance, the importance of the unavailability to the specific plant. MR. KRESS: Right. MR. HICKMAN: That's correct. MR. BONACA: Okay, so it doesn't recognize that. MR. KRESS: It may be that that plant that shoots up there has always been there and it didn't matter. MR. BONACA: Maybe there is another system behind it. CHAIRMAN SIEBER: It might have five pumps. MR. HICKMAN: We recognize that. We have had many discussions about this. There's actually four indicators per plant on the safety system unavailability and we're undertaking a major effort to kind of overhaul this. And of course, as George keeps reminding us, we're aiming towards the plant specific PIs, the plant specific thresholds. That's the goal. MR. BONACA: This is a good effort there. MR. HICKMAN: It's going to go from the beginning. [Slide change.] MR. HICKMAN: Let me go back to this slide again. Now Garrett can talk better about this because he did this and I'm not a PRA person, but basically what he did was to take some generic vendor models that we had. He used the old SPAR models, not the new rev. 3 models, but the old one. And there were just a limited set of those, I think about a dozen or so and those were essentially vendor types of models for the various configurations of the vendors for Westinghouse two loops, three loops, four loops, etcetera. He then ran this parameter, varied the parameter that we were monitoring to get a change in CDF of 10-5 for the white/yellow threshold. And he did that for each of the models and if you look in Appendix H of attachment 2 to SECY 99007, that's Garrett's appendix where he describes how he set these thresholds and there are tables in there and it will show for various plants representative of each of these models what the numbers were. And essentially what he did was to take the most conservative number, the smallest number. MR. KRESS: That's how he got around the plant specific part of that. MR. HICKMAN: Right. So to make sure that essentially every plant was covered. If you read it carefully, you'll see there's a few holes in there and there's still work to be done on the thresholds, but that was the basic approach. The same thing was done for the yellow/red threshold, but adjusting the parameter to get a delta CDF of 10-4. MR. KRESS: That's using the old SPAR models? MR. HICKMAN: Yes. Right. Right now we have -- MR. KRESS: It's kind of group plants by vendor type? MR. HICKMAN: They're just vendor models. MR. KRESS: There's one that's treated as one type of plant? MR. HICKMAN: Right, they're pretty generic vendor models, but there's a particular plant, I guess that it gets modeled after and they're listed in the tables in Appendix H. MR. KRESS: That represents these plants? MR. HICKMAN: Yes. CHAIRMAN SIEBER: I take it you couldn't do that, use that technique for the green-white threshold because almost all plants would be white then, right? MR. HICKMAN: With green-white it would be more difficult. CHAIRMAN SIEBER: You would have -- all you'd have to do is have one failure and you would be white, a CDF at 10-6, right? MR. HICKMAN: But as Garrett points out in Appendix H, this method worked well because you'll see that there is still quite a gap between the green-white threshold and the white-yellow threshold. So by going by outliers from industry norm, we think we have a pretty good threshold. It gives us a decent green band for the licensees to operate in and it gives us a white band for us to react and to try to prevent further degradation of performance. So it did work out pretty well. MR. KRESS: This is a one time fixed event, threshold and it won't be adjusted later? MR. HICKMAN: We set the thresholds this way prior to the pilot program. At the completion of the pilot program we looked again and we did make some adjustments. Actually, it wasn't based on the pilot data because we only had 13 plants, but when we got the initial input from the entire industry, giving their historical data, that's what we looked at and we did make some adjustments based on that. In some of the safety system unavailability indicators, in the security equipment performance index indicator an din the occupational radiation exposure indicators. Also, in safety system functional failures and scrams of loss of normal heat. MR. SHACK: I also suspect that the finer you make that delta CDF the more the plant specificity makes a real importance, that is, if you did that at 1 times 10-6, you really would almost have to do it on a plant specific basis. By the time you get to 10-4, you're probably not terribly sensitive to -- MR. KRESS: I think you're exactly right. MR. SHACK: Minor variations. So there's a certain rationale to doing it that way. MR. COE: That's a good point. I would also point out that some licensees, because these thresholds or these thresholds for unavailability in this case are generic, may find that their own maintenance rule, performance criteria for the same piece of equipment allows much greater unavailability for certain components that are being monitored by these PIs and this is a source of concern to them, that they're being held to this generic standard whereas their own plant design, their unique features of their plant design would allow a more unavailability to accrue for that particular component before they got to that risk threshold. MR. JOHNSON: Yes, if you remember where we were, as Don points out in 1998, we really were trying to make progress, given the tools that we could seize upon quickly, given the PIs that we could seize upon quickly. We did create some new PIs and in fact, we did end up trying to set thresholds for those and then trying to benchmark those thresholds and make adjustments to those thresholds in the pilot program. And we recognize, as we go forward, that we'll need to continue to work on and to refine the performance indicators and the performance indicator thresholds. We have a process that -- and we talked about this a little bit at the last briefing, that as a formal change process for changing PIs or changing thresholds and it's a deliberate process that has us look and pilot and benchmark before we make decisions about changes. But again, I think we agree with the ACRS that our thrust for the major improvement with respect to PIs is in trying to, to the extent that we're able to, do something with respect to being more plant specific. CHAIRMAN SIEBER: There's some slight difference in the wording of the second and third bullet. Is that just editorial or is there some meaning you're trying to convey there that I'm missing? MR. HICKMAN: Garrett wrote that. I really don't know. MR. SHACK: The rule about parallel construction. (Laughter.) CHAIRMAN SIEBER: Come to the right place, right. MR. HICKMAN: If there are no more questions on that, there was apparently a desire to see how the process works, how we collect the PIs and report them. CHAIRMAN SIEBER: Okay. MR. HICKMAN: I don't appear to have a transparency for that. You will have it in your handout. CHAIRMAN SIEBER: 43. PI Reporting. MR. HICKMAN: Yes, PI Reporting. I used here an example, again, from the safety system unavailability indicator. The PI is defined in the guidance document, NEI 99-02. And I've shown that definition here. It's the sum of the unavailable hours to plan the unplanned and the fault exposure hours. MR. UHRIG: What do you mean by fault exposure hours? MR. HICKMAN: Fault exposure hours are the hours that a train was in a failed state, but was undetected. MR. UHRIG: Before you caught it? MR. HICKMAN: He didn't know it was failed until some time later. MR. UHRIG: How do you know when that is? MR. HICKMAN: Well, if -- let's say you ran a surveillance test and it failed, but you could trace that back to some maintenance that was done some time prior to that test and if you could show that that's what caused a failure then you would count that amount of time. MR. UHRIG: Okay, what about where there are two surveillances, one, it passed, one, it failed? MR. HICKMAN: If you had no way of knowing when the failure occurred -- MR. UHRIG: Then you've gone all the way back to the other one? MR. HICKMAN: What you do is you use half the integral. MR. UHRIG: Half the integral. MR. HICKMAN: The standard statistical technique, assume it's a uniform probability of failure, divide by 2. It's good for large sample sizes which we don't really have, but that's the way it's typically done. MR. UHRIG: All right. MR. HICKMAN: And that's an issue that's been a problem in this program for quite a while. There's a lot of serious discussion about the use of T/2. We have had a number, about three failures of 18 month surveillance tests, which meant licensees had to count 9 months of unavailable hours which is -- and then that sticks with you for three years, basically. CHAIRMAN SIEBER: But that's been the fact for a long time, you know. I remember that from 20 or 30 years ago. MR. HICKMAN: That's pretty standard, pretty standard technique. So we do that. That is how we calculate a train unavailability, per train. MR. KRESS: The hours train is required, is that to differentiate shut down conditions when you don't need it? MR. HICKMAN: Well, ideally it should. What we're doing right now and what WANO does is to simply lump them together. Ideally we would have separate indicates for power operation and shut down conditions, but right now we just lump them together. MR. KRESS: This is just the number of hours over which you determine the unavailability then? MR. HICKMAN: Yes. MR. KRESS: So it's the code of thermal errors. MR. HICKMAN: And what's used there is the hours that the train is required per tech specs which means if you're shut down and tech specs only require one EDG, you can take the others out and do whatever you want to with them and not have to count the hours. Now the other thing that INPO did, INPO actually developed the indicators in the late 1990s and WANO started using them in 1995. INPO developed them in the early 1990s. They did some tests collecting actual data and then looking at easier ways to calculate unavailability that would be less of a burden on licensees with regard to the data they have to submit. And they found that by taking these -- the train unavailabilities of a system and averaging them together, they came up with a system unavailability that tracked pretty well with the real thing. The numbers weren't the same, but they tended to go in the same directions. So this is what they used and it's what we are now using. It's not right. Ideally, you would want to know when both trains were out at the same time. You'd have to have the timing information, but rather than collect all of that, they said this is close enough and it suits our purposes and that's what they were using and so that's what we're using. We recognize the weaknesses. MR. KRESS: But that other information is probably available, just harder -- more work to get it. MR. HICKMAN: Yes. And as you know, the Office of Research is developing the risk-based indicators and they're trying to get information like that into the EPIX system so they can calculate unavailability. What the licensee actually submits to us then is four numbers for each train, the planned unavailability, the unplanned unavailability, the fault exposure hours and the hours the train is required. They send that to us in an e-mail with an attached file that is actually a delimited text file. That comes into our system here and it's automatically dumped into a spreadsheet and each of those numbers is put in the right bin. It's all automated. That spreadsheet then calculates the values. That's been thoroughly checked. All through the pilot program we checked that to make sure it works properly. So really the processing is hands off. We do nothing with it until it's all in this spreadsheet. We then take that spreadsheet. We send the data back, first of all. We send the delimited text files back to the licensee to say this is what we got. Is this what you sent? That's the confirmation process that takes about a week. Then once they've confirmed that the data we've received is accurate, then we review it. We give the regions a chance to look at it and within a week then we put it out on the external web. Actually, at the end of the first week you put it on the internal web for the regions to see and a week later then we put it out on the external web. And that's really all there is to the data processing. Are there any questions? CHAIRMAN SIEBER: So basically, the way you're using performance indicators differs from the way plants use it. Plants use it as a predictive measure and they collect sometimes as many as 250 different performance indicators saying that if you have backlogs building up and so forth, that that's an indication that your maintenance program, your corrective action program or what have you is declining and so they use that to redirect resources. What you're doing is calculating and reporting changes in risk, in effect, which is more or less real time. If unavailability goes up, then the risk changes for a given plant. And if reactor trips go up, the risk from ATLAS is changed and so on down the line. So there is a different concept between the way utilities use performance indicators and the way you folks are. And I think you have to do it your way so that it matches the regulatory system. You don't want to be in the business of managing the plant the way a plant manager would do it. So I think that's appropriate, what you're doing. MR. HICKMAN: That's exactly true. There are a number of good indicators that will work if people don't know you're tracking them and that's good for plant managers to be looking at those kinds of things like backlog. For us to take them and put them on the web would not be good. CHAIRMAN SIEBER: Well, you don't have regulations that speak to backlogs. MR. HICKMAN: Right. In fact, the backlog will go away instantly if we start -- CHAIRMAN SIEBER: All you have to do is sit down and do some homework. MR. HICKMAN: That's right. MR. JOHNSON: This whole shift in the process with respect to our use of performance indicators was really dramatic from what we had done prior to the oversight process and to be quite honest, we were a little bit surprised at the industry's willingness to go forward with some of the performance indicators. By that, what I mean is we've got thresholds on scrams for 7,000 critical hours and there's no regulatory requirement that says that a plant shouldn't have four scrams per 7,000 critical hours -- CHAIRMAN SIEBER: Or 10. MR. JOHNSON: Or 10. CHAIRMAN SIEBER: Except it does change the risk. MR. JOHNSON: So what we did, what we set out to do and what we were able to accomplish is that we chose a set of indicators that we believe is indicative, now they're not perfect, they're not as risk-informed in some cases we would like them to be, but they give us insights along with inspections into issues that begin to emerge at a plant at a level where we as a regulator ought to engage as opposed to where licensee management ought to be doing its business. CHAIRMAN SIEBER: Okay. MR. LEITCH: Has this definition been the one you've used here in this whole one year, initial one year period? MR. HICKMAN: Yes. MR. LEITCH: Have any of the other definitions changed during the one year period like scrams and if so, how did they change? CHAIRMAN SIEBER: Yes, they did. MR. HICKMAN: You may be aware that we just finished a pilot program for replacement scram indicator. Are you aware of that? There were a few people in the industry who were concerned about unintended consequences, unintended influences on operators from counting manual scrams, so the industry -- industry representatives working within NEI developed an alternate indicator to replace that one and we just finished a pilot program. The intention is to count exactly the same thing and that was automatic and manual scrams, but without ever using the word scram in the definition, so it's kind of a funny thing. But we are looking at that now. We have criteria to evaluate that against and we'll use that to serve as a replacement. MR. LEITCH: That's one thing that confused me. In the pilot program you counted both manual and automatic scrams, just like the initial one year program, it's just the matter that they didn't call them scrams? MR. HICKMAN: Right, that was the intent. Whether we did that or not is still yet to be determined. We're looking at the data now. We just got it in, final, couple weeks ago. CHAIRMAN SIEBER: Well, that particular argument goes back about 10 years because the industry made the same arguments to INPO that says you're going to inhibit the operator from manually tripping the reactor and the INPO philosophy is to trip it manually before some automatic system takes you out which lessons the transient on the plant in a lot of cases and so I guess I wonder whether counting manual scrams is really the right thing to do, even though from the standpoint to causing an initiating event by the twist of a wrist does change the risk of the plant because it causes a lot of other things to happen. Is there something on either side of that question as to whether you count it or you don't count it? MR. HICKMAN: As you know, the AEOD PIs agreed to use the same definition as INPO. CHAIRMAN SIEBER: Right. MR. HICKMAN: When they started those in 1985. CHAIRMAN SIEBER: Right. MR. HICKMAN: And we only counted automatic scrams while critical for that reason. But there were people here who were concerned that operators might try to beat the PI by manually scramming it, so we monitored that we never really saw any signs of that. Manual scrams have remained relatively constant around 40 per year, up until the last couple of years. Some as high as maybe 55, some down to about 29 or so, but roughly averaging around 40. They're down a little bit now, down into the low 30 range, but of course, the automatic scrams have come way down from several hundred, down to about 50. CHAIRMAN SIEBER: Right. MR. HICKMAN: But from the very beginning and working with NEI on this, we never really doubted whether we needed to count manual scrams because the conditions in the plant that require a scram are the same and whether the operator manually scrams it or it takes an automatic scram, whatever has gone wrong with the plant that required that scram is what we want to count. CHAIRMAN SIEBER: Yes, but the technical challenge to the plant is typically less because you haven't reached the set point or the limiting safety settings. MR. HICKMAN: That's true. As the operator scrams it, he may prevent other automatic actions by not reaching -- CHAIRMAN SIEBER: And less than the excursions that the plant goes through during a shutdown. MR. LEITCH: But this new definition, revised definition, not using the word scram is separate pilot program. That is, the initial one year period, nothing has changed during that period? MR. HICKMAN: No. We're still using the same indicator that we started with in the pilot program and it says the indicator counts all automatic and manual scrams while critical. MR. LEITCH: Was there a change or is there a change being contemplated with regard to unplanned power changes? MR. HICKMAN: Yes. We're getting ready to try a pilot program on a replacement for that. Actually, there are two proposals, one from the NRC and one from NEI that we'll pilot. The concern there is that we had a 72-hour rule, basically it said if the time between the identification of a problem and beginning to insert negative reactivity is greater than 72 hours, then it doesn't count. This was something that was of concern to NEI and the industry that we shouldn't count power reductions that are planned. It was never the staff's intention. We never worried about whether it was planned or not. We used a definition that's in the monthly operating report and there, the distinction was not planned versus unplanned. It was forced versus schedule which is not exactly the same thing. CHAIRMAN SIEBER: That's right. MR. HICKMAN: And what we captured in the monthly operating report was whether they had to come down at the first available opportunity to fix it, or whether they could ride through that and continue on. At that time, when the monthly operating report was put into place, the first available opportunity was considered to be the next week. So that was the criterion. But what's happened is with the 72-hour rule, that provides an incentive for licensees to -- CHAIRMAN SIEBER: Struggle along. MR. HICKMAN: And ride it out. And in fact, we had a licensee who was very forward with us and he told us, I can't afford another power change. I'm going to ride it out and he did that a couple of times. In defense of the licensee, he didn't do it when he thought it was a safety problem, so even though it was going to cause him a problem, he did shut down and he did count it, but when he thought he could get away with it, he didn't do it. CHAIRMAN SIEBER: Well, that's one of the problems with performance indicators across the board. People know what the thresholds are and what the goals are and they will manage the plant to meet those expectations. And that's not always in the plant's best interest. MR. JOHNSON: That's right. CHAIRMAN SIEBER: And so that should be an important factor when you folks are divining what kind of performance measures you're going to use, because you might as well face it, people do manage to those indicators. MR. JOHNSON: Absolutely. MR. HICKMAN: That's true and this is a particular problem in the initiating events cornerstone and the mitigating systems cornerstone. CHAIRMAN SIEBER: That's right. MR. HICKMAN: We've had a number of successes in the program in the emergency preparedness cornerstone and in the physical protection cornerstone. If we could make all of the indicators like those in the EP cornerstone, they provide the incentive for the licensee to do the right thing, that is, we've got a drill exercise performance indicator and a drill participation indicator. And if he's having problems with either one of those, the answer is to run more drills and get more people in the drills. And we've had great success. We've had people who were not paying attention to whether there were people who were actually getting trained or not on a regular basis and when we started the PI, they realized that and they responded and they brought their PIs down to within the green band. And that's good, if everybody stayed within their green band, that would be good. The same thing happened in the security equipment performance index. We had a couple of licensees who had very bad problems with their security equipment and had just never gotten management attention and as soon as the PI came out and the manager looked at that, he said what's this all about and he immediately fixed the problem. CHAIRMAN SIEBER: Yes, but there was a practice among licensees in security to say that if I put a watchman in place or a response officer in place of the defective piece of security equipment, that compensating measure was equivalent to having that piece of equipment in the service, so they would sit down and calculate it's going to cost me $25,000 to fix a TV camera, how long can I keep a watchman there to watch that zone? And will I, in effect, make out economically by doing that? Okay, so what you've done there is change the economic balance of supply/demand situation for the management. MR. HICKMAN: Sooner or later though they'd have to fix it, but I mean at some point the cost of the guard is going to exceed the cost to fix it. CHAIRMAN SIEBER: That's true. It all depends on whether you have capital money or operating money to spend. MR. HICKMAN: That's true. CHAIRMAN SIEBER: Some day I'll have a meeting to explain the power plant economics, but some plants don't have capital money. You know, they just don't have a rate base, so they don't want to spend it. MR. SHACK: Has anybody objected to any of these PIs as a backfit? MR. JOHNSON: Not to my knowledge, no. MR. COE: There has been some discussion at high levels regarding the earlier question, the earlier point that was made is that these aren't based on regulatory requirements and therefore there's a question out there about de facto regulation. But I think that those haven't been, there hasn't been a unified chorus of individuals out there that are complaining about that. I'm speculating, but I think it's primarily because they see greater net benefit, you know, the disadvantages as they perceive them are offset by the benefits of the program. So they're willing to work with us and continue to evolve the program to what they hope would be better in the future. MR. JOHNSON: I actually think we could be more positive. There was concern early on about whether we needed a regulatory requirement to collect these, regulation to collect these performance indicators and NEI said God forbid, don't do that. And we said okay, we'll have this voluntary PI program and if you guys don't give us PIs, we'll go do baseline inspection to get the insights. Well, we've not had licensees not give us performance indicators because they don't buy the program. Now having said that, we work very closely with the industry and other stakeholders and the public meeting to refine the reporting criteria, to make sure they're reasonable and understood. So it's been a lot of work for us to be able to implement this voluntary aspect of the ROP. But there's not a course there. MR. SHACK: I hear that plants collect 200 PIs and whenever the risk-based PIs are mentioned, oh my God, the burden is incredible, can't believe it and it just somehow seems like a mismatch here. Again, maybe there's a difference between collecting the data for your own purposes and swearing to the NRC that this data is accurate and I'm ready to go to jail if it's wrong. MR. JOHNSON: Yes and those are some of the issues. In fact, the last time I sat in on the risk-based performance indicator talk that you all were given by Research and that is what licensees tell us. I think what we heard from licensees of late is we've got this new oversight process. We've got PIs associated with that process. Why don't we live with that for awhile and why don't we consider very carefully adding additional performance indicators that could result in additional burden. So there is definitely that theme that we're getting. And again, when we go to collect performance indicators, I sort of am remembering now how that last risk-based performance indicator briefing went and some of the issues that came up that were discussed and I think we have an IOU, as a matter of fact, to the ACRS that came out of that briefing, but again, remember, the performance indicators provide a valuable piece of information. Now the performance indicator program is a voluntary program. It turns out there are OMB clearance requirements, requirements with respect to collecting data from more than nine licensees. So if we go to do that, we've got to make the case about burden and about benefit. And so we're -- we think we are appropriately cautious with respect to adding new PIs to make sure that they give us the benefit that we need, but at a cost that is appropriate. That was some of the sense that we discussed last time. You're right. You do hear the industry say hey, don't give us a whole bunch more of performance indicators when what we have is okay for now. CHAIRMAN SIEBER: I think the other problem that comes up sometimes is the fact that if NRC comes out and says I want this performance indicator and I'd like you to send it to me, but my definition is different than WANO's, then the licensee sees that as a whole new indicator because they have to engage somebody to produce it every month for you. I think on the other hand, the industry appears to prefer risk informed and performance based regulation to deterministic regulation and if it adopts that kind of preference, they have to cooperate and I think that's what you're seeing. MR. HICKMAN: And you hit on one of their big concerns and that is if they have to calculate unavailability one way for WANO and another way for the maintenance rule and another way for us -- CHAIRMAN SIEBER: That's right. MR. HICKMAN: That's a burden. CHAIRMAN SIEBER: It's confusing too, because it's usually the same person who's doing all the calculations and to keep all that stuff straight for a whole bunch of different indicators is troublesome. MR. HICKMAN: Especially if you're going to be held to 50.9 requirements for sending it to us. CHAIRMAN SIEBER: That's right. MR. HICKMAN: The other aspect of that -- I just lost it. Oh, the other aspect to the more indicators is in their view it's just more ways to go white and why do we need more ways to go white if we've got 18 already that work. MR. JOHNSON: Okay, that captures the discussion we plan to have on performance indicators although I do note that Garrett is in the room. CHAIRMAN SIEBER: Too late. Unless one of the Members has a question that they would like to direct to Garrett. [Slide change.] MR. JOHNSON: Okay, the last section that we want to cover and we've just got a couple of slides is there were some selected issues. Two of the SECY issues I think we've already talked about, and that is we talked about thresholds and the threshold for green to some extent. Hopefully, you're satisfied and we don't need to talk about fire protection any more, because the fire protection people are no longer in the room and I can't even spell fire protection. CHAIRMAN SIEBER: Well, I'm the chairman, but the one who asked the question isn't here. So I'll take it upon myself the duty to go over it with him. MR. JOHNSON: Okay, the last topic that we wanted to talk about was the topic of cross-cutting issues because we know there has been some interest with respect to this topic and for that Jeff Jacobsen is going to talk very briefly about cross-cutting issues. [Slide change.] MR. JACOBSEN: Okay. I guess where we left this, just a little brief history as cross-cutting issues is something that has come up throughout our engagement with the public and internal stakeholders with regard to how cross-cutting issues are treated in the new oversight process. And cross- cutting issues we defined originally as three issues: human performance, safety conscious work environment and problem identification and resolution. So when we talk about cross-cutting issues, those are the three things we're talking about. The fundamental assumption when we designed the framework for the revised oversight process was that these cross-cutting issues would show up either in the performance indicators or in the baseline inspections, in a sufficient time frame to allow us to engage before a real safety issue arose. We consciously did not design a program to specifically go after human performance, for instance, because we thought that if human performance was weak, it would show up in one of the performance indicators, reactor trips or unavailability if it was maintenance related to human performance, etcetera. With regard to safety conscious work environment, a similar analogy was thought that weaknesses in that area where people are afraid to bring problems up or there's retribution, our experience has been that those facilities performance has suffered as a result of that and we would see it. CHAIRMAN SIEBER: You would also see that as allegations, would you not? MR. JACOBSEN: Right, which we also monitor kind of outside of the performance indicators and baseline inspection, but it is part of our overall process. We do, however, have a significant portion of our inspection program that's directed at problem identification and resolution because we believe that is a very important part of the process, so we look at that. We were looking at it annually. We recently made a decision to change that to a once every two year inspection. So we do look at that. CHAIRMAN SIEBER: How do you determine whether the licensee for any given plant has set a low enough threshold for formally identifying problems? MR. JACOBSEN: Our experience has been that each licensee's program is somewhat unique. CHAIRMAN SIEBER: That's right. MR. JACOBSEN: We don't have a go-no go, per se, for what's a low enough threshold. What we would use would be if we, for instance, in our other inspections identify problems that we think are significant, that the licensee didn't get into their corrective action program for whatever reason, we would infer that they do not -- they either don't have a low enough threshold or they aren't looking in the right direction. If we're finding stuff or other external organizations are finding issues, and the licensee isn't finding them, then that's either a threshold question or a question that they just aren't looking in the right areas. CHAIRMAN SIEBER: Well, how do you weave that into the regulatory system? I mean you could determine that through observation and inspection, but how do you bring that -- MR. JACOBSEN: How do we act on it? CHAIRMAN SIEBER: Well, how do you relate that to the requirements of the regulations? MR. JACOBSEN: Well, Appendix B has -- is really the appropriate regulation. CHAIRMAN SIEBER: You can cite anybody for anything through Appendix B. MR. JACOBSEN: Right, well, most things. There are some areas that Appendix B isn't applicable and that has actually come up in this process, emergency preparedness, for instance. CHAIRMAN SIEBER: Right. MR. JACOBSEN: Appendix B is not applicable. The way we deal with it is if we were to have an inspection finding that turned out to be a significant finding and if we found out the root cause of that finding was related to threshold issue or improper evaluation of a previous issue, we would deal with that in that manner. CHAIRMAN SIEBER: Okay. MR. JACOBSEN: It would be on a for-cause basis for the most part. MR. LEITCH: What's the basis for moving that inspection module from annually to semi-annually? MR. JACOBSEN: That was a very general statement of what we're doing. In addition to changing the frequency, we've done some other things. We've beefed it up a little bit so although we're going to do it less frequently, we're going to add some resources to it because we think that the look every two years in a deeper way is more effective than doing it annually in not as deep a way. The basis for it in our experience, licensees' programs such as this will not change significantly on a one year basis. We've seen declines in corrective action programs, trends, but we believe that a frequency of every two years will be sufficient to pick that up and if we went and did an inspection at a facility and found they had a good, corrective action program one year, it would be highly unlikely, in our opinion, that it would decline significantly in one year. It's more of a cultural -- it's almost analogous to plant culture. And that's something that you know takes a long time to turn around. It also pretty much takes some time to go down. So that's -- we're also adding some additional requirements where we're going to instead of doing a team inspection, we're going to look at some limited samples throughout the two years on a per inspector basis. So every so often, one of the inspectors is going to pick something in the corrective action program and do an in-depth inspection of that one item. And then every two years the thought would be that you would integrate all those insights that you got throughout the year, as well as the insights you got while you're doing the team inspection into a more broad assessment of the corrective program. MR. LEITCH: Okay. MR. JACOBSEN: And the last thing we're going to is an we'll talk about this a little more when we get to the action matrix discussion next time, in July, the other element that we're adding is is we're beefing up the role of this PI&R inspection or I guess I should say if a plant would end up in the action matrix in the degraded cornerstone column, we would, in fact, consider -- the regions would consider doing a problem identification and resolution inspection. We think that provides a better opportunity for licensees, for the NRC to look at the performance of the licensee and the performance of the PI&R program in a specific event where they've crossed some thresholds. So we think, in balance, even though we say we're going from a single year to a biennial frequency, we've done some other things of PI&R that we really believe make it more, a much more effective inspection. MR. FORD: Just for information, what does move out of the licensee response band, they don't correspond with it? What does it mean? MR. JACOBSEN: The second item? MR. FORD: Yes. MR. JACOBSEN: Okay, I'll go into that. Our experience with the first year of implementation of the revised oversight process has pretty much supported the first assumption and what we mean by that is plants that we've looked at and we have concerns about in the cross-cutting areas and primarily they've been in the problem identification and resolution area. For instance, if we did our annual team inspection and we had a lot of green findings, we haven't had any white findings or greater as a result of the corrective action inspections. We've had very few white inspection findings overall. But in the PI&R area we haven't had any. But we've had a lot of green ones and if you look at the plants where there's been a lot of green findings and where the inspection team came away with concerns about the adequacy of the program, in all cases those plants have moved out of the first column, that licensee response column of the action matrix, either to a degraded cornerstone column or a regulatory response column which has allowed us to engage further and to look in a more programmatic sense at the corrective action program. A good example of that is Kewaunee where we had concerns with their performance during our problem identification and resolution inspection. They had a yellow performance indicator and when we went out and did that, we identified broader concerns with the corrective action program as well. As a result, they totally revamped their corrective action program. So these four facilities are examples of facilities where we had concerns after doing the baseline inspection and they also -- we had opportunities to look further as a result of our supplemental inspections. The contrary to that is we have not identified any plants where we have significant concerns in the cross-cutting areas that have not moved out of the licensee response. So it's been a very close tie between the performance and actually crossing the thresholds that allow us to engage further. The third bullet, no significant precursors caused by cross-cutting issues, well, in fact, the definition of significant precursors, I believe, is an event that's defined as having a 1 in 1,000th greater chance of leading to a reactor accident. There haven't been any of those period. Really, if you were to look at the fundamental assumption and the basis of the ROP is we would be concerned if we had, for instance, one of these significant precursors and found out they were caused by a cross-cutting issue and we didn't have an opportunity to go after it and prevent it. That hasn't occurred. The way we're going to deal with that is kind of on the next page. We're going to look at things at a threshold actually lower than significant precursors. We're going to look at ASP events and inspection findings that come out yellow and red and we're going to look and see in those instances whether cross-cutting issues were one of the root causes that caused the event or the inspection finding to occur. And if so, would our program have at least given us the opportunity to identify those type of issues. So I guess the bottom line is we believe our fundamental premise of the ROP with regard to cross-cutting issues still appears to be true. However, we still have some on-going actions to continually challenge that and ensure that, in fact, we are focusing our resources in the right direction, as we do with all areas. It's not limited to cross-cutting issues. That's pretty much all I wanted to go into. MR. JOHNSON: Very good. CHAIRMAN SIEBER: Thank you very much, appreciate it. I'd like to take a few minutes to ask if any Members have any comments that they'd like to make based on what we've heard today? MR. UHRIG: I just have a question. This was handed out. I don't know if you handed it out or this came from somebody else. MS. WESTON: I passed it out and my only question is what are the titles of the codes. MR. UHRIG: Among other things. (Laughter.) CHAIRMAN SIEBER: Okay, there are the seven cornerstones. MR. UHRIG: The other question had to do with there's a number after, for instance, white 3. Does that mean three findings? MS. WESTON: That's the inspection summary findings for the first quarter. MR. UHRIG: That would be third quarter, 3 would mean third quarter? MS. WESTON: That's what he's looking at. This is the first quarter. This is all the first quarter. These are cornerstones. MR. UHRIG: What does the 3 mean? MS. WESTON: I don't know. CHAIRMAN SIEBER: Since we're still on the record, maybe we could have people speak into the microphone. MR. JOHNSON: What you're looking at is one of the web page printouts and we've got a number of these employees to summarize the results for all of the plants, in addition to be able to pull up any individual plants, these are the performance indicators and the inspection results. So Don is going to try to answer the question. MR. HICKMAN: What you see here is for each plant and each cornerstone, you see the inspection finding results. What they show there, the ones with the numbers, the color is the color of the highest, the most significant one and the number is the total number. It doesn't necessarily mean there are three whites in that block, but it means there are three and the highest one is a white. MR. UHRIG: Okay. So I take it where there's no number, there's only one finding? MR. HICKMAN: Yes. MR. UHRIG: For example, most of the greens are that way? MR. HICKMAN: Yes. MR. UHRIG: Okay. CHAIRMAN SIEBER: The other question was there are a large number of no findings. That means simply that this is the first quarter and during that first quarter there was no evaluation? MR. UHRIG: No. MR. HICKMAN: They had -- they conducted inspections and had no findings. CHAIRMAN SIEBER: None at all. MR. JACOBSEN: They may or may not have done an inspection in that area. In either case, there were no findings. MR. BONACA: And green means simply -- CHAIRMAN SIEBER: That there was a finding. MR. BONACA: Yes, but for example, the initiators, the first category, a green would mean simply that it was -- CHAIRMAN SIEBER: Well, it means there was a finding which means there's a deficiency, but it's within the licensee's prerogative and control to fix it. MR. JOHNSON: Exactly. CHAIRMAN SIEBER: Without additional enforcement emphasis. MR. UHRIG: Notice in some cases, sister plants for instance here, Peachbottom 2 and 3 had whites both -- is that a common failure? Is it a site failure? Is it just the individual plants happen to come out that way? MR. HICKMAN: It depends. MR. UHRIG: Same with quad cities. MR. JOHNSON: Every unit has -- the ROP is specific for the unit with respect to the performance indicators and the inspection findings. And so it's entirely possible. MR. UHRIG: Is that cornerstone emergency preparedness? MS. WESTON: Occupational radiation safety. It's the sixth column. MR. UHRIG: Internal rad. MR. JACOBSEN: And in some cases the finding can affect both units. In other cases, it may be two separate independent white findings of a unit. MR. HICKMAN: For the cornerstones and site white programs like EP and occupational radiation safety and security, they both get account. MR. BONACA: The question I have is you pointed out that there is a correlation between plants that has a problem and the effectiveness of the corrective action program and I always believed that. But a question I have do you have any specific set of indicators on the corrective action program being used or is it so, is it again considered subjective by a licensee, a judgment he may express on that? I'm going to the fact that more and more we are speaking about objective evidence and when I look at this data, I mean I can interpret it and it tells me something. But I still believe that the corrective action program tells me much more than anyone of these boxes. That's my personal belief, if I could get into it. And so the question I have is when you do the inspection, since there is no quantitative assessment that is translated into a caller, do you use some specific indicators and are they agreed to by the licensees? MR. JACOBSEN: I'll answer it a couple of ways. First of all, we have some indicator and that is if we have findings we do run those findings through the SDP so we have either so many green issues or so many white issues. That's a very crude indication. MR. BONACA: Okay. MR. JACOBSEN: The second, I guess, answer to that is every licensee has their own set of indicators that they're using to measure their programs. The problem is that eery one of these programs is different and every licensee has a different set of indicators with different thresholds. The third answer is we understand it would be a big improvement if we could develop some more objective ways of assessing these corrective action programs. Because our assessment right now is largely qualitative and not quantitative. So we do have a task group that we're working towards and it may not be performance indicators as we think of them today, but we are looking at developing a more objective way of assessing the corrective action programs. And if we were to come up with indicators, we would have to get industry to buy in. It gets back to the question that we raised, how much burden do we want to add for what gain? We might have to develop site-specific thresholds, for instance, and then you have to validate the indicators. MR. BONACA: But typically, you do have some -- like threshold level, is it low or high? And you have some way of -- agreed to by the industry. I mean I've seen, I can go from one site to the next and I've been there looking at corrective action programs and I can see they all speak the same language, pretty much, because there is a lot of shared information today. The other one is categorization. Okay, what do you lump into category 1, 2, 3? Do you have the right percents distributed there? What is the time of response? I'm just pointing out that maybe, by now, there is more consistency among the programs than not. MR. JACOBSEN: Well, they're becoming more consistent and we're -- and the industry has done some work in this area and INPO has some inspections that they do. Nobody has been able to come up with any joint performance indicators. MR. BONACA: True. MR. JACOBSEN: But we're looking and we're going to continue to look at that and the types of things you mentioned are good. WE do have our procedure broken down into areas that we look at. We look at threshold and we look at prioritization and we specifically have attributes that we look at in each of those areas, but to take those and quantify them is a whole -- I know two plants, one that has 10,000 items they put in their corrective action program a year, and one that has 1,000 and they both may work real well. It's just how those two programs are managed. It's very hard to say to somebody you need to have so many thousand items in your corrective action program or your threshold is not low enough. You don't want to do that. You have to be real careful. MR. JOHNSON: John, did you have anything you wanted to add? I know you like to talk on this topic. MR. COE: Only that your comment is a very good one and it's been one that I know that I've been thinking about a lot for several years, because the process of these inspections is as Jeff indicated, very qualitative. At one point, as -- in my previous existence as an analyst, I actually went out and tried to do some more quantitative look at corrective action programs by taking the current open issues and gauging them according to their functional impact and then also gauging them in accordance with their risk importance, and then essentially combining those two elements for that each item to come up with kind of a composite list of those issues which were both functionally, had functional impact associated with it and had risk significance. And that might be one way of assessing whether or not the licensee is applying the correct priorities, okay, and investing the right level of resources, if they're grading their resources in a manner which makes sense from a risk standpoint. In addition, there's a question out there that could be asked about what about the accumulation of lower level issues that in the risk kind of sense combined together, synergistically, to provide a greater risk impact than each one looked at individually. So these are the kinds of issues and the kinds of questions you're raising are very good ones and they're ones we've been thinking about. MR. BONACA: Well, the reason why I raise this is also some, for example, some licensees are more aggressive because they have been having problems and they tend to do more cause analysis. Others, who believe they are very good or they believe that, they tend to say we do too much and so they go now to apparent cause in many more cases because there is some kind of complacent setting. You'd be surprised how the first type of individual finds more things. Therefore, you tend to say he has more problems. And the other one doesn't find that much because he does all the apparent causes and very few causal evaluations and on the surface he has less problems. And so you tend to think the other guy is better off. I've seen these cases and compared them and you are surprised on how you can get truly the wrong conclusion. And so that is the point I was making. Maybe there are some indicators that can be determined to help in that process because I think it's such an important area. MR. JOHNSON: Very good. I think it's a good point. The last thing I would point out with respect to that is that we raised this issue with the industry, continue to raise this issue with the industry and the last time we raised it with the industry, you might appreciate that the industry doesn't feel like we need to do more with respect to performance indicators particularly in this area. MR. BONACA: Somehow I'm not surprised. (Laughter.) But they don't mind that you are looking into it, right? They can't do anything about it. That's the fundamental area of inspection. MR. JOHNSON: That's right. CHAIRMAN SIEBER: I guess I'd like to ask if any other Members have questions or comments that they would like to make at this time? MR. SHACK: I guess just the one I'd make is it seems to me like such a fundamental area, that is one you wouldn't want to back off on the inspections and that's always the price that industry is looking for. Yeah, we'll give you a PI if you back up. But this certainly seems like about the last inspection you want to back off on. MR. JOHNSON: Yes, absolutely, and that's why I was careful to say we don't believe that we're backing off on PI&R. We think that what we're putting in place is a more effective PI&R and that's really the focus of our changes in that particular area, although I think there is a net decrease of 25 hours a year or something. MR. JACOBSEN: Yes, it about 5 or 6 percent. That's on paper anyway. What actually gets implemented is -- CHAIRMAN SIEBER: Actually, it seems to me that the number of modules in their rigor has increased under this new program from what it was before which has pluses and minuses and the pluses, of course is more directed inspection and the minuses, that there's less abilities for the region and individual site inspectors to use their discretion to respond to special situations in the plant. And I guess that as you gain more experience in the inspection process, you'll be able to judge whether the balance that you now have is appropriate compared to something more akin to the past practice which seemed to have more flexibility in it than the current system. MR. JACOBSEN: Actually, the change we're making in PI&R responds to that very comment, that one of the regions felt strongly about having this part of the program where we could look at things in a more real time basis than they thought. So rather than doing it as a team once a year, we're going to pick these things throughout the year. So that actually responds to that flexibility question. So we are looking at that and making changes as appropriate. CHAIRMAN SIEBER: Right. Okay. Any other questions or comments? MR. LEITCH: I'm still just perhaps a little confused about the expectations for the predictive nature of this reactor oversight program. If you had a hypothetical plant that was running along with basically green performance indicators and no color inspection findings, and then it's had a track record of that for several months, a year and then you come through some self-reviewing event, you find that the plant has a lot of problems and winds up in a regulatory shutdown, would you be disappointed with the reactor oversight program or would you say well, this is not a predictive program, we had no way of knowing that? I'm still groping for what the expectation is there. MR. JOHNSON: Yes. If we saw a plant -- we as an Agency, we constantly look at these situations and we do a lot of hand wringing and soul searching and we try to make decisions about whether the process, the performance that results is a process failure. And if I saw a plant that was in the licensee response band that ended up in the degraded cornerstone corner, that doesn't mean that we've had a programmatic failure. Now in our self-assessment matrix we look, we will look, we continue to look at jumps in plant performance across multiple columns of the action matrix to see if there was something that should have been in the process that was not in the process. But the process hasn't failed because again, we haven't built a process that we guarantee predicts that kind of thing. If you tell me, if you're painting a picture of a plant that was in the licensee response band today, that tomorrow we have to issue an order to remain shut down, that is their performance is unacceptable, then yeah, I think we have to really step up to the plate and talk about whether we need to do something drastic with respect to the program. MR. LEITCH: I mean admittedly, I have not seen such a thing. I'm not saying such a thing exists. I just don't understand your expectations. Thank you. CHAIRMAN SIEBER: Any other questions? Since there are none I would like to comment to you, Mike, and to all the speakers today that I think you have been very responsive to the questions that we asked. I thought your presentations were well prepared. And I think that you're on the right track, but you've only been in this business for a short time and I'm sure you're still in the learning process and as time goes on you, for sure, will make some adjustments in what you're doing today, but it just seems to me this is a step forward and I want to thank you for putting in the time and effort to give us good presentations and well thought out responses. So with that I think we can conclude, unless anyone else any comments or statements to make. We can conclude with today's meeting and again, thank you very much. MR. JOHNSON: Thank you very much. (Whereupon, at 2:40 p.m., the meeting was concluded.)
Page Last Reviewed/Updated Tuesday, August 16, 2016
Page Last Reviewed/Updated Tuesday, August 16, 2016