471st Advisory Committee on Reactor Safeguards (ACRS) - April 6, 2000
UNITED STATES OF AMERICA NUCLEAR REGULATORY COMMISSION ADVISORY COMMITTEE ON REACTOR SAFEGUARDS *** 471ST ADVISORY COMMITTEE ON REACTOR SAFEGUARDS (ACRS) U.S. Nuclear Regulatory Commission 11545 Rockville Pike, Room T-2B3 Two White Flint Building North Rockville, Maryland Thursday, April 6, 2000 The committee met, pursuant to notice, at 8:30 a.m. MEMBERS PRESENT: DANA A. POWERS, ACRS Chairman GEORGE APOSTOLAKIS, ACRS Vice-Chairman THOMAS S. KRESS, ACRS Member MARIO V. BONACA, ACRS Member JOHN J. BARTON, ACRS Member ROBERT E. UHRIG, ACRS Member WILLIAM J. SHACK, ACRS Member JOHN D. SIEBER, ACRS Member ROBERT L. SEALE, ACRS Member GRAHAM B. WALLIS, ACRS Member ALSO PRESENT: JOHN T. LARKINS, ACRS Executive Director MEDHAT EL-ZEFTAWY, ACRS Staff HOWARD J. LARSON, ACRS MICHAEL T. MARKLEY, ACRS Staff NOEL F. DUDLEY, ACRS Staff PAUL A. BOEHNERT, ACRS Staff SAM DURAISWAMY, ACRS Staff CAROL A. HARRIS, ACRS/ACNW Staff PATRICK W. BARANOWSKY, NRR STEVEN E. MAYS, NRR . C O N T E N T S NUMBER PAGE 1 OPERATING EXPERIENCE RISK ANALYSIS BRANCH PROGRAM OVERVIEW 308 2 COMMENTS ON APPENDIX 2.B. "STRUCTURAL INTEGRITY SEISMIC LOADS" 380. P R O C E E D I N G S [8:30 a.m.] CHAIRMAN POWERS: The meeting will now come to order. This is the second day of the 471st meeting of the Advisory Committee on Reactor Safeguards. During today's meeting the committee will consider the following: special studies for risk-based analysis of reactor operating experience; a report of the Materials and Metallurgy and Thermal Hydraulic Phenomena Subcommittees, future ACRS activities, report of the Planning and Procedures Subcommittee, reconciliation of ACRS comments, recommendations, proposed ACRS reports. The meeting is being conducted in accordance with the provisions of the Federal Advisory Committee Act. Mr. Sam Duraiswamy is the Designated Federal Official for the initial portion of the meeting. We have received no written statements or requests for time to make oral statements from members of the public regarding today's session. A transcript of a portion of the meeting is being kept and it is requested that the speakers use one of the microphones, identify themselves and speak with sufficient clarity and volume so that they can be readily heard. I do want to bring to your attention that subsequent to our discussion of the spent fuel pool accident risk for decommissioning plants on Wednesday, April 5th, 2000, the Nuclear Energy Institute has provided comments on this matter via an e-mail to Dr. El-Zeftawy. Nuclear Energy Institute comments have been distributed to the members this morning. They will be made part of the record of our meeting. CHAIRMAN POWERS: I also will remind members that it is time for your ethics training and we have set it up so that you can quickly go down, grab some lunch, come back up here, and Mr. Szabo will come in and give us the ethics instructions that we need to keep ourselves on the right side of the legal requirements. That does mean, however, that I am going to be holding schedules fairly tightly in today's presentations and discussions, so that we can meet, as scheduled, Mr. Szabo. Do any of the members have comments they want to make before we begin today's proceedings? DR. SEALE: Have we seen that e-mail from NEI? CHAIRMAN POWERS: You have a copy in front of you. DR. SHACK: About an eighth of an inch thick. DR. SEALE: Oh, good. CHAIRMAN POWERS: Seeing no additional comments, I will move on to the provisions of the agenda. Our first topic is special studies for risk-based analysis of reactor operating experience. Dr. Bonaco, I believe you will lead us through this? DR. BONACA: Yes. Good morning. We met with the Operating Experience Risk Analysis Branch on December 15th, and we had an overview of the program that they have put forth and we heard about the improved availability of different sources that is a major improvement in the availability of information that comes and is being used by this branch now. We also heard about a number of the activities that they support with this information. We believe that meeting and its presentations were very informative. We also believe that some of the programs that these activities support are very important to the future of the agency, particularly to risk-informing the regulations and because of that we invited the branch to come and give an overview to the whole committee. We asked them to put particular emphasis on the activities they support, although it is very important that we understand what the databases are and how they gather the information. It is also very important to understand who the users are and the community they support and activities they support and I believe that the presentation this morning will be focused on those items, so with that I will let Mr. Baranowsky to give us an overview. MR. BARANOWSKY: Okay. For the record, I am Patrick Baranowsky, Chief of the Operating Experience Risk Analysis Branch, and I have Steven Mays here, who is the Assistant Branch Chief, and thank you. We did give our presentation to the subcommittee in December on this, and it was about a half a day long, so I have cut this down and tried to put some more information in here on some of the uses of the work that is conducted in my branch. The first viewgraph I have here talks about the purpose of this presentation, which is to give you the overview of the activities, discuss our role in the regulatory process, and provide some typical results of the kinds of things that we finally performed, this risk-based analysis of reactor operating experience. Why don't we just go on to the next viewgraph? Yesterday we talked about risk based performance indicators and what we have in front of you here is a chart which shows how all the activities that are conducted in my branch are organized logically and hierarchically, and information from one set such as the data flows into analysis of special areas where we do industry-wide analyses, such as system reliability, component and initiating events. That information can then be used for plant-specific type analyses with some enhancements or it supports things like the Accident Sequence Precursor Program methods and models, and all of these things can be pulled together and have been pulled together as we discussed yesterday as elements or at least learning exercises, if you will, as to how we might go about constructing plant-specific, risk based performance indicators. One other point I want to make is that in addition to each of these elements providing information that flows up this chart. There is also a horizontal utilization of the information at each level as we go along here, so various NRC activities that are interested in the more raw data that might come out of either LERs or the EPIX system, which we will talk about in a minute, can have access to those data sources, and they do. We get many queries for each of those data systems, plus the industry-wide analysis have results that in and of themselves are important that get fed into the regulatory process either for generic issues or for risk inspections and things like that, and of course the Accident Sequence Precursor Analysis Program provides insights on the most significant events that occur, some of which result in fairly immediate regulatory actions or they could result in information notices after some deliberate study, and then finally we have the risk based performance indicators, which we discussed yesterday. DR. APOSTOLAKIS: Bob, the box there says operator error probability studies. MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: We haven't discussed this, have we? MR. BARANOWSKY: Correct, and that is more of a place-holder than anything else. There is another branch that has the primary responsibility for the operator error studies. We put it in there just for completeness and we tried to factor in whatever we learned from other sources on that into these programs, but I guess it's stashed in this particular block to show that we are -- DR. APOSTOLAKIS: I don't recall it being part of the Human Performance Program we saw yesterday. I mean that is the other branch. MR. BARANOWSKY: Yes, that's Jack Rosenthal's branch. DR. APOSTOLAKIS: Jack Rosenthal's. I don't remember. CHAIRMAN POWERS: My recollection is he had something on this subject, but it seems to me you also made the point that the latent errors outweighed the active errors by four to one. DR. APOSTOLAKIS: Right. CHAIRMAN POWERS: And I am wondering if this -- DR. SEALE: At least. CHAIRMAN POWERS: I mean it is labelled operator error. What about the latent errors? MR. BARANOWSKY: Okay, that is a good point. We can pick that up right noww. We think that the studies that we perform on reliability of component systems, initiating events and common cause failure capture the human error input to the availability of those functions or the likelihood of those initiators because that is just a causal factor and what we end up doing is collecting the data and performing analysis that allows us to organize that information in terms of its impact on systems, so it doesn't show up as a human performance analysis per se, but we know that it is an important contributor and it shows up in our various system and components studies. CHAIRMAN POWERS: The guys that are doing -- looking and trying to model human performance need a database on latent errors as much as they need one on active errors, don't they? MR. BARANOWSKY: Yes, and I would say we are not producing that database and that is probably being done by the other branch, but if they want to have a perspective on the risk significance of those errors and how they fit in to impact on safety functions, they can go and look at any of these results, and that is what I mean by a cross, horizontal utilization of this particular information. DR. APOSTOLAKIS: So you are really proceeding forward -- MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: -- given that there is an error, what happens next. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: And the other branch will investigate the causes of the error. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: Or the latent part. MR. BARANOWSKY: Right. If you want to make corrections to the root causes, then that is the level that you have to work on, and what we try to do is provide the risk perspective as to what is important. So there's really two aspects of it. Once you know what is important, then you can spend your resources making the fix. I would not want to go the other way around, which is the traditional intuitive way. DR. SEALE: Are you going to tell us what you are going to do or what you are doing to validate or verify the breadth of applicability of your SPAR models? MR. BARANOWSKY: I will talk a little bit about it. I wasn't going to cover it in too much detail because it is just one fairly modest piece part here, but I can talk a little bit about it. DR. SEALE: But it is kind of crucial. MR. BARANOWSKY: Yes. We have some activities ongoing. I will go over that. DR. SEALE: Okay. MR. BARANOWSKY: It's a little later on. We do have the fundamental mission of performing this analysis and when we do it, especially our intention has been over about the last year to make sure that our work supports the four safety goals of maintaining safety, improving regulatory effectiveness and efficiency, reducing unnecessary burden, and improving public confidence. Now originally I looked at, oh, gee, let me see if I can pull out specific sub-bullets on each one of these things, but what I found was that most of the activities that we do support these things across the board. As an example, when we find what is important and insights on incidence or in special studies, that information could go towards maintaining safety by providing, say, an information notice that would go out to licensees on the insights. It could affect regulatory efficiency by providing insights on how to focus either a risk informed inspection or determining if a generic issue that has been or is being implemented is showing improvements that were forecast when the issue was claimed to be resolved, for instance, and we have done those kinds of analyses on station blackout or ATWS and things like that. Reducing unnecessary burden -- again, if you focus, as I said earlier, on the important factors before you go into the root causes, that is sort of the optimum way of determining how to expend your resources on things. Lastly, I think, improving public confidence -- we do provide a relatively independent cut, what we think the operating experience is showing. We also take a look at how the operating experience supports or has differences with respect to licensee PRAs, and we have several examples in the studies that we have done on systems, initiating events, and also on the Accident Sequence Precursor Program. CHAIRMAN POWERS: We have an industry that is producing a variety of analytic tools to assist them during periods of low power operation and shutdown, particularly shutdown, configuration analysis and what-not. We have a large number of events that take place during those periods of shutdown. Do you do comparisons between those models and these events? MR. BARANOWSKY: There aren't too many shutdown models -- CHAIRMAN POWERS: Well, industry seems to have a proliferation of them. MR. BARANOWSKY: Okay. Well, I am not too familiar with them -- DR. SEALE: Maybe that is the reason he asked the question. CHAIRMAN POWERS: Maybe. MR. BARANOWSKY: If they have a proliferation of shutdown models, I am not familiar with them. CHAIRMAN POWERS: They've got EOS and they've got, what is it? -- ORAM. MR. BARANOWSKY: Oh, okay. CHAIRMAN POWERS: And they have got -- DR. APOSTOLAKIS: Sentinel. MR. BARANOWSKY: Okay, I'm sorry -- CHAIRMAN POWERS: San Onofre -- MR. BARANOWSKY: -- these are the risk management models. I am thinking of something different like a shutdown PRA. CHAIRMAN POWERS: Well, Seabrook claims to have a shutdown PRA. South Texas, SONGS. MR. BARANOWSKY: I am familiar with the software packages that you mentioned and that they are using those for risk management type decisions. DR. SEALE: And apparently very effectively, and it would be nice to know are there special elements in those packages that contribute to that effectiveness. CHAIRMAN POWERS: Well, I wonder how effective they are because I seem to see an awful lot of incidents occurring during low power and shutdown operations. MR. BARANOWSKY: That is an interesting point. That's one that we haven't looked at that I know of. MR. BARTON: I think what we are seeing is the human element of it. DR. SEALE: Yes. MR. BARTON: You look at configuration and the defense-in-depth and the risk analysis. That seems to all be in place, but then somebody goes and screws it up, and that is what we are seeing, I think, in the shutdown. DR. BONACA: I think what we are seeing is also the acceleration of the shutdown for refueling activities. That is really where the challenge comes from the human factor in many ways, and that challenges any ability of predicting. MR. BARANOWSKY: We do have a couple of things that relate to that area. One is that we still are doing Accident Sequence Precursor Analysis including shutdown events. The Wolf Creek blowdown event was done under the Accident Sequence Precursor with insights from the shutdown there. There was a similar event that occurred recently at Waterford which we are analyzing. In addition, we have been pushing in the EPIX database to get unavailability information on key components and stuff at shutdown as part of the EPIX database. So we're putting that kind of stuff in place, and we recently did an analysis of the kinds of events that involve loss of offsite power, loss of heat removal, loss of level control in events for NRR for their uses in their shutdown significance determination process model. So we are involved at that level of trying to gather the information, put information available for people to do that, but our Branch isn't doing the development of shutdown models in that case right now, although we do have an ongoing task in SPAR model development, which Pat will talk about later to try to figure out what kind of SPAR models we need to have for our regulatory uses. So we are involved in it. DR. APOSTOLAKIS: How do you do the accident sequence precursor analysis for a shutdown event if you don't have a model? MR. BARANOWSKY: We develop models on the basis of what the particular event is on each case that comes up, and we use information that -- for example, there are shutdown models that were put together for the shutdown rulemaking that was done, and we used that and general PRA practice to develop the strategies to deal with those things. We documented those in the Wolf Creek analysis, and there was one documented in the Vogel analysis when they had the offsite power mid-loop event, so we do it on a somewhat ad hoc basis, but we still do it. DR. POWERS: How do you -- you construct these ad hoc models, and I know you're extremely skilled, but surely you're not the only people in the world that can produce error-free models on a demand basis. What I'm struggling with is, we have so much trouble with getting risk models that have been around a long time to be accurate and have great fidelity to the plant. How do you do it on an ad hoc basis and have any assurance that the product you get is -- or the predictions you get out of the model are of useful quality? MR. BARANOWSKY: The reason that it works for us is that the models are very limited. We don't have to go and search for all the possibilities and put them in some sort of pecking order, which is what you do when you're developing the full-blown PRA. We already have enough information that tells us what functions and systems have been impacted. And then we ask ourselves, what are the ways that one can find success paths beyond that? It's a much simpler problem. It's an order of magnitude simpler, to be honest with you. DR. SEALE: But aren't you flying blind, in a way? The utility is using some sort of management program, ORAM or something like that -- MR. BARANOWSKY: Right. DR. SEALE: -- when it does its planning. It has a Wolf Creek or a Waterford or whatever, and they used something like that, I assume, in leading up to that and somehow there was a failure. Somehow you didn't catch that particular problem and they got themselves into the corner where that happened. It strikes me that that's a whole ensemble of questions that ought to be followed pretty carefully, if you really want to find out whether or not these management models are being useful. MR. BARANOWSKY: I think we're not really making the assessment of whether the management models are useful. That's a point that is probably worth us thinking about, because when we look at the full power PRAs, we say to ourselves, are they capturing what they're seeing in the operating experience, or not? In many cases, we find that at least on the failure mode level, things that are not being captured, and sometimes at a higher functional level. When it comes to the shutdown models, we have less access to the specific types of models that they are using. ORAM and RISKMAN and all that stuff, those are tools. When I say model, I mean the model for Waterford that has all the logic in it. We don't have their shutdown model that they're using, if they are processing it with RISKMAN or whatever. So it's a little bit more difficult for us to make the comparison of the shutdown analysis that we do with the way that they did their analysis. Probably the only way that that could be done right now is if an augmented inspection team went out and we asked them to look at the model and compare it to what we found in our own risk analysis. It's an interesting thought. I just hadn't -- I don't think we thought about that before. DR. APOSTOLAKIS: As a point, I don't think that the PRA would have included in it, what happened at Wolf Creek. It's something that normally we don't investigate, and this Committee, in a letter in the past has asked the ATHENA people to think about how normal operations can lead to initiating events. DR. SEALE: Well, yes. DR. APOSTOLAKIS: This particular event, I don't think would -- DR. SEALE: Well, the Committee has recognized the fact that the big problem with shutdown operations is that they are so diverse that it's impossible, practically, to have a stylized system like a PRA that would cover everything. And yet there's a regulatory responsibility that exceeds the scope of the models that are being used, and I think this is a legitimate set of questions that ought to be asked. DR. BONACA: One other thing, at least for some examples I have seen of some events that took place, was a typical event was, they predicted what was supposed to happen, and then changes were made to the schedule just at the last minute. There were a lot of changes of those schedules, and there was even some assessment from some PRA person that said, well, it looks okay. But there wasn't a full quantification or an evaluation of new possibilities introduced by the new schedule, and that's really what came out. And so that's why they couldn't predict what happened, because -- DR. APOSTOLAKIS: Again, it depends on what you mean by "what happened." MR. BARTON: You have an event, is what he's saying. DR. BONACA: You have an event in the original evaluation with the timing and the activities that were planned. It couldn't have happened. But then because there were changes to the schedules and to the activities and they were put in different orders -- DR. APOSTOLAKIS: I think that the way PRA is used in managing the shutdown risk is really to prohibit or to not allow certain configurations. I think it's a high level use. It doesn't go down to details like how this was initiated. And I don't think the state of the art allows you to figure out some of these, to predict some of these things -- some of these, but some others are there. It's really a combination of defense-in-depth ideas, and PRA insights. Basically what they are saying is that if I'm in this configuration, do I have alternate means of achieving certain functions? And some of the insights come from the PRA, others from just experience, defense-in-depth and so on. I don't think there is any detailed analysis. But even that is very, very useful. There is no question about it. MR. BARANOWSKY: I may be jumping the gun a little bit here, but since we're on this, I might as well get to it. You know, we are finding that 15 or 20 percent of the accident sequence precursors don't look like what we see in the PRAs. But that the implied core damage frequency is about what we would expect. DR. APOSTOLAKIS: I mean, you have to be careful when you say that. MR. BARANOWSKY: I am careful. DR. APOSTOLAKIS: Parts of it are not in the PRA. MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: Typically what I think is not in the PRA is the detailed manner in which something evolved, like take TMI, for example. I mean, ultimately it was a small loca, right? But the details of how that happened, the valve getting stuck open and so on, was not in the PRA. But at some point, you do get into the PRA. Isn't that true? MR. BARANOWSKY: Sure. But the point is this; that you can't make detailed procedural corrections on some of these things that you can't forecast very well. MR. BARTON: Right, true. MR. BARANOWSKY: That means you're going to have some residual level of risk and error, and the question becomes, is that an acceptable thing? I think we have evidence that will probably add a reasonable residual level of risk. DR. APOSTOLAKIS: Yes. DR. BONACA: What I meant to say before is that if you look at a shutdown, typically the issue you are dealing with is the removal and restoration of certain systems from service. The order in which you do that is a fundamental issue, because it allows us to assess, in fact, what equipment you have available, and what equipment you have to restore first in order to cope with certain conditions. Now, if you have pressures on the outage that will force this order to be continuously changed, and we seek this more and more with the faster shutdowns taking place, then you are challenging the ability or predicting because you are just simply -- then there are these screw-ups, as we call them. But really they're not screw-ups; they're the consequence of not allowing people the proper process, and it becomes very hard to make those predictions. MR. BARANOWSKY: Okay, the next chart I have is -- DR. SHACK: You can come back tomorrow if you finish this one. MR. BARANOWSKY: I'm going to get through on time. This is some of the uses that these programmed elements -- some of the things that these programmed elements are used for. I took the programmed elements and put them on the left-hand side. Originally, we were going to try to take each one of these uses on the right side and say, well, which of these program elements are used there? And then we realized, they're pretty much all used, when we started going through it, so I had a redundant chart put together. And one way or another, all of these program elements on the left side of this chart -- data sources, reliability studies, common cause failure, ASP, risk-based performance indicators -- fit into a number of these applications that are listed on the right side. Risk-informed inspections, that's an area that we're working pretty hard right now with NRR to get common cause failure, system and initiating events insights into what they're inspecting when they do the risk-informed inspections. The operational events: NRR has, for instance, a daily cut at what's important, and so does the operations center, and we work both with them, and, in addition, provided tools for them to perform those analyses. They review various licensee applications. Many times, they will go to the system studies or initiating events, and see what it says about that plant versus what the licensee's claiming on that thing, and really so on and so on as I go down here. It's really pretty much the same; they're looking at either the methods or the insights that are derived from each of these areas for application to analysis or for judging whether an application by a licensee has covered as best we understand today, the insights from the operating experience. And so this is sort of the list that we were able to come up. DR. BONACA: Just one note, during the presentation in December, you showed us a SMUG group. MR. BARANOWSKY: Yes. I'm going to cover that. DR. BONACA: I think that would be interesting to the Committee to have an actual owners group there that is driving the -- MR. BARANOWSKY: I actually have these kinds of groups on a lot of the activities that we undertake when we're trying to produce a product that has usage in NRR and the Regions. We get someone from NRR and usually a representative for two or more of the Regions to get on some sort of a users group. And the philosophy is that they have to define what they need before we can produce it. It's like ordering a car; you've got to tell me a little bit about it first. Okay, so the first level of risk-based analysis of operating experience elements that we have is the operational data. And I have three of them highlighted here because they're the primary three that we use: The sequence coding of search system captures LER information. There are over 47,000 LERs coded in there. And it's a pretty significant source of information for us, NRR, our contractors, and it's fairly trustworthy in terms of the quality. It's been maintained and coded consistently, and licensees report information in a fairly consistent manner. But as you know, it's more at the system level or higher level of events and it doesn't capture component and train-like information. So, the reporting system that used to capture that was called the NPRDS, the Nuclear Plant Reliability Data System, and that is now defunct. And replacing it is the Equipment Performance and Information Exchange System that INPO is developing. It's been under development for a few years. It's actually operation now, but I wouldn't say it's at a high enough quality level that we could go in and extract the data and use it primarily because of our concern about the completeness of the data, and maybe some errors in the way it's been coded in there, but mostly the completeness. MR. SIEBER: Did the NPRDS old data go into EPIX? MR. BARANOWSKY: No, the old data has been archived and can be extracted using software that's available, but this is quite a different system, and it would cost a horrendous amount of money to backfit it. MR. SIEBER: Okay. DR. POWERS: If I wanted to know something about the frequency of fires at a particular site or industrywide, which of the datasets do I go to to interrogate? MR. BARANOWSKY: Okay, we have some special studies. I'm not sure how that shows up on here, but we did -- have done one on fire, and we're probably going to update that. There is also some proprietary fire data at EPRI, I believe, which sometimes we negotiate to get them to let us use, which includes information of lower significance incidents at plants than are normally reported through LERs. But through LERs, we can get all the fires that meet the reporting requirements, which I think are the ones that are greater than ten minutes in duration where the affect the safety function in some way. And those we have available and update -- I think we do it in our initiating event report, right? MR. MAYS: Yes. MR. BARANOWSKY: Yes. DR. POWERS: The people that do fire risk analysis are put in the position frequently of having to extrapolate a database to get a fire big enough to pose some threat to the plant. Is the database that one derives from the LERs, suitable for that extrapolation, since it reports only fires that meet a reporting criterion, missing some kinds of fires? So I'm wondering if that affects an extrapolation. MR. BARANOWSKY: You can get the number of large fires, but as you said, there aren't that many, and certain ones that affect numerous or even more than one train of safety systems are rare or probably nonexistent. MR. BARTON: Really rare, yes. MR. SIEBER: Yes. MR. BARANOWSKY: I guess that's a good insight that we're getting, and that is that some of the Appendix R protective features seem to have worked in reducing both the frequency and the consequence of fires, at least that's the result of the study that we completed a couple of years ago showed. I don't know that there has been any indication that it's changed since then, although there has been interest in folks in NRR for us to update that work, people that are working on new fire protection rules and so forth. DR. POWERS: It seems to me that we're getting a mixed message here. When I look at some of the preliminary information that has been released on the IPEEEs, I see fairly -- what strikes me as surprisingly high core damage frequencies, given that I have Appendix R and all this restriction, in fact, despair of ever seeing a fire big enough to affect more than one train. But it just doesn't seem -- the two results just don't seem to square with each other. MR. BARANOWSKY: Yes. We've talked about trying to take the fire data and comparing it with the IPEEEs. I think when we first started doing this, we didn't have all the IPEEEs, for one thing. MR. MAYS: Plus, the bigger thing we found was that there weren't so many differences between what we were reviewing in the fire frequencies; the big differences are in the assumptions about the ability to detect and suppress before you get to a big enough fire to have the adverse consequences. So we don't have a lot of data on detection and suppression capabilities, so our ability to compare operating experience to what's in those PRAs is very limited by that. And I think that's where the big gap is, quite frankly. If you look at a fire PRA, it compares the frequency of the fire, the probability of non-detection and suppression, and the probability that the remaining system is not affected would work. It's that middle block where there is the limit on our operating experience capability to review. DR. APOSTOLAKIS: There is one more, although I agree with you that that's a major one. But also the probability that the fire will become large enough, that's something that -- MR. MAYS: Yes, the loading in the area is the deterministic part of that. DR. APOSTOLAKIS: Yes. MR. MAYS: The thing we've seen is that at the larger scope level, the bigger picture, since other than the Brown's Ferry fire, we haven't had a fire that both causes a trip and takes out more than one train of anything. We haven't seen that. We've seen a couple of fires since then that would cause a trip and take out one train, but not two. So at the high level, from a performance standpoint, we're not seeing degraded performance at that level. What we're seeing in the IPEEEs, I think, is cases of particular plant configurations that are more susceptible than others. For example, the IPEEE on Quad Cities had a major fire area where feed pumps, which have large oil reservoirs, if a fire were to start in that area, there was also cabling in that area that would affect offsite power, that would affect DC power, that would affect AC power and would affect HPSI and RCSI. Well, that's kind of a plant-specific configuration thing for which operating experience data is not likely to be able to detect, and that's the reason why you need to go do a fire vulnerability study in the first place. DR. POWERS: I guess my point that I'm making in an elliptic fashion is that I think this is a part of basic data, more boxes that need to show up at this level here, and not only the database that you mention on detection and suppression. I think there's a database that's needed on fire effects. And I think we've got extremely conservative approaches to that which say that essentially a fire in a fire area takes out anything that you might want in the area in the worst possible way. MR. MAYS: That's the assumption. DR. POWERS: We just don't have a lot of data to tell us about that sort of thing, whereas I see a proliferation of data on fire frequencies. There is an insurance industry one; there's one sitting up in International, and I don't see people setting up databases that say things like what a fire affects. I see people struggling with what you've talked about, suppression and detection, because you don't want to start the clock on those things. MR. MAYS: It's the denominator problem. DR. APOSTOLAKIS: I really don't know what that database would be. MR. BARANOWSKY: Well, I'm not sure that we're the right people to put the database together. I think that one of the points I would like to make is, remember, when we talk about EPIX, we're not putting the database together. Actually, the main tool that we have is the next one that I was going to talk about, which is the Reliability and Availability Data System. We pull information out of these databases. I mean, we're the primary contact, maybe, with INPO for access to EPIX, but they're the ones that are designing it and filling it. We're just saying we need these pieces and parts of data, and we pull that out and put it into our Reliability and Availability analysis system. DR. BONACA: But I had a question on that. Yesterday when you showed us the RBPI process, you showed us in the chart, a main new element. MR. BARANOWSKY: Right. DR. BONACA: And the element ran right though EPIX. MR. BARANOWSKY: Correct. DR. BONACA: In fact, I had a question on that because you're telling us that by next August or so you'll have already some deliverable, but now you're telling us that EPIX is not usable yet. MR. BARANOWSKY: Correct. That is an issue. There are a couple of aspects of EPIX that are probably worth noting: One is the business that it's not fully supported at this point by all the utilities. The second one is that it has some limitations in it that require us to do extrapolations and estimations that we think would be better done if they would provide the information directly to us. We're not getting all the demands for all the systems that we need in order to estimate the demand failure rates, for instance, and so we have to make some extrapolations. And it's the same thing for some of the down time on some of the equipment. DR. APOSTOLAKIS: Is it because they don't collect the information or they don't release it. MR. BARANOWSKY: No. The licensees collect all this information from what I can tell. They don't have it in the form that fits in EPIX. And what INPO is trying to do is put together some processors that will allow licensees to collect the information the way they do, but according to the definition rules that everybody agrees are correct, what's a failure, what's a demand, and then have it in any form they want and let the processor pull it altogether into a common form that's called EPIX. And if that can happen, that can be done fairly efficiently. DR. BONACA: In December, we talked about the need for having proper definitions of these terms, and also the importance for the industry to provide the information that you need for this kind of work. MR. BARANOWSKY: It's about a half-FTE per plant to support this kind of data need that we're talking about. DR. APOSTOLAKIS: Are you providing input to the INPO people as to what your needs are? MR. BARANOWSKY: Yes, there are two groups. There is a working group which we provide technical input to, that works with folks from the utilities who say they have needs for certain things, and we identify ours. And there is an Executive Oversight Group that Bruce Boger from NRR is the primary member and I'm the backup, I guess you would say, sort of the technical arm, who are looking over the Technical Working Group's proposals to make sure that they make practical sense. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: And that's ongoing over the next couple of months. We expect to come to some resolution. The important thing will be whether or not NEI and INPO can get the utilities to buy into making this a complete database. We could probably live with it without it being perfect, but it can't be one of these things where it's supported 50 percent by one utility, and 20 and 100 by another. It won't work. DR. BONACA: That's right. DR. APOSTOLAKIS: Exactly. DR. BONACA: But there is a wealth of information there, potentially. MR. BARANOWSKY: By the way, even if it's only partially supported, there's a wealth of information there. DR. BONACA: Right now, however, the fidelity of it, so far as -- MR. BARANOWSKY: Well, for doing quantitative analysis, it's a little difficult if the variation in reporting of information is very wide. I certainly couldn't make performance indicators that made any sense. DR. APOSTOLAKIS: Now, who decides what is a failure? Is it the licensee? MR. BARANOWSKY: Yes, but the big thing is definitions. We have common -- we have people working on definitions, sometimes page after page, depending on the type of system and component. DR. APOSTOLAKIS: That's the most difficult part of data collection. MR. BARANOWSKY: I know. People think data collection is just put it into a spreadsheet, but that's not it. DR. APOSTOLAKIS: So are you going to have a chance in the future to confirm that what they're doing is, in fact, reasonable? MR. BARANOWSKY: Yes, that's a good point. There are two ways that it will be confirmed. One is, INPO goes out to the plants themselves, and looks for consistency with the way they're collecting data with regard to the rules. The second thing is, if these performance indicators that we were talking about yesterday become reality and we use that data, then our own V&V activities will look at the fidelity of the data also. DR. BONACA: I thought that in December you committed to write the white paper that, in fact, Dr. Apostolakis was recommending, where you would identify some of these definitions, as well as the kind of raw data that has to be collected to -- MR. BARANOWSKY: Actually, we talked about writing a paper on reliability and availability, and we have started doing that and have got a pretty good cut at it. DR. APOSTOLAKIS: Let me give you an example now that I remember. Many, many years ago we were collecting data on fires. And at one facility there was a cabinet which switch gears that had had five fires within a few weeks. And finally the utility did a more serious investigation and they concluded that there was a common cause that was causing these fires, so they just replaced the whole thing. Now, what kind of data do we have here? We have zero fires, as they claimed, because they replaced it? In other words, the database should have nothing? We have five or we have one? MR. BARANOWSKY: The right thing is to capture the information in a database accurately, and don't confuse the analysis of the data with the factual. DR. APOSTOLAKIS: Let's say that something like that happens again. Will that information reach you or you will just see one fire in five weeks? MR. BARANOWSKY: These are the kinds of things that a technical group argues about. DR. APOSTOLAKIS: Okay, because that's an extremely important point, of course. MR. BARANOWSKY: I completely agree. DR. APOSTOLAKIS: And the typical attitude from the licensees, at least at that time, was that we fixed the problem so it can't happen again; therefore, it shouldn't be part of the database, which, of course, in a bigger scale we saw in the ATWS controversy with the German failure, the Kahl reactor. Was there one failure to scram or not? MR. BARANOWSKY: But if the data was collected in a factual manner, I wouldn't want to conclude that the data was wrong, as much as maybe the analysis or the inference from the analysis might be questionable. DR. APOSTOLAKIS: But you really have to go into the rationale sometimes. MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: Just the numbers don't do it. MR. BARANOWSKY: That's why these people need over and over. Okay, let me just mention one more thing about RADS. I think that's going to be a fairly interesting analysis system. Just by itself, it's going to provide some interesting component, train, and system reliability results of both generic and plant-specific nature. The analytics are pretty much done on that. What we're doing now is testing it, using what we know is slightly faulted data, but at least we're seeing if all the routines work on it. DR. APOSTOLAKIS: But you have seen a lot of PRAs, you and Steve, done by different people and so on. Overall, based on your vast experience in analyzing data, could you say that the distributions for failure rates that people are using in their PRAs are reasonable, consistent with operating experience? MR. BARANOWSKY: In fact, that's sort of what we're going to be showing you. We have two viewgraphs here. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: In fact, the next area that I wanted to talk about was reliability studies, but it's really this middle band of industrywide analyses and the first is reliability studies and initiating events, in which we're trying to use as much as possible, actual demands, failures, unavailabilities, analyze trends, quantify uncertainties, compare findings with what's in the PRAs, which is what you just said, and identify either plant-specific or industrywide insights to feed back into the regulatory process. And compare with regulations like station blackout and ATWS to see if the analyses indicate that the risks are as we've stated they were going to be in some of the backfit analyses that we did. So let's go to that chart on the Summary of System Reliability Study Results. These are reliability systems that we have performed reliability and unavailability analyses on, and what this chart shows of course is the name of the system and the dates from which we collected the data, the mean unreliability, which from a terminology point of view it's probably what we are referring to as unavailability before but it is like unreliability on demand, if you will. Whether the unplanned demand trends are going down or undetectable -- as you can see for most of them the unplanned demands are going down, and these are unreportable LER demands, okay? Failure rates -- whether they are declining, increasing or we couldn't tell, about half of them seemed to be declining in failure rate. Is there a trend in the unreliability? In most cases, we can't tell. Now one of the reasons we can't tell is this is not a data rich system that we are working with on the LERs, even though we spanned five years, so from a performance indicator point of view the LERS aren't going to really let us see plant performance changes in a timely manner, and we wouldn't be able to satisfy the criteria we talked about yesterday. DR. BONACA: But wouldn't you have to have, for example, a common cause failure of a HPSI system to have an LER? MR. BARANOWSKY: No. DR. BONACA: If you have an individual train failure, it's not being reported. MR. BARANOWSKY: No, but you probably get a reactor trip or -- DR. BONACA: I see. MR. BARANOWSKY: -- or a couple of failures reported, maybe one train failed and one had a degradation. How for HPSI, which is the BWR high pressure coolant injection system, it is a single train system so whenever it actuates or fails it is reported. MR. BARTON: Right. DR. BONACA: All right. MR. BARANOWSKY: But for auxiliary feedwater, that is another story. We have to go to places where they had initiation of auxiliary feedwater. Luckily we had quite a few of those, or maybe not luckily but -- [Laughter.] MR. BARANOWSKY: -- in the data we had quite a few and we could get a pretty good picture on auxiliary feedwater systems. Now in the interest of time I am not going to run through -- DR. APOSTOLAKIS: Let me understand this. Consistency with PRAs -- MR. BARANOWSKY: I'm sorry. DR. APOSTOLAKIS: -- let's take the HPSI. The PRA is three times lower than operating experience. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: Which PRAs are these? MR. BARANOWSKY: Several of them. DR. APOSTOLAKIS: Really? MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: And if they report a distribution, what do you mean "lower" -- the whole distribution is -- MR. BARANOWSKY: No, we looked at whether or not our mean was outside of their 90 percent bounds or whether their mean was outside our 90 percent bounds. We had pretty wide bounds, by the way. DR. APOSTOLAKIS: You are also saying that the failure rate is going down -- MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: -- with the operating experience. Is it possible that the PRA guys had tremendous foresight and eventually it will stabilize three times lower than the current operating experience? MR. BARTON: I doubt it. MR. MAYS: I think what we found in that case, George, was something that we do on all these operating event studies that was a critical thing. We looked at the unplanned demands, which were more like the real demands for these things to work under accident conditions, and what we found was that in some cases the test demands that were being done were showing a different failure probability than the unplanned demands were, and what the people at an individual plant were doing was using their test demands to figure out their probabilities because they didn't have access to all of the other information from all the other plants. So we have seen in some cases where we couldn't pool the test demands and the unplanned demands because they belonged to a statistically different population and that was probably the basis for why they were using the numbers that they were using and were coming up with lower failure probabilities in their PRA. DR. APOSTOLAKIS: Because I find that a bit surprising. The PRAs I am familiar with, the distributions were based on plant-specific data -- MR. MAYS: But I mean it's like the nature of the test demands were not producing the same kind of results as the nature of the unplanned events. DR. APOSTOLAKIS: Which brings us back to the earlier point -- MR. BARANOWSKY: That was true on diesel generators -- DR. APOSTOLAKIS: -- what is a demand and what is a failure are the key issues here. MR. BARANOWSKY: Right. The definitions are not the same so what happens is a licensee comes in and says this tech spec should be approved because I have got data that shows it is, and the NRC says well, I don't agree with you -- so what we are trying to do is bring this all together, so that the facts aren't argued about anymore. DR. APOSTOLAKIS: I wonder whether a statement PRA is three times lower than operating experience is really valid because there may be cases when this is not valid at all. I mean right now there is a difference. I think that is the accurate statement between your views and their views. MR. BARANOWSKY: Yes, I think that's true. That is a fair point. DR. APOSTOLAKIS: Because, you know, again there was Zion and Indian Point that I am very familiar with. They spent tremendous amounts of time and dabates and all that as to what is a failure and what is a demand and those are reflected in the distributions, so that would probably be unfair to those guys to say that. MR. BARANOWSKY: To some extent it might be, but to be honest with you, we've had some phone calls from utilities that we have talked to and they usually come around to our way of thinking. Moreover, this stuff has gone to the owners groups and we take the owners groups' comments and we factor them in. We don't poo-poo them, and now they are going and CE and Westinghouse in particular are trying to standardize amongst their groups, and they are using these reports as a baseline. DR. BONACA: And some of it will in fact have to do with what they call what, I mean if they count in a different way than you count? MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: Yes. DR. BONACA: It is going to give you different numbers, and that is again one of the fundamental issues here of people, what kind of counting they do. MR. BARANOWSKY: Well, if we can get more data from EPIX I think this problem will go away for sure. DR. APOSTOLAKIS: Now when you say failure rate trend is down, and yet on the left you say that the mean unreliability is .07, was that calculated using methods that assume a constant unreliability? MR. BARANOWSKY: I think so. DR. APOSTOLAKIS: So -- MR. MAYS: But we went back and tested -- DR. APOSTOLAKIS: What? MR. MAYS: We went back and tested over time on a year by year basis based on the data to see whether we were seeing a change in the unreliability associated with that. When we said the failure rate trend here was going down, we listed in that column all the failures that were reported whether or not those were the failures that were grouped together with the demand, so there's a little bit of a difference there. DR. APOSTOLAKIS: No, but my point is if you know that the failure rate is going down, maybe in your statistical calculations for the numbers that should be part of the calculations. MR. MAYS: Yes, we would go back and check to see whether that was a factor in the calculations and do that. We did that as part of the analysis. MR. SIEBER: Has anybody taken this reliability dataset and put it into something like a SPAR model to see -- MR. MAYS: Yes. MR. SIEBER: -- see what the change in risk would have been? What does the change in risk look like for various reactor types? MR. MAYS: Well, we haven't gone back and compared the various reactors at that level. We have used this information as input into the SPAR models -- MR. SIEBER: Right. MR. MAYS: -- and when we do things with the SPAR models such as accident sequence precursor analysis, those all go to the licensees for their comparison with their PRAs. We get information back from them about how well our SPAR models match up with their PRAs and where the differences are, so we found that the SPAR models have been pretty consistent with the plant PRAs but maybe for different reasons. Maybe we have higher failure probabilities in some systems but they have higher initiating events, so some of it is, you know, they may agree with the bottom line but agree for different reasons. MR. SIEBER: Yes, it comes out in the wash. MR. BARANOWSKY: Yes, but the important thing for us is to also have some insights that have to do with modes and causes so we are trying to be as careful as we can with the data available about whether or not we are getting phony insights or real ones. MR. SIEBER: This phenomenon is not introducing a systematic bias into the system of reported risk numbers, CDF type numbers, that the industry is using, right? -- just because of the data? MR. MAYS: We haven't done that level of analysis of all the PRAs. First off, these comparisons here were made against the IPEs. Subsequently there have been numbers of changes to the IPEs and to the way plants operate, since 8820 was put out. MR. SIEBER: Right. MR. MAYS: So this was just our first cut to say do we have a huge difference or a little difference, and what is the nature of it, so that when we deal with plants on an individual basis we will be able to focus on where the potential differences are. MR. SIEBER: I asked the question because you could have -- the Commission has a set of safety goals or safety goal policy. Maybe they would come up with a risk goal policy. The question is how reliable is the data in the models, both from the NRC standpoint and from the licensees' standpoint to be able to stand up against a risk goal policy statement. MR. BARANOWSKY: I think what we have seen here is that the PRAs don't have substantial differences from our own independent analysis looking at it with some different tools. A factor of three on this system, higher. I can go down here to diesel generator failure to run, which is important in the blackout sequences, and the utilities are using conservative failure to run rates. We have pretty good statistical information that says they are using very high failure to run rates, so if they have some that are higher, some that are lower, but generally with one or two exceptions on a system basis they are all pretty much in the ballpark. I am not sure that the insights are exactly the same. I know on auxiliary feedwater systems we didn't think they were modeling some of the suction path potential failure modes that could cause AFW to become unavailable. MR. MAYS: We saw a similar thing on HPI. When you get very redundant trains, the things that are going to dominate the unreliability are not train level faults. They are going to be things that go back to where the trains connect like the suction sources. We did see, for example, in the high pressure core spray system the PRA numbers and our numbers matched up pretty well, but their numbers were -- excuse me, in the isolation condenser -- their numbers were based on failure of the return valve to open, and our experience was -- the causes of failure was inappropriate isolation of the suction line, so while we got about the same number, we got completely different causes. DR. BONACA: Okay. We need to move on. We have a little bit less than 20 minutes. MR. BARANOWSKY: Okay, let me move to -- let's just mention quickly the initiating events, I think. Do you have that one? MR. MAYS: I've got it. MR. BARANOWSKY: So we did do a fairly extensive updating of initiating events that have been pretty much used by everyone in the PRAs and we found that, as you wouldn't be surprised, that the initiating event frequencies are declining pretty much across the board by about a factor of four to six lower than the IPEs and there are a number of them, over half I think, that show statistically significant declines in their frequencies, and that the risk significant initiators like the loss of feedwater and I think the small LOCAs and things like that -- MR. MAYS: Loss of heat sink. MR. BARANOWSKY: -- loss of heat sink, they were declining at a faster rate than the average, say, drop in reactor trips or something like that. The other thing that we did was take a look at the data on loss of coolant frequencies for small, medium and large breaks, and we looked at some worldwide data including work that was done by SKI, and working with our piping and fracture mechanics experts both at the labs and at NRC, concluded that the failure rates for pipe break type LOCAs, not transient-induced ones, are conservative, and we have now come up with a lower failure rate, which we think is still conservative but in light of the uncertainty it is about as far as we are going to go until I think we have a more extensive analysis by the piping and fracture mechanics people on this. DR. APOSTOLAKIS: But this is different from the other kind of work that you are doing in the sense that you also did calculations. MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: I mean you didn't base it on any experience or -- MR. BARANOWSKY: Well, we did actually. We -- DR. APOSTOLAKIS: What, a large LOCA? MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: The only experience is zero. MR. BARANOWSKY: But no, what we did is we took worldwide experience and working, as I said, with the piping and fracture mechanics folks, the whole power industry, if you will, we asked what is the applicability of this and how can we translate that information to use it as sort of a prior information, if you will, for the nuclear industry. DR. APOSTOLAKIS: That is not the same as collecting failure data for components. MR. BARANOWSKY: It is not the same, you're right, and if we did collect it like you said, well, then we wouldn't be able to say anything. DR. APOSTOLAKIS: Well, it is consistent. If you say zero failures over, you know, two, three, four thousand reactor years, it is consistent with the current estimates, but you are saying the current estimates are high anyway. MR. BARANOWSKY: Our job is to analyze operating experience any way we can. We don't have to use one technique. DR. APOSTOLAKIS: Yes, but this is not operating experience. It is -- MR. BARANOWSKY: Yes, it is. MR. MAYS: Not quite, George, because what we did was when we have a situation like aux feedwater we had zero failures in a thousand demands for aux feedwater. We didn't stop there and say therefore the probability is some big number with uncertainty. We went down in the aux feedwater study and said we know information about failures in the aux feedwater system that we can put together logically in a model to develop the probabilities. In a similar way, for large LOCAs and for medium LOCAs, what we did is we went back to the basic of how does a break occur from the fracture mechanics and physics. We went back and said, well, the first thing you do to break a pipe, you have got to get a crack that goes through a wall, and then the crack has to grow, and those things have probabilities and things associated with them, so what have got data on was the through-wall cracks, the natures, the causes and the frequencies of the through-wall cracks and then applying the understanding from fracture mechanics in a conservative way went from the frequency of through-wall cracks to the frequencies of pipe breaks in the same way that we went from the frequency of AFW pumps not working to the AFW system not functioning. DR. APOSTOLAKIS: So you had -- MR. BARANOWSKY: So we used a similar operating experience technique. DR. APOSTOLAKIS: So you had evidence -- MR. BARANOWSKY: Right. DR. APOSTOLAKIS: -- on the crack level. MR. BARANOWSKY: Absolutely. They have a great database. DR. APOSTOLAKIS: The models themselves evolve too, don't they? They improve. The models themselves? MR. BARANOWSKY: The models have improved significantly from what we knew 10, 15 years ago. DR. APOSTOLAKIS: So how much -- what is the large LOCA frequency now? MR. BARANOWSKY: It is close to -- MR. MAYS: We had two values, one for PWRs and BWRs and they were in the 10 to the minus 5 range. DR. APOSTOLAKIS: One order of magnitude. MR. MAYS: Yes. DR. APOSTOLAKIS: And the small LOCA? MR. BARANOWSKY: About 10 to the minus 3 for a small LOCA. DR. APOSTOLAKIS: Another order of magnitude. MR. BARANOWSKY: Now the transients -- MR. MAYS: It's still about at 10 to the minus 2-ish. The ones where you have transients and stuck-open PORV or something like that, those are still in about the nine, ten to the minus 3 I believe was the number that we came up with for those. DR. APOSTOLAKIS: So ultimately it will tell us h ow the core damage frequency trends in time, based on all this information? You will do something like that? MR. BARANOWSKY: Well, the only thing we have to show something like that is really the ASP results. DR. APOSTOLAKIS: And what are the insights from there? MR. BARANOWSKY: Apparently the risk is declining in time and may be -- DR. APOSTOLAKIS: Well, that would be an extremely useful insight. Don't do it as a side -- MR. MAYS: No, that is in the ASP report as well as in the ASP SECY paper we sent up every year. The frequency of ASP events in each of the bins, 10 to the minus 5, 4, 3 have been going down in a statistically significant fashion and continue to decrease. The rate at which precursors in the 10 to the minus 3 or greater range occur is about one every three years or so and so that seems to be kind of the residual level at which we are seeing those, and if you look at the core damage index for comparison with PRAs as a rough approximation of core damage frequency, you are seeing reasonable agreement between what those are. DR. APOSTOLAKIS: Is there a single document where you guys put your insights regarding risk and these trends and document them a little bit? MR. BARANOWSKY: No. DR. APOSTOLAKIS: Wouldn't that be a very useful thing to have? MR. BARANOWSKY: Yes. In fact, that is sort of my last viewgraph that I am going to mention that a little bit. DR. APOSTOLAKIS: That really would be great. MR. BARANOWSKY: Why don't we just flip, go to ASP. I am going to skip the common cause failure -- DR. APOSTOLAKIS: Yes, we have seen this. We have used it in our letters. MR. BARANOWSKY: Actually Steve has already talked about -- let's go to -- MR. MAYS: Evaluation of trends in ASP? DR. APOSTOLAKIS: 14? MR. BARANOWSKY: Yes, 14. There is a slight error in this viewgraph. This 10 to the minus 3 over here should not be there. DR. APOSTOLAKIS: It should not be there. MR. BARANOWSKY: Right. What we do with the ASP, every year we put out a report to the Commission as requested and we identify the insights from the current year and also the trends. As Steve said, the trends are going on in all the bins except the 10 to the minus 3 bin, which means 10 to the minus 3 -- greater ASP results. But we are pretty sure it is going to show a statistically significant decline after we finish the 1999 events because we don't believe there are any 10 to the minus 3 events in 1999. We have preliminary results or at least review of all the events that have occurred. They are not all finalized but they all seem to be less than 10 to the minus 4. In the year 2000 we have some 10 to the minus 4s already so, you know, you can't get too cocky just because you seem to have a slight decline in that group. That doesn't mean that it has gone away, and we have had a couple of interesting events already in the year 2000. Let's see -- what else do I want to say? Oh, and the best we can tell, the occurrence rate of these accident sequence precursors matches up with what we would expect with the CDFs predicted in the IPEs, so even though some differences exist in the modes and causes, and I think I have a couple listed -- DR. APOSTOLAKIS: I am trying to digest this first comment. What you are saying is, in the first bullet, that most of these ASPs are going down, except the bad ones. MR. BARANOWSKY: Right. DR. APOSTOLAKIS: So the frequency of real screw-ups has not been changed. MR. BARANOWSKY: It is a low frequency. DR. APOSTOLAKIS: Why is that? MR. BARANOWSKY: Well, that is because they don't have as much experience. DR. APOSTOLAKIS: Ah. MR. BARANOWSKY: If you look at all these declines, you see a learning exercise going on, and the more information you can get to people about what is not good, the more likely they are to make some corrections, so we very rarely get information on these 10 to the minus 3s, but we are getting some, and we are starting to feed that back into the system and I think we are going to see some dropping off here. Now people are aware of things like Wolf Creek. Let's have the next viewgraph -- also LaSalle, where they had a failure mode of the service water system that I would have never thought about, which they were pumping these concrete sealants, just loading it into the system -- MR. BARTON: Through the floor. MR. BARANOWSKY: Who would have thought of that? That is just not anywhere, but now it is in our common cause failure database. That is another reason why we are, sometimes we had some differences with licensees. Until recently the common cause failure data base wasn't made available to the industry. We had industry. We had it but it wasn't made available, so we could do pretty good common cause failure analysis. DR. APOSTOLAKIS: I don't know what kind of information you are going to give them to reduce that. I mean the Wolf Creek event is a rare event. It is not something we see all the time, and what are they going to learn from that? That when you make changes in your work plans, make sure that somebody knows? Well, they are supposed to do that anyway, so how would that affect the CCDP? MR. BARANOWSKY: I think it just gives a heightened focus on these kinds of things. We know all this stuff. There is nothing that occurs -- we are not supposed to pump sealant into things without knowing where it is going. I knew that. You knew that. But so did they. They did it. DR. APOSTOLAKIS: But that is what worries me. I mean okay, higher sensitivity? All right. MR. MAYS: The point is this, George. If you look at the PRAs as a predictive model -- DR. APOSTOLAKIS: Which I don't. Which I don't. MR. MAYS: If you were to look at the PRAs, it says this is a measure of the baseline risk. DR. APOSTOLAKIS: Yes, that's good. MR. MAYS: Okay? Then the Accident Sequence Precursor Program would say, well, take one step back from core damage frequency based on that model. How often should I see things that are getting me close to core damage frequency? The ASP program is a way of measuring whether the expected non-core damage frequency events from your PRA are occurring at a rate that is comparable to what your PRA would have predicted, and what we are seeing is that in large part that is true. There may be some differences in the particular modes, and the more that information gets fed back to the licensees the better position they are in to make the frequency of them happen less. It won't mean they will go away. It means that we can get an opportunity to reduce the occurrence rate and that is what we are seeing. MR. BARANOWSKY: And that seems to be happening. DR. APOSTOLAKIS: No, but I am still trying to understand what it means that the really bad events have an almost constant rate of occurrence while the others are declining. There must be some fundamental reason -- maybe the latent error business. I don't know. MR. BARANOWSKY: I don't know. There is not enough of them for us to draw a conclusion yet. DR. APOSTOLAKIS: Yes. No, I understand that. MR. BARANOWSKY: Let me briefly mention the work that we did on D.C. Cook, which of course you all are aware of we had significant issues with. We performed a special ASP study on that because of the heightened public awareness and the Commission's concern about Cook and analyzed all the licensees in the NRC's inspection findings and we came up with 141 issues, which we analyzed individually and integrally, and as a result so far we have determined one of these issues produces a conditional core damage probability of greater than 10 to the minus 6, which is the ASP criteria. That involves a high energy line break that has the potential for causing loss of component cooling water trains which could lead to a reactor coolant pump seal LOCA and you wouldn't have a high pressure injection available without component cooling water trains. There's four others that could meet the criteria for being a precursor. They are being sent out to the licensee for review and comment and to NRR to make sure that the facts that we have are correct on them. If they are, then these four would also be. They are also high energy line breaks, but this could affect either AFW trains, vital AC buses, or emergency diesel generators, ultimately leading to blackout type sequences, pressure locking in some motor operated valves required for recirculation functions, and two seismic issues. One has to do with the potential for failing emergency service water due to inadequate anchorages that were found, and the other one has to do with the failure of block walls that were put up, I think for fire protection but I am not sure, which could fall down on equipment because they weren't seismically designed. They are trying to determine how bad the damage would be if those block walls fell down, but any of those could be up in the 10 to the minus 4 range of the Accident Sequence Precursor Program. We should be done with that analysis in about another month. As I said, we are going to the licensees for comments now. DR. APOSTOLAKIS: That is a probability, 10 to the minus 4. MR. BARANOWSKY: Those are conditional -- DR. APOSTOLAKIS: CCDPs. MR. BARANOWSKY: -- CCDPs and since they are over one year, they are like the CDF. That is in essence what the CDF was at that plant for the time period these conditions existed. DR. APOSTOLAKIS: But it wouldn't really be right to compare it with a goal of 10 to the minus 4. They are two different things. MR. BARANOWSKY: I am not sure about that, to be honest with you, because I mean if a plant takes away a system and it doesn't exist, and the core damage frequency is above the goal with that system not being there, and it operates like that -- DR. APOSTOLAKIS: For how long? MR. BARANOWSKY: A year, two years, three years -- for that period of time, I question whether they were -- DR. APOSTOLAKIS: You are right. MR. BARANOWSKY: Okay. So I think that is a pretty significant finding. DR. APOSTOLAKIS: Yes. MR. BARANOWSKY: SPAR model development. This is what we talked about. We are working on several areas to develop accident sequence precursor models but these are the standardized plant analysis risk models. They are really fairly simplified in comparison to what the licensees have in their PRAs, although they are getting more complete every time we work on them. We have established this SPAR Model Users Group called SMUG which has both NRR and Regional folks on it. [Laughter.] MR. BARANOWSKY: And we are trying to detail out exactly what kind of models and capability are required for regional analysis, NRR analysis, or just to perform ASP calculations. This would Support things like the Significance Determination Process and so forth. And at this point we have a pretty good step-up on doing all the Level 1 models for the plants, I forget how many we have done so far. We have Rev. 2 SPAR models for every plant. We are working on improvements to those in Rev. 3, for which I believe we have -- MR. MAYS: Nineteen. MR. BARANOWSKY: Nineteen. Nineteen of those initials models completed, and we are doing them at a rate of about 20 a year. DR. POWERS: When you produce a model, one of these SPAR models for a plant, is there the capability to do uncertainty analysis on the results? MR. BARANOWSKY: Yes. It is included. DR. BONACA: In the December meeting, you, at some point, mentioned that you would consider doing a systematic comparison of the SPAR models with the IPEs. MR. BARANOWSKY: Oh, yeah, that is what you wanted. We were talking about validation. In fact, there is a couple of things going on with validation. One is we do compare with the IPEs and ask ourselves why there is any differences in our SPAR models. The second thing is there is a simplified checklist type of screening tool that has been put together for the Significance Determination Process that NRR is trying to validate, and we are tagging along on those validation get-togethers with the licensees and asking questions that allow us to clarify some concerns that we had about SPAR model details. And then, lastly, we are looking at whether or not we need to have even more thorough, maybe multi-day plant visits working with the resident inspectors to get even more accurate information into the SPAR models. So that last part is still a little bit up in the air, but we are doing the first two things. And that's -- I think we will have a pretty good simplified model. It won't be able to do everything in a licensee's PRA, but it will be easy to use and the assumption will be standardized, which is important for making comparisons. DR. APOSTOLAKIS: Now, the issue of peer review has come up in the past of these models. Do you plan to do anything about it? MR. BARANOWSKY: Well, I don't know about peer review in the sort of traditional sense that we would do it for a PRA. DR. APOSTOLAKIS: No, no. But some independent evaluation, some kind of an independent evaluation. MR. BARANOWSKY: Yeah, I don't know that we had plans to do that. DR. APOSTOLAKIS: Do you think that that is something that you may consider? MR. BARANOWSKY: It is probably something we should think about. DR. APOSTOLAKIS: Okay. MR. BARANOWSKY: These models are different and simpler than the PRAs, and I don't want to try and say they are a PRA, but within their limitations, they should produce consistent results. DR. APOSTOLAKIS: No, but it would be nice to know that, say, a couple of practitioners, say, from the industry spend some time actually looking at it. MR. BARANOWSKY: Yeah. It might not be. I think maybe we would put together some kind of a report describing this. DR. APOSTOLAKIS: Some group. MR. BARANOWSKY: And we are sending them to all the licensees, too. MR. MAYS: Right. The licensees, now that they have known we have these, have been asking for us copies of them at an increased rate, and they are indicating, depending, of course, on their PRA sophistication and desire to do that, we are finding more and more licensees want to see what we have in our model so they can understand the difference between ours and theirs. And when they give us feedback, as we did through the ASP program, about, well, we have this system or this capability, or something that you don't have modeled, and we would go back and verify that, we would change the model. DR. APOSTOLAKIS: Eventually, your models will be PRAs, won't they? MR. BARANOWSKY: Well, they are PRAs. DR. APOSTOLAKIS: Why not? MR. BARANOWSKY: They are PRAs. DR. APOSTOLAKIS: Steve, why not? MR. MAYS: I am not sure what you mean by the statement, they are. I mean they are risk assessments. The question is, at what level of detail and to what thoroughness and completeness are they relative to what is the plant's -- DR. APOSTOLAKIS: But I mean you are not making some of the very drastic assumptions that were being made in this program 15 years ago. MR. MAYS: Absolutely not. DR. APOSTOLAKIS: Like the operator will do this or that. MR. MAYS: We have found that these are more consistent and/or realistic in terms of PRA sequences, numbers and contributors than what was done before. And that is one of the goals, is to make it a more realistic process. MR. BARANOWSKY: They are primarily standardized, to be honest with you. DR. BONACA: That is why how it compares with the IPE is important, because you may find that one system you did not model makes a big difference. DR. APOSTOLAKIS: Yes. DR. BONACA: If you don't find that, then, you know, that confirms that your modeling is adequate enough, even if it is at a high level. So I think that that kind of comparison is most useful because it speaks about the structure and the level of details you have to go to. MR. MAYS: The other thing that it is very useful for is when we get licensee applications under Reg. Guide 1.174, for example. That is going to be based on some PRA result. The key thing about these that will be good is we can compare our results independently with theirs, and instead of going back and having to review their entire PRA in its entirety, down to the last level of detail, we can use our information, which is going to be based on operating experience and standardized ways of looking at risk sequences, and say our assessment is different from yours in this way, and we can go focus on where the differences are, rather than having to spend, in an efficiency standpoint, going out and looking at everything that they did. DR. APOSTOLAKIS: But in that context then, I mean you told us that you will develop SPAR models for low power and shutdown models? MR. MAYS: That is correct. DR. APOSTOLAKIS: So would you be then -- DR. POWERS: I guess I don't understand that. I think that when your management discusses this with the Commission, they say you have all the capability you need. MR. BARANOWSKY: Well, they are probably thinking that when we have these models in place. DR. POWERS: Oh, forward-looking individuals. I hadn't thought of that possibility. MR. BARANOWSKY: Our management is forward-looking. DR. APOSTOLAKIS: So, Pat, would you then say that these models could be good enough to given to senior reactor -- what do we call them? MR. BARANOWSKY: Absolutely. DR. SHACK: SRAs. MR. BARANOWSKY: They are on the SMUG group. They are on the SMUG group. DR. APOSTOLAKIS: They are SMUG group. MR. BARANOWSKY: They are some of the people who are telling us what capabilities they need. DR. APOSTOLAKIS: So the agency is in the process of developing these tools? DR. BONACA: Oh, yes. MR. BARANOWSKY: Right, we are. And I think there is a big difference between, you know, no knowledge and perfect knowledge, and what we are is in the middle and moving. DR. APOSTOLAKIS: Who told us yesterday that incorrect knowledge is worse than no knowledge. DR. POWERS: Their sister organization. MR. BARANOWSKY: No, we are not saying incorrect. We are saying if you have some knowledge, and you understand its limitations, that is better than having no knowledge. DR. SHACK: Which your sister organization said, too. MR. BARANOWSKY: Good. Why don't we go to the last viewgraph because I don't think we need to say any more about risk-based PIs, which we covered yesterday, and I wasn't going to. DR. APOSTOLAKIS: That's fine. MR. BARANOWSKY: I need to tell you sort of where we are heading with this stuff. We are going to try and streamline the things that we do and make more current and consolidate the work on initiating events, system and component reliability studies, common cause failure analyses and so forth, mainly because we went through a learning exercise in producing this stuff the first time around or so,, and now we see that there is a way to bring this stuff together and still get system component and common cause failure insights out, but probably integrated them a little bit better. So that will be an efficiency type thing. The other thing, we want to make ASP more current. Sometimes now it takes six months to get an ASP result out, because it needs to be more coordinated with the revised reactor oversight process, where they are making findings, and, in particular, the Significance Determination Process. DR. APOSTOLAKIS: Good. MR. BARANOWSKY: And I can tell you the reason why it usually takes a long time to do the ASP results, usually the ones that we, ourselves, get tagged with doing, there is a lot of questions about the engineering capability of equipment or the thermal-hydraulic response of the plants. The ones that are easy, where the equipment was broken and fell on the floor, and you just punch a little button on the SPAR models and make it fail, I can do those in, you know, 30 seconds. But going and figuring out whether a pump that had a degraded condition would have actually provided enough flow to satisfy the thermal-hydraulic requirements in a plant, that sometimes can take months. So it is going to be difficult to get this to be too fast if we don't have those kind of capabilities worked into the system in some way. Then we want to try and prepare this annual report which -- DR. APOSTOLAKIS: Which we discussed earlier, right? This could be the insights and the transient risk. The report we -- MR. BARANOWSKY: Annual report? DR. APOSTOLAKIS: Yes. MR. BARANOWSKY: Yes. DR. APOSTOLAKIS: Good. MR. BARANOWSKY: And we would pull it all together also. DR. APOSTOLAKIS: Yes. MR. BARANOWSKY: Instead of having it in four or five different locations and reports produced throughout the year. We might still produce individual reports, but we would have one report. I don't want to say it is reviving the old AEOD annual report, because it would be a little different, but it is along that idea. Again, we would like to try and make that a current kind of thing, not be years old. DR. APOSTOLAKIS: When do you think you are going to have this one? MR. BARANOWSKY: I think the first cut at it will come around March of 2001. Well, the reason is we can't make all this stuff current until then. There is a lot of work to take these system initiating event studies and bring them into some currency. DR. APOSTOLAKIS: When you say we, how many people are involved in this? MR. BARANOWSKY: Well, that is the other problem. DR. POWERS: Well, I think that is really a management issue. DR. APOSTOLAKIS: I know, but it is information. MR. BARANOWSKY: We have like -- DR. POWERS: I am going to give you some information, that I will take the time from now on out of your hide. MR. BARANOWSKY: We have 12 staff and $4 million in contractors. And we won't be able to do all this stuff in a timely manner that I am listing here, because budgets keep on changing. I have had some staff reductions. But we will do as much as we can. SPAR model, we identified what we wanted to do that. We want to implement the databases and continue developing the risk-based PIs. You also made some suggestions on looking at licensee shutdown analyses, risk analyses and comparing them, and some comparisons with fires. I am putting those down as -- DR. APOSTOLAKIS: Seismic? MR. BARANOWSKY: Seismic. All possibilities, questions whether we have the resources to do these things, or whether it would be so spread out in time it wouldn't be worth doing. We would make that decision to. DR. APOSTOLAKIS: Peer review, think about the peer review. MR. BARANOWSKY: All these things are in the hopper, but I have limited resources. I have the smallest branch in the Office of Research. Thank you very much. DR. BONACA: Thank you. Any other questions? DR. APOSTOLAKIS: I just want to say that I always enjoy the presentations of those guys, it is always informative. DR. BONACA: Well, I think this is most of the intelligence that comes within the staff. Not because the rest of the staff is unintelligent, just because it is a lot of information that comes from operating experience which is so important. Okay. With that, thank you, and I will give it back to the chairman. DR. POWERS: Thank you. I will echo my appreciation of the presentation. DR. SEALE: High information density. DR. POWERS: I think it, however, has raised more possibilities for work than the smallest branch in Research can tolerate. At that point, I will recess for 12 minutes. [Recess.] DR. POWERS: We will come back into session. I want to begin immediately with seeing if any members have feedback they would like to offer Mario Bonaca in preparing a letter on the material we have just heard. I have provided Mario some comments. My comments are, I see needs for more and different databases than the agency has. I see this as yet another indication that the agency needs to completely redo its PRA implementation plan in the light of the rapid progress that we are making toward risk-informed regulation. I think updating the existing plan is no longer sufficient, that we need to take a broad and holistic look at the entire plan so that we have a comprehensive development of the kinds of materials that these gentlemen, Mr. Mays and Mr. Baranowsky, told us about today. I also find it striking that the SPAR users groups, including the senior reactor analysts, are asking them to develop their models to include low power and shutdown assessment capabilities at the same the management is telling the Commission that they have all the models they need in this area. I think that is a striking difference in opinion. I ask are there other members that have comments they would like to have included in this -- or considered for inclusion in the report on this work we have just heard about? Dr. Apostolakis, you made mention of points on peer review. DR. APOSTOLAKIS: Yes. DR. POWERS: Anything else that you think needs to be commented on? DR. APOSTOLAKIS: I think we should say something to the effect that this is one of the most useful activities of the agency. MR. BARTON: That would be nice. DR. POWERS: I think we ought to hearken back, I can remember, since I have been on the committee, Professor Seale hammering on people that mining the operational database is going to be the source of information that is unavailable from other sources. Is that roughly correct, Bob? DR. SEALE: Yes. DR. POWERS: And I think we might just point out in our letter that this has been a long-time interest in the committee, of mining the operational database for things to compare against PRAs and have some validity to the PRAs. And that might be an excellent lead-in to, I think, another point that Professor Apostolakis suggested, that there be some sort of a summary report on this comparison. DR. APOSTOLAKIS: Yes. And it is something we suggested in the past, and I am not sure anything has happened, is, again, I don't think that the community out there at large is really aware of the work that these guys are doing, and taking advantage of it. Now, I don't know what they can do. I mean they do go out and present papers and all that. DR. BONACA: I tried already to have a draft which is complimentary enough, and I tried to strength that portion upfront of the importance. Really, this is the lifeblood of the agency, I mean. DR. APOSTOLAKIS: Yes. DR. BONACA: And that is why, you know, I keep harping on EPIX, because they keep saying we are going to do these wonderful things, and then if you look at the chapter from yesterday, there was -- it ran right through EPIX. And without it, there is not going to be any of these wonderful things they want to do. DR. SEALE: I think you want to be very careful when you say that though. DR. BONACA: Oh, yes. DR. SEALE: We want to say it -- you want to point out that the people who are in the dark are the rest of the agency. It is not the utilities, the utilities are following this EPIX data very closely. They know what it is in there. Is the rest of the agency that doesn't know what is in there. And I think when we say, you don't know it, it is an agency problem more than industry problem. DR. BONACA: No, I agree with that. And one comment I would like to make, Dana, would be your comment on a different type of database is well taken. I wonder if this is the place to make it. You know, if I look at the thrust of your comment, really, it is more on the PRA implementation plan needs a rework, and maybe, my suggestion would be that we put this comment in the letter that we have next month. I believe we have a presentation on the PRA implementation plan. Because if we put it, you know, narrowly in this letter, it undermines to some degree the other message we are trying to give, that these people are important, what they do, it is important. I don't know if you have any thoughts about that, but I am afraid that putting it here, it will somewhat undermine the interaction they are having right now with INPO to strength the quality of EPIX. DR. POWERS: Of course I am never opposed to the idea of focus leads to resolution, as opposed to having a shotgun approach, and, so, I am perfectly willing to defer to your judgment on that matter. I think it is -- I am concerned about whether we are having the strong support information and technology development that this move toward risk-informed regulation is going to take. I think that when we look at things like performance indicators and Significance Determination Processes, you see a crude nature to them at this juncture. We have the hopes that they improve with time. And if we don't have strong support efforts going on to develop technology and develop better indicators, and develop better processes, those things may become entrenched and they can't be changed. So I am concerned about that. But I offer you my comments simply for consideration, and if you think it is better to wait till another time, I think that is fine. DR. BONACA: I can weave in the need for consideration of an expanded database without reference to the PRA implementation plan. I would like to stay away from it right now as far as that portion. But let me see what I can do. DR. POWERS: Well, I defer to your best judgment on that matter. Any other comments? [No response.] DR. POWERS: Seeing none, I think at this point we can go off the transcript. [Whereupon, at 10:18 a.m., the open portion of the meeting was concluded.]
Page Last Reviewed/Updated Tuesday, July 12, 2016
Page Last Reviewed/Updated Tuesday, July 12, 2016