109th ACNW Meeting U.S. Nuclear Regulatory Commission, May 11, 1999
UNITED STATES OF AMERICA NUCLEAR REGULATORY COMMISSION *** ADVISORY COMMITTEE ON NUCLEAR WASTE *** MEETING: 109th ADVISORY COMMITTEE ON NUCLEAR WASTE (ACNW) U.S. Nuclear Regulatory Commission Two White Flint North Room T-2B3 11545 Rockville Pike Rockville, Maryland Tuesday, May 11, 1999 The committee met, pursuant to notice, at 3:50 p.m. MEMBERS PRESENT: B. JOHN GARRICK, Chairman, ACNW GEORGE HORNBERGER, Member, ACNW CHARLES FAIRHURST, Member, ACNW RAY WYMER, Member, ACNW STAFF PRESENT: ANDREW C. CAMPBELL, ACNW LYNN DEERING, ACNW HOWARD J. LARSON, ACNW RICHARD K. MAJOR, ACNW JOHN SORENSEN, ACNW Fellow PARTICIPANTS: CHRISTIANA H. LUI, NMSS KEITH MCCONNELL, NMSS. P R O C E E D I N G S [3:50 p.m.] CHAIRMAN GARRICK: The meeting will now come to order. This is the first day of the 109th meeting of the Advisory Committee on Nuclear Waste. My name is John Garrick, Chairman of the ACNW. Other members of the Committee include George Hornberger, Ray Wymer, and Charles Fairhurst. The entire meeting is open to the public. During today's meeting the committee has already met with the ACRS-ACNW Working Group on Risk-Informed Regulation, where we discussed a framework for risk-informed regulation in NRC's Office of Nuclear Materials, Safety and Safeguards, and we have also had some prior discussion on committee activities and future agenda items. What we want to do now is listen to a description by the Staff of the strategy for converting the issue resolution status reports for the proposed high level repository at Yucca Mountain into a review plan for the repository license application. Howard Larson is the Designated Federal Official for today's session. We are conducting the meeting in accordance with the provisions of the Federal Advisory Committee Act. We have received no written statements or requests to make oral statements from members of the public regarding today's session. Should anyone wish to address the Committee, please make your wishes known to one of the Committee Staff. As usual, it is requested that each speaker use one of the microphones, identify themselves and speak with clarity and volume. Before proceeding with the first agenda item, there's a few items of current interest that we want to mention. A couple of Staff issues -- Michelle Kelton and Ethel Barnard of the ACNW-ACRS Office received an Achievement Award for their contribution to the Y2K application renovation efforts at a March 12, 1999 awards ceremony. We want to thank them for their efforts, and the others in the agency. We are told now that our computers are ready and that we may not have to unplug them after all. Mary Thomas's six-month rotational assignment has recently ended, and she has returned to the Office of Nuclear Regulatory Research. Mary was a major player in our recent working group session on the effects of low-level ionizing radiation. The Nuclear Waste Technical Review Board issued its report on the viability assessment. "Moving Beyond the Yucca Mountain Viability Assessment" was the title of the report. The Board notes that, quote, "So far it has not identified any features or processes that would automatically disqualify the site but that DOE should give serious attention to alternatives to the VA reference design including changing from a high temperature design to a ventilated low temperature design below the boiling point of water." The Board also notes that DOE's plans to determine the suitability of the proposed repository by 2001 is, quote, "very ambitious and much work remains to be done." The House Commerce Committee approved the Nuclear Waste bill, H.R. 45, that will provide for interim storage of spent commercial power reactor fuel at Yucca Mountain, Nevada. The bill passed on a 39 to 6 vote and now moves to the House floor. In an April 16 order, a Federal Judge sided with the utility low-level waste generators and site developer, U.S. Ecology, in their lawsuit claiming political bias caused Nebraska regulators to deny a license for a disposal facility last year. The Judge noted in the order that, quote, "There is good reason to think that a license denial was politically preordained." The utility, U.S. Ecology, and the Central Interstate Low Level Waste Commission sued the state and its regulators last year, blaming politics for delays in the licensing process. The licensing denial will be repealed, and in that connection -- DR. HORNBERGER: Appealed. DR. GARRICK: I'm sorry, appealed -- appealed, yes. Thank you. In connection with that, Nebraska's legislature has approved L.B. 530, a bill to remove the state from the Central Interstate Low Level Radioactive Waste Compact. On May 6th the bill passed the third and final round of debate by a vote of 33 to 11 with 5 Senators excused and not voting, and it now proceeds to the Governor. The Governor is expected to sign it into law, and if enacted, the legislation will take effect on August 29th. As provided in the legislation, the Governor may then write to the Governors of the compacts under member states to notify them of Nebraska's withdrawal. Of course, under the terms of the compact agreement withdrawals generally do not take effect until five years from the date of such notification. One other item of interest, maybe two others. County Commissioners in the Las Vegas area have made it very clear that they plan to fight the transportation routes chosen for moving radioactive waste through the Las Vegas area from DOE's Fernald site in Ohio. This is of interest given the discussions we had earlier in the day on transportation. One other perhaps note of interest is that New Mexico Attorney-General, Patricia Smith, late last month pulled out of litigation challenging the Environmental Protection Agency certification of the Waste Isolation Pilot Plant's long-term disposal standards. The standards are aimed at protecting public health and safety for 10,000 years. Madrid's motion to withdraw from the case was granted Wednesday by the U.S. Court of Appeals for the District of Columbia, which also decided to terminate oral arguments between the State of New Mexico and EPA that were scheduled for Thursday. I think this is -- I guess this is dated May 10th, so I guess they are talking about this week. I think that is all the items of interest that we want to cover. Now we are going to turn to the Staff for a discussion of Yucca Mountain review plan and I guess Keith McConnell is going to kick it off and then introduce Christiana. DR. McCONNELL: Thank you, Dr. Garrick, members of the committee. We are here today for the first of several interactions we intend to have with the committee as we develop the Yucca Mountain Review Plan. Today's briefing is basically at the concept level, defining how we intend to approach the development of the Yucca Mountain Review Plan. It doesn't get into specific details, although there is an example that Christiana has as backup. As we get into the details of specific issues such as defense-in-depth, how we intend to implement defense-in-depth in the Yucca Mountain Review Plan, we will be back to the committee as well as other details as we go along in this process and part of Christiana's presentation is a schedule, at least a proposed schedule, for all of the elements of the work. In fact, I think we are right now scheduled to talk to you briefly about defense-in-depth at the June meeting you intend to have in San Antonio. So with that I will turn it over to Christiana Liu, who is the lead for the development of the Yucca Mountain Review Plan. MS. LUI: Thanks, Keith. The title of my presentation today is "The Framework for the Yucca Mountain Review Plan." This is a work in progress and we thought that as we proceed along the development of the review plan, we will come back to the committee to talk to you as we develop more and more details. Today is the part one of the series. I am Christiana Liu, and I work for Keith McConnell in the High Level Waste and Performance Assessment Branch in the Division of Waste Management in the Office of NMSS. Basically, there will be pretty much three parts to my presentation today. The first four bullets cover part one, which I will give you a brief introduction to the framework for the Yucca Mountain Review Plan before we jump into the more detailed part of how we are intending on integrating the material we have published in the IRSRs into the Yucca Mountain Review Plan. And the last part of the presentation, I will conclude with the advantages of the approach that we intend to implement and, also, like Keith has mentioned, a schedule for the proposed work. There are basically four very high level principles before we jump into the framework. The staff is principally responsible to defend the conclusion of our review of any potential license application for the Yucca Mountain site, and DOE, who is the licensee, is responsible to make sure that an adequate case is made in the license application. We have recently published a proposed performance-based, site-specific rule, Part 63. In the out for public comments, basically, we thought that a performance-based, site-specific rule should be accompanied by a performance-based, site-specific review plan, and the focus of this review plan is for NRC staff evaluation of DOE safety case, including how the site characterization and experimental work has been conducted to support DOE's safety case. You are going to see later on with more detail of what we mean by performance-based and site-specific approach, which is basically a top-down approach to include all the work that has been conducted. The strategy for licensing the Yucca Mountain site has been published in SECY-97-300. That was the strategy paper prior to the staff's work on Part 63, and in that particular SECY paper, the staff talks about how we intend to develop the Yucca Mountain Review Plan and the work here basically reflects that particular strategy. And the last point is the review should really be done in an integrated fashion, that integration should take place as at the technical staff level. Again, this is speaking to a top-down approach, recognizing that there will be a natural tension between implementation of a performance-based rule and the need to prepare and guide the staff in performing the review and writing the final Safety Evaluation Report. The Yucca Mountain Review Plan should be formulated based on the staff's current understanding of DOE's approach and all the iterative performance assessment work that the staff has been doing in the past decade to build up our own capability. Given that, the framework should be sufficiently flexibility to accommodate changes in DOE's approach. In all works in the performance-based approach, the licensee is given the flexibility in designing how they will address their safety case and the staff should be prepared to be flexible in accommodating whatever approach DOE has decided to use in their license application. Next, I will turn quickly to the features of the review plan. As in all the other review plans, they are five standard components, the areas of review that basically describe the scope of the review, in other words, what is being reviewed. Acceptance criteria delineates the criteria that can be applied by the reviewer to determine the acceptability of the compliance demonstration. Review procedure discusses the appropriate review technique to determine whether the acceptance criteria have been met. Evaluation findings basically presents the general conclusions and finding based -- resulting form the staff's review, and that will be the material that will we make into the Safety Evaluation Report. Finally, references, that will list any applicable references that the staff has used in its review and the review plan. The next part is a very high level outline of what the final Yucca Mountain Review Plan will most likely look like, abstract, executive summary, introduction, and there are basically three elements in the introduction that we would like to cover. Part I is basically the principles in formulating this performance-based review plan that will give the background that I have presented to you at the beginning of this talk. And structure of progression of NRC high level waste program, basically, this will describe how we have used the KTIs for pre-licensing consultation and issue resolution and how we are transitioning from the KTI approach into an integrated team approach for reviewing the license application. And the third part, we will attempt to provide a clear relationship between how the Yucca Mountain Review -- the Yucca Mountain license application is going to be reviewed and in what context the requirements under paragraph 63.21, that is the content of license application, are to be reviewed. Chapter 1, review of general information, which is basically a requirement in the proposed Part 63, paragraph 63.21(b). What I am going to spend more time in the subsequent part of this presentation is really to talk to you about Chapter 2, which is the Safety Analysis Report. And there are three components to that Safety Analysis Report Review, the pre-closure part, the post-closure part and the administrative and programmatic requirements. The IRSR work has been done, mostly on the post-closure part which is basically Chapter II.B on this outline. We may come back to this outline later on as we go on with this particular presentation. The next three pages basically gives you more information of what Chapters II.A, II.B and II.C will look like. Chapter II.A and II.B are organized by performance objectives and the associated technical criteria. Under areas of review for II.A, the compliance demonstration to meet the pre-closure performance objective and the requirements for an ISA and Subpart F, and the review chapters are basically formulated based on the pre-closure performance objectives. Basically, we have required DOE to use an integrated safety analysis to demonstrate the pre-closure safety during operation and for design basis events categories 1 and 2. Another performance objective for pre-closure is retrievability plan and alternate storage. And the last objective for pre-closure performance -- for pre-closure portion is the performance confirmation program. For each of the review chapters, we will clearly identify what part of the license application is to be reviewed. For example, taking Chapter II.A.2, the retrievability plan and alternate storage, the content of Yucca Mountain license application to be reviewed will be in paragraph 63.21(c)(19), it is the retrieval and alternate storage plans. And evaluation findings basically is to conclude, after reviewing all the parts in the license application as they have been specified here, whether DOE has met the pre-closure performance objectives and the technical requirements associated with it. If they have, then the staff's -- the conclusion that will go into the Safety Evaluation Report is that DOE has successfully demonstrated they have met the pre-closure performance objective. Because of the parallelism that has been building to the proposed Part 63, the pre-closure and post-closure have similar approaches. On the next page, page 6, II.B, repository safety after permanent closure, again, the areas of review here is the compliance demonstration. To me, the post-closure performance objectives which are delineated in paragraph 63.113 and the technical requirements for doing a performance assessment, the technical requirements for critical group and, again, performance confirmation program. And the review chapters are performance assessment and performance confirmation. Because we fully expect DOE to use performance assessment to demonstrate that they have met the post-closure performance objectives, namely, the mean peak dose will not be greater than 25 millirem per year, and also the multiple barrier requirement, and the license application that will be reviewed to determine if DOE has successfully demonstrated post-closure performance in the performance assessment are identified under Chapter II.B.1. We will review site description, the material that has been used in construction, especially for underground tunnel, the EBS design, of course, performance assessment itself, the stylized human intrusion analysis, use of expert elicitations, and there are probably also other parts that I have not listed here. But as we work on the development of this review plan, all the different parts and detail will be fleshed out. Again, the evaluation findings is to determine if DOE has successfully demonstrated they have met the post-closure objectives outlined in paragraph 63.113 and met the technical requirements in 63.114 and 63.115, and also the performance confirmation requirement. Administrative and programmatic requirements. Right now, these are -- these contain Subpart D, that is the records, reports, tests and inspections; Subpart G, quality assurance; and Subpart H, training and certification of personnel. For each of those subparts, there will be a review chapter associated with it and the evaluation findings is to determine if DOE has demonstrated they have met the requirements under these subparts. One thing that I would like to mention is existing guidance, such as Regulatory Guides, NUREGs and the other Standard Review Plans will be used or modified to the extent applicable. In terms of when we develop the Yucca Mountain Review Plan. In other words if there is no need to reinvent the wheel, we will not do that, especially for the administrative and programmatic requirements and also some of the preclosure portions. Page 8 and page 9 basically gives you a preliminary idea of what part of the license application is going to be reviewed where. Since each of the chapters will contain the acceptance criteria and review methods, these will clearly be for guidance to the Staff and indirectly provide information for the DOE on what we expect to see, how we are going to determine what is acceptable or what is not acceptable. This particular linkage may change as we finalize Part 63 and we further develop the Yucca Mountain Review Plan. One thing I would like to mention, that as we are going through this more or less systematic process we found that some of the entries currently under 63.21 will probably be modified. For example, the use of expert elicitation. The Part 63 right now only requires DOE to supply its use in the post-closure portion and we know that DOE is using or is planning on using expert elicitation for some of their design work for the preclosure part, such as probabilistic seismic hazard analysis, so we will modify 63.21 to require DOE to supply information for expert elicitation use for both the pre and post closure part. There may be some redundancy in 63.21 and we will most likely consolidate some of the requirements without changing the essence of what is currently under 63.21. Some of the requirements under 63.21 are really technical requirements rather than content-related requirements, so those technical requirements will be moved to the appropriate technical requirement portion under subpart (e) leaving 63.21 strictly content. Finally, the sequence may be rearranged to reflect a more logical structure in the final rule. This basically concludes Part 1 of the presentation on the framework of the Yucca Mountain Review Plan. Now I would like to provide you more information on how we are approaching it in terms of integrating the IRSRs into the Yucca Mountain Review Plan. You may want to keep in mind that for the post-closure part, the outline is on page 6. DR. GARRICK: Maybe the committee would like to ask some questions before you make the transition. MS. LUI: Okay. DR. GARRICK: One of the things that I was trying to track here was in the framework for the, in the framework for the Review Plan was how you are going to use some of your tools and what those tools are and this table kind of gets in it, but it seems to have apples and oranges, as you kind of explained. It has methods of analysis and it has physical features. MS. LUI: Right. DR. GARRICK: For example, the site is a physical feature and the performance assessment and integrated safety analysis are methods of analysis. I guess site description could be considered. I was just having a little trouble trying to figure out what you were attempting to do here. MS. LUI: Okay. I believe we are on Slide Number 8, is that correct? DR. GARRICK: Yes. MS. LUI: Okay. The left-hand side of the first column is really the very abbreviated description of what is currently in paragraph 63.21, the content of license application. Again what is currently under 63.21 is more or less a flat structure, so yes, you are correct, Dr. Garrick. For performance assessment in order to really evaluate a complete performance assessment there are a lot of other pieces that will need to be used to support that performance assessment but they are currently laid out in the linear structure rather than in a hierarchy structure. One of the possible things that might happen when we go towards finalizing Part 63 is to organize the content of application in a more logical fashion that would correspond to, more or less correspond to our review process. DR. GARRICK: Okay. The other thing I was looking for as how your TPA, now the NRC's TPA really enters into the framework, into the review process. You do talk about performance assessment with respect to permanent closure but again I was trying to figure out how the tools you have are going to be employed in the review process. MS. LUI: Okay. You will hear more on that during Part 2 of this presentation. DR. GARRICK: Oh, okay. MS. LUI: But one of the immediate results that we have used in terms of applying TPA code is to help us focus on what we need to focus in terms of reviewing the performance assessment. Basically I think I mentioned this when I talked about the principles is that the Staff's IPE effort has been used in terms of formulating this particular performance-based approach but if you are looking for a specific use of TPA code we'll talk about it a little bit later. DR. GARRICK: Okay, well, yes. I think sooner or later we want to learn a little more about the Independent Safety Analysis process too. I don't know whether we are going to get into that here or not but that has come up in other issues that we have discussed today and we have got input coming on that, so we can assume we will learn more about that later. DR. McCONNELL: Yes. The preclosure -- this is Keigh McConnell -- the preclosure safety case review plan is lagging behind the postclosure because we have over the past couple years focused on the postclosure capability within the Staff and also that is the focus of the KTIs, but we are now developing our concept towards developing an ISA and what would be in a review plan that we would expect for DOE, what we would use to gauge an acceptable ISA that DOE would submit, so I guess what I am trying to say is we're a little bit earlier in the process with that part of the Review Plan than we are with the postclosure. DR. GARRICK: Right. DR. McCONNELL: So this briefing is just going to focus on the postclosure. DR. GARRICK: Okay. Any other comments? [No response.] MS. LUI: Okay. Now I'll move on to the second part. To focus on how we are integrating all the IRSRs into the Review Plan. Again, just to reiterate the strategy, SECY-97-300 describes staff strategy in developing the proposed Part 63 and Yucca Mountain review plan. You will find on your next page, that's page 11, the flowdown diagram that we have included with the strategy paper. And integration of all the IRSRs for the postclosure work is really taking place at the third tier of this flowdown diagram. Up until this point the total system performance assessment integration IRSR uses this particular framework that you see on page 11 here to basically set up the IRSR material, and all the KTIs have done the crosswalk to the lowest tier of this particular flowdown diagram to identify and indicate how the subissues are contributing to the total system performance assessment. From this point on to avoid duplication and keep a consistent set of acceptance criteria and review methods, starting from FY2000, all the acceptance criteria and review methods will be in the Yucca Mountain Review Plan, and integration if we intend to use this lowest tier as the integration tool to help us systematically integrate all the information from the various key technical issues and the plan. The status of issue resolution will continue to be documented in the IRSR. DR. McCONNELL: If I could, Christiana, this is where we are implementing the TPA effort into the Yucca Mountain review plan. It says this lower tier on this diagram, which you should be pretty familiar with, is where we used the information from our reviews of DOE documents as well as our own IPA efforts to identify those elements of the performance assessment that are most important. So here's how we're integrating the TPA effort into this process. And you'll see more of it as Christiana goes on. MS. LUI: Well, actually, Keith, you have jumped ahead -- DR. McCONNELL: Sorry. MS. LUI: To Slide Number 12. DR. McCONNELL: Okay. MS. LUI: Okay. The bottom tier here, we have retermed them as integrated subissues. It seems everybody's a lot more comfortable in terms of talking about subissues. You may also know the lowest tier as key elements of subsystem abstractions or KESAs. But in the review plan we will not name them KESAs, rather, we would like to call them integrated subissues, because here it is really integrating all the efforts that have been taking place by the staff up until this point. And what these integrated subissues are, they are really developed from a top-down approach, and developed based on as Keith has indicated review of DOE's TSPAs, knowledge of current design options and site characteristics, and the staff's IPA work, which basically relies upon TPA a lot. And they are integrated processes, features, and events that could impact system performance. This particular framework provides KTI an integration tool to describe their contribution in the context of total system performance assessment, and also helps us to facilitate integration at the technical staff level. For example, many KTIs require interaction with other KTIs in evaluating repository performance such as for waste package corrosion, which is basically the very first box on the lowest tier on page 11 of the flowdown diagram. The current KTI division is not -- the waste package corrosion issue cannot just be addressed by any one single KTI. It's a concerted effort between thermal effects on flow, the near field environment, the container lifetime and source term, and of course we use the TPA code under the TSPA KTI to help us evaluate the waste package corrosion issue. The next page, after giving you some background information now I would like to give you some more detail in terms of the performance assessment review, which is basically Chapter 2(b)(1) under the outline for the Yucca Mountain Review Plan. In this particular chapter we have basically four subgroups. This is some description and demonstration of multiple barriers, scenario analysis, model abstraction, and demonstration of the overall performance objectives. We fully expect that DOE will do analysis to show how they are taking credit for the various features, events and processes or a combination of them to satisfy the multiple barriers requirement, therefore upfront we would like them to summarize what they have done, what they are taking credit for in terms of demonstrating the multiple barriers requirement. That will also help the Staff to focus on our review in terms of what DOE is relying upon. The next step is scenario analysis. Basically here we will expect DOE to clearly identify what they have taken into consideration in their post-closure safety case and what they have excluded and what are the technical bases for inclusion or exclusion. In other words, this part will set up the boundary condition for the performance assessment. After that part is done, then we look at the model abstraction, what has really gone into the calculations and here is 14 integrated subissues that we have shown on the lowest tier of the flow-down diagram. We basically help the Staff in focusing our effort and focusing and integrating the information that we have published in the IRSRs. Lastly, after reviewing all the previous three parts, we will make a determination to see if DOE has truly demonstrated they have met the overall performance objectives in terms of dose requirement and the multiple barriers requirement. Here is also the place where we will be developing acceptance criteria review methods in terms of the transparency and traceability of DOE's analysis. I know this is a very, very busy slide but I just wanted to keep everything together rather than have you flip back and forth. DR. GARRICK: I understand. I like to do that once in awhile myself. I am just trying to interpret it in terms of, say, the multiple barriers and where you really address those. MS. LUI: The multiple barriers would be the very first group. There are three technical requirements associated with multiple barriers. The part is to identify, is to require DOE to identify what barriers they are taking credit for and the second part is to provide the technical basis, and then -- sorry, I forgot what's the third part. Tim? MR. McCARTIN: The middle one is the capability and the last one is the basis for the capability. MS. LUI: So basically the technical requirements are laid out in the proposed Part 63. What we will do in the Review Plan is to develop acceptance criteria and review procedures and most likely we will also provide guidance in terms of what DOE can do to demonstrate multiple barriers. DR. GARRICK: Okay. MS. LUI: Okay? The next three pages basically highlights what I have talked briefly about five or 10 minutes ago is the crosswalk that all the KTIs have done to relate their subissues to the integrated subissues. These are the existing relevant KTI subissues. As we go through this process ourselves, we may find that there will be knowledge and expertise needed to bridge the gap in the integrated subissues that currently do not have a KTI subissue associated with it. This will help us to focus our efforts by using a top-down approach. You probably have seen this particular crosswalk in one form or the other when you received the IRSRs from us, because all the KTIs have basically done this particular crosswalk and here it is just a summary of all the information that had been published previously. The next two pages, page 17 and 18, as I have mentioned previously because we are trying to implement this performance-based approach we need to provide the licensees the flexibility in terms of how the licensees intend to address their safety case. In the viability assessment, in TSPA VA DOE basically has identified 19 principal factors that are of greatest importance to post-closure performance. I just want to demonstrate that the integrated subissue approach will allow us to review DOE's safety case that's built on these 19 principal factors. Even though there is not necessarily one-to-one correspondence, we do have every single principal factor covered in the proposed approach for the Yucca Mountain Review Plan. DR. McCONNELL: Can I just add that we intend to work with the Department of Energy over the next couple of months in trying to come closer together perhaps on matching their elements with our integrated subissues. In fact, we have talked to them briefly and they seem to be willing to do that, so there may be more coalescence as we move on. DR. GARRICK: But for now they have all been accounted for. MS. LUI: Yes. I would like to point out that the 19 principle factors do not address disruptive events, but in DOE safety strategy, they do have attributes that talk to disruptive events and we do have integrated subissues that will review the disruptive events process such as mechanical disruption or waste packages and igneous activity integrated sub-issues. So even though you don't see the disruptive events here, they will be included. And, finally, the third part, advantages of the approach. Review of both the pre-closure and post-closure safety cases are performance-based because from the very beginning we are set out to determine if DOE has met the performance objectives or not, and we use the top-down approach to evaluate the license application. It will encompass all the related activities such as site characterization, experimental work and all the way up to performance assessment. By using the top-down approach, the iterative cycle of performance assessment, data collection is clearly and closely maintained. And in this particular framework we can clearly indicate why DOE's supporting data is acceptable or deficient in the context of how that piece of information has been used to support DOE's safety case. And as we go through this particular process, as we go through the integration of all the acceptance criteria and review methods currently in the IRSRs, we will attempt to minimize any duplication and also modify or eliminate, possibly -- I would like to emphasize the word possibly, overly prescriptive acceptance criteria currently out in the IRSRs. And the requirements under the content of license application, which is paragraph 63.21, and the requirement for addition information, RAIs, are clearly justified in this particular context. And we certainly -- and we hope by using, by implementing this particular framework, it will lead to a streamlined, transparent and integrated review plan. And on the last page are the seven activities that we have identified related to the development of Yucca Mountain Review Plan that will be of interest to the committee. We are planning on a technical exchange in the last week of May to talk about the Yucca Mountain Review Plan with DOE and our attempt is to get understanding from them, and also to work on possibly having a consistent approach between the review plan, between the license application and also leading to a consistent Safety Evaluation Report at the end. Currently, the staff is working on Revision 2 of the IRSRs and they are -- all of them will be schedule to go out of the NRC by September 30th, 1999. And the materials in the IRSRs will be appropriately integrated into the Yucca Mountain Review Plan, or be referenced by the review plan. The acceptance criteria and the review methods will certainly be integrated into the review plan. However, some of the technical basis section will probably be left in the IRSR and will be referenced by the Yucca Mountain Review Plan. Activities 3, 4 and 5 are inter-related. We are currently scheduled to provide to the Commission the final Part 63 package and also an annotated outline of the Yucca Mountain Review Plan by November 30th, 1999, however, there is a possible postponement to February 2000 which are currently being talked about. But in any case, once we have submitted to the Commission the final rule package and after the Commission has approved that, we intend to have public meetings with the state and county and also interactions with DOE to present and clarify the final Part 63 and also the accompanying Yucca Mountain Review Plan. The Rev. 0 of Yucca Mountain Review Plan is currently scheduled to go out on March 31st, 2000, but be aware that, as Keith has mentioned, the pre-closure part is not as mature as the post-closure part at this point. So in the Rev. 0 Yucca Mountain Review Plan, it will contain some "to be determined," or "to be developed" sections. But as we proceed along the line, those sections will be filled in. And we will publish future revisions of the Yucca Mountain Review Plan before the key DOE milestones such as site recommendation and license application. And during this entire process, we will come back to the committee to brief you on our progress and seek advice as appropriate. This basically concludes the formal part of my presentation. DR. McCONNELL: Okay. What Christiana does have at the back is some backup slides to kind of go through one of the integrated subissues and defines how we are starting to bring in the acceptance criteria from the various KTIs into the integrated subissues, and I'm not sure if you want to go through that in this forum or not or just have it for your reference. DR. GARRICK: Well, I think for now we'll use it as reference. One of the things that the Committee has been anxious to resolve, and this begins to address that, is the interrelationship of the KTIs and the TPA, and the ability for us to be convinced of what's really driving what. And I suppose that we're going to have to, you know, see some detail in order to understand that. But the one concern we have had is that the KTIs would assume a sort of a sacredness of level with respect to being an issue, and whether or not the iterations of the TPA would indeed be able to map to the KTIs in a way that would importance rank them, for example, and keep their importance in some sort of order in relation to a risk perspective, which kind of brings me to some overarching observation. You do put a lot of effort in this presentation with respect to it being performance-based. Not much is said about it being risk-informed except indirectly as it relates to the performance assessment. And I guess I'm just raising the question, is the risk perspective really in the review plan? MS. LUI: I believe it is, because we have put a lot of focus in terms of what's the expected evolution of the repository. DR. GARRICK: Um-hum. MS. LUI: Basically staff's position is this is what we anticipate to happen sooner or later. So in that perspective the probability is close to 1, except for the disruptive events which we do have screening criteria set up in Part 63 to talk about the probability part. DR. GARRICK: Um-hum. DR. McCONNELL: I'd just add that implicitly -- I think what Christiana's pointed out is implicitly it is in what you've seen today, and it's explicitly -- implicitly in the sense that in the integrated subissues basically the approach we've been taking in evaluating DOE submittals like the VA was to look at the integrated subissues, look at their contribution to the risk in the sense of the PA, in the PA calculation, and use that as a guide for where we would concentrate our review. So in our own sensitivity studies where we look at the risk sensitivity of the various integrated subissues, we've used that to focus it so implicitly it's in this diagram and in the review plan. We are I think now embarking -- and the KTIs are embarking on a path where they are using risk-informed to look at their own subissues within a KTI, and that then will flow up. This is one case where things will flow up into the PA where we will concentrate our effort when we look at our own code and also when we look at perhaps the identification of other integrated subissues. So I guess the message I'm trying to say is it's implicitly in here, and as we go along this process we'll probably do more to explicitly bring it into the review plan. And I think you'll see that. DR. GARRICK: Well, maybe the other Committee Members can see it more clearly than I. I'm still struggling with trying to understand how the risk perspective really is a management tool for the review process. You talk a lot about a top-down approach. My idea of a top-down approach would be a series of scenarios that would characterize the risk of the repository and then a ranking of those scenarios in terms of their contribution to the overall risk, and then the fallout from that of these items in these lower boxes in terms of their contribution to first the scenario, the risk of that particular scenario, and, second, the risk of the aggregation of scenarios. I'm still looking for something that gives me a comfortable feeling that there's a real genuineness here in the implementation of a risk approach to determining the performance, the safety performance of the repository. And I hear and see a lot of things that look like you're trying to do that. DR. GARRICK: But as I say, it is not sufficiently transparent to me yet to really be convinced that that kind of thought process prevails, and I yield to the other members to comment. DR. HORNBERGER: Well, I am just curious whether or not except for disruptive events we have a scenario driven approach. DR. McCONNELL: Well, I think we do, but maybe Tim can -- MR. McCARTIN: Well, I don't know if this will scratch the itch or not, but let me try. I think what Keith and Christiana have both said, I mean the KESAs or those subissues are the areas that we believe, based on our analyses of the site with TPA as well as all the KTIs, the information they have had, that need to be addressed, the risk part of it gets tricky because DOE is the one that is going to demonstrate performance and they have a variety of ways to deal with those subject areas. There could be some of those subject areas -- we are going to do a bounding analysis here or use a bounding parameter because it doesn't have much impact on the repository performance but that is their demonstration and we have to -- they have that flexibility in looking at the risk, with the rigor that they will attach and the depth of detail and that part, other than they have to address those issues, how they address them I think is getting to your risk part, and it is their demonstration, and our review will have to be flexible enough to, well, if it is a bounding analysis there is one way we will look at the information versus detailed information, et cetera, and hopefully we will try to capture that in the acceptance criteria. DR. McCONNELL: But we do try to use a risk basis for judging where we place the emphasis in our reviews and that is based on our own sensitivity studies, notwithstanding the fact that DOE does have the responsibility for making the safety case. DR. HORNBERGER: No, I mean I think that is -- that much is clear to me and I think that you are doing exactly what you said. I was referring to, I was trying to get my hands around what John suggested in a risk approach, which was a down-down approach where you started by defining a whole suite of scenarios that would lead to a failure, if you like, and I am not sure that we have that. I think that as Christiana said, this is going to happen and we have sort of one scenario and the scenario is that it is going to happen, that we are going to get this solution and migration and transport and everything else. Well, correct me if I am wrong. MR. McCARTIN: Well, DOE will have the job of pulling together all that. We will review that for completeness but I don't know -- DR. HORNBERGER: I'll rephrase it. From what you have seen DOE do so far, do you anticipate having a scenario-driven approach to review? MR. McCARTIN: Yes. Okay, and give me some examples of the suites of scenarios that you anticipate, excluding disruptive events. I understand that one. @@ MR. McCARTIN: Excluding disruptive events? Oh, you mean a base case that is primarily -- has more than one scenario? Well, to date, I would say the base case that is being analyzed has the uncertainty being involved be it seepage, be it the amount of fracture flow, is somewhat encompassed in the uncertainty of that base case. Now I guess if you want to call that one scenario, I suspect that -- DR. HORNBERGER: I'm just trying to understand. I don't mean to argue with you. MR. McCARTIN: Right, right. DR. HORNBERGER: I am just trying to see if I understood John's question and your response in terms of it and I understand what you are saying. MS. LUI: I think the closest that we have come to in terms of separating out scenarios probably would be in the area of alternative modelling approach, because there are various -- well, various approaches in looking at for example how a waste package could fail. In our Review Plan we will talk about, we will require DOE to look at all the credible alternative approaches but I don't believe we will ever get to a point of assigning probabilities to those models. The best we can do is to look at the technical basis that has been provided with those alternative approaches and look at which particular approach DOE has decided to use in supporting their safety case and the Staff's evaluation will be looking at how defensible that would be when we do our safety evaluation. I think that's probably what Dr. Hornberger was driving at. DOE at one time was going to give us everything together. In other words, they were not going to separate out if they have two or three different model approaches. They were just going to combine everything and give us the final results altogether and we have stressed to them over and over that we want to see the results being presented separately and also the technical basis that had been used to support them, so we will see that in the Review Plan. DR. McCONNELL: Where do stand on answering your question, Dr. Garrick? DR. GARRICK: Well, all of it is somewhat helpful. I am sure it is a question I am going to continue to struggle with for awhile and we will discuss it in future sessions. I think that one of the problems is that I have a certain notion of what constitutes a risk approach and I am having to back away from that considerably to appreciate fully what you are doing here, and it is going to take a little time to be convinced that the mapping that you have done is indeed done from the point of view of the right logic engine being in control, and the logic engine that I am most interested in is the one that is driven by what is most important with respect to the performance parameter. In this case it will be a dose standard of some sort, so I think that what you have attempted to do here in decomposing the KTIs into subissues and matching them up with both your own and DOE's descriptions is an important step. And it's part of the overall puzzle. Whether or not it's enough to really let the important contributors drive the process, I don't know, I'm going to have to convince myself of that. DR. McCONNELL: Okay. And in succeeding briefings we'll try to emphasize that. DR. GARRICK: Charles? DR. FAIRHURST: Yes, it's quite a lot to swallow at one time. Let me ask a separate question almost. I'm particularly interested in what factors control and what certainties there are, seepage into the drift. Nothing going beyond, but if I wanted to know, for example, infiltration rates, your model for fracture flow, matrix flow, et cetera, thermal-mechanical effects around the excavation. It would help an awful lot in understanding the approach to the engineered barriers if one could first understand that subset of the problem. Would you have through this, if I were to ask you, would you have an independent way of presenting, you know, a response to that question without waiting for DOE to give you sort of its read on it and then what you're going to do is to say it looks reasonable to us? I mean, how independent is your assessment going to be? I've got a lot of other questions besides, but I don't know whether I've indicated clearly. It would help me if NRC rather than just DOE could give some input to that earlier on rather than later. MS. LUI: Okay. I think I've mentioned that. We all recognize that there's a natural tension in between who's leading whom in a way in the performance-based -- I mean risk-informed approach. Our job is really not to do licensees' work for them. Rather, our job right now is to basically develop staff's capability in reviewing what DOE might come in. So in our TPA code, that's sort of our attempt to utilize what we know, and staff's relatively independent thinking in what's required to be in the assessment so that you'll be defensible to help us down that path. DR. McCONNELL: So I think the bottom line is we have the capability, but it's right now in the existing TPA code to look at that particular factor. DR. FAIRHURST: That's right. DR. McCONNELL: But we're waiting for DOE to demonstrate what they intend to do, and then we'll review that using this capability that we have. If that isn't too jargonistic or bureaucratic. MS. LUI: And all the interactions that take place between the staff and DOE is our attempt to learn as quickly as possible what's DOE thinking and what's DOE's approach, and develop our own toolbox in order to be ready. DR. FAIRHURST: Yes. MS. LUI: For that job. DR. McCONNELL: To give you another example, I think we have the capability, and Tim probably could correct me if I'm wrong, to mimic things like drip shields and other factors that might be important in water contacting waste or the waste package or things like that. So we have the capability, and if DOE tends to go that way, then we will probably move our program in that direction, so we'll be in a position to review it. But we're not going to be probably out ahead of DOE on these particular issues. DR. FAIRHURST: No, I agree, not out ahead, but at least giving some insights into weights or various -- I don't want to call them risks or what, but just where the potential is. For example, to reduce uncertainties through engineered barriers or through similar -- just to get an idea of what level, for example, of infiltration would be such that from that point on wouldn't likely violate, you know, the regulation or standard. DR. McCONNELL: I think we have that capability. Maybe we need to demonstrate what we have. DR. FAIRHURST: I would like to see something indicating what that is, for that specific part of it, because I think that from that point on if, for example, you can show that you can avoid a critical rate of inflow, then things beyond that become a little less -- I'm not saying that they're not important, but that becomes a particular cutoff. Let me ask another couple questions. At some point -- right now in a probably very constructive way you're having a dialogue with DOE about how they're moving forward and so on, but at what point, and I also see that there's a mention of 2001, I thought it was 2002 that you're going to get the license -- MS. LUI: Okay. Are you talking about the last entry on the schedule? DR. FAIRHURST: I'm talking about safety -- I'm talking about license application as you'd get it in 2002, right? MS. LUI: Right. Right. DR. FAIRHURST: You didn't put 2001. I heard it earlier as a -- MS. LUI: Okay, 2000 -- okay, September 30, 2001, that's when we plan to publish Rev. 2 of the Yucca Mountain Review Plan. That's about five or six months ahead of the expected license application. DR. FAIRHURST: Okay. So it's the review plan that you're going to concentrate on. MS. LUI: Right. DR. FAIRHURST: Okay. MS. LUI: Right. DR. FAIRHURST: And at what point before that do you feel that you have to sort of cut off dialogue with them so that -- at some point they have to move independently from you, present something that you are going to then independently review; right? DR. McCONNELL: Yes, I think that -- I think our position is we're doing that now, that they have to do their work to demonstrate safety. We're doing our work to build our capability to have the independent review there. But this can't proceed without some interaction. DR. FAIRHURST: Right. That's what I'm getting at. DR. McCONNELL: And that's where we are. And I don't think at least, I could be corrected, I don't think we anticipate cutting off the prelicensing consultations at any point right now. DR. FAIRHURST: Okay. DR. GARRICK: Ray. DR. WYMER: I've been waiting to see some of the criteria that you use to evaluate the license application sharply enough drawn that anybody who looks at them can decide exactly what they have to do in order to satisfy the criteria. I presume they'll come sort of in a general sense from the KTIs and the related types of things. MS. LUI: Okay. I'm not sure if you are familiar with the work that has already been done by KTIs such as effects on flow and container life and source term. DR. WYMER: In a general sense. MS. LUI: The attempt is to integrate the acceptance criteria and review methods currently in those IRSRs as the first step towards what's going to be in the review plan. In fact, if you spent a little bit of time looking at the example, as the backup slides to this particular package, you will probably get a flavor of the path that the staff is taking. DR. WYMER: Yes. It seems to me those criteria are fundamentally important. MS. LUI: Right. Those acceptance criteria are basically formulated to address the technical requirements in the rule. You will find a lot more detail in the review methods or review procedures that are associated with each of the acceptance criteria. DR. FAIRHURST: With the intrusive events, we'd surely be talking about vulcanism and things of this kind, but how are you going to deal with -- I heard expert group, but with human intrusion? MS. LUI: Human intrusion is a stylized analysis -- DR. FAIRHURST: Of a single hole, right? MS. LUI: Right, a single bore hole, right. And in the -- I have a place in mind. I mean, that's me speaking, for where it is going to appear. In fact, you can probably see that on page -- I forgot which page -- it's page 13 to B-1. It did not show, it did not say human intrusion but 63.113(d) is a post-closure performance objective. DR. FAIRHURST: 63.113(d)? MS. LUI: (d) as in "David" is performance objective for evaluation of human intrusion. DR. FAIRHURST: And does that stylizing involve penetrating a waste package or something like that? MS. LUI: Yes. DR. FAIRHURST: Okay. MS. LUI: So it is incorporated but again this is a work in progress so we don't have the detail for you at this point but as we develop more detail we will come back and talk to you about it. DR. FAIRHURST: But presumably it would involve bringing cuttings to the surface that are radioactive? It didn't involve any flow? DR. McCONNELL: Tim? MR. McCARTIN: Well, to an extent but the critical group is still the same as what is done for post-closure and we would not expect any drill cuttings at the surface to have any significant impact at 20 kilometers away. DR. FAIRHURST: Right, okay. MR. McCARTIN: There isn't -- the stylized calculation uses the biosphere critical group assumptions as is done in the PA, so there isn't another -- there isn't an attempt to look at for example the well drilling crew. There is not a dose to the well drilling crew. DR. FAIRHURST: Maybe I am naive, but what impact could human intrusion have --? MR. McCARTIN: Well, as the NAS recommended, the desire of this calculation was to see the resiliency of the repository for the critical group and not necessarily trying to look at, say, what would happen to the well drilling crew, whatever, so it is looking to see would a single penetration of the repository so degrade the performance that the critical group would be adversely affected and -- DR. FAIRHURST: I see, okay. MR. McCARTIN: My understanding that part of that was a reaction to the experience at WIPP, where a well penetration was the only significant source of getting consequences and that is pretty much where the recommendation came from, but it was looking at the overall repository performance in the context of the overall performance assessment. DR. FAIRHURST: As I recall, in the TYMS report, there was a suggestion that one consider a second well if the effects were not additive. In other words if some mechanism for release were generated which would not be there by the one, which is again coming back from WIPP. MR. McCARTIN: Yes. DR. FAIRHURST: E-1, E-2 thing where two wells causes -- MR. McCARTIN: Right. DR. FAIRHURST: But it seems like it is a very different kettle of fish. DR. GARRICK: When you talk about a stylized scenario, how far can you take it? Drilling a hole into a canister is one thing, drilling a hole into a canister and bringing cuttings up to the surface is another thing and drilling a hole and bringing cuttings up to the surface and having a rain-out and a stream is a third thing and so on and so forth. MR. McCARTIN: Right. Well, the way we have proposed it in Part 63, and we will be interested to see what kinds of comments we will get, which we have not gotten any to date, but on the stylized calculation but we are assuming a somewhat typical drilling event where the cuttings would go up to the surface. We would not suggest that the cuttings, say, go down to the water table, but we would expect a penetration through the waste canister all the way to the water table so you now have a relatively fast path of potential ingress of some water from the surface to the waste package and to the water table, and you would be examining that particular event. DR. GARRICK: So in principle you could have a couple of pathways. You could have a pathway -- I am thinking of to the critical group -- you could have a surface pathway and you could have a groundwater pathway. DR. HORNBERGER: A surface pathway airborne? DR. GARRICK: Well, either airborne or liquid. DR. FAIRHURST: Where, on the surface? DR. GARRICK: Well, what happens when it rains? DR. HORNBERGER: In the Amargosa Valley there is no stream. MR. McCARTIN: We are not anticipating any impact from the surface cuttings but you're right. I mean there is a potential. DR. FAIRHURST: There is another big difference between WIPP and Yucca Mountain and that is what you put the waste in. We're having what? -- 10 to 20 centimeters stainless steel and then C-22 alloy, so for a long period of time the probability of a drill ever getting through that is extremely small. If you take standard drilling practice, the moment anything hits a metallic it wrecks, so -- DR. HORNBERGER: Titanium shields. DR. FAIRHURST: There's titanium shield, the lot, you know? DR. HORNBERGER: Steel, thick steel. DR. FAIRHURST: These are 55 gallon drums at WIPP -- DR. GARRICK: Yes, but the concentrations between the two are widely different. DR. FAIRHURST: I agree. MR. McCARTIN: It certainly is conservative to assume at an early time someone can drill through that container. It would take a fairly -- DR. FAIRHURST: Aggressive? MR. McCARTIN: -- dedicated well drill team. [Laughter.] DR. HORNBERGER: A laser drill in the next century. DR. GARRICK: All right. Any other comments, questions -- any questions from the Staff? DR. CAMPBELL: Yes. One of the areas when the committee reviewed the viability assessment we also looked at the IRSRs in Rev. 1 to the IRSRs and you mentioned the alternative models that are part of the review methods as DOE considered alternative models. When I read that in -- there is some version of it in each of the IRSRs -- it appeared not to be the intent of the Staff to say that they had to use the most conservative conceptual model, so in reviewing whether or not DOE has considered appropriately the alternative conceptual models obviously you have in your mind some concept of probability for the different conceptual models, and I was wondering how that is factored in and where it may come from. MS. LUI: That probability or the nonexistence of that probability basically depends on our current understanding, our current belief of what is most likely going to happen. You can always assume the worst case scenario -- DR. GARRICK: You're already got it -- MS. LUI: -- like for example take, say, wind direction, wind rows. It is not likely that a wind row always blows toward the critical group but there are aspects in DOE's analysis that they can decide, okay, we are just going use that or we are trying to defend what is going or what is not going to happen, let's just take what we believe at this point is going to be the most conservative approach, so it's going to be a knowledge dependence there. DR. GARRICK: It isn't in here? DR. McCONNELL: Yes, I think DOE if they want to can take a bounding approach to particular issues or factors including conceptual models and so we are not telling them they have to do that, but we are also not limiting their flexibility in doing that in particular areas where they may not want to narrow the uncertainty. They may just want to bound the problem. I don't know whether we are addressing your question or not. DR. HORNBERGER: I was just going to say it ties in with that question, I can ask it in this context, that Andy did. Of course there are similar questions that one can ask about all the acceptance criteria, which are a lot of them are you are going to judge that DOE has presented sufficient data to support the conceptual model or the model for seepage into a drift, et cetera, et cetera, and it is one of these things where it is I don't know how to define it but I will know it when I see it, you know? I will know it is sufficient when I see it, and the question is, as Andy said, with different conceptual models somewhere back there you are either going -- you either have a notion that yes, an equivalent continuum is reasonable for this place, and if they have that as their primary model it is okay, or you say, boy, an equivalent continuum model just won't do and if DOE comes forward with that, that is not going to be sufficient, so you have in your mind already -- you have to have a subjective probability associated with all these things. I think this is probably okay as long as the license application does in fact come through in 2002 because there will be corporate memory here and you people have interacted. If for some reason the license application doesn't come through until 2012, which we may laugh at now but it is not all that inconceivable given the history of the program shall I say -- There will be limited corporate memory and some of these ideas may change. Maybe that's okay, too. MS. LUI: Well, I hear what you're saying, Dr. Hornberger, but we all have to understand that the performance assessment, why is it kind of not as transparent to outsiders, because there are so much interdependency on the various parts. DOE may elect to do a conservative analysis for one part because they are being, quote unquote, as realistic as possible to other parts, or they may change their approach. So there will be billions of permutations out there. It will be extremely difficult for us to write down every single thing. DR. McCONNELL: Yes. And I think we're trying to use the issue-resolution status report too as a mechanism for documenting some of the thought process that's going along. So hopefully if it does -- DR. HORNBERGER: Which is more important, because, as Christiana said -- I agree; I wasn't suggesting that you could write all this down. You can't. You're doing what you can. But it's still this philosophical question you're left with. And I think that that's right. If you write down the thought process, that's the best record you can have. DR. CAMPBELL: Are you guys going to still brief the Committee on sensitivity studies from TPA 3.2 in the near future? Is that still on the books for San Antonio? DR. McCONNELL: I believe it is; yes. We intend to do that. We have an interchange with DOE at the end of the month, and then depending on I think the Committee's agenda, I think we'll try to fit in as much as we can. But I think we'll have to talk to you and find out what your schedule's like. DR. CAMPBELL: Because it might be very useful. DR. WYMER: We have been after that for a long time. DR. McCONNELL: Okay. DR. GARRICK: Any other questions or comments from Committee staff? Anybody from the audience? If not, we want to thank you. I don't think at this point you're asking for a letter. Unless we in our discussions of it find some issues that we feel very concerned about and want to call to the Commission's attention, we'll take this as information for now, and expect to hear from you again. Okay. MS. LUI: Okay. DR. GARRICK: Okay. Thank you very much. MS. LUI: Thank you. DR. GARRICK: Let's see what's on our agenda now. DR. FAIRHURST: That's it. DR. GARRICK: It's up to Charles. Do you want to do your reports as we had planned tomorrow when you do the rest? DR. FAIRHURST: Sure. Yes, because I'll have copies of the white paper too, but we can talk about that later. DR. GARRICK: In view of that, and unless there's further comment, questions, we will adjourn. [Whereupon, at 5:20 p.m., the meeting was concluded.]
Page Last Reviewed/Updated Friday, September 29, 2017
Page Last Reviewed/Updated Friday, September 29, 2017