121st Advisory Committee on Nuclear Waste (ACNW) Meeting, September 20, 2000
UNITED STATES OF AMERICA NUCLEAR REGULATORY COMMISSION *** ADVISORY COMMITTEE ON NUCLEAR WASTE *** 121ST ACNW MEETING PUBLIC MEETING *** Ballroom B Crowne Plaza Hotel Las Vegas, Nevada Wednesday, September 20, 2000 The Commission met in open session, pursuant to notice, at 8:00 a.m., B. John Garrick, Chairman, presiding. COMMITTEE MEMBERS PRESENT: DR. B. JOHN GARRICK, Chairman, ACNW DR. RAYMOND G. WYMER, Vice Chairman, ACNW MR. MILTON N. LEVENSON, ACNW Member DR. GEORGE HORNBERGER, ACNW Member. STAFF AND PARTICIPANTS: DR. JOHN T. LARKINS, Executive Director, ACRS/ACNW MR. HOWARD LARSON, Acting Associated Director, ACRS/ACNW MR. RICHARD K. MAJOR,, ACNW Staff MS. LYNN DEERING, ACNW Staff MR. AMARJIT SINGH, ACNW Staff DR. ANDREW C. CAMPBELL, ACNW Staff JAMES E. LYONS NEIL COLEMAN, NMSS WILLIAM REAMER, NMSS DR. JOHN TRAPP, NMSS. P R O C E E D I N G S [8:00 a.m.] MR. GARRICK: Good morning. The meeting will now come to order. We welcome Jim Lyons, joining us on the staff side of the table, the new Director for Technical Support for the ACRS and ACNW. MR. LYONS: Thank you. I'm glad to be here. MR. GARRICK: Good. This is the second day of the 121st meeting of the Advisory Committee on Nuclear Waste. My name is John Garrick, Chairman of the ACNW. Other members of the committee include George Hornberger, Ray Wymer, and Milt Levenson. The entire meeting will be open to the public. Today, the committee will hear a project overview from the Department of Energy; a Department of Energy representative on Yucca Mountain will give a status report on the site recommendation consideration report from DOE; discuss major aspects of the total performance, system performance assessment, the site recommendations version; hear an update on chlorine-36 from DOE and the M&O; discuss the results of ongoing studies on the fluid inclusion issue from a panel composed of representatives of University of Nevada-Las Vegas, DOE, and the state; and, review relevant activities and status of site tour scheduled for tomorrow. We'll also continue to work on preparation for the meeting that the committee has with the Commission, originally scheduled for October 17, now rescheduled for December 18. At that meeting, we intend to discuss such issues as site sufficiency review, risk-informed regulation in NMSS, Part 71, ACNW action plan and priority topics, and we will also discuss the recent trip taken by ACNW to England and France. Andy Campbell is the Designated Federal Official for the initial portion of today's meeting. This meeting is being conducted in accordance with the provisions of the Federal Advisory Committee Act. We have received one request from a member of the public regarding today's session. Mr. Corbin Harney, who spoke to us yesterday, wants a few more moments. I'm suggesting that we maybe do that just before the break this morning. And if others of you wish to make comments, please contact a member of the staff. Also, please use one of the microphones, identify yourself, and speak clearly. Okay. With that, I think we'll jump right into the first presentation. I'm going to ask each of the speakers to introduce themselves, tell us what their positions are and organizations and what have you. MR. DYER: Good morning. I'm Russ Dyer. I'm DOE's Project Manager at Yucca Mountain and I can't tell you what a delight it is to be here this morning. I just finished a three-week detail in Washington, got in last night. Next slide, please. I'm going to give a quick overview. There are about six topics that we're going to go through, look at a little at the project accomplishments over the past year, design evolution, the role of the site recommendation consideration report, the progress that we're making toward the closure of the NRC key technical issues. My understanding is I think that was gone over in considerable detail yesterday, so I'm just going to hit the highlights. Integrated safety management system is an initiative that is going on across the Department of Energy throughout the complex. We are finishing up our last phase validation this week on that. That is something that has taken a lot of energies, just to tell you a little bit about that, and then a quick project outlook, what lies ahead of us here. Next slide, please. Accomplishments. Next slide. If we look at the general schedule of things here, running out through the site recommendation and its associated final environmental impact statement and then further on out to the license application and eventually operations. We have put the draft environmental impact statement on the street. We've had public hearings on that. The next major activity that's called for in the Nuclear Waste Policy Act is the site recommendation. We are putting out a site recommendation consideration report this fall/winter which lays out the technical basis at this point in time to support and underlie that site recommendation decision. Next slide, please. The things that we got done so far in 2000. Underlying the site recommendation consideration report are a series of process level model reports and then below those lie the what are called AMRs, analysis and model reports. I'll show you a diagram in a little while which shows the cascading hierarchy of technical supporting documents, all of which get summarized up into a site recommendation consideration report and then finally built into the site recommendation. There are nine process model reports. We have accepted seven of them unconditionally. The remaining two PMRs have been accepted with conditions. There were 121 supporting lower level, more detailed analysis and model reports, 119 of those have been completed. On the design side, the pertinent document there is called an SDD, the system description document. There are 24 of those that lie within our plans. All of those have been -- 24 that are needed to support SRCR. There are others that will be needed eventually to support the license application, and those have a further finer level of detail in the engineering analyses and 34 of those have been completed. If you have been following the repository safety strategy, about a year ago, we determined that we had focused entirely on the post-closure. We needed also to put in a pre-closure element of the repository safety strategy and we've completed a preliminary cut at the pre-closure safety assessment. The total system performance assessment for the site recommendation has been completed. Bob Andrews will talk to you much more about that, and about the base case calculations for the TSPA-SR and the sensitivity calculations that have been completed and are also in progress. Next slide, please. In the testing arena, and Mark Peters will talk to you in considerably more depth about this and we'll see more of it tomorrow, we've started hydraulic testing in the luvium single well pump test, that's been initiated in one of the Nye County holes down in the south of Yucca Mountain. We should start multi-well testing and tracer testing in '02. Within the frost drift hydraulic properties and seepage testing has been initiated in the lower lith, lower lithofizal unit. That's primarily in -- that would be in niche five. The drift-to-drift seepage test is underway in alcove eight. And yesterday, I believe, although I didn't hear confirmation, we turned on a test over at the Atlas facility, over on Locie Road, an engineered barrier system ventilation model test. It's a scale model of surrogate waste packages that have heaters in them and we're looking to validate our models to look at how ventilation can remove heat from that system. And the geo technical investigation of north portal surface facility is underway. This is to provide us more information in the seismic design arena. Next slide, please. Data and software qualification progress. The commitment was that we would have 80 percent of our data and software qualified at the time of the SRCR. We're well on the way there. We're actually over the target with the software qualification. We've got a little bit to go with the data qualification. Our draft rule for the replacement to 10 CFR 960, 10 CFR 963, we went through the public hearing process on that. There was a draft final rule that we've submitted to NRC for concurrence back in May and, of course, that's all tied into the 197 and 63 issue. Planning process. Of course, like most -- well, every other DOE agency, we're anxiously awaiting Congress to work on the appropriation. There is supposed to be a conference committee scheduled for today and I'm hopeful that by the end of today, we'll have an idea of what our budget for 2001 will be. Of course, we've done planning based on assumptions. We have a lot of stuff that we would like to do. It does not look like that we will be able to bring in the entire 437.5 million dollars that the Administration requested for the project and the program for this year, but it will be reasonably close, I hope. Of course, that's the caveat at the bottom. We have a lot of plans, but you can only do what you can afford to do. Next slide. Design evolution. The design has evolved somewhat since the EDA-2 design of a couple of years ago. One of the major changes is as we looked at the waste stream inventory coming in, the waste stream, in general, is somewhat hotter because there's higher burn-up fuel as part of the waste stream. Whenever we reevaluated the heat that was in the system, of course, one of the requirements that we have on the system is a maximum 350 degree Centigrade centerline temperature in the waste packages, so that the cladding is not subjected to unzipping. And when we looked at the effects of the blanket, the thermal blanket that would be put in if you put in backfill, it looked like we would violate that. So part of the exercise that we've done over the past year is looking at updating the design to remove the backfill, which, of course, gives you a system which then could be ventilated for a considerable period of time. This is just a schematic of what we're looking at now with, again, the horizontal cylinders emplaced on pylons. We now have a drip shield that is part of the system. The drip shield is of titanium, a titanium alloy. The material for the waste package is still a nickel alloy corrosion resistant material on the outside, with, I believe, a stainless steel, essentially a structural material on the inside. Other things, we've gone away from a concrete lining in the emplacement drifts which we had several years ago. Right now, the ground support is just structural steel. Next slide, please. The design, of course, has not been static, continues to design, continues to evolve. Looking at the underground, there are a couple of variables that we can play with reasonably simply. Whenever we talk about design, it's really a combination of design and operational parameters that you can look at juggling. For instance, you can take the same design, that is, waste package design, the same emplacement drift specifications as far as diameter spacing, ground support that you use, and you can operate that system in different ways by either varying the spacing between the waste packages, so tailoring the line thermal load within the emplacement drifts, or by using combinations of active and passive ventilation for various periods of time, you can adjust and control how much heat the system will put into the rock mass. That's been, as you're undoubtedly aware, an issue that the Nuclear Waste Technical Review Board, for one, has been concerned about. It's also one that the U.S. Geological Survey expressed concern about in the Director's review of the viability assessment a couple of years ago. So looking at different ways that one can approach the management of heat in the system is something that we're doing now. The current reference design has moved the emplacement drifts somewhat further apart, about twice as far apart, about 81 meters apart, and would allow the rock mass, the drift walls to heat up above boiling for some distance away from the surface of the emplacement drifts, perhaps maybe ten to 15 meters. And looking at what it takes if we were to try to bring this whole system to a below boiling design, we're still evaluating what it would take to do that and what you would gain from such a change. Next slide, please. TSPA-SR, the TSPA to support the site recommendation, will incorporate the no backfill reference design in the base case analysis. That's a change that we're moving to, and I believe that either Bob or Abe will talk about some of this, looking at the sensitivity studies that are associated with that. We will -- the current plan is to maintain the reference design, an above boiling operational mode and design is the base case, and look at sensitivity analysis from the TSPA to try to understand what are the pros and cons of various approaches. Next slide, please. The site recommendation consideration report. This is not a pitch for pyramid power or anything, but it is an attempt to try to lay out, at least graphically, the hierarchical structure of the thing that sits at the top, which is eventually a national action, the Secretary's action, and then the supporting documents that make up part of the package that go to these higher levels, the Secretary and the President, to support a decision. The site recommendation consideration report currently consists of two volumes. Volume I is an update of what we know and how TSPA works. Volume II is a preliminary suitability evaluation. Eventually, the full site recommendation must have many other things in it that are called out by law, including comments from the state and affected counties and Native American Indian tribes. Sitting below these two volumes of the site recommendation consideration report are what we would call the technical basis report. I talked earlier about the PMRs, the nine process model reports, and then below those nine PMRs, there are, I believe, it's 122 of the analysis and modeling reports, the engineering analysis. All of that makes up hundreds of thousands of pages of reference material that is referenced or pointed to by these higher level documents. Our intent has been to try to put this together in an integrated manner. The way we chose to do this is we're going for a web-based approach on this, with hypertext links. So one could go to a document and eventually be able to go from a citation in the document, directly plunge down into the reference citation. It's going to take us a while, I think, to get to that. Next slide, please. The role of the site recommendation consideration report. It is not a document that is called for by legislation. But the Nuclear Waste Policy Act does call for the DOE to hold public hearings to receive residents' comments on DOE's consideration of a possible site recommendation action. And we wanted to put something reasonably recent, the last big document we had was the viability assessment, we wanted to put something reasonably recent on the street that could inform this dialogue and comments for the public hearings, and that's where the site -- the idea for the site recommendation consideration report came from. We are still on schedule for releasing it in late 2000. The intent is to summarize the technical basis and to facilitate the public review and comment process. The technical basis documents, all the bottom of the pyramid, almost all of that is going to also form the technical basis for the license application. So we have encouraged certainly the Nuclear Regulatory Commission, as they are looking at the basis for sufficiency comments on the body of work that DOE has done, to not confine themselves to just looking at the SRCR, but to also look at the other documents, the PMRs, the AMRs, and down to the individual lower level reports. Next slide, please. And that was a proposal that we proposed to the NRC back November of last year. NRC agreed in principal to this proposed approach. So far, we've provided seven of the nine PMRs to the NRC staff and 119 of the 121 supporting AMRs, and 11 of the SDDs have been provided. Some of the documents that we have currently in review within DOE, the TSPA Rev. 0 and the Yucca Mountain site description, an update of that. So we will get those, I hope, on the street here within the next four to six weeks and those will also support this technical basis that we're building. Next slide, please. Progress toward closure of NRC KTIs. Next slide. We've had three meetings, each on one of the KTI areas, since August, got a couple more planned. I know that we've got five more that we need to get planned. Those have been -- I'm very encouraged by the progress on those. We've been able to either close or close pending virtually every issue that has been put up, which means that between the information knowledge that's known now and the plans that are provided essentially as a promissory note for things that will be done over the next period of time before the license application, those have been acceptable to NRC staff as a reasonable path forward. Of course, there is no guarantee that the answers that you get and work that has not been done yet is going to turn out the way you think it will now. Next slide, please. The sub-issues. As I said, this has allowed us to take almost all of the sub-issues that were examined in the three meetings to date and either close them or close them pending. The agreements that have been reached help us define a path forward for issue closure and we're -- my understanding is that we proposed dates for the technical exchange on the remaining KTIs. The intent, of course, is to get those done. I think the last meeting that we have scheduled is February on TSPA, if I remember right, and the intent is to get everything in hand well before that date. Next slide, please. Integrated safety management system, and this is a very large effort within the Department of Energy, and it is an attempt to bring what I would call disciplined process and integrated management across the entire spectrum of things that DOE is involved in. Instead of just having nuclear quality assurance and nuclear safety and having a different culture and a different program for other aspects of safety, the intent is to bring one safety system umbrella over all that goes on within the department. For years, within the department, we have been sort of victimized by a series of almost independent programs that have competed for resources, rather than putting a single program in place that really helps you define across the whole spectrum of your programs, everything that's going on; helps you prioritize it, helps you identify where the safety risks are, and dedicated appropriate resources to the most important things. So this has been a very good exercise for us and we are in the last phases of that and I would hope by the end of this week we'll be able to put our name on the wall and say that we are moving forward in this. This is not a one-time endeavor, however. It's one of these things that is based on the concept of continuous improvement. So once you have essentially your basic system in place, the charge and challenge is to continue to improve that system year after year after year. Next slide, please. Phase one verification. We had no discrepancies. We had six opportunities for improvement. Those are listed here. One noteworthy practice, the worker involvement. The workers, especially at the site, have really embraced this program and have made it the success that it is. When told that they could stop work if they thought that there was an unsafe environment or a lack of adequate planning prior to implementing some activity, this really energized them and they have assumed absolute ownership of this program. It's great and you will see that tomorrow, I think. You will see the pride and the ownership of the safety program on the part of everybody who works on the program. The phase two verification, the way it breaks out, phase one is that you have the paper in place. Phase two is that you can demonstrate that you are implementing the program that's on that paper. So the phase two was completed back in July. We are currently in the middle of -- I'm sorry -- the phase one completed back in July. Phase two, we're currently in the middle of that activity right now and should get that report from an independent review team this week. Next slide, please. Well, the big thing looming on the horizon, funding uncertainty. The appropriation, as I said, has not been set. We are holding our fingers to see what comes out of conference committee today. There is a considerable discrepancy between the marks that the House took into the conference and the Senate. The Administration request was 437.5 million. The House marked 413, the Senate marked 351. So probably it's going to be -- the final mark will be somewhere between 351 and 413, I would hope much closer to 413, but any of those marks is considerably below the mark that the Administration supported. As we prioritize the work for the next physical year and coming physical years, that work supporting the site recommendation schedule has the highest priority and what we told the appropriators and the Congressmen was that at lower funding levels, the site recommendation schedule may be at risk. The schedule for licensing milestones is uncertain. Of course, our highest priority is for the site recommendation work, and what that does to the work that's needed to meet these commitments that we're making as part of the KTI resolution meetings remains to be seen. Some of those may be stretched out in time somewhat. Next slide, please. The SRCR and the technical basis documents below it continues to have almost all of our energy put into it. There's an enormous amount of work that has been done to put together this integrated effort. We learned an awful lot in the viability assessment effort and trying to put things together, put the defensible basis together, and make sure that there was consistency through all of the program. But translating those lessons learned into firm accomplishments has taken a lot of energy. We're putting a high priority, obviously, on the interactions to close the KTIs and to support the NRC sufficiency review. Looking forward to better understand the sufficiency review process, of course, we're waiting, as most people are, for 10 CFR 63, 963 and 197 to hit the street. The same caveat about funding limitations applies here. Actually, I think this is almost exactly the same thing we said on the last slide. Depending on the funding level, the SR schedule may be at risk and, of course, it can impact everything downstream from that, including the LA. Next slide, please. And I believe I've gone through my last slide here. So let me, if we have time, I'd be happy to answer any questions, Mr. Chairman. MR. GARRICK: Thank you. Would you be willing to comment on the things that would have to happen for the design to stop evolving? MR. DYER: Stop evolving. MR. GARRICK: Or to fix the design. MR. DYER: Even if we were to select, say, a point design concept, design is an evolutionary thing and I can't imagine it stopping, per se. MR. GARRICK: What I'm getting at is there is obviously, in the minds of the experts, some things that you're looking to achieve, some things that have to happen before you feel comfortable with the design. We keep talking about evolving design. As a matter of fact, the National Academy of Sciences recommended many years ago in their re-thinking report that you remain flexible, that we remain a little more flexible in the way in which we're going to manage the high level waste, up closer to the time that we have to really do something. But nevertheless, there comes a point beyond which you have to make decisions. You have to make that decision about the design that you're going with. And I guess what I'm asking, from a design standpoint, not necessarily from a regulatory standpoint, but from your own requirements standpoint, what are some of the key things that you think have to happen in order for you to be happy with the design? MR. DYER: If we look at the requirements that a design must satisfy, first, making clear what those requirements are, there are technical performance criteria that must be built in, but there are other criteria that may revolve around economics, they may revolve around I'll call it technical credibility. And some of those we're having a hard time putting our arms around those about how hard and fast those should be as requirements, how do you build those into a requirements document, which you normally thought of as pretty much performance specifications. But there are other considerations that may be just as important as those hard and fast performance specifications. But that's an area that we're exploring right now. A program that is technically immaculate, that costs hundreds of billions of dollars is probably not very useful for this country. What's the tradeoff between a program that adequately protects health and safety of the public and the workers and a program that the nation can afford and support? I suspect I didn't answer your question, because I'm not all together sure that I know or that anybody knows what the right answer is. MR. GARRICK: What I was getting at is where the focus of attention is. Sometimes I think in the preoccupation with the regulatory milestones and the various reports that you have established as goals, it obscures some of the understanding of what's going on in fact with the design in terms of what really is important to safety. Now, we'll get into this more with the performance assessment, of course, but, also, I think that people, after you've spent three-plus billion dollars, are beginning to think that the time ought to be getting close to when a design ought to be surfacing that you're quite comfortable with or if you're not comfortable with it, you have a pretty darn good idea of what it is that it's going to take to meet these fundamental requirements of safety and performance. I was just pushing that a little bit in terms of what were the main issues from the perspective of the project manager. Any questions? Milt, have you got any questions? MR. LEVENSON: No. MR. GARRICK: Ray? MR. HORNBERGER: Yes. Russ, if your budget does come in at, let's say, 413 million, will the SRCR be released in calendar year 2000? MR. DYER: That's our current intent, yes. MR. HORNBERGER: So when you mentioned it might be winter 2001, that was under a budget constraint scenario. MR. DYER: Yes. That would be a fairly strong budget, constrained budget. MR. HORNBERGER: So for the NRC, you're still looking for May of '01 to have the comments back. MR. DYER: I think that's right. I believe -- yes. Right now, we haven't changed anything in the way of our interactions. That's what the KTI meetings have been geared for and that is our preferred path. MR. HORNBERGER: Also, just one small clarification, if you can clarify for me on the technical exchanges. Of course, you held the one on the unsaturated zone in Berkeley, I believe, and you have one coming up in Albuquerque on the saturated zone. And your little footnote said that unsaturated flow is also covered in the saturated zone flow. Is this just a follow-up from the unsaturated KTI or is there some part of the unsaturated zone flow that gets lumped into the saturated zone? MR. DYER: I'm going to get some help here. I thought they were lumped together and we just treated them separately. MS. HANLON: There are two aspects of that, Dr. Hornberger. There are two aspects of that. First of all, the unsaturated zone in Berkeley did not include the matrix flow for the saturated zone. So that part will be moved forward to Albuquerque. The second thing is we will be discussing our evaluation of additional information on infiltration regarding the unsaturated zone in Albuquerque and how we're going to proceed forward to hopefully clothespin that item. So those are the two things that are put forward. MR. HORNBERGER: Thank you. MR. GARRICK: Any questions from the staff? [No response.] MR. GARRICK: Thank you very much, Russ. Very good. MR. DYER: My pleasure. MR. GARRICK: Our next presentation will be on the site recommendation consideration report. It says here Steve Brokoum, but you don't look like him. MR. SULLIVAN: No. You have Tim Sullivan instead. Good morning. MR. GARRICK: Good morning. MR. SULLIVAN: Actually, this will be a two-part presentation. When I'm done, Carol Hanlon will give you a brief overview, from DOE's perspective, of the sufficiency interactions that have been conducted to date. I'm the Team Leader for the Site Regulatory Products, including the site recommendation. This is an outline of what I'm going to discuss here. I'm going to focus here mostly on the contents of Volume II of the SRCR and I will explain why in a moment, and then some further schedule information. Some of this is redundant with what Russ has presented, so I'll move over those parts quickly. Current status. The report itself is in DOE review and it's on schedule for release in late 2000, calendar year 2000. An overview of the contents of the site recommendation consideration report. Volume I of the two volume report was built around the requirements of the Nuclear Waste Policy Act, Section 114(a)(1), which requires three things; a description of the proposed repository, a description of the waste form and packaging, and a discussion of the data relating to the safety of the site. So on the next slide, the Volume I is organized as follows. It includes, in addition to the information I just referred to, as a part of Section 4, the discussion of the data relating to the safety of the site. It includes TSPA-SR results. It also includes a section on pr-closure safety assessment. Both of those support Volume II. And then Volume II itself -- Volume I will be similar in form and content to the viability assessment; that is, it is descriptive material and analytical results. Volume II is somewhat different in that here, DOE will make a preliminary evaluation of suitability in accordance with the proposed 963 siting guidelines. So Volume II is, in fact, organized around the proposed regulation. Now, I'm going to spend a few minutes describing to you what the contents of that volume are. It's actually three parts. The first is an introduction, the second is the preliminary pre-closure suitability evaluation, includes a description of the assessment approach used to achieve safe operations before closure of the repository, and it will also include a discussion of the suitability criteria identified in the regulation. The focus here is to ensure that the repository systems limit releases and that we have adequate emergency planning systems in place or response systems in place. Then Section 3 is the preliminary post-closure suitability evaluation. Here, the proposed rule includes requirements for the TSPA methodology that I will describe. It calls for DOE to identify natural and engineered features important to isolating waste, and we will do so. There are a series of suitability criteria or characteristics that I will describe. And, finally, it has release limits to which we will compare the TSPA results for the 10,000 year compliance period. So, first, I'm going to describe in a little more detail Section 2 and then Section 3. So in Volume II, Section 2, the pre-closure methodology, the approach that we're taking is to apply established nuclear technologies, proven technologies, and using those technologies and established methods for design and operations; that is, accepted codes and standards. The assessment approach itself, fundamentally, is to reduce releases to workers and to the environment. The approach starts with a systematic identification of events based on standard hazard evaluations. It then follows with a screening or a determination of which events apply to the systems that will be built at Yucca Mountain and then, finally, these events are categorized into category one or category two, depending on their associated probabilities. Then the consequence analysis is complete, the preliminary consequence analyses are reported. These determine the dose for comparison with the limits specified in the regulation. Finally, the repository design establishes criteria for the prevention and mitigation of repository -- repository design establishes criteria for the use of features and controls important to safety. These are, of course, to prevent and mitigate consequences. Now, the suitability criteria in the regulation, page eight, in these sections, DOE must demonstrate the ability to contain radioactive material and limit releases of radioactive materials. Here, the burden is on DOE to demonstrate that the radiation doses are below the limits. Secondly, to implement control and emergency systems to limit exposure to radiation. Again, DOE will describe the preliminary program for the emergency systems and this program will rely on industry standards and proven technologies. Ability to maintain a system and components that perform their intended safety functions. In this section of Volume II, we will record analyses of structures, systems and components to ensure that they are performing as intended. And, finally, the system design will ensure the option for retrieval up to 300 years. In Volume II, the preliminary post-closure suitability evaluation. Part 963 has a series of requirements for the TSPA methodology that will be used to assess post-closure performance for the 10,000 year compliance period. These requirements are that data related to the post-closure suitability criteria or characteristics are incorporated in the TSPA, and I will describe those characteristics in a moment. The methodology also must account for uncertainties both in information and in modeling. It must demonstrate that alternative models have been considered and in Volume II, we will summarize the consideration of alternative models in the TSPA-SR, in the PMRs, and in Volume I. It must provide the technical bases for input parameters, for the FEPS analyses, the features, events and processes analyses, and for the models used or abstracted into the TSPA. Finally, the TSPA must conduct appropriate sensitivity studies. So Volume II will address each of these requirements for the TSPA methodology, on page 10. The rule also calls for DOE to identify the natural and engineered features that are important to isolating waste and we will do so. Six natural and engineered features will be described. These are summarized from the barrier importance analyses that are in the TSPA-SR that Bob and Abe will describe in a moment. First is surficial soils and topography, reduce the amount of water entering the unsaturated zone for surficial processes, infiltration is less than precipitation. Unsaturated rock layers overlying the repository and host rock unit reduce the amount of water reaching the repository. Here, seepage is less than percolation. Drip shield and inverts surrounding the waste packages. The drip shield that Russ described, the titanium drip shield protects the waste package from any seepage that may enter the drift and there is a ballast placed in the invert below the waste package in the bottom of the tunnel that serves to limit advective transport into the host rock. On page 11, the waste package prevents water from contacting the waste form for thousands of years based on the corrosion resistant material that's been selected and the corrosion testing that the department has completed to date. The spent nuclear fuel cladding delays or limits the water from contacting the actual fuel pellets. Finally, the waste form itself serves to limit the contact of water with the nuclear fuel itself, both in the commercial and the DOE high level waste glass form. Each of these will be described in Volume II of the SRCR. Okay. Another requirement of the rule is for DOE to evaluate a series of suitability criteria or characteristics. There are a total of nine here. These closely parallel the process model reports that the department has prepared to support the TSPA-SR and Volume II of the SRCR. For each of these characteristics, Volume II will describe the technical basis -- that is, the data -- that's used in this evaluation. It will describe the models and it will include a regulatory evaluation of how this information is represented and incorporated in the TSPA-SR and will do so for each of these post-closure suitability criteria. On page 13, similarly, disruptive events will be evaluated as suitability criteria, also. Then, finally, in Volume II, the results of the total system performance will be evaluated. The methodology described and the criteria will be evaluated -- will be used for a preliminary evaluation of the suitability of the site. This is done by comparing the release standards to the TSPA-SR does results both for the pre-closure and the post-closure requirements. You've seen page 15 before. The only point here is that I'm going to briefly describe the technical basis documents, some of which Russ has already described. I want to emphasize here that we have assembled the site recommendation consideration report such that it is fully traceable to the underlying technical basis documents. By that, I mean, specifically, that no new information, per se, is presented in the site recommendation consideration report. Everything that is presented and summarized in that report is fully traceable to the technical basis documentation that you see in the pink and the purple areas of this pyramid. Page 16, Russ mentioned the analysis and modeling reports used to document the analyses and models of individual FEPS using site characterization data sets. They cover both the natural and engineered features of the site. The PMRs then synthesize and integrate groups of AMRs to describe and model general categories of features important to post-closure repository performance. They serve as an intermediate level, the equivalent of the TSPA-VA technical basis document, in which the component models for the TSPA were described. The SDDs are the engineering analyses to document surface, sub-surface and waste package designs, and then the TSPA-SR uses abstracted results from AMRs to analyze the performance of the repository with a focus on a 10,000 year compliance period, but it presents results for longer time periods, as well. Page 17, I think we've been over this. On page 18, I provide the current status of the PMRs, the process model reports. In fact, they've all been now accepted as Rev. 0 by DOE. However, further updates and modifications to the process model reports are underway in the final column. There are two main drivers. The first is that we are -- the M&O is now completing the revisions to AMRs to incorporate the no backfill design that Russ described and those final updates are underway now and they will result in updates to the PMRs between now and December. And number two, based on internal comments, the FEPS AMRs, the features, events and processes AMRs, and there is one for each of the PMRs except the integrated site model, based on internal comments, those are also being revised and updated and those results will also be incorporated in the PMRs that will be available at the time of the release of the SRCR. You will also note here, on the right, that full revisions of two of the PMRs are underway, the UZ PMR and the engineered barrier system PMR. We will have, though, at the time of the release of the SRCR, all of these PMRs in their final form available. Page 19, additional technical basis documents. The preliminary pre-closure safety evaluation is a report that's been prepared to support our preliminary evaluation in Volume II of the SRCR. It's an analysis of the radiological safety for a repository at Yucca Mountain. The site description is a comprehensive compendium of site information, including chapters on natural resources and natural analog studies. The repository safety strategy, which will soon be in Revision 4, is a general plan for the identification and prioritization of the factors important to repository system performance and it formulates a safety case where DOE will present the essential aspects of the performance of a repository system. On page 20, I just want to make a couple points here. The first is that the process that we have used to assemble the TSPA-SR and the SRCR includes a foundation built on the left side of AMRs and PMRs, some of which, as I mentioned, are being updated, in the red box, to reflect the no backfill design and those support the TSPA-SR and the SRCR. We'll go through a similar process and we'll do another update to the TSPA in the spring and we'll go through the same process again of updating and revising AMRs and then PMRs to develop a subsequent iteration of the performance assessment for the license application. There are a couple of new boxes in here. In light green, you'll note the FY-00 technical update. This is a report that the department will also release at the time of the SRCR. Its intent here is to provide an update on testing and design work that has occurred subsequent to the freezing of the inputs for the TSPA-SR to provide all interested parties the most current information that we can. And we are currently contemplating another such document for the time of the SR currently scheduled for July of '01 to, again, provide as current information as possible to support these documents. On the next page, page 21, DOE has developed the site recommendation consideration report to inform the public of the technical basis for the consideration of the repository at Yucca Mountain and to facilitate the public comments during the SR consideration hearings. It's to promote the dialogue. The SR, the site recommendation, that will follow will provide a comprehensive statement of the basis of any recommendation that the Secretary will make to the President. It will include additional information as required by the Nuclear Waste Policy Act. It will include the views and comments of the governor and legislature of any state or the governing body of any affected Indian tribe, together with the response of the Secretary to such views. It will include, on page 22, the preliminary sufficiency comments from the NRC. Carol will discuss that a little further in a moment. It will include a final environmental impact statement, any impact reports submitted to the department by the State of Nevada, and any other information that the Secretary considers appropriate. On page 23, the act does require that the Secretary hold public hearings in the vicinity of Yucca Mountain for the purpose of informing the residents of the area of such consideration and receiving their comments regarding the possible recommendation. Current planning calls for these hearings in early 2001. Location and number of the hearings has not yet been determined. Finally, here on the schedule chart, I won't go through each milestone, but I will identify several key ones. The first star there in late 2000 represents the release of the SRCR. The vertical red dashed lines and the horizontal arrow identify the comment period. Our current planning is for 90 days. And subsequent to that, the gray star would represent receipt of the NRC sufficiency comments, Secretarial decision on whether to proceed with the site recommendation in June, and then DOE will submit the SR to the President in July of '01, waiting 30 days after notification to the state of the Secretarial decision, a minimum of 30 days. On the bottom are some EIS milestones which culminate in the submittal of the FEIS to the President at the same time as the SR and its subparts. So that's all I had to present this morning. I could entertain questions now or we could have Carol do her piece and have questions later. I'll leave that to you. MR. GARRICK: Let me ask the committee. Are there any questions? MR. WYMER: I have one sort of a general off-the-wall question. This is going to be a large and comprehensive document with a lot of stuff. MR. SULLIVAN: About 1,300 pages. MR. WYMER: Who will review this thing? It seems that most of the people in the country who are competent to review it have been involved in the preparation of it. Do you have any idea how this thing will be reviewed and where will they find the people to do this? MR. SULLIVAN: Well, I can't speak to that specifically. It will be accompanied by an overview, which is a much slimmer document, similar to the VA overview. We have targeted the SRCR to a general audience. We have tried to make develop the document so that it will be understandable to people who are not expert in individual disciplines of engineering or science. MR. WYMER: But you have no feeling on that, and I don't really know why you should have, but you have no feeling of who actually will do the review, where they will get the people from. MR. SULLIVAN: DOE has received 11,000 comments on the DEIS. So we expect people will review and comment on portions or the entire document. MR. WYMER: So it's just people. MR. SULLIVAN: Stakeholder organizations and individuals, interest groups of various kinds. MR. WYMER: But the President, in quotes, is the person to actually say this looks like it's the real stuff and it will work, so let's go ahead. MR. SULLIVAN: Right. First the Secretary, then the President. MR. WYMER: So somebody has to advise him. MR. SULLIVAN: Right. MR. WYMER: And it will not be the public. MR. SULLIVAN: It will be the Secretary. MR. WYMER: Okay. MR. SULLIVAN: Taking into account all of the comments that have been received on the document and the views of the affected governors and legislatures and the preliminary sufficiency comments of the NRC. MR. WYMER: So it's kind of a closed loop here. Okay. MR. HORNBERGER: Tim, in your pre-closure comments, you mentioned that retrievability would be assured for up to 300 years. MR. SULLIVAN: Correct. MR. HORNBERGER: So I take it you're still holding to, what, a 50 to 300 year possibility for pre-closure. MR. SULLIVAN: Yes. We identify the period up to 300 years as the period in which the repository potentially could be monitored and we would have decommissioned the surface facilities and entered into a monitoring phase. But we would have retained the -- not the possibility -- we have retained the capability to retrieve the waste up to 300 years, meaning we would maintain the ground support within the repository and if and when needed, we would commission surface facilities to handle the waste retrieval. MR. HORNBERGER: And then your analysis also then includes the possibility of a tunnel collapse and things like this and still have the capability to retrieve. MR. SULLIVAN: Yes. You will see degradation analyses in the SRCR and in the supporting documents. MR. GARRICK: Any other comments? Staff? MR. LARKINS: A quick question. Yesterday, we heard from a representative of the State of Nevada who discussed how his views on the importance of the performance confirmation plan and DOE strategy, and we didn't hear any mention of it this morning. Is this a part of your -- how doe the PCP fit into this? MR. SULLIVAN: We will describe, and I omitted that, the performance confirmation plan in the SRCR and there is a plan, a stand-alone report, that will be available also. So that was an oversight on my part. MR. GARRICK: Thank you. Thank you very much. Carol? MS. HANLON: Perhaps if I just begin my presentation, and Andy will go ahead and he will adjust it and you can hold up your hand if you don't hear. You will notice that we spent a good portion of this meeting discussing different portions of the key technical issues, technical exchanges. I think that that's an aspect that warrants a good deal of emphasis. I appreciate the presentations that we had yesterday from the Nuclear Regulatory Commission, Bill Reamer, John Trapp and Neil Coleman I think did an excellent job of setting the stage and identifying some of the things that have gone on to date and discussing for us the path forward that we've looked at. I would like to discuss a few of those things from the Department of Energy's perspective to clear a couple areas I think that remain perhaps not fully understood, to help you all with your understanding of our document and perhaps understand Dr. Wymer's question a bit more on review and how the reviews are going. One of the things I'm going to talk about is I'm going to talk about process for these meetings, what we're looking for, what we're going to do with upcoming meetings, and also how we're handling the agreements on items that have come out of the meetings, the technical exchanges that we've previously had and will continue to come out, because there will be agreement items and we're watching those very closely. So as Tim and Dr. Dyer have both said, we have a requirement from the Nuclear Waste Policy Act that we provide sufficiency comments. We include sufficiency comments provided to us by the Nuclear Regulatory Commission as part of our site recommendation and those are in two important areas, at depth site characterization, waste form proposal, and I think the words that they are sufficient, they appear to be sufficient for inclusion in any application to be submitted. That's an important concept. That's a forward-looking concept and it helps us focus on where we're at with regard to the site recommendation consideration report, the SR and forward to the LA. So obviously to the Nuclear Regulatory Commission staff, a very important portion of their sufficiency review has been the resolution of their key technical issues, non-key technical issues, and those have been documented in the issue resolution status reports. Our purpose, our goal is that the technical exchanges will result in a clear understanding of the status of each of these and where there is not resolution, a path forward for reaching resolution. The first general kind of overarching technical exchange was held April 25 and 26 in Nevada. That had a dual purpose; first, the purpose of discussing how sufficiency would be approached and, secondly, to discuss in full the status, both perceived by the Commission and as perceived by the department, with the new information that we had as we were preparing analysis and modeling reports and we were preparing the process model reports and going forward to the total system performance assessment, what our reflection of the status was. Based on that, we subsequently set up a series of technical exchanges on specific topics. Now, this is where a bit of the unclarity, I think, still remains for your committee and I would like to just take the opportunity to clarify that a bit. You will recall the November 24 letter from Dr. Brokoum to John Greeves and in that letter, we proposed an approach for providing additional information which would support the Commission's ability to review our technical basis documents and make sufficiency comments. Based on that, we had suggested a series of meetings, specifically nine meetings, focused on process model reports. You've heard a good deal this morning on process model reports, analysis and modeling reports and other documents, both from Dr. Dyer and from Tim, and it perhaps is a bit daunting. Our approach in setting up these meetings was to discuss the particular category of a process model report, the analysis and modeling reports that contributed to that, the impact to the total system performance assessment, and, of course, very importantly, the contribution to the specific key technical issue that was addressed by that process model report. To make our series of meetings and our technical exchanges more effective, we adjusted that approach and I think that may be something that was a bit confusing and left you all a bit in a lurch on that. We are now focused specifically on the key technical issues and that's so that we can very specifically take that same information, the process model reports, the AMR reports, the portions of the TSPA, and address it very specifically to an individual key technical issue. So on this, you will see the set that you had previously seen, I think that we did this schedule in late July and you've seen it since then. I briefed you, I believe, in July on this. And that includes our completed unsaturated zone technical exchange in Berkeley. The igneous activity, also completed, was held in Las Vegas, and the container life and source term that were completed last week in Las Vegas. Two that you're still familiar with are the structural deformation and seismicity technical exchange that will be held the week of October 11, actually the 11th through the 13th, and saturated zone flow, as we mentioned this morning, will be held in Albuquerque. We've added an additional day to that. It's October 31st, November 1st and 2nd. I also might just make another comment, to make sure that we have ample and adequate information for full discussion, full vetting of the issues and full presentations, basically we have three days for all of these technical exchanges. In order to facilitate the Commission's ability to do their sufficiency review and have their comments ready, as we've asked them, in by the end of May, the Commission asked us to go forward in setting up the series of meetings on remaining key technical issues and to try and have those done by the end of January. We moved out in February a bit, but I think we're pretty close. These have not been fully agreed to with the Commission staff yet, but I have put months down here so that you can see what remains before us and get an idea not only of our busy schedule, but consequently, I think, your busy schedule. The in-package criticality, which is subissue five of the container life and source term, will be held October. It's actually, I think, October 24 to 26, as it is currently scheduled. In November, we will have the technical exchange on thermal effects on flow. December we'll have the technical exchange on radionuclide transport. Also, in December, we have identified the possibility of a briefing on the total system performance assessment results, since we had previously had the meeting in June in San Antonio. Now that the TSPA will be coming out, there is a potential that there will be a briefing on that, not solidified yet. In January, we will have two meetings; earlier January, evolution of near field environment; later January, the repository design, thermal mechanical effects. And the last meeting, the 6th through 8th of February, is the total system performance assessment integration. So that's our new set and when we have the formally agreed to list of meetings, we'll make sure that you get a copy of that. You've heard quite a lot about the status of these meetings and I just want to recap for you a bit. The three meetings we've had are unsaturated zone flow. The number of subissues with regard to that were five. The sixth one will be handled, as we've said, at the saturated zone in Albuquerque. That's on matrix flow in the saturated zone. From that KTI technical exchange, we have four issues which were closed -- subissues, excuse me, which were closed or closed pending and one remains open. That is on the infiltration and that came up in some of the discussions yesterday. There was some information, new information from the center that was presented during the technical exchange in Berkeley that our staff had not had the chance to fully evaluate. So what we have come up with is, regarding that subissue, a proposal to evaluate that new information and identify what we do need to include and what would explain or justify what we may not think is relevant. So that's the approach that we are going to present to the saturated zone. If you have any specific questions, we have Martha Pendleton here, who can talk to the specifics of the infiltration. Another issue that came out of that unsaturated zone flow meeting was the discussion of the importance of fully understanding the subsystems as you are also understanding the performance of the entire system. I think that's had a little bit of confusion here. The point is not that because of a very strong component in this case, a long-lived waste package, you're unable to understand the subsystems. The point is that in order to understand the full system, we must understand all the components, we must understand all the portions of the subsystem, the natural system, the engineered system. So our particularly strong portions should not obscure the issue nor should it be a penalizing component, but we must understand fully all of those. That's what we are committed to doing. Bob Andrews, in his discussion of total system performance assessment, is going to go into a bit more of exactly how we take those subsystems, full understand them, as we move forward to an understanding of the total system. With regard to igneous activity, we had two issues there. One is closed, one is open. I included our score card, and so did John. This is actually we're a little better than that, we're about 90 percent closed on the subissue for consequences and the reason we're not fully closed on that is that we have some AMRs that will be completed late this year, this calendar year, and provided to the Nuclear Regulatory Commission, addressing the no backfill issue, and it was believed that to fully be able to understand the consequences and close that issue, they needed those AMRs. We fully agree with them and we're making every effort to move forward to get those AMRs and they will be in their hands in January. Container life and source term, we had six issues, we've closed five. The issue, subissue we did not close was not addressed in this particular technical exchange. It's the issue of criticality and it will be addressed in a few weeks at the second CLST meeting in October. It's not in my handout, but one of the things I did want to mention to you is that as you can see, and I've tried to create a complete record here, you can move through a few of these, John, so that you can see what the title is, what the status is, and what the agreements are. Now, it's very important, as we make these agreements, to track them very clearly and make sure that we get the exact information back to the Commission on schedule, as we've promised. An example of that comes again from the unsaturated zone meeting in Berkeley, where the Commission staff asked to see a copy of the test plans for Alcove 8. We committed to providing those within a week. We did provide them within a week. Neil Coleman made sure that the review occurred and he's got our comments back. So those agreements have been handled. We are now in the process of evaluating those and seeing if there are modifications that we can make to our test plan to accommodate the comments we received from the NRC. To make sure these things are happening, we are tracking them very closely, both formally and less formally. I won't say informally. We have a condition identification reporting and resolution system, which is a formal system that's in place, and that's a system, it's important for you to know that that is in place and it will transition as we go through the period of having a new contractor. So these agreements on the items regarding the closed pending will not be lost. They are in a formal tracking system and they will transition. Also, on an operating basis, to make sure that we are staying very much on top of this, we've instituted weekly or biweekly briefings on the status of items that are coming out of these and where they remain and where we are with regard to providing information to the Commission. So we are watching those very closely. Another one that's much on our mind is the infiltration issue that's coming up for the saturated zone. I think we can probably move forward, leaving you with these things to review at your leisure. I think we can move forward to 14. From our standpoint, the department has spent a great deal of effort on these. We understand and appreciate the effort, also, that the Commission staff has dedicated to these technical exchanges. We feel that they have been extremely productive. The three meetings only have resulted in four subissues changed from open to closed pendings. Four subissues I've discussed remain open, and two of those subissues were not addressed in the interactions. They were the saturated zone and the criticality. So we are moving forward. We believe that these technical exchanges have been extremely effective in establishing either the status or the path forward to a closure status. I feel that our teams are working closely and very well together. There's a good exchange of information and that's very productive, I think, to working toward providing the Commission with the information that we believe they will need. And the management commitment, both by the department and the Commission staff, has certainly been a definite asset. So we're continuing to hold and move forward with these technical exchanges, supporting the staff's approach. We were a bit disappointed that we weren't able to year the Yucca Mountain review plan approach and that the formal sufficiency approach. We're concerned that we really are definitely on target and that we're not missing something that the Commission will need. So the sooner that we can know those specific details and know that we are on target, we can move forward or we can adjust our path as we need to. So we'd like to confirm the path that we're currently engaged in and moving forward to both creating information that will be sufficient for the license application docketing and, also, the sufficiency comments that the NRC will provide. May I answer any questions for you? MR. GARRICK: Any questions? John, you have a question? MR. LARKINS: I was just curious. What happened to the biosphere PMR technical exchange? MS. HANLON: That was included in the igneous exchange, John. That was very much a part of the second day of the igneous activities exchange. MR. LARKINS: Has that been combined now? MS. HANLON: It's been completed. It was the second one that was completed. I believe John Trapp, Dr. Trapp spoke about it yesterday. It was believed that in order to fully understand those components of the igneous activity, the dose consequences and so forth, that the biosphere aspects had to be brought in. So I think we had either a half-day or a full day. MR. LARKINS: Because there was a previously scheduled separate technical exchange and we had committed to have somebody participate, but I didn't realize it had been combined with the igneous activity. MS. HANLON: Dr. Hines was there, so you did have support there. And I hope that in this I've been able to partially explain Dr. Wymer's question about who and how is this review going to be done. There is a great volume of information and Dr. Dyer spoke about it a bit in our hyperlink text system on the internet. So if you have a specific question, hopefully we'll be able to click on that and go down through that. But these meetings are -- we are attempting to set them up so that we are explaining the technical basis that's supporting various aspects of our TSPA, of the site recommendation consideration report, and specifically focusing them on the KTIs. So hopefully that's assisting the review somewhat. MR. WYMER: Well, I had naively assumed that there would be some sort of independent initial review, but that isn't going to happen. MR. GARRICK: I guess just to extend that thought a little bit, the committee was quite impressed with the reports that were prepared by the peer review group that was put together by DOE on the TSPA and I just wondered if that particular model was going to be applied to any of these other key reports. MS. HANLON: I don't think we've implemented at this time, but we certainly can keep that in mind. MR. GARRICK: George, you had a question. MR. HORNBERGER: Carol, these technical exchanges, and now that you have some experience with them, obviously take an awful lot of effort, involve an awful lot of work. I gather, however, even given your experience, you're confident that the very ambitious schedule you have you're going to hold to. That is, you haven't fallen behind in any of this as a result of having the ones that you've had. MS. HANLON: We're doing pretty well so far and you are right, Dr. Hornberger, it takes a tremendous amount of effort, both on the part of the Commission staff and the center and the part of DOE and our supporters. So they are extraordinarily intensive. However, we've done pretty well in staying on target. We've looked at things carefully. The Commission has been extremely good if something needs to be moved and we've also tried to accommodate needs that they may have had to move a meeting. But we're shooting for this target and I'm very optimistic about it. We did leave ourselves Christmas and New Year and we managed to leave ourselves Thanksgiving. If there are no other questions. MR. GARRICK: Any other questions? I want to thank you for presenting an excellent scorecard for a very complex process. That's very helpful. MS. HANLON: You're certainly welcome. Thank you. MR. GARRICK: Okay. I think that in spite of the fact that a break is not noted in the program and if this goes into the Federal Register, I'm going to accept the risk of violating that piece of information, and declare a 15-minute break. [Recess.] MR. GARRICK: Let's come to order. We're not going to turn our attention to the often referred to total system performance assessment site recommendation. I think, Abe, you're going to lead this off, is that correct? MR. VAN LUIK: Last week, I had a reminder that as we become more effective in life, we take on more risk. My oldest grandchild got his driver's license last week and my youngest grandchild decided she could have a better life without diapers, and I'm happy to report that in the ensuing five days for both of them, neither one has had an accident. I'm Abe Van Luik. I'm the Senior Policy Advisor for Performance Assessment and I'm going to give you a short introduction to the TSPA-SR, and Bob will give you the details under the heavy lifting. I wanted to talk a little bit about the regulatory requirements that we're addressing right at this moment, the objectives of the TSPA, summary of the major improvements since the viability assessment, mention a little bit about the barrier design and the basis for process models. If you look at the regulatory requirements, you know that we have proposed regulations on the street right now, and to make this talk very short, what we are doing is addressing all of the nuances of the proposed regulations from EPA, NRC and DOE. When these are finalized, they will become simpler because the NRC will incorporate the final provisions of the EPA and we will not have basically dual nuances on definitions of our MEIs and that kind of thing. If we look at the individual protection or dose, which is the primary performance measure that we are concerned with, we have to include probable behavior, as well as potentially disruptive events. If we look at the objectives, they have changed over the years. TSPA-91, for example, was just to show that we could do one. In '93, we began to get serious about using TSPA to align the project, to look at what's important and what's not. TSPA-95 and the viability assessment got stronger in that department and then TSPA-SR is to support the national decision-making process. TSPA integrates underlying models of individual process components. We're looking at several performance measures, individual dose, ground water protection, the human intrusion standard, and peak dose for the final environmental impact statement. We are looking at the significance of the uncertainty in the process models and Bob will talk a little bit about some of that evaluation process. Major improvements. We've had both technical and process improvements. The process improvements, I think, are easily underestimated in terms of the effort that they have taken. But everything is under quality assurance procedures at this point. We are using the analysis and model reports that Russ Dyer talked about as the thing from which we trace our data and the information flow. We have explicit evaluation, a comprehensive evaluation of features, events and processes. We are using traceable data sets and the TSPA model itself can be used to move down into the data sets themselves, and we are tracking the quality status of all data, models and software. Technical improvements. Some of you mentioned a while ago, I think it was John mentioned that we did have a review on the viability assessment. That review, of course, came after the completion of the viability assessment. So the TSPA-SR and the TSPA-LA will be where we respond to that review in terms of improving our modeling. Models with major enhancements, looking at those comments, and also comments from others, such as the NRC, in the exchanges, as you've heard Carol talk about, of course, we have learned a lot from the NRC about what their expectations are and some of these improvements also address those. But climate and seepage has been greatly improved. A couple thermal processes are lot farther along than they were in the VA. Waste package degradation, we're looking at stress corrosion cracking and initial defects in the welds. Saturated zone transfer and volcanism, all of these models have seen major enhancements since VA. The engineered barrier. TSPA-SR is based on the site recommendation design, no longer the VA design. We're looking at an average thermal load of 62 metric tons of heavy metal per acre, which is lower than the viability assessment. We're looking at at least 50 years of ventilation, it may be more. This is some of the operational mode adjustments that Russ was talking about. We're looking at blending of fuel at the surface to levelize the thermal load. The engineered barrier design considers the titanium drip shield, non backfill, waste packages placed end to end, an average line load of 1.4 kilowatts per meter. The waste package itself, still 21 pressurized water reactor assemblies or 44 BWR assemblies, and co-disposal of Defense spent nuclear fuel and Defense high level waste. The outer layers, alloy-22, 20 millimeters of it; the inner layer, stainless steel, 100 millimeters. The inner layer of stainless steel is not taken credit for in the TSPA. It is a member that gives structural support to the waste package. There is a dual alloy-22 lid closure weld. The outer lid closure weld, the stress is mitigated by solution annealing. The inner lid closure weld, the stress is mitigated by laser peening. These turn out to be very important to long-term performance. This is just a listing of the process model categories and the process model report. On the right are ones that were mentioned by Russ in his talk and, of course, some of these reports, like the near-field environment and the EBS degradation flow and transport reports come into one basket when it comes to the actual modeling, which is the engineered barrier system environment. I don't think I need to spend any time on this. It's just I may make the boast that this is the best integrated performance assessment we have ever done. In the past, you could find principal investigators that said, well, I handed the data over to PA, but I don't know what they did with it. This is no longer the case. The principal investigators are intimately involved in taking their data, abstracting it and putting it into the total system performance assessment. So we've come a long way since the '93-'95 days. An issue that we have to be aware of is that we have to have some statement of how confident we are in whether or not these results that we come up with are useful in the decision-making process. Demonstrating confidence requires a lot of things, but it requires showing a sufficient understanding of processes, determining system behavior. Carol mentioned this in her talk just a few minutes ago, that the NRC staff is very concerned that we show that we understand the processes that we're putting into our modeling. Systematic applications of the features, events and processes, screening. It's a way to show completeness of the arguments, that there's not something big out there that you've just completely forgotten about. This is another way of showing that you have a reason to have confidence. Systematic evaluation of component process models and their importance. You can have a haphazard evaluation and do some neat calculations that say, oh, look how important this is, but a systematic evaluation is what's important to building confidence. And you have to show that you have properly incorporated the important uncertainties. And, of course, the TSPA, as your documentation, has the challenge to make these points clearly and traceably, and in my review of that document, I find that it's a very good read. It's a good document and it's well on its way to illustrating these points quite nicely. Happily, the NRC staff and we see things the same way when it comes to a risk-informed performance-based approach, and I think some of the success that Carol was talking about in having issues closed pending, the delivery of the final products so they can verify that we actually did what we said we're doing. Some of that is based on them and us agreeing that some things are more important than others. Risk-informed means that the entire uncertainty distribution, not just mean value lines, are being used to inform. Performance-based means that the outcome is -- you know, whether a feature or a system or a process is important to swinging the outcome one way or the other, is an important criteria for judging performance. And, of course, something that -- as a corollary to that is something that we mentioned on the previous page. You also have to show that you understand what you're talking about, because if you don't understand it and model it wrong, then your conclusions here don't mean much. So all of these things are together to give us confidence. We use these types of considerations to prioritize science, engineering, design and modeling, and if the budget comes in lower than we would like it to, we will invoke these types of results to say, well, this is more important than that and adjust the funding for science and engineering accordingly. And we also use these considerations to rank key technical issues and decide on the level of effort to be devoted to address them and I think on this, we are in synch with the NRC staff. They agree that this is the right thing to do. The thing that we have to do is to make decisions, even in the modeling of this system, in the face of uncertainty. The basic rules that we have applied so far is that where something is extremely complex or the quantity of data is just insufficient to develop a meaningful distribution, that we take a conservative or a bounding approach. We have several coordinated activities underway now to evaluate how these decisions, which were made as part of the process of creating the TSPA-SR, affects the actual performance measure of dose. We are looking at unquantified uncertainties and these are the conservative values I talked about in the first bullet or, in some cases, maybe even optimistic values. This is in the eye of the beholder or the reviewer, in some cases. We are identifying those and then we will do a trial study, coming up with a distribution for that parameter and running sensitivity studies to see how important it was to have assumed that, or should we go to a more detailed approach. On the next page, we are also, at the same time, looking at the quantified uncertainties, the ones that are documented with data distributions, CDFs, et cetera, which are sampled into TSPA. And we're looking at the uncertainties that were considered at each modeling level and to get a better feeling for just how uncertainty has been rolled up, how well or how poorly it has been rolled up, all the way from the data interpretation process level modeling into the TSPA. And depending on the outcome of these activities, a given activity may be expanded or a new activity defined. This is work in progress, in other words. The goal is to increase understanding of the basis for the TSPA results and improve the basis for judging if there is an appropriate level of confidence for the current stage of the societal decision-making process. The current status of TSPA-SR analyses, and some of you may not believe this, but some of the results presented today are preliminary and subject to change. We are still in checking and this got a chuckle at the technical review board meeting from some people saying, yeah, sure, you're just saying that. But you will see in what Bob presents that some of the curves have changed since the TRB because of the checking process. This is serious. You are seeing draft material that is still being worked. And so they are certainly not suitable at this time for making regulatory compliance judgments. They are intended to be used right here, right now, for general discussions of sensitivities. The calculations that you're seeing are going into Revision 00 of the technical report. The repository safety strategy Revision 04, which is also in draft right now, and the SRCR. We expect to make minor updates, not major revampings, of all of these calculations for TSPA-SR Revision 01, which is coming in next spring, which will support the site recommendation and the final environmental impact statement. So you're seeing a work that's close to being done for one stage and then there will be another smaller stage for the SR, and then we have not looked in this presentation to the LA. And, of course, you'll want to save your questions for Bob Andrews, who is now going to show you the technical side of things. MR. GARRICK: Committee? Milt, go ahead. MR. LEVENSON: I've got a couple of questions. One has to do with I suppose you could define it as a decision-making process. As you go through here, items are either included or you decided to leave them out, like not taking any credit for the stainless steel canister, et cetera. Whether it's to include something or to leave it out, at what level, who makes those decisions? How much review is done of whether you should include something in the TSPA or not include it? This has nothing to do with the technical part of how you treat it, but who decides, in essence, the scope? MR. VAN LUIK: That's a good question. I think there is a process that's not as well defined as you might think it is, but it's in the process model TSPA abstraction interactions where it's actually decided that there's enough information to go forward with this, to bound this, or to say, for the sake of conservatism, we will not take credit for this, although it has a definite purpose, which is to maintain the integrity of the waste package. So these types of decisions are documented in the AMRs that describe the abstraction process, for example, and in some cases, they are documented in the analysis and model reports for the particular example you're talking about, for the waste package lifetime, where it was discussed that, yes, you can get credit for hundreds, maybe even thousands of years, but because of the way that the system is functioning now, this does not really add much except in the terms of the doses of two, 300,000 years out, when it would not perturb things at all. So it's all documented, but the decisions were made, in some cases, at a lower level, at some cases at the abstraction level, and in some cases, DOE might walk in and say do this differently. So it's a decision process we're all made aware of, but it happens at different levels, but hopefully it's our intent that all of these decisions are documented in the AMRs and then rolled up into the PMRs. MR. LEVENSON: They may be documented, but are they carried forward in evaluations such as uncertainties, because it seems to me they could have a significant impact, particularly things that are left out. MR. VAN LUIK: The task that I'm very well aware of, because I'm part of the DOE oversight of it, on the uncertainty evaluations that I was describing, these are some of the issues that we're looking at; should we have done what we did or should we bite the bullet and go into more detail in the modeling of this particular issue. So it's definitely one that we will address in that process, but that's still work in progress. MR. LEVENSON: In the consideration of uncertainties, are they all treated equal? I mean, some uncertainties are symmetrical or something is plus or minus. In other cases, uncertainty is all plus or all minus. In looking at the overall uncertainties, are things being carried with a sign as well as a quantity? MR. VAN LUIK: This is actually the topic being addressed by our uncertainty task, because what we did at the beginning is we put out some general guidelines on how to treat uncertainties. What we're doing now is verifying whether those were followed or not and we're finding, in some cases, that what you're describing is basically a judgment on what the degree and sign of the uncertainty is, that it was not done. So we're going back to fix those types of things. But in every case, the analysts thought that they were making a conservative assumption, except in one or two cases. Some other analysts disagreed with them. So we are getting at the bottom of those types of things. But I think this should not be confused with the idea that we did not capture what we know are the major causes of uncertainty and I think those are very well wrapped up in this TSPA. So we're looking at something that's a second order correction, basically. MR. LEVENSON: One other question, and I'm not sure you're the appropriate one to ask, but you're standing there. One of the very, very useful outputs of the work you're doing here, but I'm not sure it's being done, and that is the design has been evolving over the last, say, couple of years, a fair amount of time and money has been spent on the design evolution. The TSPA could tell you, probably better than almost any other method, how effective those design improvements have been or are they improvements, do they really reduce dose to the public at the end. Do you have any feel for how much change in the dose to the public has occurred because of so-called design improvements? MR. VAN LUIK: We have done considerable work in sensitivity analyses, a few of which Bob will show in a few minutes, after I sit down. But the basic point is my answer would be yes, we have a very intimate relationship with the designers and we evaluate what they do in those terms. However, there's other things besides dose. There is defense-in-depth considerations. There's considerations of confidence. So because there is uncertainty in the modeling, we also do things like add a drip shield that the TSPA shows that for 10,000 years, since the waste package is intact with or without a drip shield, the dose comes out about the same. It's not exactly the same. But there is another consideration. Do you have defense-in-depth? Do you have reliance on only one barrier? That drip shield gives us two barriers. When you take away one or the other, as you will see when the repository safety strategy four is issued and also you will see in TSPA-SR documentation, you will get the feel that there is actually some backup in this system and that you can have confidence that even though have uncertainty here, that 10,000 year dose number is a pretty good number. MR. GARRICK: Ray? MR. WYMER: No. MR. GARRICK: George? MR. HORNBERGER: Again, just a quick follow-up on Milt's question. Have you done a catalog? We're talking about these bounding values or conservative values or on-off switches. Do you have a catalog? Can you give us an idea of how many of them you have to deal with? Is it ten, is it 100, is it 1,000? MR. VAN LUIK: I think it's over 100. We have an initial catalog, but this is work in progress. And so it may expand and it may decrease when we see that we have double-counted some items. MR. GARRICK: Abe, this is a process question and it is an extension of Milt's question. It's one I will lay out that we may come back to with the other presenters. But I want to get it out on the table. We're seeing a lot of language now about quantifiable and non-quantifiable uncertainties. It's language that's also popped up on the nuclear power plant risk assessment world. And I think that it's a situation where there is a great deal of opportunity, it seems to me, for a lot of mischief and I'd like to be enlightened a little bit more on the whole issue of non-quantifiable uncertainties. You say that the way you're handling these, for the most part, is to take point estimates and take and use conservative values or bounding values or whatever, and, of course, part of Milt's question and George's question is some sort of a taxonomy of the impact of these non-quantifiable uncertainties and how they, in the aggregate and through the propagation process, impact the overall credibility of the analysis. But the whole concept bothers me a little bit, because it's a little bit contrary to the notion of what we mean by quantitative risk assessment. Not that you can quantify things that you can't quantify, but the whole thrust of doing a risk assessment is not the manipulation of statistics and information nearly so much as it is establishing the logic between what you're trying to calculate, which might be an event frequency, about which you may have no information, the logic between that and a level at which you do have information. Then the thrust of the evaluation, the review is on the credibility of that logic, not so much the credibility of an unquantifiable piece of information. So I'm just wondering, the way we got around this a lot in the nuclear power risk models was to spend a great deal of time establishing that logic and that, in essence, became the focus of the creativeness, if you would, of the risk assessment as opposed to what often comes up on people's minds when they think of a risk assessment as being a game of statistics. I've said many times, statistics may be five percent of a risk assessment, but it's not a very big part of it. The real effort is in establishing the answer to the question what can go wrong and what characterizes the logic of things going wrong. Are you doing anything specifically to get to these non-quantifiable contributions to uncertainty of that nature? Are you really -- to me, that's what the breakthrough of risk assessment was all about. The rest of it is old technology, came about by way of reliability analysis and general modeling. What's really creative and the aspect of risk assessment that constitutes a major step forward is this modeling process or this logic development process that you go through in going from what you want to learn about, about which you have nothing, down to something that you have good information on and you clearly understand the connection between that and what you're interested in. Isn't that the way to address non-quantifiable uncertainties and are you doing that? You don't have to answer that completely now. MR. VAN LUIK: I can give you a partial answer. We're very well aware of this issue and that's why I mentioned we have a long list of candidates. What we want to do is with some outside expert help, pick out the most likely importance items, perhaps half a dozen to start with out of that list, and say, well, if we have that uncertainty quantified in stead of bounded at this point, what would that buy us. And what we'll do is for that six or maybe eight items, we are talking about invoking an expert elicitation process, with people who would have a feel for this subject from the outside, as well as one or two from the inside and establishing a PDF for that particular parameter or modeling option, and then looking at the importance to the outcome from that and then depending on -- and that's why I said, you know, depending on the outcome of that, if that shows, whoa, this is a bigger thing that we originally thought it was, then we'll move on to the next five or six and by the time of licensing, we should have a much more solid story. But this is something that will take time and it's a sizeable investment. We're aware of the issue and this is the approach that we're piecemeal putting into place for dealing with the issue. MR. GARRICK: Thank you. Any other comments? Yes, Milt. MR. LEVENSON: I have sort of a follow-on question to George's about this catalog of things left out and are bounding. Is it possible to get a copy of that list, recognizing that it's much more fragile and work in progress, but just to get a -- at the moment, I find that I have no real feel for the scope of this issue and it's a fairly important part of the credibility of the TSPA. MR. VAN LUIK: The TSPA-SR document itself, when it comes out, will have a list in it of places where we use conservative assumptions and that's the starting list and that's a pretty short list. In the meantime, some of our other people have dug way into the hinterlands of the process models and the data interpretation reports and come up with a much longer list. The principal investigators looking at these things feel that that's -- the one approach may be a little bit short, the other approach is a little bit overboard. But you will see the very first version when the TSPA-SR document comes through the DOE review and becomes available to you. MR. LEVENSON: That will include lists of all of the things left out, like the stainless steel in the container, et cetera. MR. VAN LUIK: Those kinds of things are the more obvious ones and they will be discussed in the document, yes. But I'm talking about where we use the bounding value rather than a PDF, because our internal expert judgment was that to go beyond that, since we already know that it makes very little difference to the performance measure of interest, would be money not well spent. So those kinds of analyses also will be reflected in that table in the TSPA-SR, and then you go to the AMRs and PMRs and see what the actual logic was. So that is pretty well in there. It's pulling all that together and we have several different people doing it, coming up with different lists, and our job, our task force's job is to come up with one list that we all agree on. But as soon as it's a little bit less flaky than it is now, you can have it. MR. GARRICK: Okay. Andy? MR. CAMPBELL: It sounds like what you're focusing on, though, are parameters, where you're not sure what the range ought to be, so the analyst picks a value that they believe bounds what that range of values would be. And in that analysis that you're doing, are you going to also look at what the shape of the distribution might be on the results in terms of that parameter? That's one question. But more importantly, how are you going to factor conceptual model uncertainty, because you're dealing with parameter uncertainty in that case, but you also have the whole issue of do you have the right conceptual model and then you get into type one and type two type of errors, and that kind of stuff. Do you have a plan for dealing with that? MR. VAN LUIK: That is included in what we are supposed to be looking at. Obviously, the easiest thing to do is to look at parameters. That is another order of magnitude more difficult, but that goes to the interpretation of data, what interpretations does it allow. So it definitely is part of the plan, but whether we get that done in the first phase or the second phase, I would guess it would be the next phase. But we plan to do this to the point that at the time of licensee application, we have our ducks pretty well in a row. But I have to re-emphasize that I think and I think most of the PIs in the program think that these are second order corrections and we have captured the major uncertainties and in several instances we have captured, by analyzing, different conceptual models, for example, and chosen either the more conservative one, which is another source of conservatism, or we have somehow combined them. So there is already conceptual model uncertainty addressed in the TSPA-SR, and the supporting documents, itself. It's not like it's something that we say, oh, gee, we forgot that, but there are the major ones, like in the unsaturated zone, and then there may be some other minor ones, too, that we will want to address further in this exercise. MR. GARRICK: Okay. No further questions, we'll listen to Bob Andrews. MR. ANDREWS: Thank you very much. As Abe said, this is work in progress. The documentation of the analyses and calculations and the model and the report itself are internal to the M&O review and comment resolution as we speak. It's been an incredible team. Most of you are aware of them from the VA. It's been a fairly stable team, I'm very thankful for, and that team is very hard at work still, that team in Albuquerque and here in Las Vegas. I had the joy yesterday of looking at the cover page of the technical report and in the cover page, we're putting all the contributors to the documentation of the TSPA-SR and it was just kind of -- by the time you looked through the 20 or 30 names on there, you were well aware of the hard work and incredible work of pulling this thing together and documented, as Abe said, in as clear and concise and traceable and transparent a fashion. And sometimes, some of those adjectives compete with each other, as though of you who have prepared large documents are aware. You try to make it traceable, but in so doing, you might have lost some transparency, or you've tried to make it transparent, but in so doing, you might have lost some traceabilty. But I think we have a happy medium between those and the issues that were raised by the questioners in Abe's presentation I think have been addressed. My objective today is to kind of walk through the TSPA-SR as it stands right now. The various attributes, look through the system, how the system is connected or the components of the system are connected, and then go to the results. The objective in the hour or hour and a half that we have here is not to go through each individual component part, starting with climate and going through to biosphere and to disruptive events, but to talk about it as an integrated system. If you have questions about an individual part, I'll do my best about how that part was implemented in the TSPA model. As Abe and Russ and Tim and Carol told you, the whole building blocks of this TSPA are those 121 AMRs that provide the technical foundation and the data, in fact, that support those AMRs. Those AMRs, those analyses, model reports have used site-specific data, analog data, as appropriate, in situ data, laboratory data, literature data to develop their technical bases and the technical bases reside in those AMRs. So with that, let me go on, John, to the next slide and talk about process. The next two slides kind of go hand in hand. I think this is a fairly well defined process. This is a pictorial representation essentially of the requirements of TSPA as they are defined in Part 63, where the first step is to identify those features, events and processes that may significantly affect the performance of the repository system, both the engineered features, events and processes and the natural system features, events and processes. We've looked through, starting with an NEA database, which the department and NRC were both a part to in the development of that NEA database, as well as the international community at large. WE added Yucca Mountain specific features, events and processes to that, to the point where it became something like 1,600 features, events and processes that then had to be either evaluated and screened into a model or, with a basis, screened out of the model. Once we've done that and I think the NRC is reviewing the analysis and model reports that relate to the FEPS screening process, we developed two basic scenario classes using the definition in the TSPA-I IRSR of scenario class. Those scenario classes are what we call a nominal scenario class and, in this particular case, of a panic or disruptive events scenario class. IN the disruptive event or volcanic scenario class, there are actually two. I think the parlance in the TSPA-I IRSR is an event class; two event classes, one an intrusive event class and one an extrusive event class. Given that I have those scenario classes, now I have to have the individual component models and the integration of those individual component models and the scientific technical underpinning for those individual component models. In the TSPA, if you come around the wheel of the figure, there are essentially nine of those component parts. There's subcomponents, that I'll get to in a second, that are in a series of backup slides to this presentation, but we start first with the unsaturated zone flow, the things above the repository, climates, air infiltration, et cetera. We then get to the engineered barrier system environments. The environments in the drift, around the drift, in the rock, that can affect the degradation characteristics of the engineered materials that are placed inside the drift, with the proposed design that Abe talked to you and Russ talked to you about. We then get to the degradation of package and drip shield, their performance over time. Next, the waste form. Once the packages are degraded, the waste form starts degrading, the cladding can start degrading, the internals of the package affected by the environments inside the package once they start degrading. We then have transport through the package, through the invert materials, into the rock, and then transported in the unsaturated zone to the water table, transport through the saturated zone, and, finally, we have a biosphere, where the nuclides that are released to the community of individuals -- let's not get involved whether it's a group or maximally exposed, but we'll talk about that a little later -- that group of individuals is exposed to those nuclides. So we end up with a volcanic dose, dose induced by the low probability volcanic scenario classes, and the nominal dose, those that are not impacted by these low probability volcanic scenario classes. As Abe pointed out, there's two other principal performance measures, regulated performance measures, the ground water protection, concentration, however that's finally implemented in Part 63, assuming that it still exists in 197. It does exist in the proposed 197. And the human intrusion dose, assuming, again, that it's implemented in the way Part 63 has it currently in the draft. As you're well aware, Part 197, in its proposed regulation, allows the applicant to potentially exclude that scenario from consideration if the applicant so decides. We have, for the purposes of the TSPA-SR, included that particular scenario class into the assessment. It's not weighted. It's a stylized calculation to evaluate the robustness of the system. The one that's not shown on there is the peak dose, the requirement for the final environmental impact statement. So that is the process of developing the TSPA-SR. If I skip over the word slide, because that's in there more for completeness, and walk now through the various component parts that feed into the TSPA-SR. This set of attributes is very familiar to this board, I know, from the repository safety strategy Rev. 3 and also the viability assessment. The viability assessment volumes three and four talk about the major attributes that affect the long-term performance of a repository at Yucca Mountain. All we're trying to do here is trying to put it into a little more general construct and use some icons to point the reader through where they are in the system, starting with the natural system, with the water above the repository, as it gets into the drifts, then the package itself, then the mobilization and release; finally, transport and whatever the consequences and risks associated with disruptive events might be. As a general view, the next slide takes the TSPA wheel for the nominal scenario, a similar one for disruptive scenarios, but this is for the nominal, and shows you the individual component parts and the individual process model factors or subcomponent parts or process models, whatever you want to call it, that feed into that TSPA-SR. This is kind of shorthand notation. In the last, I think, nine slides of your handouts, they're in the backup, I didn't feel it was necessary to go through it in the presentation, but I think it's to provide you a road map, if you will, from the feeds into TSPA, which are shown here and the individual component parts that are shown there, to show you what analysis model report is the final analysis model report providing the input parameters, the input parameter distributions, discussion of alternative conceptual models, discuss the technical bases for those parameters, what those ones are. So it's a tabular mapping that takes the individual parts, piece parts shown on this wheel, tells you what parameters those piece parts are generating. There's on the order of several hundred parameters that are being generated for input into the TSPA, and shows you where the supporting documentation is for that. It gives you the analysis model report title, the analysis model report number. So I believe the Commission has, I think, as Russ said, 119 of those. So I believe all the ones that are indicated there, the Commission has. That's more of a traceability completeness kind of backup presentation than for you to necessarily do anything about, unless you want to review those AMRs. So the next set of slides just walk through those process model factors, the individual component parts that feed into the TSPA model, starting, first, with those that affect the attribute. That's the water contacting the waste package. So climate, infiltration, UZ flow, the effects of thermal hydrology on the in-rock processes and, finally, seepage into the drifts. The next slide shows those principal factors that are affecting the environment inside the drift, the environment that the engineered barriers are likely to see as they change with time after the wastes are emplaced. In particular, the chemical environment and the thermal hydrologic environment. Also in that physical environment are the stress environments associated with the degradation of the drifts themselves. The next slide looks further into the drifts and this is essentially looking at those component parts that relate to the degradation characteristics of the drip shield and the degradation characteristics and projected performance of the waste packages themselves. Next slide gets into the internals of the package. We have two basic types of packages, one being the commercial spent nuclear fuel packages and one are what have been termed the co-disposed packages. It's co-disposed glass, logs with DOE spent nuclear fuel rods going down the center. In addition to these, though, there are other kinds of packages, special packages, for example, the Naval wastes have specialized packages because of their size and handling requirements. But these are the two principal ones and any other special type package, we've done a special off-line analysis of the consequences associated with that kind of inventory and that kind of waste, if it's different from these kinds of wastes. The next slide, it should be pointed out that there are on the order, for DOE type wastes, there are on the order of several hundred different specific waste forms. There are not data on every one of those specific waste forms, so the DOE spent nuclear fuel program, for our purposes, has lumped those waste forms into 13 individual types of waste forms, with similar types of characteristics and similar types of inventories and similar expectations about the degradation of the cladding associated with those waste forms. The next slide talks to the transport aspects, away from the engineered barriers. In particular, transport through the unsaturated zone, transport then through the saturated zone, and ultimately the uptake of these contaminants by the biosphere group and whatever dose consequences are associated with the uptake and use of the water that may have been contaminated at some point in time by the other degradation processes and transport mobilization processes. The next slide shows that when we have a volcanic event, with the probabilities that are currently being estimated, and I believe you talked about those yesterday, when you talked about the KTI meeting on igneous activity. Right now, we're sampling that probability from the expert elicitation that was performed on the probability of an igneous event, but we have then two types of consequences. So, therefore, two types of event scenarios that are being assessed. Type one is the event occurs, intersects the repository, degrades the package and the event conduit continues to the surface and you have a cinder cone and an ash associated with that. The ash is redeposited with the wind over the member or members of the critical group and there's a dose associated, potential dose associated with that release pathway. That one we've called an extrusive volcanic event or an eruptive volcanic event scenario event class. The other possibility is that the dike intersects the repository, degrades the packages sufficiently so they've lost their containment possibility, degraded the drip shields, degraded the cladding, and then the normal processes of the nominal scenario take place; i.e., all the slides that I had in there earlier about radionuclide mobilization, alteration of the waste form, release from the waste form, transport through the engineered barrier system, transport through the UZ and transport through the saturated zone, and then uptake in the well and biosphere dose consequences associated with that. So we have two very different pathways, all with the same initiating probability, but very different consequence models from that initiating probability on. They are then, of course, combined at the end for the same event. If I go to the next slide, the regulation currently requires stylized human intrusion scenario. This is that stylized human intrusion scenario. Somebody drills inadvertently, goes through a package, goes through the waste form, goes back through the package, back through the EBS and down, continues down to the saturated zone, and then radionuclides can be mobilized, released from the package through that degraded -- what now is a degraded engineered barrier, through what now is a degraded unsaturated zone barrier, and through into the saturated zone and then the other processes take their normal course. The next slide tries to summarize, I think, what Russ has told you, what Abe has told you, and I'm going to go into a little more detail on some of the following ones on. The technical bases -- turn back to the viability assessment. The viability assessment, volume three of the viability assessment, which I know this panel reviewed, had a large technical basis document that went along with it, essentially nine chapters. That provided the individual bases for the individual component models in the viability assessment. In the SR or the SRCR TSPA that we're talking about here, that nine got expanded to nine process model reports, which are very similar, slight differences between the technical basis document and how they've lumped the process model reports. But the fundamental science is embodied in those 121 AMRs, analysis model reports, developed by the labs and the GS and M&O participants to support the TSPA-SR. Of those 121, 40 of them are direct feeds into the TSPA. So those 40 provide a direct data set or model or conceptual model or equation or something, are a direct feed into the TSPA. You say, well, what about the other 80. Well, the other 80, probably 15 of them relate solely to screening arguments, features, events and processes screening arguments. The other 65 are process models. They are alternative models. They are supporting models that feed into those 40 that ultimately support the TSPA model itself. So all 120, we have several family trees of these 120 and how the information flows in all 120 AMRs takes about 15 figures to show all of that, and we've put that in the appendix of the TSPA document, to show where did all the information come from to support the final feed into the TSPA. So one can very, very quickly and very easily pull the chain, pull the string, and go back to the AMR that gave the technical basis for its inclusion in the TSPA. And I might add, based on the discussion that was going on before, the technical bases for the assumptions involved in those analyses and models and if there were assumptions on degree of conservatism or degree of complexity or something that's going to be treated as a PDF or something that's not going to be treated as a PDF or something that's not going to be treated as a PDF and the reasons for that, it's in those supporting analysis model reports. The 40 that feed directly into the TSPA-SR model are shown on this slide. This is the one you need in addition to Russ' pyramid, the one you need the magnifying glass for. Actually, the next one, too, you'll need a magnifying glass for. In the actual document, they appear a little bit bigger. But the color coding is color coding the AMRs to the corresponding process model report in which that analysis model report is summarized and the arrows are showing where in the TSPA model are the individual component analysis model reports input and then where in the middle is the process of doing the TSPA, starting with the package degradation, going to waste form degradation, going to EBS transport, UZ transport, SZ transport, and, finally, the biosphere. That middle part is the guts of the TSPA model, if you will. All of those inputs are what's being integrated within the context of the TSPA model. If you don't like boxes, then the next two slides try to walk it through from an information flow point of view; what information, what kind of, if you will, intermediate result is passing from component to component within the TSPA model. One of the objectives of the TSPA model itself and how it's been constructed and how it's been documented is to show, in as transparent a fashion as possible, given that I have 40 AMRs that are feeding it, as transparent a fashion as possible, how the information flows from component to component; how does climate information flow to unsaturated zone flow; how does unsaturated zone flow information flow to seepage; how does thermal hydrology information flow to the degradation of the package, degradation of the cladding, degradation of the drip shield; how does thermal hydrology information flow to the characteristics of the invert, the thermal and hydrologic characteristics of the invert. So there are placeholders within the TSPA model where we go in and we've done this in the technical report and in the model document, where we go in and stop the results, if you will, and look at what information is passing from component to component; what flux, it's general energy kinds of things. It's heat that's passing from component to component, it's mass that's passing, it's volume of fluid that's passing from component to component, and ultimately it's -- and it's chemistry passing from component to component, and, ultimately, it's radionuclide activity that's passing from component to component. You can say that from barrier to barrier, you can say it from feature to feature, but it's traceably and transparently showing how that system evolves based on the models available through time. Some of that information, as a backup, I've included, in the first part of your backup slides, I was not going to go over it here unless it gets into that level of detail. But as we go from flow to seepage to package to water mass movement, mass release, activity release, and ultimately activity release at the 20 kilometer point of compliance, that ultimately leads to a dose. That's all in the backup and it's in greater detail in the technical report. So these two slides show that how information flows from part to part. Same thing on slide 17. It shows once I get internal to the engineered barriers, that is internal to the drift, how the environments are propagated through time and what downstream processes impact, and then how the degradation characteristics of the engineered barriers are propagated through time, and then ultimately the release, mobilization and release of nuclides from those engineered barriers and back into the natural barriers. The next slide, Abe alluded to this. Uncertainty and variability, both primarily quantified, has been directly incorporated in the TSPA-SR model. The third bullet there is a bullet that Abe had on his slide and that is where the individual analysis model report originator, the reviewers, the checkers, felt there was either a large degree of complexity, large degree of uncertainty. The goal was to be defensible and in being defensible, they were probably, in some cases, a little conservative. They documented that in their analysis model report. Are there alternatives they could have chosen? Yes. Did they document what those alternatives were? Yes. Did they give a rationale why they didn't choose that alternative? Yes. Could we propagate those alternatives back through the rest of the system in sufficient time? Maybe, because some of those alternatives require alternative representations. They require alternative data. They require alternative analyses and they require alternative abstractions into the performance assessment. It's not a simple flick the switch. It is in some cases. If it's down at a parameter level, I think Andy's question is very well taken. If it's down to the parameter level, it is a very simple aspect. If you don't have the process model for how you think -- take the example that was alluded to earlier, the stainless steel degrades and that degradation characteristics are sufficient complex and the amount of information available in the environments that we have is sufficiently uncertain, it would have taken a big effort to incorporate stainless steel into the TSPA model. First, the process model, then the abstraction, and finally into the TSPA model. So a conscious decision was made by all concerned, including the Department of Energy, in that particular case and others, of which things to include and which things to exclude and carry those excludes. MR. GARRICK: Bob, just to pick up on that a minute. Obviously, you have to have some evidence in order to assign a number or a parameter that you think represents an upper bound or a conservative estimate. It just seems that it would be easier to hedge your bets with a discreet probability distribution, for example, than it would be a single number. In other words, you talk about an unquantifiable event or number or parameter, on the one hand. On the other hand, you say you assume conservative values. So you've got to have some basis for justifying it as a conservative value. MR. ANDREWS: I think they do. I think -- MR. GARRICK: And my point is this. It's easier, as a matter of fact, to justify some sort of a distribution than it is a single value. MR. ANDREWS: You're right. To justify the distribution, in some cases, say I think I'm pretty sure it's within this range. I have evidence to say it could be at low values, pick the parameter, pick the model, it's the case all across the board. There's evidence to support it down here. There's evidence to support it up here. There's evidence to support it in the middle. It might be analog evidence. It might be direct field observations, whatever the evidence is. However, when faced with the requirement of being as defensible as possible and to not be optimistic with how performance may evolve in this system over time, the analysts, I believe, personally, correctly, went with the conservative approximation. There have been words in Part 63 or in the statement of considerations, perhaps, I'm not sure where they were for Part 63, that if the applicant -- it might have been 60, in fact, I forget where they were -- if the applicant has uncertainty in a particular aspect or alternative representations or alternative models, it is okay to include those alternative models and alternative representations in your performance assessment, but we still want to see the effects of that more deterministic, more conservative representation. MR. GARRICK: That's not a particularly unreasonable approach. My only point is you don't want to get yourself trapped into saying, on the one hand, that you have no information, and, on the other hand, you're using a number that's characteristic of information. MR. ANDREWS: That's well taken. That's a point well taken. The next three viewgraphs summarize some of these aspects in a kind of Consumer Reports sort of fashion of what uncertainty was directly incorporated in the TSPA-SR model, what variability, and this is generally spatial variability, was included in the TSPA-SR model, and a brief set of comments. The comment, if I didn't check uncertainty or variability, those are generally the ones where a more conservative representation was taken within the analysis model reports in order to avoid some of the complexity associated with that particular process or coupled process. The very first one on there, on page 19, is the coupled effects on seepage. This was one that also was noted in our review of the viability assessment, it's been noted several times by the Nuclear Waste Technical Review Board, noted by NRC staffers and center staffers on their review of the viability assessment, as well. And that is that the actual seepage into a drift following emplacement and the perturbations that are caused by emplacement, the mobilization of water, the changes in the chemistry, the changes in the mechanical stress around the drift, the degradation of the drift support system itself, all of the thermal mechanical hydro chemical coupled processes are -- I think everybody would acknowledge are very difficult to quantify with a high degree of defensibility. There are data, yes. There are data from the -- specific data, even, associated with seepage tests that have been conducted by Berkeley in various niches at the site, the actual thermal hydrologic drift scale test at the site, smaller scale thermal hydrologic tests at the site to support and plus modeling and analyses of coupled thermal chemical processes. There are data to support a range of possible effects associated with thermal hydro chemical mechanical stress evolution at the site. But it's fairly broad what the possible outcomes are in terms of their effects on the rock properties, the permeability, the fracture aperture, the capillary suction in and around the drift. Therefore, because of that complexity and because of that uncertainty in what is the most reasonable, the most defensible model for seepage in the effects of all these coupled processes, a conservative assumption was taken that take the flux five meters above the drift, I think the bullet says there, and put that flux into the seepage model. Don't try to capture all the nuances of what happens in the first ten centimeters or the first millimeter of the drift wall as the drifts are degrading with time. So that's not uncertainty, it's not variable in the model. Is the actual seepage into a drift uncertain? Yes, because the seepage model itself is uncertain and the flux itself is very uncertain and highly variable. But the coupled processes and the complexity associated with those was, in this particular case, eliminated by making this conservative assumption. I should point out that that was, going back to the TSPA-VA, that was, in fact, a recommendation of the peer review panel to say in order -- the complexities associated with this one are so large, you, department, may be better off simplifying it and that's what we've done. George is biting at the bit here. MR. HORNBERGER: Just a quick question on that. The TRB has been critical of this particular aspect and one of the contentions is that some of this uncertainty may, in fact, be diminished or go away if the temperature of the repository were lower. Can you give me the basis of that? Do you agree with that contention and if so, what's the basis of reducing the uncertainty? The coupled processes are still there, is that right? MR. ANDREWS: Yes. Even at 70-80 degrees C, you still have coupled processes. You still have mechanical processes at zero degrees C. As soon as you open that drift, you have mechanical degradation processes that will kick into gear and then you have the thermal hydrologic coupled processes and the thermal chemical coupled processes. I can't speak for the board and their beliefs about reducing uncertainty with cooler repositories. I believe it's not totally tied to this particular one. I mean, this is what gets the focus, but it's also on the degradation characteristics of the engineered barriers in a cooler -- where cooler now is 70, 80, 90 degrees C, environment rather than 120, 130 degrees C. So it's not solely in the rock that they're after. MR. HORNBERGER: I'm not asking you to speak for the board. I was just asking, in your experience in doing TSPA, would the uncertainties be significantly reduced at these lower temperatures? MR. ANDREWS: No. We got enough uncertainty. MR. GARRICK: It just may be displaced. MR. ANDREWS: It might be displaced in time. MR. GARRICK: Right. MR. ANDREWS: I don't know if you want to pick up any other of those that don't have checkmarks, but if I go to the second page, for example, I think it's a useful one on the waste form characteristics and the waste form degradation. There are a number of fairly complex processes that occur once water, in whatever form it exists, whether it's humid air or actual liquid water, when it comes into contact with the waste form, there are very good data collected at PNL, at Argonne, at Lawrence Livermore Labs, and internationally about the degradation characteristics of the waste forms. Those data have been used, but they have a fairly broad range and the applicability of that range under the exact environments that we're expecting is uncertain. We could have used that range or we could have used a more bounding value. What's there right now in Rev. 00 that we're talking about is that more bounding value for the waste form characteristic degradation. I don't know if there's any -- in package transport, that's another nice one. What the internals of the package look like after the package has degraded, as a function of time. The characteristics of the basket material, the degradation of the basket material, the stresses involved inside the package, the hydrology in the package, the chemistry inside the package are very uncertain, very uncertain. And for anybody to confidently predict the internals, with the exception of some basic fundamentals like temperature, would be -- well, it would be difficult. So a more bounded type approach was taken of bring that mass, bring that activity, once the waste form is altered, to the edge of the package, the inner edge of the package. So transport inside the package, transport along fuel rods, transport through a degraded fuel rod assembly was not taken credit for in the Rev. 0 TSPA-SR. And that would require, obviously, a different model, different representation, alternative conceptual model of how you think the internals of the package perform over time. It is much simpler, much more defensible to take the more conservative approximation in this case. By the way, most of these have limited effect, where limited is small, effect on the post-closure performance. WE did not make the collective decision of where to add conservatism based on things that were highly important to performance. So it was generally focusing on those that were less significant to performance. One exception to that is this secondary phase issue that's a bullet on the waste form degradation and solubility limits. There's no other -- well, there's other interesting ones in here, but maybe that -- I think those examples, I think, are probably sufficient to walk through the process, the collective decision-making that was done. As Abe told you, each one of these are being examined with this small group, evaluating is there -- should we look at quantifying this uncertainty and if we do, should we then run a calculation to see what the impact of that particular aspect on the system performance was. I think taking Dr. Garrick's point to heart of now that we have the tool, it's running, it works, we believe giving reasonable results, now it's kind of time to exercise the tool with alternative representations, gain additional understanding. MR. WYMER: It's a little disturbing to me that most of the areas that are unchecked, or many of them, are chemistry related areas and to a chemist, those don't look any more difficult than some of the other complexities which have been taken into account. And in some of them, like second phase formation, you may have quite a profound effect on the release of radionuclides, where it would be significant in the dose to the person at 20 kilometers. So it seems to me that, from a prejudice point of view, what I see is that reflection of the knowledge and backgrounds of the people doing the study rather than reflection of the difficulty of doing the analysis. MR. ANDREWS: Well, there are a lot of good geochemists working on this project and who have supported, through their analysis and model reports, supported the development of these inputs. The one example you pointed to is a very near and dear example to many of our hearts. It was an example that we used in the VA, in the viability assessment. It's an example we've had extended discussions on with the folks at Argonne who are doing the detailed fuel testing and characterization of the alteration phases of the fuel. It's a point that I think has been made by NRC and center staff with respect to review of Pena Blanca and utility of Pena Blanca as a very valuable analog for waste form degradation and mobilization and transport of some of the actinides that we're talking about here. MR. WYMER: There are other kinds of chemists than geochemists, you know. There are inorganic chemists and physiochemists who could give a good deal of insight here. Geochemists have their own point of view. MR. ANDREWS: Well, I'm not going to get into that debate. MR. WYMER: Unfortunately, I have one sitting next to me. MR. ANDREWS: But the point is well taken and I think there are complexities associated with the controlling phases as these materials degrade. There is complexity associated with some fundamental thermodynamic information on these controlling phases. Not so much on the uranium side, but for -- when I put neptunium in them or plutonium in them, those fundamental thermodynamic information, I think, is -- I'm not a geochemist, so excuse my bias, but I don't think some of that fundamental thermodynamic information is available. I think there was a presentation to the technical review board in Reno by Dr. Glassly from Livermore and he pointed that uncertainty out as well, that some of the basic thermodynamic data is just not there, and I think the board -- I forget -- one of the board members asked him, well, are we pushing to get that kind of information, and his response was, well, I hope so, but it sounds like fundamental university type laboratory research to come up with those data. And in the absence of some of those thermodynamic constants and time stability constants, it was difficult for -- in the high defensibility role that we wanted to be in, to take credit for some of those aspects. Is there more that we could gain there? Sure as heck is. And I think Dr. Ewing was a member of our peer review panel or DOE's peer review panel, I should say, and he, I think, shared your kind of comments on the VA. I think they are throughout the VA comments. It's pretty complex. MR. WYMER: Yes. No more complex than some other aspects, though, that are dealt with, in my judgment. But, okay. I'm not going to beat a live horse. MR. ANDREWS: Okay. MR. GARRICK: By the way, Bob, these are excellent exhibits. I must like matrices, because these are very helpful. MR. ANDREWS: Good. Okay. Now, we've set up the stage. It's time to get to some results, preliminary results. The first result slide, John, why don't you go ahead with the VA, puts it into context of what the TSPA-VA result that's most comparable to the results that I'm going to be showing to you. In the VA, of course, we did not have Part 63 or 197 on the street as proposed regulations. I think they came very shortly after the summer of '98. However, there was enough discussion, I think, with ACNW, with NRC staff, that we had fairly good knowledge of what was going to be in the draft regulation when it came out, which was a slightly different way than we were doing most of our plots in the VA, quite frankly. The how one does the expected, quote-unquote, the dose -- just the mathematics of doing that calculation was slightly different than the way we were proceeding in 95 percent of our plots that we presented in the VA. But we had one set of plots, shown here, figure 4-28, that most closely represents for the nominal -- this is nominal scenario class in the VA -- most closely represents the way NRC ended up writing the proposed Part 63. What I've shown here is the 95th percentile mean, i.e., expected median and the 5th percentile is actually off the curve. The 5th percentile was zero out to 100,000 years. So it gives you an idea, backed with VA models, VA assumptions, whether they were good, bad or indifferent, VA design, this is the comparable VA, TSPA-VA result for the slides that are going to be following. I did not change the time access to be logged like the subsequent ones are. There have been a lot of changes. As Abe pointed out, there was a number of changes. There was hardly a model that didn't change between the VA and the SR. The design changed from the VA to the SR and in some cases, the approach changed between the VA and SR. So the following slides all relate to those 121 AMRs and their incorporation in the TSPA. So let's skip to slide 24. What I've shown you here or have up here are 300, the skinny little lines are 300 individual realizations of the total system performance based on those 40 AMRs, cranking them through with their uncertainty and their individual parameters and models, et cetera. So each one, each one of 300 has an equal likelihood of occurring. We then superimposed on top of those 300 four basic statistical measures of each one of those, essentially done in a per year basis, where per year really means, in this case, per time step and time step is about 25 years. So it's per time step slice and I've shown the 95th percentile, the mean or the expected value, the median or the 50th percentile value, and the 5th percentile value. Several things to note on this slide. One is there is quite a wide spread, variance, if you will, of the dose as a function of time. If I'd look at it at any particular time, there is a wide variation of dose. If I look at any particular dose, it's a wide variation of time. So it's quite a wide spread. Understanding that spread, that distribution of the results is an important component of performance assessment, and we're going to spend a little more time talking about that, why the broad spread. If you look at 100,000 years, that spread is going over seven or eight orders of magnitude. It's very broad. Another point of observation is in those 300 realizations, for the data that are contained within the analysis model reports, for the models that are supported in those analysis model reports, none of them had a failure of the engineered barriers prior to 10,000 years in this nominal scenario class. We'll look at some examples in just a second where that's not the case. It's not the case in the volcanic event and it's not the case in human intrusion, but for the nominal performance, the expected performance, with its uncertainty embedded in there, none of them failed before 10,000 years. Your logical question is, is it impossible to have failures before 10,000 years, and the answer is clearly no. There is a probability of having a degradation prior to 10,000 years, that we'll get to here in a second. But for the nominal performance over that many realizations, there were none. You might ask how did you pick 300, why didn't you use some other number. Well, our goal was to get a stable mean, not a stable 5th percentile or a stable 95th percentile, but a stable mean and we ran 100, 300 and 500 and 300 was stable enough, 500 was sitting right on top of 300. So we stopped at 300. Just totally economics related to why 300 and not 500. They're one on top of each other. The difference between 100 and 300 is not that much, it's probably 20 or 30 percent, and we have those plots in the technical report as backup. Other points of information are probably best explained by looking at the next slide. MR. LEVENSON: Bob, before you leave that one. Is that correct that the 95th percentile line crosses the mean? MR. ANDREWS: Yes. And you'll see it also when we get to volcanic, in a way, that the 95th percentile can exceed the mean; i.e., the mean is driven by the top two or three percentile of the distribution, which is what you're seeing right there. The next slide talks to the principal nuclides that are controlling this dose rate and all I've done is taken, for the expected value case, this is now expected, I have to be careful here, this is expected value of the output. So this is based on -- still based on 300 realizations of the probable performance of the repository system and I'm just looking at one expected value from that, and then what are the nuclides that are controlling that expected value. And what you see here, quickly, is that up till 40 or 50,000 years, we are dominated by the very high solubility, unretarded, diffusing radionuclides, like iodine and technetium, same thing we had in the VA. That at early times, those things that are unretarded, those things with very high solubility which can diffuse through thin films or long thin film boundaries, they are released first. After a while -- in this case, it ends up being about 50,000 years -- it's the lower solubility, retarded, but not completely retarded nuclides. They diffuse a little bit, but they're more controlled by how much water advects through the system, how much water advects through the package and through the invert, neptunium and the colloidal plutonium. This is -- we have two colloidal plutonium species that are being tracked in the TSPA-SR and those are the ones that are coming out at longer times. I should point out that a large fraction of the inventory, of the total inventory is still retained either in the waste form or in the waste package or in the EBS, simply because of the very low solubility or very high retardation characteristics of those other nuclides. So what we see here is things coming out and trying to understand why they come out and what is the order in which they come out, but there's a lot of other things, and I did not bring that plot, but we show the plot in the technical report, that are retained and that are essentially delayed significantly before they come out and provide any kind of dose consequence. There's a little bit of understanding that can go into these plots, but that understanding really resides in the backup slides. Part of it's clear that until you have the primary containment barrier, i.e., the package, degraded and you get water into the package, you have no release. Even when you have the package degraded, and, generally, and there's backup slides to point this out, for your benefit, once the package degrades, it's generally degrading at the closure welds. It's generally degraded by stress corrosion cracking at those closure welds. Those cracks are micron size. So liquid water is not getting into those cracks. There's an analysis model report to support that. But humid air can get into those cracks and nuclides through humid air condensing on the waste form itself and that other conservative assumption that we talked about earlier, about the innards of the package and how they're being treated, nuclides can diffuse through those cracks in the stress corrosion cracked welds of the earlier package failures. So what you see here for iodine and technetium is diffusion through very thin cracks, while the drip shield is intact, but the packages have degraded at the closure welds. At later times, the drip shields are degraded. Liquid water for that fraction of the repository that has seepage, which is depending on the timeframe and the climate state and the percolation flux, et cetera, the fraction of packages that see seepage changes with time. But for that fraction that sees seep and for that fraction where the drip shields have degraded and the packages have degraded, liquid water can get into the package and liquid water can take away nuclides from the package, through the package, through the invert, and back into the rock, and that's when you see neptunium and plutonium taking over. MR. GARRICK: Have you somewhere cast these into a dose exceedence CCDF form? MR. ANDREWS: No. MR. GARRICK: That would be very -- that communicates very well. I just wondered if you had considered that, especially for the disruptive events which have a recurrent cycle to them. MR. ANDREWS: I appreciate that, John. We struggled internally on communication of this and, quite frankly, in polling people, obviously, we didn't poll you, we concluded that CCDFs, although explanatory for some fraction of the audience, were obfuscating for a large fraction of the audience and, therefore, we -- including people who were on our peer review panel before. So reasonable people. So there was a transparent -- MR. GARRICK: These are important curves, but from a risk perspective, the CCDF is the risk curve, whereas these are not. MR. ANDREWS: Right. MR. GARRICK: These are just probabilistic dose curves. MR. ANDREWS: These are risk when you multiply them by .999, which is what these are. MR. GARRICK: Yes. But I would think that especially for the disruptive events, it would be a useful, a very useful presentation to be able to -- MR. ANDREWS: That's a good suggestion. MR. GARRICK: -- answer the question in one diagram what the risk is. MR. ANDREWS: That's a good suggestion. We had a little dialogue with some other review groups on how to prevent the disruptive event work, and that would be, I think, a good suggestion. We have not implemented that yet. MR. GARRICK: Okay. MR. ANDREWS: Let me go on to slide 26, the one looking at longer-term performance. This particular slide and the one that follows, which talks about the key nuclides associated with this slide, is the case of extrapolating those models used in the 10,000, 100,000 year performance results, extrapolating those models on out to a million years. We understand, as we read proposed 197, that that maybe isn't exactly what they intended, but this shows what the impacts of doing that would be. We stopped at a million because of 197 saying that was the time period of geologic stability, which is also the time period the NAS thought was of geologic stability. I think it reasonably captures the peak. You can see the peaks are coming down as you go out past two or 300,000 years, coming down primarily because either things are slowly but surely decaying. Some of these nuclides have very long half-lives, so they're still in the system or they're being absorbed or they've come out of the system. One thing -- I think that's probably enough to say on that particular curve and the following curve. Let's go to the disruptive events, the igneous activity scenario class. This one. This is the one that I think Abe said things were in checking or in review. This is the curve that's different a little bit from the curve that we presented in the beginning of August to the Nuclear Waste Technical Review Board, and I'll walk through that difference here in a second. But before I do that, let me talk to the form and structure of this particular set of curves. At the left-hand side, you see some nice smooth responses. The curves all more or less look the same. You just have a distribution around a mean or around the median for those curves. Those curves are all the extrusive igneous event, whereby the pathway is -- the event goes through the repository, intersects the package, takes a fraction of the waste up, entrains it, and then the wind blows south and the waste is deposited over the landscape along with the ash and the primary pathway is an inhalation type pathway associated with that release. So there is a wide uncertainty on those, being driven by the number of packages that are hit, by the probability of occurrence. These particular curves are still sampling the probability of occurrence over that distribution from the expert elicitation, uncertainty in the wind speed, et cetera. So there is a distribution around that. Another thing I should point out on this is these are probability weighted dose rates. These are not deterministic doses, given the event occurs, what's the dose. These are already factored into the analyses for comparing apples and apples to make them comparable to the other slides, have already factored that probability into the dose rate. So this really is a risk. So that 95th percentile is 95th percentile not of the total distribution, I think this is getting to your point where I think, John, it's a really good suggestion, this is 95th percentile given the event. So it's conditional. And I think if you showed the complete, either as a PDF or probably better, as you point out, as a CCDF, the full distribution from probability one to probability ten-to-the-minus-eight of dose consequences, you probably would have a clear picture of the overall system, whereas by probability weighting it, which is the way Part 63 asked us to do it, but by probability weighting it, you've lost that consequence time probability factor, because it's already in there. The third thing to point out, which is the reason the slide changed from what we did before, is the right-hand portion of the slide. What I have shown here is actually 500 realizations, individual realizations, but there's actually 5,000 realizations that are behind this particular plot. You say, well, my gosh, why did you go to 5,000 in this case. Well, the reason is in order to get a stable mean on the dose consequences associated with the indirect intrusive event, so the volcano goes -- well, it's not a volcano, now it's a dike, the dike goes, intrudes the repository, degrades the engineered barriers, and then the natural system takes over. The timing of when that event occurs, of course, is very uncertain. So it's being stochastically sampled in here. But the consequences are a function of the uncertainty and all the other aspects of the system, uncertainty in the unsaturated zone, uncertainty in the saturated zone, uncertainty in colloids, uncertainty in seepage, all the other uncertainties take over or get a distribution of dose responses for the indirect intrusive volcanic event. And simply put, in order to get a stable mean, we had to go to 5,000 realizations over a 50,000 year time period. That's why the plot stops at 50,000 years. In order to get a more reasonable stable mean, that's the red line there, for the risks associated with the indirect intrusive volcanic event. So this curve is slightly different in the presentation that we gave to the TRB back in the beginning of August. We had, I don't know, 500 realizations, not 5,000 realizations, for this particular plot. So the plot is a little different. Let's see. The next plot essentially just combines the two expected cases. So looking at the means of the distribution of the nominal performance and the means of the distribution of the igneous scenario classes, we essentially just add them. And these are now correctly weighted by the probability. The probability, when I combined them, is one. The probability of one is .999. And you see for this case the predominance in the first 10,000 years is driven by the low probability, but high consequence igneous event scenario classes, something I think the NRC has pointed out or alluded to in their review of the VA and I think in earlier documentation that they have produced. MR. WYMER: Before you leave these curves, which are very interesting, it occurs to me that with the exception of the igneous activity curves, the other curves all start at post-10,000 years, and that's because, I presume, C-22 is considered to last that long before you get any significant failures. Ten thousand years is a long time and the database for the corrosion of C-22 is not too large and you tend to believe extrapolations like that when you have a good understanding of the basic processes involved as opposed to just measurements. You understand mechanisms. How comfortable are you that you can believe that it will not corrode for 10,000 years? How much do you know about the fundamental corrosion processes of C-22? MR. ANDREWS: Well, I'm probably not the best person -- I'm not a C-22 expert. MR. WYMER: This is fundamentally your whole analysis. MR. ANDREWS: Well, in the first 10,000 years, for the nominal scenario class, yes, the degradation characteristics of that waste package and the welds and the stresses at the welds are what dominate the performance. The analysis and model reports and the data and their justification incorporating the uncertainty, and there is uncertainty in the degradation characteristics, there is uncertainty in the stress stage at the welds, there is uncertainty in other mechanisms, other process mechanisms that can lead to degradation of those engineered materials, all of those are in the analysis model reports, plus a description of their uncertainties and their basis. In order to evaluate kind of the what if sort of scenario that you're alluding to, I think, is we've done several things. One, that I'm going to get to here in a second, is that we looked at this barrier, as a barrier, and looked at all of the component parts of that barrier that, first off, are included. So the things that we have included and their uncertainty and pushed them to their -- all of the key ones to their 95th percentile in order to see what happens, if you will. That's one thing that we've done. The other thing we've done is we've done, in support of the repository safety strategy, we've done -- and looking at this potential vulnerability, we've looked at juvenile failures, quote-unquote, juvenile failure of the package, and then done the same analyses off of those juvenile failure packages that we've done for the nominal packages. MR. WYMER: Where you assume a limited number of failures. MR. ANDREWS: Right, assume a number have failed. They fail with a patch kind of opening under a drip shield and then we've sometimes failed the drip shield and the package at the same time, just to gain an understanding for how the system performs in the absence of that barrier, if you will. The other thing, of course, that we've done is -- MR. WYMER: What results do you get when you do that? MR. ANDREWS: Those are -- for a singular -- I don't have them, at the top of my head. They're in the document that's undergoing review right now. My recollection is for a single package, it was, at 10,000 years, it was on the order of ten-to-the-minus-two or ten-to-the-minus-three millirems for a package. Abe, do you remember what the numbers were? MR. VAN LUIK: I don't, but it sounds about right. MR. ANDREWS: It's in that ballpark, anyway. MR. WYMER: Okay. Thanks. MR. GARRICK: While we're on that curve, which is a very interesting one, given that the igneous event is now controlling the risk through the time of compliance, does that not bring up the whole issue of the design of the repository in terms of how much you should invest in trying to design a 12,000 year waste package? In other words, aren't you in a situation here with a much more conventional waste package wouldn't change the risk? MR. ANDREWS: Well, if I had a much more conventional waste package, I'm not sure that line on the right would start appearing after 10,000 years. MR. GARRICK: No, it wouldn't. MR. ANDREWS: It would be significantly before. MR. GARRICK: That's the point. It doesn't matter. MR. ANDREWS: For this. MR. GARRICK: If you moved everything -- yes. MR. ANDREWS: If I moved it significantly to the left. MR. GARRICK: It doesn't matter from a risk perspective. You could get by with a lot less extravagant repository design. Given that the risk is being driven by something that you can't design out. MR. ANDREWS: It's possible. MR. GARRICK: Just something to think about. It seems to me, as far as my paying for this, from a risk standpoint, I'm paying a heck of a lot money, perhaps, perhaps if the calculations haven't been done, to achieve a level of performance that, from a risk perspective, is kind of irrelevant. MR. ANDREWS: I agree and that's very possible, and I think Abe alluded to some design related sensitivity analyses that are to be done. Right now, we have close to a point design. It's a flexible design, but it's close to a point design. We are, this fall, probably going to do some limited simplifications of that design, just to see what if, I think to address questions like that. Having done that, that's one step of performance assessment, generating the curve and a series of curves. Another, and, in fact, as important, maybe even more important aspect of it is to evaluate why those curves are the way they are. Do the sensitivity analyses, do the statistical analyses, do the barrier importance analyses, to gain an understanding of what's going on and why is it going on. So there's been a wide suite of these done in support of the SRCR, first, and documented in the technical report and documented in the repository safety strategy. These start first off with basic statistical evaluations, what drove the variance, what drove the mean, what drove the top ten percentile, et cetera. There's a wide variety of statistical techniques available and we've used the ones that are most appropriate for our intended purpose. We've done then a large number of sensitivity analyses, taking an individual factor, an individual parameter and doing 95th-5th percentile sensitivity analyses on that factor or parameter, trying to gain an understanding, by factor, what is its contribution to the system performance, always running out to the 100,000 year time period. Why 100,000 years? Well, 100,000 years was to go sufficiently beyond 10,000 years to the point where the engineered barriers were degrading, go sufficiently beyond 10,000 years to assure ourselves and the reviewers that there was no significant degradation of any aspect of the system beyond the 10,000 year regulatory time period. So all the sensitivity analyses and barrier importance analyses are also done out to 100,000 years, with the exception of that 50,000 year volcanic event one. The final set are what we've termed barrier importance analyses and those are going barrier by barrier and looking at, in two different ways, looking at either 5th or 95th percentiles of the key aspects of that barrier, 5th being on the good side of performance, 95th being on the bad side of performance. If that particular parameter, a low value, meant bad performance, then we flipped it. So it's looking at good or bad within the distributions that we have. Finally, the last bullet there is in some cases, for particular barriers, they were neutralized; i.e., the function of that barrier was removed from the analysis to evaluate the contribution of that barrier and all the other barriers to the system performance. Those were mostly used in support of the defense-in-depth and multiple barrier determinations that are embodied in the repository safety strategy. The next three -- I'm going to go a little faster here -- slides just talk through using the same orientation I did in the previous table, which is the same table that's in the back of the document, go through what's the barrier, what's the barrier function, and what was the importance analysis done on that barrier, where that performance analysis here was that 95th-5th percentile aspect. So you can read what aspects of the system were pushed to their limits in that particular table. The next slide, which is a results slide, walks through some regression analyses, and these are supported in a lot of different ways. I'm just showing you the regression analyses of the nominal performance scenario to what drove the variance. So I'm trying to understand what drove that six-eight orders of magnitude of variance of system response for all of the uncertainty inputs that were put in there. I think Abe already alluded to this one, but the top four relate to the package. They relate to the welds of the package and they relate to essentially stress corrosion cracking at the welds of the package. So that is a point of continued investigation, continued discussion. I believe this issue came up, although I was not able to attend, at the container life and source term KTI meeting, about the importance of the stress state at the welds and degradation mechanisms at the welds, which includes the defects at the welds. The other one is the saturated zone flux. Having done that, it points to a few parameters of the several hundred that are in the total system model to do a barrier importance analysis on. The next slide, the results slide, John, if you'd just slip to that one, assumed -- you know, I took 95th percentiles and 5th percentiles of those stress factors, those corrosion rates, the defect distribution, the defect orientation, the defect size, all of which are parameters in the TSPA model, and pushed them to their 95th percentile, and then ran 100 realizations, we only did 100 when we were doing barrier importance analyses, 100 realizations and redo a mean or expected value off of those 100 realizations. And you can see that if we sample some of those parameters at their close to extreme values, obviously not the 99th percentile, but at their 95th percentile, we generate early failures and when we generate early failures, we get early releases through those stress cracks at the welds. These are all driven by the weld and degradation at the weld and those are early and prior to 10,000 years. With that, I have a few slides to just summarize, but they more or less repeat what Abe had, first talking about the major technical improvements and the major process improvements and then, finally, summarize with where we are. As I've said, the results are done, but in checking and review, the document has not been delivered to the Department of Energy yet for their acceptance. They will, of course, do an acceptance review and we'll make whatever changes in the document, in the analyses, in the report as required, as they did with all the process model reports. So with that, I'll entertain any questions you might have. MR. GARRICK: Good. Thank you. Thank you very much. That's a very interesting presentation. Committee, any comments? Milt? MR. LEVENSON: I had one comment, or a couple, on the same issue. You discussed the degradation of the waste container and it was very clear and I think probably is what happens. You get a crack in a weld and et cetera. But unfortunately, in one of the pieces of paper we got a month or two ago, it said that the assumption was made that when the container lost the ability to be helium leak-tight, the assumption was made that 50 percent of the container disappeared. Now, that's quite in contrast to what you've said and I can't give you the specific reference, but I know the other committee members have read the same thing. So I was a little -- what you're telling us is what's really in the TSPA. MR. ANDREWS: I don't know this helium leak test issue. MR. LEVENSON: Well, the problem is when -- that they assumed when failure occurred, when it cracked through, the 50 percent of the container disappeared, so the fuel was immediately immersed. MR. ANDREWS: No. I do believe -- well, I don't know where that -- if it was a criticality -- that sounds like an extreme assumption for criticality analysis purposes, but I have no idea if that's the case. MR. LEVENSON: It was supposed to be release. The other question I have, which is somewhat related, is that even at 10,000 or 15,000 years or whatever time we want, the fuel is really the only heat source in the mountain and, therefore, if humid air gets through a crack, why would you expect condensation on what is the warmest thing in the mountain? MR. ANDREWS: That was one of those -- I think I alluded to it, maybe not as directly as your question is pointing out. The innards of the package and the hydrology, other than the temperature, the hydrology and chemistry of those innards of the package, that was simplified and conservatively so for this particular iteration of the PA. There have been analyses, supporting analyses of the amount of heat that the waste form, in particular, the commercial spent nuclear fuel waste forms. This is not so much an issue with the glass or the DOE spent nuclear fuel, which are much, much, much cooler. The heat output is much smaller in those packages. But for the commercial fuel, your point is well taken and there are some analyses that talk about more or less an eight to 15,000 year time period, the time period is a function of burn-up and age-out of reactor and storage and things like that. One would expect the system that you just described to occur, that the water, humid air, would never condense on the waste form itself during that kind of time period because of the heat that's being generated. After that time period, the temperatures are so evenly distributed across all aspects of the waste form, the package, the innards, the basket materials, et cetera, that taking any more credit beyond that time period would not be very feasible. MR. LEVENSON: Because there is no time period that you can go to when the fuel is colder than the mountain. MR. ANDREWS: That's true, that's true. MR. GARRICK: Ray? MR. WYMER: No, I've said my piece. MR. HORNBERGER: Bob, Abe was -- we got into that discussion about the qualified -- I think you called it qualified uncertainties, the issues that you were just discussing. How will you work that in after you do these analyses? Do you have any idea how you'll work that into the presentation? I mean, the idea would be, on some of the issues, the kind that Milt was just referring to. These are one-sided. So they're not symmetric. MR. ANDREWS: There's a large number of them that are not symmetric. Given that we've -- in fact, most of them are probably not symmetric. We made -- and not just PA doing the final crank turning, but we, the project, in these various areas where those assumptions were made, made that conscious decision to be on the conservative side of a distribution. Doing -- so that means we're on the -- whichever side is conservative, I don't know if that's right or left, but we're on the right side of the distribution and everything is to the left side of that distribution, in which case putting in the whole distribution, if it's distribution, or alternatives, A/B kind of alternatives, means we're pushing it to the left, pushing it to lower doses, if you will. And they'll probably be treated as sensitivity analyses off of what we say here is a conservative representation of how we think this system will evolve. So there will probably be, in terms of the documentation, and we haven't discussed this with the department and they might have different ideas, and so Abe probably should say something, but we probably would put it into this alternative model representation. The other alternative, of course, is -- maybe I shouldn't go here, but we have EPA and NRC have slightly different definitions that go after the word reasonable. One talks about reasonable expectation, one talks about reasonable assurance. It may very well be that some of these issues that we're talking about fall into the reasonable expectation kind of case, rather than a reasonable assurance case where you really want very high defensibility of your models and assumptions before you go forward with a license. And that kind of discussion is going on right now within the department, but whether it's a sensitivity analyses, whether you use it in this reasonable expectation, whether they are used for peak dose kind of considerations. You've been conservative during the time of regulatory concern and let's say appropriately conservative during the time period of regulatory concern, but after the time period of regulatory concern, when you're looking at geologic stability timeframes, perhaps you should put this reasonable expectation argument in. It's for the final environmental impact statement, which, of course, goes along with the license application, but it's serving a different purpose. Those peak doses are serving a different purpose for a different audience. Maybe that's an idea. I don't know. Abe, do you want to -- it's kind of a policy question. MR. VAN LUIK: It's a policy question, but I think your answer was correct. We were discussing exactly that issue. MR. GARRICK: Abe, can you go to the mic and repeat that? MR. VAN LUIK: Your discussion is essentially, in a nutshell, what we are trying to determine within the department should be our approach between the case that will actually evolve into the licensing case and also the larger case for the peak dose, which is to satisfy NEPA and satisfy 197's reasonable expectation. And they go out of their way to say that what we're interested in is what you expect, the expected value case, that they're not interested in details of our distributions for that particular purpose. So that's some of the internal discussions that we're having and we'll make a decision on how that goes forward before the time of SR. But what you will see in the SRCR is basically what Bob has shown you. MR. HORNBERGER: Second question, Bob. Can you summarize for me how information, let's say, derived from analog studies makes its way into the TSPA analysis or is that considered separate from the TSPA? MR. ANDREWS: There's a couple of ways. That's a good question. A lot of it, and most of the time, it's confirmatory type information to lend support to the process models that are the underpinning analysis model reports. Most of the time, it's that, to help defend, if you will, the quote-unquote, validity of those process models. So in that sense, it's not a direct parameter or direct data set that's fed into PA. There are examples where those analogs, in addition to providing support, actually are used to support some of the data. It's non-site data, not DOE data, but literature, available information that's relevant and applicable to be applied. So it's incorporated within the process model itself as a -- if you will, down at the process model, a direct feed. By the time we see it in PA, by the time it's gone through an abstraction and incorporation into the TSPA, you probably have lost a little bit of that direct one-to-one traceability from that data set, that analog data set into the process model and through the process model into the TSPA, but it's there. So it's confirmatory for most of the cases, kind of add confidence, confidence-builder. MR. GARRICK: The committee has a lot of questions that they would like to explore one day on the TSPA and I think we're going to have to do that at a later time and when we get down to maybe a lower level of the analysis. I think this has been an excellent overview. There's a great deal of interest here in how some of these algorithms are actually applied and how some of the analysis is done, and we'll look forward to covering those in other meetings as we get more involved in the PMRs and the AMRs and the input documents into the TSPA. So we'll look -- and I have a lot of questions, but we'll look to another time to do that. I've had requests from two people. Andy, you have a question? Go ahead, I'm sorry. MR. CAMPBELL: Just a quick one, Bob. Do you feel, does the program feel that in the transition from TSPA-VA to TSPA-SR, the modeling and the approaches and what gets incorporated into the models in terms of parameters has become more conservative or do you think it's become a little more realistic? MR. ANDREWS: I think, in toto, it's become more realistic. There's the few examples where a conservative -- if a conservative assumption was made in the SR, it was probably also being made in the VA. MR. CAMPBELL: The reason I ask is it looks like the dose curves have increased. At any one time, you get higher doses in the TSPA-SR by, in some cases, almost a factor of ten, if you look at the mean or the 95th percentile. If you're becoming less conservative, shouldn't the doses be going down? MR. ANDREWS: No, more realistic. MR. CAMPBELL: If you're becoming more realistic. MR. ANDREWS: I think the one aspect of the system that's driving -- or a significant difference that drives that is the solubility distribution, in particular, for neptunium, because neptunium is what's driving the 100,000 year and, in fact, the peak dose. In the VA, we tried to broaden that uncertainty to incorporate both the waste form degradation characteristics and the data from Argonne and PNL and Livermore, but mostly from Argonne, as well as the direct observations of solubility from Los Alamos and Berkeley. So we had all the labs doing varying parts of neptunium solubility work, which made it a fairly broad distribution, had some lower values, had a few higher, but mostly, on toto, was lower than what we have in the SR. There were comments on the VA saying make the solubility more chemistry dependent and we made the solubility more chemistry dependent. And in those analyses and model reports that relate to the chemistry dependence of neptunium solubility, the currently available data and fundamental thermodynamic data pointed to a very strong pH. dependency which is incorporated in the model, which meant, by doing so, the solubilities ultimately were increased from what we had in the VA. So I think it's reasonable, defensible, but higher value. MR. LEVENSON: As a follow-up question, I understand the technetium solubility is from real data. What about technetium retention as it moves through the ground? There is an assumption that is not related to, I think, experimental data. MR. ANDREWS: Now, there are technetium, neptunium and technetium, there are retardation data for both of them from Los Alamos, with site-specific kinds of materials. The technetium retention in the alluvium is being considered in the TSPA-SR, as is neptunium retention through the whole system is being considered in the TSPA-SR, based on best available data on those retention characteristics. So if they're retarded, they're with data that support them, they are in there. MR. GARRICK: Okay. Speaking of intrusion, we're getting into our lunch period. But as I was about to say, we've had two people ask for an opportunity to make a few comments and I would like to get those in and I would like to limit the comments to three to five minutes. Is Mr. Harney here? He's not here. All right. Then Judy. Judy Treichel, of the Nevada Nuclear Waste Task Force, would like to make a comment. MS. TREICHEL: Judy Treichel, Nevada Nuclear Waste Task Force. You said that you were coming here so that you could hear from us and I want to find out how that works, because you are hearing. I've got questions on the pyramid that I guess you have in color and the rest of us got in black and white, that shows the -- where we're at and where we're going and all of that sort of thing. I want to know, because the SRCR is supposed to be a featured event in our lives very soon, how big is that? Are we talking telephone book, something like this, are we talking site characterization plan? How tall is that thing expected to be? It's two volumes. MR. SCOTT: I believe it's about 1,300 to 1,400 pages. MR. GARRICK: You're going to have to use a microphone and announce who you are. MR. SCOTT: Mike Scott, M&O. I believe Tim Sullivan, with DOE, said this morning it would be about 1,400 pages. MS. TREICHEL: Okay. So probably about like the draft EIS. Then am I correct that when the site characterization -- when the site recommendation report itself comes out, you just slap two additional volumes on there? Those first two go as they are and you put two more, or are you going to rewrite the first two? Here's one and two, and then three and four come in. Does all of this in this hunk stay the same or are you making changes before you go to the Secretary and the President, after we've had our chance to comment? MR. VAN LUIK: The idea is that we would actually listen to the comments, consider them, and make changes in the documents before they go to the Secretary. MS. TREICHEL: So it's up to us to get this site recommendation all ready to go, because you were asked if it was going to be reviewed and it appeared that was kind of iffy and you weren't too sure about that. MR. VAN LUIK: Ninety days. MS. TREICHEL: For us. MR. VAN LUIK: Uh-huh. MS. TREICHEL: And that's interesting, too, because during those 90 days, which, of course, we're pretty used to, you have all of the major holidays and you still got -- and we -- and in addition, we have no EPA rule, we have no NRC rule, we have no DOE guideline. So I'm not sure why, if we were doing a bona fide job of reviewing this thing, we wouldn't review it against the existing rules, that we wouldn't just take 60, we wouldn't take 960 and see if it passes muster, because that's what's there. We won't have anything new. At the same time, while we're doing this exercise, we're going to have eight technical exchanges that are going on that are trying to bring about sort of resolution or clarification or whatever you want to call it, changing the status of your KTIs and the questions that people have. So those are going on and we do those all at the same time. In addition, the other myriad of meetings that happened in which everything changes. There was information you got today from DOE that's a little bit different than what we've seen before and the design is always different in one degree or another. It was a little different today because of a difference in the stainless steel part of the package. And we know that that's not going to get frozen. So I guess what I'm trying to figure out is whether or not I'm spinning my wheels. The public doesn't get paid. So we have to kind of save on what we spend, and that includes our energy. And I just don't want to wind up out there with something that means nothing, and maybe there's other more important things going on. So thank you for listening. MR. GARRICK: Thank you. Thank you very much. All right. Unless there's comments from anybody, I think we will adjourn for lunch. Try to be back here as close to 1:00 as possible. [Whereupon, at 12:05 p.m., the meeting was recessed, to reconvene this same day at 1:00 p.m.]. AFTERNOON SESSION [1:05 p.m.] MR. GARRICK: We'd like our meeting to come to order. For the benefit of those on the panel or on the committee that are chemically inclined, this is their time to shine. We're going to, this afternoon, spend a good deal of time on chemical related issues and to actually lead the discussion and questioning on this will be Dr. Wymer. Our first speaker is going to be Dr. William Boyle, and I think, unless there's anybody that wants to make any preliminary comments, we'll get right into it. MR. BOYLE: Good afternoon and thank you for this opportunity. You can see the names on the sheet here. I'll be very brief and just mainly provide an introduction as to why DOE wanted to make these other measurements of chlorine-36. I will be followed by Dr. Mark Peters, of Los Alamos, who will provide the technical details, and the actual work is being done by the principal investigators who are listed at the bottom, most of whom are here, Mark Caffee from Livermore is here, June Fabryka-Martin from Los Alamos is here, Mel Gascogne from AECL is not, and Zell Peterman of the USGS is here, and Robert Roback of Los Alamos is here. So if there's questions on the details, the investigators are here. For those of you who were present in Pahrump in May for the NWTRB meeting, I've got the same sheets, so it will be repetitive for you. I assume most of the audience knows why the project has measured chlorine-36, but just in case, I will give a non-expert synopsis. Chlorine-36 is one of many naturally occurring radioisotopes used for age dating. Its abundance was changed my nuclear weapons testing in the South Pacific in the 1950s, creating a bomb pulse. Measurements of chlorine-36 at Yucca Mountain have been interpreted to have this bomb pulse. These bomb pulse data, at depth, are then used as evidence that there are vast flow paths in the unsaturated zone at Yucca Mountain. That's the synopsis, and now I'll go on to why we did the validation measurements. The project's original measurements for chlorine-36 were done by Los Alamos National Lab. As you can see, there are other organizations involved now in the validation measurements. And why were these validation measurements made? Well, about two years ago or even a little bit longer, a series of reports were written by the United States Geological Survey that seemed to describe a comprehensive history over geologic time about the unsaturated zone at Yucca Mountain. This history was based upon the integration of many independent data sets. Not surprisingly, not every data set that was used to develop the integrated history flanged up perfectly. One of the data sets that did not flange up as well as some of the others is the chlorine-36 results from Los Alamos National Lab. In discussions about why there might be this difference between the chlorine-36 data set and the USGS history for Yucca Mountain, it was decided initially to follow a standard scientific practice and have an independent group make measurements of the chlorine-36, and those independent measurements are those by Mark Caffee and the others listed on the sheet, not from Los Alamos. So we went ahead and made the measurements and now you will get to hear the results of those. At this point, I will turn it over to Mark Peters. MR. PETERS: Thanks for having me. I do work for Los Alamos, but I'm really up here as a representative for the team that Bill mentioned is out there in the audience. So I'll go through a lot of the technical details, but I'm hoping to get through it, leave a lot of time for questions, and then the PIs in the audience I've already told to feel free to step up and help me answer some of the details, clarify, however you all want to work that. That's up to you. Bill already mentioned the participants. They're all sitting in the audience today, which is real nice. What I want to do today is walk you through, first, a lot of you are familiar with the chlorine-36 studies that we've done over the past three to four years. I want to bring you back up to speed on what that data set looks like, then go into some of the results from the validation study, where we're doing analyses of samples from some of the locations in the ESF, where we thought we saw bomb pulse, and then bring in some cross drift results and then wrap up with the path forward, because as I'm going to show you, there are some differences in the analyses for chlorine-36 chloride ratios for some of the samples in the ESF set of samples from the Sundance fault and we have a path forward that we're following to try to address those differences. As I'm going through, I want to clarify, so we don't get lost in semantics. I'm going to be talking a lot about sampling, processing, and analysis. When I mean sampling, I mean physically taking the sample from the rock and then processing, I mean what we have to go through in the laboratory to process the sample, fairly labor intensive, and then the analysis is done by accelerator; in the case of Livermore, at Livermore; in the case of Los Alamos, it's sent to Purdue University. So, first -- and I should also say that there's a lot of organization-specific references. You'll see a lot of Los Alamos and Livermore. We are working as a team, but it helps me distinguish between the data sets to point that out. So this is an integrated study that we're carrying forward with. In terms of the overall objectives of the chlorine-36 chloride program over the past three to four years, the objectives are shown here. We were looking to test alternative conceptual models for unsaturated zone flow and transport. Specifically, to look at flow and transport through the Paintbrush tuff, the Paintbrush, non-welded, which sits stratographically above the repository horizon. Then also to look at the significance of differing temporal and spatial scales. And what I mean by that is the effects of episodic infiltration and how well the PTN might dampen that flow into the repository horizon. The program focused on systematic samples. Parts of the ESF, we took samples every 100 meters, and then in some cases, every 200 meters. We also looked at features, meaning fault zone fracture sets and took samples, and there we were taking block samples, large samples, with a pick hammer or jackhammer, taking those back to the laboratory. Bill mentioned the validation study. In the validation study, we were focused -- and I'll show you the data from the ESF in a minute, but one of the locations in the ESF that we saw apparent bomb pulse was in the Sundance fault zone. I'll show that on a map, but it sits down in the main run of the ESF, down by Alcove 6, for those of you who are familiar with the ESF. We took -- and also at the drill hole wash fault, but I'll talk mainly about the data from the Sundance today. Those samples were taken by drill. We did bore holes. So there was a difference in sampling approach for the validation study than in the studies that we've been doing over the program in the past couple of years. Again, it was led by the USGS, with involvement of Los Alamos, Livermore and AECL. We were also looking at some of the samples, in addition to looking at chloride-36, chloride measurements. WE also have some tritium analyses that I will talk about briefly. But what we did is we did systematic bore holes across those two fault zones, two-inch cores, up to four meters, and, again, to emphasize, it was in contrast to the sampling approaches that we took in the ESF and in the cross drift. So we took samples as depth slices and splits were taken. So in some cases, we have Livermore and Los Alamos looking at some were core and some were hole. I mentioned I want to talk about the cross drift. I call this validation, maybe I should put it in quotes, but nonetheless, in the cross drift, we also did feature-based sampling, but here we were able to do predictions using the UZ flow and transport model for what we thought we would see in terms of chlorine-36 chloride systematics and we compared those predictions. So background for chlorine-36, Bill alluded to it to some extent. Chlorine, this is basically lifted out of the table already, nuclides, chlorine-35, chlorine-37, both stable isotopes, make up the full abundance. Chlorine-36 has a half-life of 301,000 years. In terms of sources in the subsurface, you've really got three primary sources. You have the pre-bomb ratio, the modern ratio is about 500-ten-to-the-minus-15. That has varied through time due to variation in the field strength, the magnetic field strength on the earth, you get more or less cosmic ray bombardment, therefore, more or less chlorine-36 production in the atmosphere that rains down. B bomb pulse has a much higher ratio. Then C, there's a contribution from production of it due to reactions, nuclear reactions with uranium and thorium in the subsurface. That ratio is much lower, 20 to 50-times-ten-to-the-minus-15, and that tends to be a negligible contribution. All in all, those three sources, there are some other minor sources, but they add up to what you see in the subsurface rocks and water and what we then measure. Now, to go back to what we've seen in the ESF, the exploratory studies facility. This plot is on the Y. Chlorine-36, the chloride ratio, times-ten-to-the-minus-15, has a function of construction stations, so that's thousands of meters through the ESF. The north ramp, the main drift and the south ramp are shown along the top. There are several different kinds of samples shown here. This is the Los Alamos results. So you have, in the black squares, you have the systematic samples. The open squares are the feature-based samples, then there's also core water samples, where we've done -- we've extracted water using centrifuge and done analyses of that water. Then there is also plotted on here, in the red are the Los Alamos validation samples, meaning the splits of the core from the validation work at the Sundance. But you can see, in general, where we see apparent bomb pulse ratios, which were above 1,200-ten-to-the-minus-15. It tends to be associated with structural features. So we chose the Sundance and the Drill Hole Wash to do our systematic validation study across there. So just to come back to the validation work. The goal here was to verify the presence of bomb pulse in samples taken from the ESF and the sample preparation method that was developed, that Livermore is using, is designed to detect the presence of bomb pulse. So we weren't looking to delineate the relative contributions from those three sources, but we were mainly looking for any evidence of bomb pulse chlorine-36 in ESF. This is the data for the Livermore results. Again, we did a series of bore holes across the Sundance fault. This is chlorine-36 to chloride times-ten-to-the-minus-15 on the Y; again, plotted against ESF station. We did a series -- we did on the order of over 40 bore holes across the structure and you can see the data there. So the present day ratio is shown about 500-ten-to-the-minus-15, ratios greater than 1,200, indicative of bomb pulse. You can see the Livermore analyses are all relatively low, in the 50 to 200-times-ten-to-the-minus-15 range. No evidence of bomb pulse. Now, for the Los Alamos results from the validation study. This is, again, a similar plot, same ratio plotted on the Y against ESF station, with the Sundance fault drawn in the -- the Sundance is right here. And here is the sampling interval. WE sampled across the fault with the bore holes. What's plotted here in the blue diamonds is data from the -- what I'll call the standard ESF study and then in the red squares are the validation results from Los Alamos. We haven't seen evidence -- we haven't seen ratios greater than 1,200, but notice there is a difference. All the samples are greater than 500-ten-to-the-minus-15, so you've got a significant difference in ratio there between what Livermore was getting for similar samples. You have to ask yourself, okay, what -- well, that's really the crux of why we're here. We've got two data sets now that are showing some differences. How robust are the Livermore data? When you correct for blanks, you don't really affect the final ratio. In general, when you correct the ratios for things, you tend to lower rather than raise the ratio. However, and I will talk more about this, if you correct for rock chloride, which might be related to sample processing, processing in the laboratory, it may be possible that you could cause the correction to go in the other direction, but that's what we're about in our path forward trying to figure out what's really going on, particularly in the processing techniques. That may be part of the differences. But I should also mention the Livermore -- Livermore runs samples from all over the world for a lot of different reasons and when the samples that were run at the same timeframe yielded results that were consistent with what they expected for the geologic setting. So we have no reason to believe that the Livermore data is wrong or the Los Alamos data is wrong. We need to look into the details of how we process the samples. MR. GARRICK: Were the sampling procedures exactly the same? MR. PETERS: Do you mean sampling in the field or processing in the laboratory? MR. GARRICK: I mean in the field. MR. PETERS: Yes. It was taken from the same bore hole. So they were dry drill bore holes and then the core was simply split for the validation study. MR. GARRICK: And were they governed by any kind of a national or international standard for taking such samples? MR. PETERS: They were collected, according to our QA program, like we sampled all the samples underground. So it's proceduralized. MR. GARRICK: They both used the same QA program. MR. PETERS: The collection in the field was the same program, because it was the same driller, same drill rig, same sample handlers. So where we're headed here is how one -- what's going on downstream of that. MR. GARRICK: And those same questions we'll want to talk about downstream. MR. PETERS: Right. And the details of that are better spoken to by the PIs. I'm not going to say a whole lot more than what I've already said, actually. But we do have a path forward that focuses on that part of the process to see if that's -- that therein lies the difference. I mentioned at the beginning that we're looking -- there's some related work associated with the validation study. For example, we're also looking at doing tritium analysis in some of the same samples, and here we're extracting the water and doing tritium analyses. This is a slightly out-of-date slide. The USGS has done additional analyses, but suffice it to say it doesn't change the distribution. So right now we've seen really only one sample that's even above detection limit for tritium. So for the validation samples, no events of bomb pulse tritium. But tritium and chlorine-36 are going to act very differently in the unsaturated zone hydrologic system. MR. HORNBERGER: Is the processing for tritium similar to the processing for chlorine-36? MR. PETERS: The tritium was all through centrifuge. You extracted the water with the centrifuge -- distillation, excuse me. So vacuum distillation. So basically putting -- moving it around in a cold trap in a vacuum line, basically, whereas chlorine-36, as you know, is basically running DI through some variation on that theme. AECL's participation is mainly focused on the U-series disequilibria work that they're doing. This doesn't -- it speaks a lot to the -- basically, the residence time of core water in the unsaturated zone. It doesn't speak as much to bomb pulse, but it combines real nice in with geochemical indicators of long-term percolation flux, et cetera. But we are seeing that thorium, uranium and radium are in secular equilibrium. Therefore, they haven't been moving around over the past 100,000 years. The 234-238 uranium ratios are depleted, show a five percent depletion in uranium-234, which suggests that uranium-234 may be lost to pore fluids and that's probably by alpha recoil. And therefore, we would expect the pore fluids to be enriched in 234, and they, in fact, are. So that can actually be modeled to give us an idea of residence time of the core water in the unsaturated zone. Now, getting back to the differences. We haven't finished. As Bill mentioned, it's a work in progress. It is a work in progress. It's going to continue next year. Livermore may yet see bomb pulse in some of the samples. Mark made this point in May, and it's a good one. Work to date hasn't demonstrated its absence. We just simply haven't found it. It could be that we may find it. The Livermore sample processing may have selected phases that don't contain bomb pulse chlorine-36 or maybe the samples may not, in fact, have bomb pulse chlorine-36 in them. But at any rate, the chlorine-36 concentrations seem to be comparable, but the difference in ratios may be due to, and this is a may, this is what we're going after, elevated concentrations, chlorine concentrations, because obviously we're reporting chlorine-36 to chloride ratios, so you can -- there's a lot of leverage with chloride concentration there to change that ratio pretty dramatically. So the differences, again, may be related to processing method. We're focusing on that, not sampling or analytical methods. So sample crushing, how you extract the chloride for analysis, and those kinds of things. I've mentioned, we've talked mostly about the ESF, but this data, I don't think you've probably seen this data in a meeting. This is data from the cross drift; again, same ratio along the Y for the cross drift station in 100 meter increments, and the same familiar dotted line bounds the range of current day values for chlorine-36 to chloride ratios. There's the faults noted in the cross drift, as mapped in the cross drift, and you can see there, again, and these are back to taking feature-based block samples through the cross drift, so back to the standard ESF program, and it's consistent with what we've seen throughout the ESF, that Los Alamos has seen throughout the ESF. So really what we're focusing in on is why are we getting the differences in the validation samples. So I've already said this, but we did do a set of predictions for the cross drift and, in general, they were consistent with the predictions for where we would see background levels and where we would see bomb pulse. And I said the second bullet already, but the sampling protocol that we use in the cross drift seems to -- it does yield similar results to what see in the ESF investigations, except for the validation study, as we've talked about a lot. So what about a path forward? There was a lot of discussion amongst the PIs in the April-May timeframe, particularly after the interactions with the TRB about how to go forward and address these differences. So the path forward is the USGS has prepared a reference sample. We've taken some muck from the cross drift excavations that we were doing back in that timeframe. That's been crushed, homogenized, and then aliquots have been distributed to both Livermore and Los Alamos. There's going to be a set of experiments done by both laboratories to test for the effects of different leaching procedures on the release of rock chloride, again, the chloride could be the key. That's ongoing. So this is really, again, a progress report. We will then do the laboratory work, do the accelerator analyses and see where we're at, compare notes. Once we've sort of agreed on a standard processing method, we hope there will be exchange of samples, exchange of information, apply that to the reference sample and the validation samples, and then the results will be synthesized. Now, I haven't talked too much about the conceptual model, but the ESF and cross drift data, except for the validation data, has been incorporated into our thinking on conceptual models for unsaturated zone flow and transport. So depending upon which way things go with the validation samples, that could have implications for our understanding of the conceptual model. The three bullets are meant to address the aspects of the conceptual model that will be under discussion, depending upon the results of that study, frequency of fast flow paths, roles of fault and fractures, particularly ones that cut through the PTN, the non-welded unit above the repository horizon, and then, finally, pore water ages and implications for infiltration and percolation. The chloride data is also very key. That's used heavily as a calibration tool for the flow field in the unsaturated zone. So this is all tied together. It's a little premature for us to really say much about what it means for conceptual model until we get at what the differences are in the data sets. So maybe this time next year we can tell you, give you a final answer on that. That's all I have. MR. WYMER: Milt, questions? Okay. I'll start out, then. Obviously, it is extremely important to understand these data, because they do bear on one of the most important parts of the whole repository; namely, how fast is the water moving and by what mechanisms, and you've known about the disparity quite a while. What is it about the analyses that takes so long to cross-check and confirm? It seems to me that could have been done by now. MR. PETERS: Well, the differences really were just exposed in the spring. We just really started to understand that there were some significant differences in the spring timeframe. MR. WYMER: Three months. MR. PETERS: I understand. MR. WYMER: Four months. MR. PETERS: There's been some constraints. There hasn't been as much work done at the laboratories because of other competing priorities associated with AMRs and stuff. I mean, it's project priorities, to some extent, that have caused not as much laboratory work to be done in this area. That may not be what you want to hear, but that's the reality. MR. WYMER: You're right. Okay. Well, that's my question. Milt? MR. LEVENSON: Yes. I've got a couple of questions. The numbers that you have for the ratios, one for the pre-bomb and the other for the bomb pulse, are those average for the entire atmosphere, the 500 and the 1,200? MR. PETERS: Let me be clear. The actual ratio -- that wasn't clear on the slide. The actual ratio for bomb pulse water is much higher. That 1,200-ten-to-the-minus-15 is what we see when we take a sample of rock in the ESF. We think anything above that is bomb pulse. But the ratio for the actual bomb pulse chloride is two orders of magnitude greater. MR. LEVENSON: So the atmosphere concentration, if you sample the air anyplace, it's going to be two orders of magnitude higher than that. MR. PETERS: For the bomb -- June, help me out here maybe. MS. FABRYKA-MARTIN: The ratio of 500 for background is based on samples from the Yucca Mountain area, actually the Yucca Mountain region. It's based on three types of samples; soil profiles, for one, where once we get below the bomb pulse in the soil profile, then we use those -- the numbers that are clearly below where the pulse is as part of the database. The second source of samples are packrat samples from packrat urine, that's fossilized. We carbon date the sticks that are stuck in the midden and then use the chlorine-36, the ones that clearly aren't recent, to show us what the background has been. And then the third type of sample is ground water itself that doesn't have any tritium or C-14 indicating absence of bomb pulse there, as well. All three sources have been consistent in defining that background. Now, the 1,200 is a statistically derived threshold. So we're saying that once it's above 1,200, then we're confident that it's unambiguous evidence that it's clearly elevated above natural background, and that's based on a database of about 300 results from the tunnel and it's also supported by our packrat midden samples where we never saw any ratios as we went back to as far as 35,000 years. None of those carbon dated samples were ever higher than about a thousand or 1,100 or so times-to-the-minus-15. MR. LEVENSON: But I still don't have an answer to my question as to what the source term is. What is the ratio in the atmosphere from bomb debris? MS. FABRYKA-MARTIN: Why do you care? I mean, because -- MR. LEVENSON: Why do I care? MS. FABRYKA-MARTIN: Because once it hits -- as soon as it hits -- MR. LEVENSON: Because how do I know how credible your 1,200 is if I don't know the source term? MS. FABRYKA-MARTIN: The thought I had -- well, first of all, it would be hard to collect a sample from the atmosphere that did not have bomb pulse in it. MR. LEVENSON: No, no. I don't want the background from the atmosphere. I want the present concentration of bomb pulse in the world, the source term for this. MS. FABRYKA-MARTIN: Okay. We do have samples of runoff water, would you accept that? Runoff catched from channels from USGS investigators? MR. LEVENSON: If that's the best you can do. Is that what you consider representative of the atmosphere? MS. FABRYKA-MARTIN: No. No. MR. LEVENSON: The source of the bomb -- MS. FABRYKA-MARTIN: The reason I say that is the moment that a drop of water reaches the surface, it's already being diluted by what's already there, what's accumulated there over the thousand -- or hundreds of years or tens of years. So I'm not sure what -- MR. CAFFEE: Not to disagree with anything June said, but one of the -- MR. GARRICK: Identify yourself, please. MR. CAFFEE: My name is Mark Caffee, from Lawrence Livermore National Lab. We can recreate what the bomb pulse input is from looking at ice core data. So there's been extensive studies of chlorine-36 deposition at the Vostock ice core and in Greenland ice core samples. So we can go -- we have gone back, these studies have been done by a number of groups, and reconstruct that profile of the bomb pulse chlorine-36 in the atmosphere. I'm positive that I didn't bring a viewgraph of that with me, but the chlorine-36 to chloride ratios are very, very high and what you see is kind of what you expect, that a few years after the atmospheric testing of nuclear weapons, you had a peak in the chlorine-36 and then it has fallen off since. I don't know if that answers your question, but that's kind of the -- MR. LEVENSON: What's the current day, current day ratio? MR. CAFFEE: If you just went out and got precipitation, it would be back to what it was before the bomb pulse, before the bombs were exploded. MR. GARRICK: If you want to comment, you have to do it in a mic. MR. CAFFEE: What June was going to say is someone has looked at precipitation in the environ of Purdue University and for what we believe the pre-bomb chlorine-36 to chloride ratio was, it has gotten back to that again. But it's not the case that there have been thousands of measurements of chlorine-36 to chloride ratio in rain waters all across the country or all across Africa or anything. These contours of input are drawn based on not a great deal of data points. MR. LEVENSON: What's the precision and accuracy to one significant figure of these analytical methods for this ratio? MR. CAFFEE: Okay. I'll answer for Livermore. I won't try to answer for anybody else. The measurement itself is much better than five percent, based on counting statistics. By the time we fold in blank corrections, and I can address the Livermore data here, the uncertainties may go a little above five percent, but they're probably still less than ten percent. MR. LEVENSON: Five percent is the precision or the accuracy? MR. CAFFEE: Precision. MR. LEVENSON: What's the accuracy? MR. CAFFEE: I don't know how to answer that question. MR. LEVENSON: That's the most important question here. MR. CAFFEE: Of course it is, but what we're trying to do, I can tell you that other samples, not Yucca Mountain samples, but other samples, for example, in situ chlorine-36 produced in carbonates in Greece and Italy and all across the world have reproducibilities that are comparable to our precision. So for samples where we have a long history of running things and our researchers have sent us duplicates and triplicates, we believe that our accuracy is comparable to the precision. But for me to get up here and tell you that our numbers at Livermore are absolutely accurate and there can be no other circumstances that can change those ratios, I can't answer that -- I can't do that. MR. LEVENSON: Let me ask one other question, which is a completely different type, because I was at the meeting in Pahrump four months ago when this first came up and there was discussion then that the simplest thing to do was to take some aliquots from a processed sample and run it both places and rather quickly eliminate whether the analytical method was relevant or not, and it doesn't sound like that has been done yet. Is that correct? MR. CAFFEE: No, we have not done that. I don't recall -- that was quite a while ago. I don't recall it being mentioned. I don't doubt that it was mentioned. I would, again, in answer to that, say that over a course of six years of running, we have run duplicates of aliquots many, many times, from meteorite samples, lunar samples, rock samples from all over the world, and secondary standards, NIST standards, a whole variety of things and we have comparisons with samples run between ourselves and Zurich, between ourselves and the old lab at Rochester, and I even believe between ourselves and Purdue. MR. LEVENSON: But that's not really the issue. The issue is to try to resolve why the difference and it potentially -- MR. CAFFEE: I'm not disagreeing. MR. LEVENSON: You've attempted to get rid of the sampling difference by splitting a sample. I think that's an acceptable method of doing it, and you could fairly quickly resolve whether it's processing or analysis by the two labs running the same sample. I just don't understand why -- MR. CAFFEE: This is one of the things that will certainly be done with this standard reference material, because here we have enough sample that we can precipitate enough silver chloride that we can make splits. The chemistry that's done on the core samples is difficult enough that after we have precipitated silver chloride, we generally don't have enough to make splits. We make enough silver chloride for one, perhaps two analyses. So we don't have in our laboratory a supply of silver chloride from the ESF samples. To do that would require processing much more sample than we've got. And let me make one further point. When we got into this, what I assumed was that we would measure these samples and we would see bomb pulse chlorine-36 in all of those samples. That was my assumption. Now, if we had made those measurements and observed those elevated ratios and everything, we wouldn't be here. No one would be asking questions about precision, no one would be asking questions about accuracy, no one would be asking questions about did you shake your samples a certain way and do you shake your samples a different way. We have seen a big difference, though, and I think it's important, but it's going to take time to go through and resolve those. So the initial work was not done anticipating this kind of problem. Had it been done anticipating this kind of problem, we would have had to spend a whole lot more time to do the work. MR. LEVENSON: We're not talking about prior to May, just the concern. I must say -- let me record a personal comment, that whoever sets the priorities so that the paperwork has higher priorities than resolving a problem as important as this really ought to reassess what they're doing. MR. HORNBERGER: I have just a quick follow-up. You were rudely interrupted when you were talking about your lab comparisons and I just wanted to know -- MR. CAFFEE: Did I interrupt myself? MR. HORNBERGER: That was it. I just wanted to know if Livermore -- have you done inter-lab comparisons with Purdue? MR. CAFFEE: You know, not in recent years, but I think we have. MS. FABRYKA-MARTIN: One of the first things I did when I became PI on this project back in 1990 was to do an inter-lab comparison with results from Rochester, which was still running then, Livermore, and Purdue. And I think I sent out a total of four blind samples that all four labs had done and they did pretty well. I don't remember the exact numbers, but they were within two sigma, for sure, of one another and I was satisfied. MR. WYMER: How much variation do you find in total chloride content concentration from various samples that you've taken? MR. CAFFEE: For in general or for the validation sample specific? MR. WYMER: I want to know whether there's a factor of two, three or five of chlorine concentration from one spot to another in the repository. MR. PETERS: How much does the chloride concentration vary, say, June, in your samples, from north to south ramp? MR. WYMER: Is one saltier than the other? MS. FABRYKA-MARTIN: Okay. Do you want to rephrase that so you're talking -- whether or not you're talking about pore water or total amount of chloride leached from the sample or a function of -- MR. PETERS: I would answer both. MR. WYMER: Yes. MR. PETERS: Answer pore water and then -- MS. FABRYKA-MARTIN: Pore water, it's about -- in the pore water, in the south ramp, it's about 80 milligrams per liter, on the average. North ramp, we saw much lower values, I'd say averaging maybe 20 milligrams per liter, and, also, along the cross drift, about 20 milligrams per liter. As far as chloride leached from the rocks, a very general trend that I saw with a lot of scatter was that we generally saw bomb pulse -- the samples that we saw bomb pulse in were the samples that had the lowest quantities of chloride leached from the rock, whereas the ones that had the lowest ratios tend to have higher chloride, and that would be in terms of we did a one-to-one leach, one kilogram of rock to one kilogram of water and saw concentrations anywhere from maybe 0.3 milligrams per liter, which means 0.3 milligrams leached per kilogram of rock, on up towards maybe an order of magnitude or more higher. Does that answer the question? MR. WYMER: Sort of. In your analyses, there was no normalization with respect to the amount of total chloride, or was there? MS. FABRYKA-MARTIN: No. We measure the ratio on what gets leached out and it's an important point to realize that we're not trying to maximize amount of chloride leached from the rock. In fact, it's just the opposite. We're trying to minimize how much chloride we leach from the rock, because we don't -- MR. WYMER: I understand that. It just seemed to me that if there was four times as much chloride to start with and you had the same amount of chlorine-36, you might change your ratio. MR. PETERS: I think that's what I tried to convey. June, the concentration in the validation samples for chloride, what was yours, and then, Mark, I guess, what was yours? MS. FABRYKA-MARTIN: It was about 0.3 milligrams per liter that was leached from the validation samples. MR. PETERS: I don't mean to put you on the spot. MR. CAFFEE: Sure. But my recollection, without having my raw data here, is that it's .8 to 1.5, maybe even two for a couple -- ppm, so one ppm. We're definitely a little higher than June's. MR. GARRICK: I was just curious. I'm not a chemist, so I can't ask good questions like you've been getting. But this technique of measuring for bomb pulse, that chlorine ratio tracer technique or actual measurement technique has been around for a long time, has it not? And there's bore holes all over the Nevada test site, are there not? So there must be a lot of experience in doing this. MS. FABRYKA-MARTIN: No. What's new here, what's really unique is the fact that we're leaching -- we're trying to measure chlorine-36, where there essentially is not much water, so you're leaching rock. And here, unlike those other types of sample types, you really have to watch out for diluting your sample with salt that's trapped in the fluid inclusions or grain boundaries or wherever it is. That's what is unique. I don't know of anyone else who is doing that. They do it in soils, tons of studies using soils, looking at in situ chloride in different mineral species, ground waters, but unsaturated rock, I can't think of anybody else. MR. GARRICK: So you think that the unsaturated rock and the uniqueness associated with that makes this a pretty much one of a kind sampling operation. What I'm getting at is that with all the experience that you've had at the Nevada test site and the bore holes, you must have -- have you -- you must have had either agreements or disagreements in the past and the necessity for confirmation or reconfirmation. I'm just trying to get at if you've had similar experiences in the past, why can't we attack them the same way. MR. CAFFEE: I'll answer that different, actually. First of all, you mentioned that the technique has been around for a long time, but while it's been around for a while, it's not the case that this is a routine technique, like Argonne-Argonne dating, for example. If you think about how long Argonne-Argonne dating has been around, in the first couple decades, it was probably a pretty painful start, too. So that's the first point I'd like to make. It's still a difficult measurement. We're measuring a million atoms of something. So that's not easy. The second point I guess I would like to make is that I think, to the best of my knowledge, the only person that has leached this kind of rock to get chlorine-36 out for AMS analysis is June. It's not the case that the literature is just totally, totally filled with this kind of thing. There's a lot of data on in situ produced or cosmic ray produced chlorine-36. There's a lot of data on chlorine-36 in ground water. Those are all laboratory techniques that we all know how to do and we can all do that. But you won't go open Geochemic and find a whole lot of articles on this kind of application. MR. GARRICK: I see. MR. CAFFEE: So I find this disparity in results not that surprising, to be honest with you. I mean, I think it's just the normal process of -- MR. GARRICK: Is this the kind of thing where a sample could be sent to an independent lab outside the weapons complex, for example, and -- MS. FABRYKA-MARTIN: I have sent samples to Purdue to process in the past, with mixed results, because I didn't define the protocol very well. Some samples came in right where I expected them. Others, they crushed them too finely, so they came down a little, not as low as Mark's, but down lower than expected. But I would like to point out that I'm not the only one who applied this technique. What would be a fair statement is to say Yucca Mountain project PIs have been the only ones to apply this technique. But I wasn't the first PI on this and, furthermore, my technicians don't allow me in the lab really, with good reason probably, and there's probably, over the last -- this project has been going on -- the chlorine-36 project has probably been going on more than 15 years or about 15, let's say, maybe slightly more and it's had probably five different technicians on it over the years that have processed samples and the samples have always been consistent from one technician age to another, and that's under two different PIs. At first, they used to be done at a lab in Tucson, Hydro Geochem, and then it got moved to where I am now. MR. GARRICK: That was the basis of my original question, the process that is employed in the lab to make the measurements, governed by the same more or less prescriptive process. MS. FABRYKA-MARTIN: Yes. That part is true. MR. GARRICK: And was that true for Purdue? MS. FABRYKA-MARTIN: I had them write the procedure that they used and send it to me along with the results when I really should have gotten in on the very beginning and looked over what they had proposed to do. I didn't bring those data with me, but it's still up in the hundreds, the ratio was up in the hundreds, not down in the 100 or less. MR. WYMER: Are you comfortable with the fact that fines might give you a different ratio than stuff not so finely crushed? MS. FABRYKA-MARTIN: Yes, I am. MR. WYMER: Tell me why. MS. FABRYKA-MARTIN: Because the finer it gets, the more likely it is you've broken along grain boundaries or opened food inclusions or made it -- MR. WYMER: That changes the ratio? MS. FABRYKA-MARTIN: The chlorine-36 to chloride ratio simply because then it releases that in situ produced chlorine-36 that's made by neutron capture on chloride and include an inclusion source, grain boundary, wherever it resides, as opposed to what's flowing along the micro fractures between the grains. MR. WYMER: I understand. MR. PETERS: Let me follow-up with a question here. Have you guys plotted the chlorine-36 to chloride ratio against -- on the Y axis against on the X axis to total chloride? MS. FABRYKA-MARTIN: We did for the Los Alamos results. MR. PETERS: Do you get just basically a straight line? MR. CAFFEE: Yes, sir. I do. MR. PETERS: Then what you may, in fact, be looking at, these high ratios, could very well be simply how much non-radiogenic chloride you release from the rock and then the question becomes are you, in fact, measuring any bomb pulse anywhere, because if all your ratio fluctuations have to do with how much chloride is there total, which is 15 orders of magnitude more stable chloride than there is chlorine-36, very small changes in your stable chloride amount released from the rock could very well be driving this whole ratio. So how do you resolve that problem? MR. CAFFEE: Let me make two comments about that. First of all, you asked me if we had plotted the chlorine-36 ratio versus the chloride, and I said there was no definable trend. It was just a straight array. That's plotting chlorine-36 over chloride versus one over chloride. So trying to come up with a mixing diagram. That's the first thing. The second thing is I think the thing you need to bear in mind is that there may well be two issues here. The first issue is that in the Livermore measurements, there's no bomb pulse chlorine-36 in any of our measurements, and the bomb pulse chlorine-36 ratio is substantially higher. It's higher by more than factor of ten than what we -- by a factor of ten, what we're measuring. Then there is also a difference in the validation samples, systematic difference between the Livermore ratios and the Los Alamos ratios. The Los Alamos ratios are a factor of two to five, something like that. That difference, the latter difference that I mentioned could well be some sort of an effect of diluting the signal with chloride. That's possible. We haven't proven that that's the case or proven it's not the case, but I agree with your line of reasoning and that's part of our path forward is to look into that. We don't see any bomb pulse I don't think can be explained by chloride dilution, because you would be having to dilute the chloride by an order -- or the entire chloride inventory by a lot of chloride and we would pick that up. We would know that. We would be measuring 20 to 50 or 100 ppm of chloride in the leached sample, and we're not seeing that. We're seeing a ppm to two ppm, June is seeing a couple of ppm. So I think there's two different things that you need to bear in mind. There is the difference -- there is the fact that our results are systematically different and this might have something to do with laboratory protocols. But then there is also the fact that we don't see any chlorine-36 or bomb pulse chlorine-36, and I don't believe that that can be explained by dilution with dead chloride. MR. CAMPBELL: But you're also not seeing modern chlorine-36. MR. CAFFEE: Bomb pulse, modern chlorine-36. MR. CAMPBELL: I mean pre-bomb pulse. If the ratio has been somewhere between 450 and, say, 1150 for the last 50,000 years, maybe the last million years, and you're looking at ratios of 50 to, say, about 150, you're looking at very old chloride. So it's not just that you don't see bomb pulse chlorine-36. What you're seeing is very old chloride relative to any of the samples that Los Alamos has. MR. CAFFEE: That's right. If there were no other data in this world and we only had the Livermore data and I couldn't do any more experiments, I couldn't look and see at the stepped release of chloride and I couldn't investigate these processes, and someone forced me to interpret this, I would have to say that based on the fact that the modern input is about 500 and our ratios are less than 200, that we're seeing decay of chlorine-36. So we would interpret our data as indicating very, very old pore water, older than -- comparable to a half-life of chlorine-36. Now, I think it's way too early to try to interpret that data like that, but if that's all I had, that's the way I would interpret it. MR. PETERS: But as you kind of alluded to, if you look at the Los Alamos data on the validation samples, it's comparable to what you would expect to see for background. If you go back -- right, June? I mean, the chlorine-36 to chloride ratios for the Los Alamos validation samples is in the range of five, six, 700-ten-to-the-minus-15, which is very similar to what she -- what you were seeing in background throughout the ESF. Now, granted, they haven't seen anything above 1,200 in those Sundance fault cores yet, and I say yet, that's key, but -- MS. FABRYKA-MARTIN: I don't think we will. MR. PETERS: Okay. Fair. MS. FABRYKA-MARTIN: There's not that much core left for us to process at Los Alamos. Most of the core has already been spoken for. So we may not have a whole bunch more samples coming from validation bore holes, but the ones we do have, it's true, they're background. I wouldn't call it bomb pulse in what we've had come back. On the other hand, when we did find a bomb pulse, we had a structural geologist pick our location, saying this looks like a likely flow path, sample here. We were very careful to try to sample right along the fractures, so we would maximize the amount of fracture surface in our samples, and that's not the case for the bore holes. MR. HORNBERGER: Can you -- well, you certainly have looked to see whether or not the processing is different at some coarse level. Can you summarize if there are any coarse differences? Are people looking at different size fragments, leaching times, are the leaching times different, anything different sort of at that macroscopic level? MS. FABRYKA-MARTIN: He shook his samples for seven hours and we just let ours stew in the soup pot at room temperature for a minimum of two days and stir it once a day. So it's passive versus an active extraction procedure, and that's what we're both investigating now, what is the effect of that on extracting both total chloride and then, even more importantly, on the chloride to bromide ratio, which is the indicator of how much of it is coming from the rock. MR. PETERS: So have you both looked at the chloride to bromide ratios? MS. FABRYKA-MARTIN: We're trying. The problem is that the leachates are so extremely dilute, that we can both get good numbers for chloride, but when it comes to bromide, it's really dicey. So right now, it looks like what we need to do is to up the scale of the experiments large enough that we can take a large enough volume of leachate out at each time in step to evaporate down and concentrate it. But the problem we have is, at least at Los Alamos, is my chemist is about 100 percent Busted Butte project, which means she only does analyses pretty much as a favor, because there's not really a lot of funding and this is a much lower priority than Busted Butte. And so she's processed -- measured a lot of samples, but it takes a couple months to get the analyses out. MR. CAFFEE: Let me agree with pretty much that in its entirety. One of the things that we do want to do is measure the bromine in these fractions and one of the things that we want to do, and we've started, but we haven't completed the analyses, is leach for two hours, leach for four hours, leach for six hours, eight hours, maybe go back and look at fines, I think that's a good idea, too, do all of these kinds of things. It takes time to do these kinds of things and we're working on it and we don't have the results in any state where we could talk about them today. Regarding the bromide measurements, I think that's really important. It's just a hard measurement. It's a very hard measurement and so we have a new chromatograph and we're trying to make those measurements, but we just have to get to the point where we feel confident that they're right. MR. HORNBERGER: Mark, you mentioned in your presentation that you had a chlorine-36 peer review panel. I guess that's new to me. Can you tell me how many are on it and how many people are from inside the program and how many people from outside? MR. PETERS: It was -- I wouldn't be able to recount the names of the folks. It was four folks. It was the DOE initiated external peer review committee, went through a whole series of these back at that same timeframe, in pre-VA release timeframe, and this was one of those. In fact, they -- one of their big pushes was to go to more of a different sampling approach with a cross drift, and that was actually implemented. So we did different ways of sampling for the features in the cross drift than we did in the ESF as part of their comments, but they did a full-blown peer review and that agreed basically with where we were at with the program at that time, anyway. MR. WYMER: Any other questions? Bill? Well, we're pretty much right on schedule. Thank you very much. It's a very important issue. I hope that it gets resolved shortly. MR. GARRICK: The next presentation is on fluid inclusions. MR. WYMER: Let's get started here. In light of the confusion, I think I would just as soon you would introduce yourselves as you speak. As I understand it, what we're basically trying to decide is whether the Yucca Mountain repository is a shower or a bathtub. Please, proceed. Introduce yourself and your affiliation. MR. DUBLYANSKY: I would say Jacuzzi rather than bathtub. Gentlemen, my name is Yuri Dublyansky. I represent the State of Nevada here and, also, I am an employee of the Russian Academy of Science. I am a senior scientist in the Institute of Mineral Geology, Geophysics and Mineralogy. I am very glad that I have this opportunity to introduce the fluid inclusion issue for you and the reason I am so happy, and I even flew all the way from Siberia yesterday to attend this meeting, the reason is that this issue, which I believe is very important, exceptionally important for the Yucca Mountain safety and performance, has been neglected for quite a few years. That's my personal opinion. Also, I want to say the name of fluid inclusion issue is a little bit misleading. So the real issue is not the fluid inclusion. The real issue is the real origin of secondary minerals at Yucca Mountain. Were they formed by hot water pathway through the mountain or were they formed by rain water percolating down through Yucca Mountain, this is the real issue. And fluid inclusions are just a tool, a very convenient, very robust tool, which allows you to determine temperatures at which the minerals will form. This information, the temperature formation is very important when you try to identify the origin of minerals. Now, I also want to mention that the fluid inclusion method is not something new. It's a very well established technique. It's been around for almost a century. It's widely used in many geological, fields of geology, like oil exploration, mineral exploration, and many, many different applications. As I understand, the ACNW and NRC have not been exposed to this issue for some reason. So I was asked to prepare a short overview of the fluid inclusion work done at Yucca Mountain. First, I want to make it clear, it's not something new. This fluid inclusion issue has been around for almost a decade. In 1992, the panel of the Academy of Science and National Research Council related the hydrothermal review concept which was identified by a DOE scientist as a potential site suitability issue. This review was requested by DOE, because DOE needed to have input to make a decision whether to proceed with characterization or not. So even though this panel, Academy of Science panel discarded this issue, the panel did recommend that fluid inclusion research needs to be done. It was an official Academy of Science report on the issue. As far as I know, the first fluid inclusion data were published in 1993 by Los Alamos researchers Bish and Aronson and they studied several site samples from depths of 30 and 130 meters from bore holes and they measured temperatures, fluid inclusion temperatures well in excess of 100 degrees Centigrade. While I personally have reason to believe that those data technically are deficient and the temperatures are probably not correct, they're unreasonably high, but I cannot check it, of course. In 1994, the USGS and Los Alamos researchers published a more detailed paper on fluid inclusions from the unsaturated zone at Yucca Mountain, since the tunnel ESF had not been built by that time, they studied samples from drill cores. Well, this paper leaves a little bit mixed impression. On one hand, it does report fluid inclusion suitable for thermometry, suitable for determination of temperatures in core sites from Yucca Mountain, but on the other hand, this paper does not report any numeric data. So they also just made a statement that this core site was formed at a temperature less than 100 degrees, which doesn't really help, because 20 degrees to 100 degrees centigrade is basically the whole range of water from cold water to thermal water which you can expect close to the earth surface, as understood in hydrogeology. In 1995, I had the opportunity to collect my first samples, working for the State of Nevada, from the tunnel, from the first 300 meters of the ESF which were excavated by that time, and I did fluid inclusion, preliminary fluid inclusion research on these samples and reported my finding to the State of Nevada. Later in 1996, part of this work was published in abstracts for two national conferences, Geological Society of America conference and the American fluid inclusion conference. Unfortunately, late in 1995, the State of Nevada -- the funding for the State of Nevada oversight activity was cut and during the next two and a half years, no work was done by the State of Nevada scientists, because just we didn't have money. In 1998, several events occurred. First, the Pan-American conference on fluid inclusion, here in Las Vegas, the USGS researchers Roder and Wellan went on record by stating that core site samples which they studied from ESF do not contain fluid inclusion suitable for thermometry, suitable for determination of the temperature, and, therefore, those core sites were formed by rain water percolating through the mountain. At the same conference, I presented my data, which I obtained from basically samples from the same tunnel, and those data indicated temperature. I did measure it for inclusion temperatures, have found inclusions which are suitable and the temperature was up to 85 degrees Centigrade. So there was quite a disagreement between two groups of researchers. Later, in 1998, the U.S. Nuclear Waste Technical Review Board completed several reports by the State of Nevada scientists, and one of those reports was a report on fluid inclusion. A consultant to NGGRB, Professor Bob Bodnar, from Virginia Tech, evaluated this report and he not always -- not only evaluated this report, but he also invited me to his lab and we spent some two days running samples on his equipment, and finally he was convinced that the data which I reported in my '95 report are real and they are not artifacts. So after that, the Nuclear Waste Technical Review Board concluded the fluid inclusions found in mineral deposits at Yucca Mountain do provide direct evidence of the vast percents of fluids at elevated temperatures in the vicinity of the proposed repository. So that was a letter report written by NGGRB. In September-October of 1998, I did more detailed research on fluid inclusions and since, by that time, the state still did not have money for this research, this part of my work was funded by the Institute for Energy and Environmental Research. So the results were released through a press conference in Washington, D.C., and after that, 30 copies of this report were requested by one of the U.S. Congress committees. I also want to mention that before releasing this report, we sent it for external evaluation to three well known, well established experts on fluid inclusion in Europe, specifically in Austria, in France and in the U.K., and all three experts agreed that the research was done properly and the quality of work is fine and the conclusions are probably fine, too. So although there was some disagreement between the scientists representing DOE and the State of Nevada, have eventually led the DOE to initiate, in April 1999, a verification project on fluid inclusion issue which is currently underway. Dr. Jean Cline, who led this project, will probably tell much more about this project in a few moments. I won't emphasize, I'll probably make some correction, but this project was not initiated in response to the hydrothermal theory proposed by Jerry Szymanski. It's not correct. I quote from one of the memos of this committee. Actually, I started my fluid inclusion research, just viewing this research as a means of very -- testing the hypothesis of Jerry Szymanski about hydrothermal activity. So I wasn't the one who proposed this concept. And when I reported my data and my interpretation, this data, both data and interpretation were questioned and were disputed by the USGS researchers, so that it had a little disagreement in that. So DOE must be given credit for initiating the project which must resolve this controversy and must resolve the situation where two groups of scientists, which study samples collected from the same location, report very conflicting observations and interpretations. So that's my take on this project. Now, USGS researchers and State of Nevada researchers and UNLV people are doing a parallel study basically on the same samples collected, and as far as I understand, the temperatures, elevated temperatures of fluid inclusions, they're getting basically the same temperatures from our samples. But we still do have disagreement on the interpretation of these temperatures and this probably will be the focus of the forthcoming discussion. In closing remarks, I want to say that as you can see, this fluid inclusion issue was started probably in 1995, where data was available. Nevertheless, the data have not been used in the DOE major documents, in the viability assessment of 1998 and in the draft environmental assessment statement of 1999. This leads me to suspect that these data will not be used in the site recommendation consideration report, either, which is due later this year. While it's fairly clear that fluid inclusion research is not complete and much more needs to be learned about temperatures, about ages of minerals involved, the distribution, it is my strong opinion that at this point, DOE has no evidence which would justify elimination of this issue from consideration in the TSPA. Thank you very much. MR. WYMER: Thank you very much. Are there any questions on this? That's a very fine introduction to the problem. MR. DUBLYANSKY: Just one moment. I have compiled a list of publications and they are attached to the handouts which I prepared and if you need them, I can try to compile the original publications and send them to you. MR. WYMER: Okay. Thank you. John, do you have any comments or questions? Thank you very much for the fine introduction. I think Jean Cline is next, is that right? Would you give us your affiliation? MS. CLINE: Sure. I'm Jean Cline. I'm an Associate Professor at University of Nevada-Las Vegas. Does anyone have a pointer? Okay. We'll wing it. That's okay. Could we dim the lights a little bit? I've got one dark slide. What I'd like to do -- can you hear me, before I get started? Okay. What I'd like to do this afternoon is to tell you about a project that we are doing. What I would like to do this afternoon is to tell you where we are in this project that we are now conducting to constrain the movement of fluid with elevated temperatures through the Yucca Mountain proposed site. When I put this proposal together, there were four questions that I posed and the project is constructed or designed to answer the following four questions. First of all, do fluid inclusion assemblages record the passage of fluids with elevated temperatures into the Yucca Mountain site. Secondly, if the inclusions are present, what fluid temperatures do these inclusions indicate. Third, if present, when did these fluid incursions occur, and then, finally, how widespread throughout the proposed site were these fluid incursions. The scientific plan that we put together to answer these questions involves four related studies and I'm going to quickly tell you what they are and the methods we are using in these four studies and then I'm going to go through each of these four a little bit more slowly and tell you what the results are and where we're at now. First of all, sampling. The first thing that we did was collect 155 samples from throughout the ESF and the ECRB. If you're looking for a handout, there isn't any. I didn't know there was supposed to be one. I apologize. Paragenesis study next. This is a critical part of this study. The idea was to put together a growth history of each of the samples that we collected and the idea was that if we could understand how each sample formed, we could then compare site to site and put together an overall history for the growth of minerals, secondary minerals throughout the repository site. The two main tools that we used to do this were petrography, mostly using the microscope, and then, secondly, chemical analyses using the electron microprobe. Additional tools that we are using include cathodolominuescence; also, oxygen and carbon isotopes, and we will probably also do some laser ICPMS analyses to refine the chemistry a bit. Third, the fluid inclusion study. Again, the first critical part was to place the fluid inclusions observed in each of the samples in the appropriate paragenetic contacts or growth history of each of the samples. We also looked at the fluid inclusion petrography and then we did heating and freezing studies. The heating studies will tell us about the temperature of the fluids that were trapped. The freezing studies will tell us about the composition of the fluids. The final study is dating. We'll be doing uranium lead, we're in the process now of doing uranium lead and uranium series dating, to put some absolute times on these fluid incursions, and, again, the really critical thing is to place the samples that we are dating in the appropriate paragenetic context with respect to the sample mineralogy and growth history and also the fluid inclusions. I want to tell you a little bit about the sampling, first of all. Again, 155 samples from throughout the ESF and ECRB. The goals were twofold on sampling; first of all, to collect samples of all of the different types of secondary minerals that were present and then to get a good spatial distribution. We tried to collect samples every 50 or 75 meters. In some places, the gaps are a little bit larger. That's because there's no secondary mineralization there. Moving on, I want to tell you about the paragenesis study. This is a view of one of our thin sections. It's about an inch and a half across length-wise here and what we're looking at is a very thin section of the rock as we look at it under the microscope. At the base, what we see in black is some tuff and this sample grew essentially layer by layer upward or outward from the tuff. The mineralogy is mostly calcite and silicate minerals. The blue is epoxy. We needed to use epoxy to stabilize most of the samples. So where you see blue within the sample, that reflects zones of porosity and permeability, and there's a lot of it. What we see in this sample is tuff, which is overgrown by some of the early blockier calcite. There are some discontinuous layers of silicate minerals, quartz, opal, additional tuff pieces that fell in and were encapsulated, bladed calcite that overgrows the earlier blocky calcite and silicates, and is then overgrown by outer sparry calcite, which also contains some layers of opal. This is one of the more complex samples from this site. Another section, just to show you there is some variability. Here we see, again, at the base, some tuff layers, but in this sample, all we see is bladed calcite. So this particular sample site did not see all of the events that produced all of the different minerals at some of the other sample sites. A third sample, again, at the base, some of the blockier, more malsave calcite, a discontinuous layer of quartz overgrown by bladed calcite, which is overgrown in places by, again, the sparry, more equant calcite and opal. Here is some selective dissolution of some of the bladed calcite. What I'm trying to show you here is that the samples are in detail very heterogeneous, but that when we look from sample site to sample site, we see repetition of patterns and the goal here, again, is to put together a growth history for each of these samples and then to link these samples to compare them from site to site so that we know how the secondary mineralization formed throughout the repository. The second tool that we used, I said, significantly was the electron microprobe. What I want to show you next is a probe map of this area right here. On the left, we have a back scatter electron image. It reflects atomic number or atomic weight and you can see there is mostly one color of gray, so we have one mineral here, it's all calcite. The black is micro porosity or permeability and what the permeability or porosity outlines is some our bladed calcite. So we have bladed calcite overgrown by blocky sparry calcite and on the right we have a magnesium map. The bladed calcite is black, indicating very low or no magnesium. However, it is overgrown by sparry calcite, which shows very fine detailed oscillatory zoning of magnesium, interspersed bands of magnesium, varying magnesium bearing calcite. It turns out that this magnesium bearing calcite is one of the more important discoveries that we have made in terms of helping us put together these growth histories for these samples. This magnesium bearing calcite is present in more than 70 percent of the samples across the repository site. It is always the outer-most and youngest calcite. So it's critically important in helping us link the mineralogy from site to site. I will come back to this calcite, because I also is critically important in understanding the age of the fluid inclusions. Cathodolominescence is another tool we used. This figure here is about two millimeters from top to bottom. It just shows you some detailed oscillatory zoning of calcite that luminesces. There are bands that luminesce interspersed with bands that do not luminesce. Luminescence is probably caused by small amounts of manganese and the absence of iron. The important thing is that we look for the location or the presence of this in individual samples and then its presence allows us to link, again, samples from different sites, one to another, and, again, put together the big picture for the repository site. This picture is important in helping us place contextually the fluid inclusions. Summarizing the paragenesis, we have 155 samples that are heterogeneous, yet they show consistent textural and mineralogical patterns. Importantly, the outer mineral zones, which reflect the youngest geologic events, are especially consistent; in particular, the bladed calcite and then the overgrown sparry magnesium enriched calcite. These patterns allow samples from different sites to be related to one another. What are fluid inclusions? A little bit of background, for people that aren't familiar with these. Most minerals precipitate from some sort of fluid and as they precipitate, there are commonly defects that occur in the atomic structure of these minerals and these defects may result in the formation of small holes or cavities in the minerals. As the minerals precipitate, they are bathed in this fluid. So this fluid will fill these cavities or holes and the mineral may overgrow these holes and seal off these small packages of fluid, thus creating fluid inclusions. They are important because they are samples of some ancient geologic fluid. Okay. Now, if these fluid inclusions or these systems are forming at elevated temperatures, then as the whole system cools down, the fluid is going to contract. However, the cavity does not change in size. So if the fluid contracts enough, eventually a vapor bubble is nucleated and this vapor bubble is essentially a vacuum. Now, what we can do to examine these inclusions is to reverse this process, to try to get at the temperature of the fluids that were trapped. What we do is we heat these inclusions up, we monitor their temperature. As they heat up, the vapor bubble, if present, gets smaller and smaller. The temperature at which it disappears is the temperature at which the inclusions homogenize, and that temperature approximates the temperature of the fluids that were trapped. In lower temperature systems, the cooling may not be sufficient enough to make the fluid contract enough to generate a vapor bubble. So lower temperature systems may contain liquid only inclusions. These are the two features that we'll be looking at. Here are some inclusions from Yucca Mountain. What we see here, the dark line is the wall of the inclusion. This is calcite, the mineral calcite. Here is the wall of the inclusion. The majority of it is filled with liquid and there's a small vapor bubble. Here is another inclusion, again, mostly filled with liquid and a small vapor bubble. This is what they typically look like, fairly small vapor bubbles. However, these two phase inclusions, which really, again, reflect the trapping of higher temperature fluids are far outnumbered by liquid only inclusions, which are also shown on this slide. It's hard for me to see them at this angle. These would reflect the trapping of lower temperature fluids. This is a typical data set. We have completed collecting data from all of our samples. Approximately half of the 155 samples that we collected contained assemblages of two-phase fluid inclusions. What we see here are three different assemblages located in different places in a sample. However, almost all of the inclusions homogenized over a four degree range from about 49 to 53 degrees C. This is extremely tight data. It's probably the best I've ever seen and it's quite representative of what we've been able to collect at Yucca Mountain. By far, the majority of the two phase fluid inclusions at Yucca Mountain homogenize between 45 and 60 degrees C. There are a couple sample sites in the north ramp where we obtained higher temperatures. Our temperatures are as high as 75 degrees. Others, Yuri Dublyansky and some of the USGS folks have obtained temperatures as high as 85 degrees, but most of our data are between 45 and 60 degrees. Where in the samples are these inclusions? Here, again, we see a typical crust with earlier calcite overgrowing the tuff and in this case, the earlier calcite is overgrown only by the magnesium enriched sparry calcite. So this sample site really saw two calcite forming events. This gray line here sort of is the dividing marker between these two types of calcite. The dark squares here show the location of the two phase fluid inclusions, again, that reflect the higher temperature fluids. Outboard of this line, the only inclusions which we observed in this magnesium enriched calcite were one phase liquid only inclusions reflecting passage of lower temperature fluids. This pattern is consistent throughout the ESF and the ECRB. Two phase fluid inclusions are trapped in early blocky or massive calcite and at the very base are in the cores of the earliest bladed calcite. One phase fluid inclusions are trapped in outer bladed calcite and outer magnesium enriched calcite. This slide really summarizes those facts. Again, two phase FIAs, higher temperature fluids are recorded in the earlier calcite, two phase FIAs are not recorded in the outer or younger bladed calcite or magnesium bearing calcite. The two phase FIAs are very consistently either present or absent in different mineralogical zones and these patters allow the relative timing of the elevated fluid temperatures to be constrained. Now, what we really want to do is to absolutely constrain the timing of these elevated fluids, and this is where we're at right now. Here we're looking at the same sample again. The two phase FIAs located in this area here and at the boundary or sort of outboard of this zone of two phase FIAs, there is, fortunately for us, inner-grown, intermittent or discontinuous bands of opal, which contains enough uranium for uranium lead dating. So what we are now in the process of doing is dating some of these opal samples to constrain the timing of this event. Within the magnesium enriched bad, there is a second continuous layer of opal and by dating that, we can determine the timing of the changeover from precipitation of the underlying calcite and the overlaying calcite, and then there is an outermost opal, shown here in yellow, which was deposited synchronous with this outer band of magnesium enriched opal. So right now we're in the process of dating these opal bands. Again, the samples need to be carefully constrained paragenetically in the context of the sample mineralogy. To summarize, our goals for dating are to constrain the age of the latest magnesium enriched calcite, which is free of two phase FIAs. In other words, which did not see the passage of elevated temperature fluids, to constrain the age of the earlier calcite that does contain the two phase FIAs, in other words, which saw the passage of fluids with elevated temperatures, and to determine the most recent timing of fluid passage with elevated temperatures. Concerning where we are with respect to project completion, we are on target for completion and distribution of reports at the end, about the end of March. We are currently finalizing the petrography and the paragenesis. We have completed the fluid inclusion analyses. We are in the process of doing the dating on some of the other analyses and we anticipate that they will be completed by the end of the year. We anticipate reporting significant results, including data, at the Geological Society of America meeting in mid-November. The final comment that I would like to make is that in undertaking, as we have been conducting this study, we have been meeting regularly with scientists from the USGS and the State of Nevada to discuss our procedures, our data, what we're doing, how we're doing it. We have also been meeting with a larger group that includes people from the NRC, Department of Energy, Nye County, the Nuclear Waste Technical Review Board and keeping them apprised of where we are at and what our results are and the goal here is really to inform anyone who is interested what we've done, how we've done it, with the hope that when we are through this project, there will be a broad understanding of what we've done and consensus on the data, if not on all the conclusions. Thank you. I would be happy to answer any questions. MR. WYMER: Thank you very much. The former director of the Oak Ridge National Laboratory, Ivan Weinberg, used to say there was big science and small science. This is certainly small science here. MS. CLINE: Yes. MR. WYMER: Any questions, John? MR. GARRICK: No. MR. WYMER: Do you have a comment over there? Yes, please. MR. DUBLYANSKY: Jean, by saying -- well, by kind of making -- elevated temperature and non-related temperature, ambient temperature, can you put a numeric number on that? MS. CLINE: No. It would be desirable to do so, but we can't. We have homogenization temperatures as low as 35 degrees C. So we can have two phase fluid inclusions that were trapped as low as 35 degrees C. The one phase fluid inclusions, there is nothing that we can do to get at the temperature of trapping. People have hypothesized a range of temperatures. I've heard below 90 degrees, I've heard below 50 degrees, I've heard below 75 degrees. It's not -- I don't know how to test that. So it's hard to say. There are -- I guess I'll leave it go at that. In general, I will review the process here. The idea, again, is that as the system cools down, the fluid shrinks, and if it shrinks sufficiently, it generates a vapor bubble. So those fluids that cool down a lot will generate a big vapor bubble. Those fluids that cool down a little bit or a lesser amount will generate a smaller vapor bubble. Those fluids that cool down and shrink minimally will not generate a vapor bubble. But putting a number at those brackets, you can't do it, because in nature, it's going to vary from location to location, fluid composition to fluid composition and so on. I will leave it at that. MR. DUBLYANSKY: I just wanted to comment on that, because from your presentation, seemed to be kind of very clear. We have two phase fluid inclusions, elevated temperature, we have one phase fluid inclusion, it's low temperature. It's not correct, because if you have two phase fluid inclusion, you have an advantage, you can measure the temperature. If you have one phase inclusion, you cannot actually tell the temperature. It may be 30 degrees Centigrade, it may be 50, it might be 60, but just by some reason, this shrinkage bubble did not form. So you should not probably put an equal sign between one phase fluid inclusion and low temperature fluid. MS. CLINE: That I would argue with. I think we can't put a number on it, but in general, we can say that the lower temperature inclusions are those that don't generate a vapor bubble and the higher temperature inclusions do generate vapor bubbles, and we see very, very clear patterns. We see certain mineralogical bands which contain two phase and one phase inclusions and then we see other mineralogical bands that very consistently contain only one phase inclusions, and I'm real comfortable saying that there is a temperature differential that's reflected by the presence of those, but I can't put a number on it. MR. DUBLYANSKY: That means there could be this difference, but it does not necessarily mean that the temperature was ambient. MS. CLINE: Absolutely. MR. DUBLYANSKY: That was the clarification that I wanted to make. MR. WYMER: Is it clear that the all the gas in the bubbles is water vapor? MS. CLINE: Good question. It depends on your hypothesis, on what was trapped. If you believe it was meteoric water, then it probably is a vacuum or just air or a vacuum essentially by the contraction of meteoric water. IF you believe that the fluids came from somewhere else, the source in which those fluids might be transporting dissolve gases of some sort, and in other systems, it's most definitely been shown that that can happen, then that vapor bubble could be something else. It could be carbon dioxide, it could be a mixture of carbon dioxide and other gases. There are some tests that can be done. They're difficult to do and the tests that have been done to try to distinguish that have not definitively shown things one way or another. MR. WYMER: What is the biggest ratio of gas to liquid that you've seen? MS. CLINE: We usually -- the call that we've made is ten percent, ten volume percent or less gas, 90 volume percent or more liquid. MR. WYMER: That's the upper limit. MS. CLINE: Yes. MR. WYMER: That's a lot of contraction. MS. CLINE: I should back up. What I talked about are two types of inclusions, the two phase liquid-vapor inclusions, the one phase liquid only inclusions. There are also some vapor only inclusions that are not well understood. There are not very many of them, but they are present. Then there are also some other liquid-plus-vapor inclusions that have larger vapor bubbles, but these inclusions tend not to have consistent liquid-vapor ratios and that makes you suspicious that they formed under some other circumstance than the ambient conditions at the time. So we throw those out because they are inconsistent. They probably formed as a result of leakage or perhaps what we call necking down when the inclusions were formed. We have textures we can use to sort those out. MR. WYMER: Can you say what the age of the most recently formed bubbles is? MS. CLINE: No. That's the data we're waiting on right now. We have a number of samples that have been submitted to a lab in Ontario. To date, the opal, they've been there for a long time. We thought we'd have more numbers by now. We hope to have them at any time and we hope to be reporting those at GSA. That's a big question. MR. WYMER: Yes. MR. DUBLYANSKY: Can I just add a little bit? To your previous question about gas inclusions. Yucca Mountain does contain all gas inclusions. It basically does not contain visible water. And I tried to study this inclusion by using random spectrometry, but all I got is a huge -- which indicates luminescence, which is normally interpreted as the presence of aromatic hydrocarbons there. Enough indication that that's probably the case, but this luminescence decay with time, the hydrocarbons decompose. But still I was not able to identify any particular gases. It's a very interesting subject and I have never seen such inclusion in any other environments. It's very unique, I would say. MS. CLINE: One thing that I could add. The freezing point depression gives us some information on the composition of the fluids and that freezing point depression is very small. So it indicates that there's very minimal salts or minimal gases dissolved in the fluid. Pretty close to pure water. MR. DUBLYANSKY: Can we put a number on that? MS. CLINE: The freezing point depressions were about half -- I don't know, Nick, can you help me out? Do you remember what either the freezing -- minus .6 was the freezing point depression. So about a weight percent NICL equivalent. MR. DUBLYANSKY: So it's not quite -- one weight percent of NICL is not quite fresh water. MS. CLINE: It's not pure water, but it's close to pure water. MR. DUBLYANSKY: It's brackish, I would say, in terms of hydrogeology. MR. GARRICK: Do you have an opinion about the results you might get in terms of whether or not there's a real safety issue associated with the repository? What kind of results would give you some concern about the safety of the repository? MS. CLINE: I guess I'm hesitant to answer that. What we see now -- I think everybody agrees who has been working on this that the fluid temperatures are somewhere in the neighborhood of 50 degrees C and in one area they may get as high as 75 or 80 or maybe even 85 degrees C. What does that mean in terms of engineering the repository site? I don't know that. MR. GARRICK: That's a question we have to answer. We have to answer the so what question and you haven't answered that. MS. CLINE: I think part of the answer -- well, part of what we're doing will help you answer that question, it's going to be the timing of those 50 degree plus fluids. We're waiting for these dates on these opals. They're going to tell us if that happened in the last half a million years, million years, five million years ago, nine million years ago. I think that's going to be a big part of the answer. MR. GARRICK: What if it's ten million years ago? MS. CLINE: Well, it's a blink of an eye in geologic time. MR. GARRICK: I'm talking about the safety of the repository. What's it mean? MS. CLINE: We would have to take the data that we get from this study and then put together our hypothesis of where that water came from and until I get the dates, we can't do that. If the hot water happened ten million years ago, there were hot volcanic rocks there ten million years ago. That's the obvious answer. If the hot water was there a million years ago, something else was responsible for hot water being there, and I don't know what that is at this moment. I'm willing to wait for the dates before I try to answer that. MR. GARRICK: Thank you. MR. WYMER: Any other questions or comments? MR. GARRICK: Do we have another presentation? MR. WYMER: Who is next? Would you introduce yourself and your affiliation, please? Thank you, Jean. MR. GARRICK: Thank you. MR. PETERMAN: My name is Zell Peterman and I'm with the USGS. I'm currently the Team Chief for the Yucca Mountain Project Branch Environmental Science Team. We certainly agree with Yuri's earlier statement that this is not a fluid inclusion issue, it's a calcite multiple fracture filling issue, but I have to add to that that fluid inclusions aren't the only evidence that's going to resolve this issue. In fact, we think it's already resolved. What I would like to do is step back in time a bit and if somebody will show me how to work this projector, the computer thing. The USGS has been studying these fracture minerals, first, in drill core and then later in ESF since about 1989 and prior to that, there were some early studies in the mid '80s of calcite fracture fillings in drill core. Listed at the bottom here are the people actually doing the work. Joel Whelan is the person principally in charge of our fluid inclusion work and the stable isotope work. Landon Namark and Jim Pacies are doing the isotopic dating and Bryan Marshall and I guess to some extent myself are worried more with the isotopic heavy isotope signatures in these materials. So let's have a quick look here. Can I move around here? Is that coming through? What do these things look like? The fracture minerals occur in fractures and lithofizal cavities. Calcite and opal are the main minerals, although there are other minerals present, clay minerals, zeolites, manganese oxides and probably some others. The coatings, as Jean said, range in thickness from several millimeters to several centimeters. These are just two examples. The one on the left is an example of a fracture surface and there is a scale there for reference. The small divisions are one millimeter. The one on the right is the floor of a lithofizal cavity, lithofizal cavities are gas bubbles that form at the top of the ash flow sheets as the ash flow degasses, and they're commonly lined with high temperature minerals all the way around. Typically, silica polymorphs and alkali feldspar are the predominant minerals. There are trace minerals of ohemitites, some garnet has been found, things like that, but they encrust the whole inside of the lithofizal cavities. In contrast, the low temperature minerals, calcite and opal, are typically on the bottoms of these cavities. Now, the green coloration is opal, which is fluorescing under ultraviolet light and it's fluorescing because it's a rather large uranium content, up to 500 ppm uranium, typically around 100 to 300 ppm uranium, and this is the key to the dating work. This is what we can date. We can't date the calcite directly because of it's very low uranium content. Why study these fracture minerals? AS far as we're concerned, the USGS is concerned, they are the physical records of long-term infiltration through the UZ. This is -- I think what Jerry Szymanski refers to as the USGS rain water hypothesis, and actually I kind of like that. Our conclusion after looking at all sorts of data is that these were formed by downward percolating water. We can date these things, as I mentioned, by uranium series, uranium lead and carbon-14, and provide a history of deposition and, therefore, a history of the fluids involved in depositing these materials. The calcite especially also contains an isotopic record of the source fluids and we look at oxygen, carbon, strontium and U-234-238 ratios and they contain fluid inclusions that may yield information on the thermal history of the rock mass. From 1990 to 1995, the only thing we had to look at were core samples and this is an exceptional sample here. And we tried to do some dating, some mineralogy, some isotopes. Our eyes were really opened when the ESF was constructed and we found deposits like this, which never would have survived a coring process. So we were really misled by what was available in only the cores. And the ESF materials are far superb. We can collect good samples and, again, this green coloration is fluorescing opal on the tops of these very delicate calcite crystals that are growing up from the base of a lithofizal cavity. So this is the history, that was the core. Core studies were up to 1995 and in '95 we ramped up our efforts because of ESF, did a lot of isotopes, mineralogy, fluid inclusions. Our early focus was on the history of the outer-most surfaces, because we were testing a model at that time, in 1995, that, in post-glacial times, the PTN was acting as a unit that moved the flow of water laterally rather than allowing it to come down. So we were tasked with what is the under stage on the outer-most surfaces that you can find, the idea being that we probably wouldn't find anything less than 10,000 years. A more recent focus has been the long-term deposition on the thermal history of the UZ, the compositional evolution of fracture water, and turning our age information and abundance information into some estimate of seepage flux. So what do these things look like in the ESF? And you will see some of these tomorrow. There are primarily two major occurrences. There are other minor occurrences, as cementation in fracture zones, things like that. The depiction on the left shows a fracture that opens up, and you will see a lot of these in ESF and you will see that the deposits are typically on the foot wall side of these moderately to steeply dipping fractures. The one on the right is a depiction of a lithofizal cavity and there you will see that typically the deposits are on the bottom. Our interpretation here, and this is not an interpretation that's agreed to universally, is that this indicates some sort of film flow moving down these fractures and it's moving under the force of gravity. So it's staying on the low sides of the opening. In contrast with what we see in the UZ are what you will see in the ESF in the calcite below the water table, and, again, we're looking at drill core. The calcite coats all the surfaces of fractures and commonly fills the fractures. As I say, we interpret the occurrences as indicating downward flow along fracture foot walls and cavity floors. We view this as pretty strong evidence that those cavities have not been repeatedly hydrologically saturated. Jean made these same observations. Calcite and opal are intimately associated in micro stratographic relationships. In other words, you can develop the micro stratigraphy starting with the base, lying on the tuff, going to the outer surfaces, and, in many case, you can convince yourself that this is an age progression. We have data from WT-24 which says that the average calcite abundance in the rock mass and the crystal pore member of the Topapah Spring tuff is .24 weight percent, and I will show you that data in just a minute. Calcite dominates, as you saw in Jean's slide, probably less than ten percent, ten percent or less opal and other minerals. The deposits aren't homogeneously distributed in ESF and the greatest abundance that we've measured and another way we've measured is conduct line surveys and actually stretch a line, a 30 meter line every 100 meters and measure the thickness of the calcite deposits in fractures and in lithofizal cavities. And this is an example of the data from WT-24, which we thought was another way to get an idea, and WT-24 was drilled by the LM-300. So there was a lot of cuttings, a lot of ream cuttings. If you've seen the holes drilled by that, and these cuttings were captured for us, integrated five foot samples and from those five foot samples, we ground them up, prepared a sub-sample and chemically analyzed the samples for C02, and then we assumed that all the C02 as in calcite and that's the way we get the .24 weight percent. And that's the mean and, to me, this is analogous to, say, determining the grade of an ore body. So in this case, the arithmetic mean is the appropriate measure, even though the distribution of values is highly skewed and looks almost maybe fractal in nature. There are many, many openings in the ESF that don't have secondary minerals, and this is just an example of one. This is a photo moziac of the tunnel wall over ten or so meters, 20 meters maybe, and the red coloration depicts cavities that have secondary minerals and the white that don't have secondary minerals. So, again, if these were the result of upwelling water, I think we have to ask the question why aren't all the possible depositional sites occupied by calcite. In terms of the isotopic evidence that favors descending water, we have carbon isotopes and the youngest calcites overlap those of the surficial calcrete, and that's the ultimate source of the calcium. If you've been out to Yucca Mountain or anywhere in the desert southwest of the U.S., you will see thick deposits of calcrete that are ubiquitous. And rain water comes down and periodically dissolves this and carries it down fractures. You can see this virtually anywhere you go. The oxygen isotopes in the youngest calcites are consistent with meteoric water heated as it moves down the geo thermal gradient. The strontium ratios in the youngest calcite overlap the values with calcrete, which you'll see in a minute. Work at Los Alamos has shown that the UZ calcite has pronounced negative cerium anomalies. This is not observed in the saturated zone calcite and cerium anomalies are small or nonexistent in ground water. The 234-238 ratios are identical to calcrete, values in calcrete and runoff, and are much smaller than values observed in the tertiary volcanics or ground water. And the conclusion is obvious. MR. HORNBERGER: Can you give me just a quick tutorial on the cerium anomaly? MR. PETERMAN: This is Los Alamos' work and I think it has to do with the oxidation state, the multiple oxidation states and the solubility. This is just an example of the oxygen and carbon isotope analyses that Joe Whelan has conducted. I should have put the number up there. There's a lot of measurements and Joe has also placed the deposits or determined the paragenetic sequence for the deposits, and, again, these are relative things, but he can classify things as early, intermediate and late, and in general, putting the isotopes in that context forms three discreet groups, but with a lot of scatter. But if you get enough analyses, you know that there are clearly three discreet groups there. There is an early calcite that occurs below the cal layer, I think that Jean referred to, and it has these low delta 018 values, which Joe interprets as having formed from water with elevated temperatures, and he's taken water of minus 12.5 and then just using the fractionation factor, he would guess that those could have formed from water between 50 and 80 degrees centigrade. Here is strontium data on calcite from drill core and at the top there is a histogram. Unfortunately, there is not an indication of the number of samples, but the shortest box there would be one analysis and these are -- this is strontium-87-86 values shown as delta 87, which is just the deviation of the 87-86 from that value for modern sea water. So the calcretes have a skewed distribution, but they peak around between delta value of between four and five, and you can see the shallow calcites pretty mimic that distribution and then as you go deeper into the UZ, the numbers go down and then in the SZ, the numbers are quite different, the strontium numbers are quite different than the strontium values in the UZ. We have data from core water salts, from, I believe, SD-6, and we see a very systematic change with depth and what we see is the beginning of a certain amount of water-rock interaction right around the PTN and then continuing down. So we've got a combination of source, plus water-rock interaction here. Geo chronology. We've done carbon-14, uranium series, uranium lead dating, and as I said, our first emphasis was on the outer-most mineral surfaces and these three boxes on the left are nested here or correlated. You can see they each represent different spans of time. We get carbon-14 ages as young as 16,000 years. We get uranium series ages that span from very young to the limit of the method, which is 500,000 years, and then we get the uranium lead ages that go to 1.8 million years. It was this relationship that led to the concept that we're dealing with, very slowly depositing material and that a given thickness of these things represents a long amount of time. So that the challenge here is sampling for dating. You will see in the next slide that we're looking at one to four millimeters per million years. Now, we have to physically sample these with a demo drill, so we're integrating -- our samples will integrate over a finite period of time and the results then will be skewed as a function of the decay constant of the method being used. And an example would be if you had two layers of calcite, one that was modern and a subjacent layer that was a million years old and your sample included both of those layers, that composite sample would be 50 percent modern carbon and your age would be, what, 6,000 years, your determined age, but the real average age would be a half a million years. So that's the kind of bias you can get in the age work in dealing with the shorter half-life systems. Now, when we go to the uranium lead system, it's a much longer half-life, so we're getting much closer to the real mean age of the material sample and the histogram on the left gives the current distribution of ages we have. There is this older calcite layer that's associated with this limpid quartz and some of that has enough uranium to date and the oldest age we've gotten so far is ten million years. Then it progresses up and I think the histogram is more of a function of how we've sampled what we've emphasized than it is a statistical distribution of ages. And this shows the data that have gone into the interpretation of the growth rate. In the upper right-hand corner is a cross-section of a sample and, again, the green fluorescence is opal. This has opal embedded in it and opal on the outside. You'll see an age around seven million years at the base and then 4.5 and 4.3 and I can't even read them all myself, .14 and .09 at the outer surfaces. So here is an example of what we call the micro stratigraphy and in this case, we seem to have it calibrated with ages. Right below that, you will see a series of lines associated with data and these represent different specimens for which we have these same type of data and what we've done is just taken the thickness and normalized the thickness to unites. So plotted on the Y axis is just relative position in a crust or deposit and then the slope of that line you can calculate back and get some estimates of average growth and those go from 1.3 to 4.1 per million years. So we think they're very slow growing. This was verified by some work that was recently done on what's called the SHRIMP, USGS Stanford SHRIMP, in Stanford, and this is the high resolution ion mass spectrometer, where you can actually zap a sample with, say, a 10 to 20 micron diameter beam, you have a beam of oxygen ions, and get ages directly without having to go through sampling or chemistry. So that other diagram, up at the top there, you will see a deposit of opal, one of these hemispheric opal deposits, which is in the outer-most surface and that traverse has ages ranging from 5.1000 at the outer-most zapped point to 530,000, but with a huge uncertainty. Anyway, the graph below that plots that data and, again, the growth rate there for that opal using those numbers is .72 millimeters per million years. So this substantiates our contention that these are very slow growing deposits. Fluid inclusions, Jean and Yuri have already covered this. Joe sees the same thing that everybody else does, single phase liquid filled inclusions, single phase gas filled inclusions, two phase with variable liquid gas ratios, and then the two phase was small, but consistent vapor-liquid ratios. And with certain assumptions, these may provide estimates of depositional temperatures. I think it's appropriate to comment here that when the major fluid inclusion thrust started, and I think Jean would agree that there was a lot of discussion on whether calcite was a suitable host for these fluid inclusions, whether calcite could be relied upon, because it was known from other studies that you look at the mineral crosswise and you're going to do something to the fluid inclusion. So there was a lot of discussion and I guess being a skeptic, I'm still somewhat skeptical, but I'm willing to be proven wrong. Anyway, as Jean said, we find also 50 percent inclusions contain deposits, contain fluid inclusions 40 to 80 degrees C. Many of these are in the earliest calcite. A few appear to be in the intermediate stage. None have been found in the latest calcite. We think the fluid inclusion assembly is consistent with calcite formation under vados conditions, but at slightly elevated temperatures. In other words, unsaturated conditions. Yuri would say that these all formed in the sat -- that there was saturation, that these fractures were filled with water. We don't agree with that. MR. HORNBERGER: But you would still have to have, as Jean said, a model, a conceptual model for getting the elevated temperatures. MR. PETERMAN: That's true, right. Absolutely. If we can believe the elevated temperatures. I think one has to ask that question first and foremost, and then we have to ask how could you get those elevated temperatures. We know that you can, you know, 12.7 million years and probably a few decades after that, things were very hot. We find evidence fumerolic deposits at the top of the Topapah Spring that are 12.7 million years old, no doubt about it. Just like the Valley of 10,000 Smokes, the water was getting very hot as it penetrated a little bit and the tuffs cooled and crystallized from the top down and there was water circulating and coming back out as very vigorous hot springs and steam vents and all that. There is no doubt about that whatsoever. Then the question is how long did it take to cool the tuffs, say, to 50 degrees. I don't think we have a good handle on that. There is a Timer Mountain Caldera, which you referred to -- it's usually quoted about 10.5 million years, called the Timber Mountain Event. Brian Marshall has done some thermal modeling and shows that at Yucca Mountain, 20 kilometers away from a possible buried pluton, there is a thermal pulse that could hit Yucca Mountain about eight million years ago and then that has to decay down. Now, it's a very simple model, it's conduction only, there's no advection and all that. It's something that needs to be pursued. MR. WYMER: We are starting to push our time a little. MR. PETERMAN: Okay. I'm going to finish real quick. This is just another example of distribution of fluid inclusions that Joe Whelan did. This is from one little teeny chip, 320 measurements. He's got the patience of Job, 54 degrees. There are some flyers out there, up to 80 degrees, 54 degrees, excluding those, standard deviation of three. This bothers me a bit. Two standard deviations of six, total range of 12 million years. If we can measure these things to better than a degree, why do we get this dispersion. It's telling me there are some other variables in here we don't fully understand. This is the USGS data now plotted against distance from the north portal. Those higher temperature values that Jean referred to in the north ramp around 80 to 90 degrees. Again, that's just a map showing the USGS data. Our attempt to constrain ages, this was asked and, again, so far, all we've done is try to sample opal that's immediately above a zone that contains fluid inclusions. We have ten ages now, our youngest age is 1.9 million. All of the others fall between eight and nine million, here is one that's seven, okay, 6.5 or seven. Again, they only provide a minimum age and the maximum age is controlled by the tuff. So I would say on the right, the fluid inclusions are between 6.5 and 12.7 million years old. On the left, they're between eight and 12.7 million years old, and this one sample has this bounding opal at 1.9. So as I say, I think we have ten numbers now and this is something that both UNLV and the USGS is pursuing vigorously. These are our conclusions. We have large and comprehensive data that shows that low temperature fracture minerals form from meteoric water percolating downward through the rock mass during the last ten million years or longer. I think the fluid inclusions may provide very interesting information on the thermal history of the rock mass and I think that's something that should be pursued as best we can. So that's it. MR. WYMER: Thank you very much. Are there questions? Surely our geochemist must have a question. MR. DUBLYANSKY: I have a question. Do you have conceptual model for high temperature calcite formed in the vados zone, how much did it happen? I didn't quite understand your answer. MR. PETERMAN: Our contention is that those temperatures are real and I still think we have to be a little bit skeptical, because of uncertainties, assumptions and all that, but those temperatures are real, then that means the rock had to be that hot to heat the downward percolating water. Now, these could be -- it doesn't just have to be done by the cooling of the tuffs. As I say, there is certainly a potential of an eight million year thermal coming from the buried pluton in the Timber Mountain Caldera. There are other ways to move heat into the upper crust. One way is detachment faulting, which is well known. You bring hot rocks up along shallow dipping normal faults below cool rocks and you have a perturbation of the geotherm. And we know that detachment faults are not at all uncommon in southern Nevada. There's a whopper of one just across Crater Flat over at Bear Mountain that did exactly that and that area over there, based on geo chronology, there was activity as young as eight million years. So I think there are certainly ways to get heat into the unsaturated zone without pumping up hot water. MR. WYMER: Well, John Garrick always likes to drive to the very practical end of things, so what does this all mean with respect to Yucca Mountain. I think I understand what each of you concluded, but is there any of the three of you that believe that there is a chance that hydro thermal water is going to come up and fill the cavities? That's your position, that that's a possibility. So of the three of you, there's -- you want to vote? MR. PETERMAN: I think we have to leave it up to the scientific community to evaluate these results and make their own conclusions. We've already made ours. MR. WYMER: How far off are we of having these answers in time? MR. PETERMAN: I would say less than a year. MR. WYMER: Less than a year. You have another presentation? I'm sorry, Yuri. MR. DUBLYANSKY: Just I will be trying to summarize what can be hypothesized about the sources of those fluids and the origin of minerals from the data which are available now. So if you have more data or some new data that will show up, we will probably incorporate them into the model. But I base my presentation on what was done by USGS at least what was available to me and on my new fluid inclusion research was done just recently during this year. So I'm not sure about this idea about heating the rock due to faulting or thermal process. The calculation done for Yucca Mountain and as soon as I have this more formalized model, I will be happy to make an assessment of that. We did, however, estimate the time required for cooling of tuffs and the age of the bedrock is about 12.7 million years, and our very conservative estimates show that the cooling would require a maximum of about 100,000 years. So essentially we attribute the formation of this elevated temperature to the cooling of tuff, we have to have the age of this calcite to be close to this 12.7 million, 12.8, but it cannot be -- it cannot last for millions of years. Then we have this very well known Timber Mountain Caldera hydrothermal, which is dated from 11.5 to ten million years ago, and this work was extensively studied and dated like that. So we have another time marker here. And then if we see that our sample or our elevated temperature calcite was formed at the time after that, I cannot see of any other explanations than to attribute it to the thermal activity, hydrothermal activity, because again, I repeat, I cannot see how it can conductively heat the rock to such a high temperature. Also, we will continue with further -- we not only have to heat it, but also if we find all the data, we have to maintain this temperature for very long periods of time, and I will come to this later. From the data which I had, I didn't have this new data that they measured ages of the quartz at ten million years, but these preliminary data indicate the age of the oldest -- by that, I mean the oldest ages measured from the Yucca Mountain samples. They were about nine million years old. So they're still somewhat younger than the Timber Mountain Caldera hydrothermal, and this new age is also between ten and nine million years. So I think this makes -- this shows that the minerals that we're talking about, they are younger than -- probably younger than Timber Mountain Caldera, but also they are definitely younger than the bedrock tuff, and this is the quote from one of the USGS reports and they seem to be aware about that. MR. HORNBERGER: Yuri, you disagree that those ages represent a minimum age then. MR. DUBLYANSKY: Well, minimum age, as far as I understand, talking about uranium series dating, which are more short-lived isotopes. Would you say that uranium lead ages also provide minimum estimates of that, of ages? MR. PETERMAN: The ten million years is on the cal simulator. MR. DUBLYANSKY: Right. The question was -- MR. HORNBERGER: But I understood Zell's point to be that you're sampling an interval and -- MR. PETERMAN: We're sampling an interval, but because of the long decay constant for the uranium, we're still getting an average age, but it's closer to the real age of that material. MR. DUBLYANSKY: On these graphs, I plot the thermal reconstruction of the Timber Mountain Event done by Bish and Aronson, these red arrows, and this work was done based on the transition between clay species. And the red rectangle shows the temperature, which was obtained from the ESF samples. You can see it's quite different, 85-90 degrees Centigrade at Yucca Mountain, just ten degrees short of boiling at this altitude. If you just assume normal thermal gradient, conductive thermal gradient, which should be operational in the vados zone, you will have here a gradient of about 200 degrees Centigrade per kilometer, which I don't think is reasonable. So I don't think by conductivity they can raise the temperature to 90 degrees at depths of about 50 or 30 or 100 degrees from the surface and keep this temperature for a long time and I don't see how can we do that without melting the rock. Now, let's discuss a little bit ages. First, we have this time marker -- well, mostly, the minerals which can be dated at Yucca Mountain is silica. We only can date the latest calcites which can be dated by uranium series ages, but all calcite cannot be dated because it doesn't seem to be old enough. So we have to use some indirect methods, particularly in those samples that do not contain silica phases. However, it is well known or established at Yucca Mountain that opal is always younger than ten million years, or probably eight million years, and the diagram which Zell was showing, it's clearly demonstrated that. When you talk about opal, you are talking about something between eight million years and probably a 100,000 years, and, again, I refer to the work done by USGS. So if we see fluid inclusions associated with this late opal, we have to assume that this calcite is also young. This is one example of the situation. This is calcite and this -- calcite, opal, which is typical young opal which occurs normally in the upper part of the samples, and this is one of the fluid inclusions with a gas bubble here. This is only one inclusion from a group of inclusions which can provide quite reliable temperatures. So in this case, we have to accept that this calcite cannot be older than eight million years. Here is another example and it's a very interesting sample. Again, we have opal in the crust of calcite, which is about one and a half centimeter. One of the opal from the samples collected from this area -- actually, there was extensive dating of these samples and the ages obtained range from six million down to 300,000 years or something like that. Opal from the samples collected from those cavities. Again, if we accept this rate of deposition which was determined by USGS, which is from one to four millimeter per million years, we have to assume that the time required for generating this crust would have been at least between four and ten million years. Each millimeter of this calcite requires about a million years to grow. It's quite a long time to form that. So what I did, I analyzed calcite layer by layer, above the opal and from outside layers, and here are the temperatures. So basically this part probably are secondary inclusion, but those are very consistent temperatures, 50 to 52 degrees Centigrade, and these temperatures just persistent through the all the crust. Again, it's a little bit -- well, it isn't contradiction, which Jean was showing, that fluid occurs only in the bottom. These particular samples, you can see fluid inclusions throughout the crust. MS. CLINE: You can't equate position necessarily to age. I guess what I'm saying is that I don't agree 100 percent. You can't just measure some thickness, attribute some number of years. You can have a whole crust that grew ten million years ago. You can have a whole crust that grew a million years ago. MR. DUBLYANSKY: Exactly that is my point, but I am not very comfortable with this rate of one millimeter or four millimeter per million year. I just kind of exaggerate, but still my point is here in this crust, we have consistent temperatures throughout the crust and if these rates are correct, we have to assume like eight million years timeframe of this growth. I'm not saying that that's what was happening. I'm just showing some kind of problems with that. But, again, these temperatures are very consistent and is present in all calcite. This calcite looks to me very typical bladed calcite. The next method which we can use is to use stable isotope, like Zell was suggesting, and this is a compilation of data. Well, it's kind of illustration of data which is represented in the one of the USGS reports. Indeed, this report says that early calcite, probably old calcite almost always have this heavy carbon and it's light in oxygen and late calcite almost always have the reverse, light carbon and heavy oxygen. Late calcite form will define between minus five and minus eight per mil and this calcite generally represents the age distribution over the last several hundred thousand years. That is according to the work by USGS. So if we have calcite which have these values, we probably, again, if we accept this idea of USGS, we have to accept that we are talking about calcite over the age of a few hundred thousand years. So here is one of my samples which I studied. This calcite does not contain opal, so it cannot be dated, but it does contain fluoride, not a very typical mineral in water, and it consists of three zones; base zone, granular zone, and then blocky calcite. So I did a very detailed and very careful isotopic study on this calcite, actually the same thing just turned 90 degrees clockwise. I used the in situ laser ablation stable isotopic analysis and with a spatial resolution about 300 microns for each spot. Some of this delta C-13 stays positive for the most part of the crust and the outer part, it drops dramatically to the values of minus six to minus eight, which are the typical values which were reported as being representative of the calcite with the age of on the order of a few hundred thousand years. So this essentially carbon oxygen distribution from this particular sample, one sample, essentially mimics what was shown by USGS. Again, we have this late calcite with consistently light carbon and heavy oxygen. So again, if we accept this USGS interpretation, we have to assume that this calcite was formed less than probably 200,000 years ago. Here are the results. Again, we have fluid inclusions through all the crusts from top to bottom and this early calcite, heavy carbon calcite, shows a little bit more high temperature and this late calcite has a little bit cooler temperature, but still the thermal temperature is 40 to 46 degrees, 50 to 52 degrees. So in this situation, we have isotopically light calcite, which is probably formed at early -- probably young and it does contain fluid inclusion and in this case, again, we have fluid inclusion present throughout the crust. So my point is that at this point, at this stage, we cannot claim that all fluid inclusions or all calcite contain elevated temperature. We cannot claim that this calcite is related to the cooling of tuffs because the age data does not allow us to tell that, and I think the only interpretation of that is that this calcite was formed by fluids with elevated temperature within the mountain. Thanks. MR. WYMER: Thank you. Questions? MR. PETERMAN: I just have a comment. If you look at page 12 on my handout, you will see that many, many hundreds of carbon and isotope analyses, they form a very broad trend and within each group, there is -- maybe it's not 12. It's the one that has the carbon versus oxygen. There it is. There's a huge amount of scatter there and other than saying there is a real trend, I don't see how you can use that data, except in the very earliest calcite, which is below the base cal simulator, which has those anomalies or very light oxygen values. I don't see any way how you can use that cluster of data as a chronometer at all. It's just way too much scatter. MR. DUBLYANSKY: My answer will be, first, that this graph which you represent here on your handout -- I just remembered the graph from your presentation, which brackets the temperature of 50 to 80 degrees Centigrade and those brackets are around the red dots, which are heavy in carbon and light in oxygen. It's just not correct, because I just have shown you that temperature 50 degrees Centigrade and 55 degrees Centigrade can be associated with calcite which should be on this blue. MR. PETERMAN: But you had no age information to constrain that statement whatsoever. MR. DUBLYANSKY: That's exactly, but the statement which you made in your report, they have no bearing on age. You just refer signature and age. Always, when you report stable isotopic values of this and carbon is minus five to minus eight. MR. PETERMAN: Joe's depiction there is based upon petrographic delineation of the paragenetic sequence. That's his categorization. Any one specimen, there may be age controls, but he just put together the paragenetic sequence and then he's put the carbon and oxygen isotopes in that framework. MR. DUBLYANSKY: Absolutely, but I also have a paragenetic sequence in my sample and I just have shown it to you. MR. PETERMAN: But you had no age information. MR. DUBLYANSKY: Indeed, I don't have age information, but I do have indication that -- at least I am using your information and your work, actually I quote this in my report, that -- let me just quote you. MR. PETERMAN: It's not going to do me any good. I certainly can't remember the paper quotes at this point in time. But there are typically a lot of things that are taken out of context. MR. DUBLYANSKY: Well, this work is based only stable isotopes and it summarizes stable isotope work and age dating work. So I'm just using your information. MR. PETERMAN: The same diagram is published somewhere in another report and I would suggest maybe you use that distribution of value. MR. DUBLYANSKY: I didn't understand your comment. MR. PETERMAN: I don't know what report that is. It may have been 1992? MR. DUBLYANSKY: No. This is the report, ages and origin of subsurface calcite, and what I did is I just took date from your table and quoted them. That's all I did. MR. PETERMAN: That's fine. That's a general trend. But if you put the uncertainty on that trend, it's huge. MR. HORNBERGER: I think Zell was just saying that you can't invert that trend, because if you look at his dots, the 13 on the early calcite goes everywhere from plus ten to minus six. It's not very well constrained. It's a trend, but you can't use it as a geo thermometer. MR. DUBLYANSKY: I am not intending to use it as a geo thermometer. MR. HORNBERGER: Or a geo chronometer, sorry. MR. DUBLYANSKY: Or geo chronometer, too. What I want to show, we have to have some handle on the -- well, I'll put it a different way. We probably cannot date this calcite because it cannot be dated with the uranium lead methods. So should we just throw away the data, fluid inclusion data, which we obtained on this calcite? I don't think we should do that. So I am using this stable isotope data as as proxy of the ages, again, based on the work done elsewhere by USGS. I understand that that it's not a perfect method and I will not claim that. But at least we cannot claim, by the same token, that this calcite is old. MS. CLINE: Do you have paragenetic data that show that that particular calcite that forms the outer band in that sample is consistently present and is consistently a young calcite throughout the repository site? MR. DUBLYANSKY: No, I don't, because it's blocky calcite. It is present -- well, it could be present in many samples. There's a granular calcite in the middle, which I was showing it's not very common Yucca Mountain calcite at all, also it contains fluorides, which also not all samples are -- not all samples contain. So I think paragenesis also should be variable from place to place in Yucca Mountain. For instance, fluoride, we have found that fluoride is mostly associated with samples that are collected close to the fault lines. MS. CLINE: Close to the fault lines? MR. DUBLYANSKY: Right, close to the major fault lines and we have found fluoride in many, many samples, in much more samples that we expected to find them. No, I cannot parallel this particular calcite as paragenesis and that's why I'm using -- I am trying to use stable isotope report to get a handle on where I can place this calcite. MS. CLINE: I would just say that in the work we have done on all the samples where we see the fluoride, it is in an older part of the sample. We have not seen fluorite anywhere we can constrain the fluoride as being part of the younger event, either the later belated event or the magnesium enriched calcite. It's just not there. MR. DUBLYANSKY: Yes, but I don't think you have an age constraint on the calcite either. MS. CLINE: This is relative age. MR. DUBLYANSKY: Relative age. In terms of relative age, well, this fluorite definitely is present in this zone and the outer zone of calcite, and here you can see fluorite just protrudes into the surface of calcite. It's a common calcite and fluorite. Again, basically, it's one way of addressing this issue, to calculate the thermodynamics of the calcite-fluorite system and that's what we are doing right now. MR. WYMER: According to my schedule, we're over with our break now. Is there one last burning question or comment? It's a very interesting discussion. Is there any final thing that one of you feels constrained to say? If not, well, thank you very much. It's an interesting discussion. Let's take a break. MR. GARRICK: Yes. Thank you. Before you break, I want to advise you of what we're going to do. There's two things remaining to be done. One is to prepare the committee for the tour tomorrow and the other is simply to do some homework in preparation for future meetings. I don't think we need the court reporter for the rest of the day. So I'm getting a favorable head nod from the staff, so you are excused, as we adjourn for the break. [Whereupon, the meeting was concluded.]
Page Last Reviewed/Updated Monday, October 02, 2017
Page Last Reviewed/Updated Monday, October 02, 2017