1 UNITED STATES OF AMERICA NUCLEAR REGULATORY COMMISSION - - - BRIEFING ON OPERATING REACTOR OVERSIGHT PROGRAM AND STATUS OF IMPROVEMENTS IN NRC INSPECTION PROGRAM - - - PUBLIC MEETING Nuclear Regulatory Commission One White Flint North Rockville, Maryland Thursday, February 13, 1997 The Commission met in open session, pursuant to notice, at 2:05 p.m., Shirley A. Jackson, Chairman, presiding. COMMISSIONERS PRESENT: SHIRLEY A. JACKSON, Chairman of the Commission KENNETH C. ROGERS, Commissioner NILS J. DIAZ, Commissioner EDWARD McGAFFIGAN, JR., Commissioner . 2 STAFF PRESENT AND PRESENTERS SEATED AT THE COMMISSION TABLE: JOHN C. HOYLE, Secretary of the Commission KAREN D. CYR, General Counsel HUGH THOMPSON, JR., Acting EDO FRANK MIRAGLIA, Acting Director, NRR FRANK GILLESPIE, Director, Inspection & Support Program, NRR ROY ZIMMERMAN, Associate Director for Projects, NRR . 3 P R O C E E D I N G S [2:05 p.m.] CHAIRMAN JACKSON: Good afternoon, ladies and gentlemen. We are pleased to have the NRC staff here to brief the Commission on the operating reactor oversight program. This is the first of three Commission meetings addressing different aspects of the operating reactor oversight program. The subject of this meeting is the Status of Improvements in the NRC Inspection Program. The subjects of the other two Commission meetings are Analysis of Quantifying Plant Watch List Indicators, at which time the staff will discuss the Arthur Anderson review of the senior management meeting process, and the Millstone and Maine Yankee lessons learned as they relate to policy issues affecting the regulatory process. These meetings are scheduled for February 18 and 19, respectively. I would just ask you, the staff, as you discuss areas of improvements, please provide for the Commission the results which have been achieved and/or examples which demonstrate improvement or are aimed at demonstrating improvement, so that if activities are in progress, you can provide a time table. Please provide a time table for the completion of the activities. . 4 I understand that copies of the slides are available at the entrance to the room. Unless fellow Commissioners have any comments, please begin, Mr. Thompson. MR. THOMPSON: Thank you, Chairman Jackson and Commissioners. Before we go into the presentation, I should remark for you and members of the public. In our efforts to upgrade the TV coverage of the Commission dialogue, we have downgraded. The motherboard didn't fit up and we burned it out. So we now have a camera in the corner over here. We will have a general view of the proceedings, but slides will still be presented. Hopefully the public will be able to follow along without much trouble. With me at the table this afternoon are Mr. Frank Miraglia, who is the Acting Director of NRR; Mr. Frank Gillespie, Director of the Division of Inspection and Support Programs, and Mr. Roy Zimmerman, Associate Director for the Reactor Projects. We are here to brief the Commission on the operating reactor oversight program, as you said. The reactor oversight program is continuously evolving. Self-assessments, both external and internal, have identified problem areas which are being addressed by a number of actions. These actions include improvement in inspector training and guidance as well as our performance assessment . 5 methodology and criteria. These actions, combined with initiatives that were underway prior to the more recent review, will enhance the staff's ability to identify safety issues and assess licensee performance. In addition, the recent reorganization and appointment of the Deputy EDO for Regulatory Effectiveness, Mr. Ed Jordan, will play a prominent role in future oversight of the reactor program. This completes my opening remarks. I would like to turn the briefing over to Frank Gillespie, who will do the major portion of the briefing dealing with inspection efforts, and Mr. Zimmerman will address those portions dealing with the project management. CHAIRMAN JACKSON: Thank you. MR. GILLESPIE: Since Hugh has already covered the objectives, if I could go right to slide three, which is just to outline the organizational structure of the overall program. The focus today is on the inspection program. The other presentations next week are really going to touch upon licensing aspects and other aspects. [Slide.] MR. GILLESPIE: NRR as the headquarters function develops policies and procedures, but we are by our own procedures very closely integrated comment-wise at every . 6 step with the regional offices because they are the implementers of the program and they do the chief implementation and the real inspection. The regions do implement the majority of the program. Other organizations provide independent review. The new Deputy Executive Director position is expected not only to continue doing what AEOD did with the lessons learned from DETs and lessons learned from case studies, but that will also be very much, we would expect, enhanced in our interface with them. CHAIRMAN JACKSON: Let me ask you a question. What are some of the feedback processes to which you refer, or feedback mechanisms in the second bullet? MR. GILLESPIE: The regions are not hesitant to tell us that we have made an error. We get feedback on the ability to carry out a procedure where we actually in coordination with the regions have published an inspection procedure and we are finding it's not getting an objective we need. We get feedback from the senior residents and residents. We periodically go out and talk to them individually and ask how are these aspects of the program going. A most recent example. I did a follow-up visit to Maine Yankee on the testing question: What did it take to . 7 find that one broken wire by that contact? What would it really take? How would that impact the program? Could we look in that much detail at every other facility? So we are going out trying to get direct feedback from the people carrying out the program, which will tend to lead us into maybe a more in-depth audit of what we are doing and how we are doing, and potentially then a rewrite of a procedure or re-promulgation of a manual directive. MR. ZIMMERMAN: If I can add a thought. When senior managers from the region and from headquarters also go out and visit the sites it is typical for them to have dialogues with licensee management as well and receive feedback at that time on their view on the inspection program as well. CHAIRMAN JACKSON: But you don't have formalized feedback processes as such. MR. MIRAGLIA: We have both, Madam Chairman. I think what Frank and Roy were talking about are somewhat informal. As part of the agency response to the reg impact survey, there are a number of mechanisms for getting feedback: requiring the inspectors' supervision to get out to the facility and explore with licensing managers their perspective of how the program is going, strengths and weaknesses, and there is a feedback mechanism. Is it an annual report now, Frank, to the . 8 Commission on some of those aspects as well? MR. GILLESPIE: Yes. MR. MIRAGLIA: So we have formal processes as well. CHAIRMAN JACKSON: But then you say other organizations provide independent review. What does this mean? Is this headquarters organizations reviewing the regions? MR. MIRAGLIA: I think Frank's intent and the intent on that slide is to say that we have other independent audits from, say, the GAO, our own IG, that also provide oversight, and we look at that, and that's a feedback mechanism. CHAIRMAN JACKSON: We do have this new Deputy Executive Director for Regulatory Effectiveness, et cetera, and you said that this organization will play a prominent role. Do you intend to have some kind of internal assessment? There is self-assessment by the organizations themselves, but do you intend for this organization to provide -- MR. THOMPSON: Joe Callan and I have discussed some of the ways we would do our own assessments. As you know, in the materials area we do kind of impact program review where we go out and review against a set of criteria the regional offices. We are looking at potentially using . 9 that model as an approach to go by and look at the regional offices with respect to the reactor program. We have not formalized anything specifically on that yet. I am looking forward to Joe coming up here so I can be looking at things like that. I have not had an opportunity recently to look at those type activities. MR. MIRAGLIA: But we do have our own self-assessments. I think internally in the program we would go out and look at our program and self-assessments. We have done that on a periodic basis. That would include a visit to a facility or to facilities within a region, discussions with the residents, discussions with the licensees, and looking at headquarters. So we do have those kinds of assessments as well, and I would think as a program office that we would continue to have self-assessments based upon routine self-assessments, and if there are indications from other places that we have weaknesses, that we would try to go out and look at them. CHAIRMAN JACKSON: It strikes me -- and I think Commissioner Diaz has a comment or a question -- that what you have essentially described is a mechanism where the program offices assess what goes on in the regions, but the program offices themselves do those self-assessments. The question is, is there any regularized way of having a . 10 quasi-independent review of the program offices' role in reactor oversight? MR. THOMPSON: I believe the answer to that is yes. In the materials program we do look at some of the activities in headquarters as part of the overall programmatic review, kind of doing self-assessments with those same criteria we use to evaluate the regions and the agreement states. CHAIRMAN JACKSON: All I am really talking about is consistency of how assessments are done. There is self-assessment and then there is assessment by others, peer assessments. So you have described a food chain where everyone self-assesses and everybody is assessed except the program offices in terms of someone else assessing them. MR. THOMPSON: Obviously the Deputy EDO for Regulatory Effectiveness is a -- CHAIRMAN JACKSON: That's the intent here. MR. THOMPSON: That would not be the only thing. We would obviously be doing some of our own assessment. I don't want to speak for Ed. He can obviously speak well for himself, but that is the intent. That organization would also look specifically at how the program offices are carrying out their functions, and that program is being developed now by Ed and his staff. CHAIRMAN JACKSON: Commissioner Diaz. . 11 COMMISSIONER DIAZ: Just a comment on Mr. Gillespie's statement that the regions are very candid in providing criticism of the office. I hope that is not an open loop, that you close the loop by being equally critical of the regions. In this way you actually close that loop in the feedback, but it was not implied in your statement. They actually do come back to the regions; is that correct? MR. GILLESPIE: Yes, it does. CHAIRMAN JACKSON: Commissioner Rogers. COMMISSIONER ROGERS: On that word "oversight" in the Deputy Executive Director of Regulatory Effectiveness, do you really mean assessment there? Is that what you mean by oversight? MR. THOMPSON: That is my sense of it. It is not any kind of day-to-day oversight activities. COMMISSIONER ROGERS: You really mean assessment when you say "future oversight"? MR. THOMPSON: Right. I think that's just the title. COMMISSIONER ROGERS: It's just a question of what the relationships are between the organizations. CHAIRMAN JACKSON: You mean assessment. MR. THOMPSON: Yes. [Slide.] MR. GILLESPIE: If we go to slide four, which we . 12 have basically covered, we do have a formal process of assessing the regions with a team of people going in each year with predetermined criteria. We in fact generate a written report which we supply to the region for comment. We do this every spring. We also are assessing not only the region but ourselves when we are looking at it, because inevitably you will find two types of problems when you are looking at someone implementing your program. One is they have implemented the program and it is not the program you thought you wrote, and therefore you have program problems where clarifying guidance is needed or making some adjustments. On the other part you might just have problems with implementation. We do visit, as Frank said, several facilities, interview people in the regions, the division directors, and we are getting two kinds of feedback. One is programmatic feedback and one is on the direct implementation in that region itself. And we compare across regions to look at consistency in implementation to see ourselves if we are getting the depth and scope we thought we should get. That is done once a year with every region, and we do share the results. We are very open with them on that. We do special reviews also where we see particular . 13 problems. An example would be we saw particular problems in how we as an agency were dealing in inspection space with safety assessment quality verification. So we did a look at safety assessment quality verification, how we are carrying it out, is it meeting the objectives, and we are now taking some corrective actions on that, and in fact recently issued some changes to inspection procedures where basically we felt that the program was being carried out more on a piecemeal basis. And the program allowed that. The procedure allowed different pieces to be just added up to say we meet the whole, but from recent inspection results we are finding that when you had a group of people go out and do the procedure, all at once you had a much better picture of safety assessment quality verification of a licensee, and that came through in some inspections we did focusing on this at both Fermi and Crystal River. So not only do we do the annual audit, but we do special audits where we see a problem starting to evolve or have some insights from some other source that we need to look at it. Then we also do follow-up on that to see that we are achieving the end we thought we got with it. That is done on a continuing basis. We do do direct observations with SALP, PPRs. This all comes from the same basic branch and organizational . 14 unit, and it is all integrated together. From those observations you tend to get the special reviews. CHAIRMAN JACKSON: Who actually performs the observations? MR. GILLESPIE: They are all coming out of the inspection program branch. That is our primary focus. We also have all the managers. Basically division directors and above in NRR periodically go out and sit in on SALP meetings as observers to see how they are going on. We get feedback also from that process and we get feedback because anyone who is a SALP board chairman in any region is supposed to observe a SALP board in a different region. So we also get feedback from them. We have got it covered from different layers. When the information coming back from these people starts to correlate together, then we are off taking action. CHAIRMAN JACKSON: How do you solicit feedback from industry and the public? MR. GILLESPIE: As Frank said, we have a formal process in a form. It is typically dominated by the regional managers who are going out to see what the residents are doing or visiting the site which asks them to sit down one on one with a licensee representative and basically try to ask him, does he have a problem? CHAIRMAN JACKSON: Is this form used on a . 15 consistent basis? MR. GILLESPIE: I would not want to say it is used on a consistent basis, but it is used with regulatarity. I'm not trying to evade your question, but it is spotty how we get it back. We annually put a Commission paper together and report the results of what comes out of those reported forms. At a higher level is the EDO's process, and each region has a process for very senior managers at nuclear facilities to call in if they have complaints about actions taken by residents or inappropriate actions or any problems at all with how we are carrying out our program. CHAIRMAN JACKSON: Is our solicitation of feedback from the public as formalized? MR. GILLESPIE: No, it's not. CHAIRMAN JACKSON: Are you thinking about that? MR. GILLESPIE: I hadn't been. It is not. MR. MIRAGLIA: It has been spotty. MR. GILLESPIE: It has been spotty. MR. MIRAGLIA: There have been broad procedures such as the maintenance inspection procedures. There were workshops. On some occasions we have had broad workshops on those kinds of activities. MR. GILLESPIE: Out of fairness, the feedback from the public we get here is -- . 16 CHAIRMAN JACKSON: And newspaper articles. MR. GILLESPIE: No. Actually, it's a little more structured than that. It is things like the UCS write in and we get appraisals from them. They are happy to appraise us. There is a person at the University of Syracuse who quite frequently gets computer runs. CHAIRMAN JACKSON: The real answer is that at this point in time it's not formal. MR. GILLESPIE: It's not formal. MR. MIRAGLIA: It's not formal. COMMISSIONER DIAZ: Would you say that the feedback from the licensees, the industry is consistent in the sense that if there is a problem they do come and say "I have a problem with this"? Is that done on a consistent basis? Is that dialogue established and that feedback mechanism clear and unimpeded? MR. GILLESPIE: I personally believe it's very unimpeded and it's very clear. What I would have difficulty answering is, does any particular utility have an inhibition for using it? I think on our side we hope we have destroyed any blocks to someone calling up. In fact, only last week we had an immediate call where an inspector said something on site which was probably out of context and out of line. The call came in immediately, and we have taken care of the . 17 situation. For the majority of licensees, I think it's both clear and unimpeded. MR. MIRAGLIA: As a matter of context and history, the form and the feedback mechanism of managers going out was a result of the regulatory impact survey that we conducted back in 1989 and 1990. One of the corrective actions of that is there was a concern -- industry voiced that concern -- about the retribution issues, and there is reluctance to complain because of their input to SALP and other things that affect licensee performance. So there is a reluctance. That issue is out there. As a result of that survey, we took the corrective actions of having the survey form, the periodic report to the Commission. We had training sessions for our inspectors. We have upgraded the fundamental of inspections. There were a number of corrective actions. Notwithstanding those activities, the Tower report came out in 1992 and raised the issue again of concern and reluctance to bring forth those kinds of issues to the agency. As a result of that, another level was put in, which is the management implications group where senior managers from the utilities were encouraged to call to the EDO and the EDO staff to raise issues at that level. That . 18 has been used and that has been in place for about a year or two years and has been used in a limited kind of way. MR. THOMPSON: We don't get lots of calls in that context. I think each utility executive as they have their communications links established with the regional staff and with headquarters staff probably make their own judgment as to whether they raise issues and select what issues they raise. I think if there is a truly serious issue, they don't hesitate to raise it. It is those that are kind of borderline, would be my guess, that some may have reluctance to raise. COMMISSIONER DIAZ: Are you satisfied that we have provided guidance at every level that it is important and necessary for the licensee to communicate reservations that they have and that is all we can do? MR. MIRAGLIA: The answer is yes. We say there are many vehicles. We encourage the licensees to use those vehicles, and for these other cases here's another mechanism to be used. COMMISSIONER DIAZ: Thank you. MR. GILLESPIE: Going to slide five. [Slide.] MR. GILLESPIE: There is almost a continuous, ongoing relationship between the Inspector General and the program, because there is almost always some aspect where 50 . 19 percent of the agency is being reviewed. So there is actually a very good relationship in the suggestions made by the Inspector General on the program as they look at different aspects and to try to understand the interrelationships between one recommendation that they make on one more limited scope investigation or audit as it might be applied across the board. The General Accounting Office, that was only yesterday. We always seem to have a GAO audit of some kind going on. In fact, one of the people working on the one that is going on right now was also on the South Texas one. So we have periodic feedback almost of a constant assessment nature coming from GAO also. Industry and public feedback. Predominantly industry feedback. Public feedback is very indirect. It tends to be different groups that come in. It is not systematically collected. [Slide.] MR. GILLESPIE: Going to slide six, this list in bullet form are the issues I was going to try to cover today to give some sense of what we are doing to improve the program in both a problem statement in each of these areas and some idea of short- and long-term action, in some cases a short-term action having already been achieved. [Slide.] . 20 MR. GILLESPIE: Slide seven, the identification of design issues. We have broken down the problem into several pieces here to somewhat try to align with the actions we are taking. Insufficient design considerations in the inspection program and lack of design of expertise in the staff. I will split that as I discuss it into both a long-term and the front part of that problem that we are trying to deal with a little more in the short term. The last part of that sentence is a longer term aspect. Lack of systematic incorporation of design inspections into the core or our regular inspection program so that we have a repetitive nature of going back with some periodicity and looking at design. CHAIRMAN JACKSON: Is there an inspection module that is in fact focused on design inspection? MR. GILLESPIE: Yes, there is. MR. MIRAGLIA: There is one in development. MR. GILLESPIE: We have one in development. We do have an inspection module, several in fact, that are specifically focused on design, but they are not part of the mandated core program. Now we are looking at how to make that digestible and actually get it into mandatory program so that we come up with a periodic and systematic look at it. It is right now a procedure that is used more in a . 21 reactive nature when we have some insights that there is a problem. The next one is insufficient information from licensees on the status of design basis reconstitution. It goes down to the 50.54(f) letter which we have recently put out, and we are starting to get responses back. And lack of availability of design information, which is something we see when we do visit the sites, the ease of even the licensee coming up with his design information for us to review. CHAIRMAN JACKSON: Who performs engineering inspections in the regions? MR. GILLESPIE: We start from the top down. There are two divisions in each region. There is a Projects Division and then there is the Reactor Support Division, which is basically the engineering division. So engineering inspections are really focused on regional specialists. Our residents are more generalists. We are focusing and are continuing to keep them focused on day-to- day operations where the idea that avoiding a transient, the idea of stable operations is a very important aspect of what we do and is a significant contributor to the improved industry performance as we look at the performance indicators that have come down so much in the last ten years. So we are really dealing with the engineering . 22 inspections being done out of the regions with regional specialists in this area. That will get me to my last bullet on long-term actions, the basic tools that we have been able to give these people. The other piece is we tend to supplement these people with contractors, which I will discuss as I go through the short-term actions here on the first bullet. CHAIRMAN JACKSON: We have decided that is the best way to proceed? MR. GILLESPIE: I think as a short-term decision we have decided we definitely -- lessons learned from everything we have seen is, if we don't look at something, then there is a chance that the industry is going to back off and not look at it. So initially, and not wanting to lose the gains we have made in operational safety, we have maintained that and we are trying to maintain that resident core of generalists focused on operational safety. In addition, the types of problems that have come out of Haddam Neck, that have come out of Millstone and Maine Yankee and Crystal River really were identified by the narrow specialists taking the time to really dig into something. So the nature of the problem that we are finding in the design area calls for a different kind of person than the resident also by background, training and by current . 23 capabilities. COMMISSIONER DIAZ: Using the language I will use with the licensee when we assess an area, we find strengths and weaknesses. Would you quickly tell us what is the strength that we have in our identification of the sites? What are the main strengths where we say we can do this? I know you have identified the problems, but what are our strengths? What are our capabilities? MR. GILLESPIE: I think our strength is definitely operational safety. CHAIRMAN JACKSON: No. He's asking what are the strengths in the design area. COMMISSIONER DIAZ: In this design issue. MR. GILLESPIE: Let me jump to some of the results. I believe the EDO recently sent a short note up on the result of the architect/engineer inspections. Actually the results are very consistent with results of past safety system inspections, which are very design-oriented, that we have done. They are not atypical type results. The strengths we are finding -- CHAIRMAN JACKSON: Identify whether the strengths you are finding are licensee strengths or our strengths. He's interested in our strengths. COMMISSIONER DIAZ: You said we have these problem areas. What are our strengths? . 24 MR. GILLESPIE: Our strengths right now as we are carrying the program out, I think is using the, quite honestly, limited number of people with design background that we have as team leaders to organize how we leverage ourselves, which is my first bullet here on the architect/engineer inspections. We have leveraged ourselves tremendously with contractors on these inspections. MR. MIRAGLIA: I think maybe your question is more our strengths and our focus has shifted to operational safety and that is where our strengths are. The residents have that operational focus. Their training is in the operational safety focus. At one point in time, when we were doing lots of licensing, lots of construction, we had more design specialists within the context of the agency, and we are drawing on those now. CHAIRMAN JACKSON: I think the simple way to answer the question is, do we have any strengths in that area? MR. MIRAGLIA: In design? CHAIRMAN JACKSON: Yes. MR. MIRAGLIA: Yes. CHAIRMAN JACKSON: What are they? MR. MIRAGLIA: I think we have a large number of qualified engineers who have worked at AEs, but they are smaller in numbers than they were perhaps years ago. The . 25 question is, how can we use that smaller amount and how could we augment those? COMMISSIONER DIAZ: Are we getting them somehow so they become a strength? MR. MIRAGLIA: Yes. And the longer term issue is, if we need to get more, how do we -- the training. You will hear we have to get there from here. CHAIRMAN JACKSON: Maybe a strength is that we have a few good men or women. MR. MIRAGLIA: Yes. CHAIRMAN JACKSON: Is that what you are telling us? I don't want to put words in your mouth. That you have learned how to leverage them through these architect/engineer enhanced inspection teams? Am I putting words in your mouth? MR. MIRAGLIA: That's absolutely correct. CHAIRMAN JACKSON: I don't know if that answers the question. COMMISSIONER DIAZ: That certainly answers my question. CHAIRMAN JACKSON: He wanted it from you. MR. MIRAGLIA: Yes. COMMISSIONER DIAZ: The identification of problems must come from a strength. I think it is very important that we assess our strength even before we assess our . 26 weaknesses. MR. MIRAGLIA: Yes. As an indication of that, Commissioner Diaz, some of the special teams that we sent had those strengths on it to look at the design areas. In terms of the special teams that we sent to Millstone and Haddam Neck, special teams that we sent to Dresden, and the special teams that we sent to Maine Yankee, they had NRC specialists in that area and we augmented those as well with contractors. So, yes, we have. We are not totally absent of that strength within the context of the staff. MR. ZIMMERMAN: I would add that in addition we guide those contractors and take their findings that may be raw and put them into regulatory context and determinations on whether enforcement is appropriate or not. We deal a lot with the raw findings that the contractors provide. COMMISSIONER DIAZ: I see that as a strength. It should be clearly spelled out as a definite strength. COMMISSIONER ROGERS: Before we leave that, you say we put these into regulatory context. Does that involve a relative risk consideration? MR. ZIMMERMAN: I would say yes, and we are getting better with the training that we are providing to our staff, our ability to use risk insights and also know their limitations are improving. So we try to always look at things from a risk perspective, take the findings and . 27 marry the two, but there is clearly room for us to continue to grow in that area. MR. MIRAGLIA: And I think risk is a factor in a number of areas, Commissioner Rogers. In terms of deciding what systems to inspect, we try to use the risk informed insights to look at those issues. In terms of regulatory space, one has to try to say safety significance, risk significance. That is a part of the enforcement policy, and we are looking at ways of enhancing the enforcement policy with respect to risk significance. So risk is a factor and an element at the various stages of the regulatory process. CHAIRMAN JACKSON: I think Commissioner McGaffigan has a question. COMMISSIONER McGAFFIGAN: One of the things Mr. Gillespie just said, and I took it down, is if we don't look at something, industry will tend to back off and not look at it. That's a fairly profound insight and it says something about risk. Maybe you could say something about the compounding of things. If we only look at the most safety-significant, the most risk-significant systems and they were working perfectly, can things fall apart by looking at non-risk-ignificant things that compound and turn out to be more safety-significant? . 28 MR. MIRAGLIA: That's a challenge. The challenge is another balance. I think we cannot focus strictly on risk-significant things. As an example, spent fuel pool cooling, which was an issue that got pulled at Maine Yankee. From a risk- significant point of view, we perhaps should have been paying more attention to that area, and we were focused on other areas. That's a lesson learned. There has to be a balance not only between operations and safety, but you have to have enough in your program that is sampling enough areas, both risk-significant and perhaps less risk-significant, but providing insights to the robustness of licensees' programs and their ability to comply with the regulations. We have to look at both, and that's another balance that is going to have to be looked at in terms of the inspection program. It can't be absent and just risk dominant only. You have to look at a little bit of everything. MR. GILLESPIE: We've got a bullet that isn't in my viewgraphs. I will try by way of example to illustrate. One of the things that we are looking at is the basic inspection procedure that most of the resident activities are covered under in operations. We require the residents to do some things, and we will hear anecdotally back from the residents: Why am I looking at this? I never find . 29 anything wrong. We have tried to maintain this balance that Frank talked about in the program of touching all the necessary activities and relate that same philosophy to the engineering of the facility and just take a risk analysis and recognize that some of the major weaknesses in our risk analysis are the cumulative assumptions that went into it. Every resident and every senior resident basically some place in their office will have the IPE chart that has been generated on risk-significant system, least to most to most to least risk-significant system. Yet when you look at that, if you spent all your time on the top two systems, you might have a highly risk-significant problem because the support systems that are assumed to work are going to be flawed. It is clearly not, to me, a simplistic argument. In achieving this balance people sometimes take us to the extreme. Well, that is not a risk-significant system. Yes, but HVAC happens to be a very important support system, if you have an accident, to maintain the environment in the control room. Other things that we might be looking at, for example, the maintenance rule, is looking at the reliability of a particular component. Well, there are some assumptions in the risk analysis that reliability of that component is . 30 X. If the reliability isn't X, then it trickles through the whole thing. It may in fact be almost as important to look at the underlying assumptions in the risk analysis and make sure the assumptions are valid as it is to look at the specific most risk-important system it identifies. There is a real balance there. So we have got to be real careful that we approach this with the same type of balance of what is the information telling us, what are the assumptions. I'll use the example of the sleeping operators at Peach Bottom. If we weren't going into the control room basically each day with a resident or a senior resident as part of their tour observing staff turnover, we might not have seen that. You will find very few items of noncompliance or statistics generated from control room observations, but it's a very important element of what we think an inspector does each da, particularly the resident. That same principle needs to be carried over into engineering. Right now we are basically looking at the most risk-significant systems, and quite honestly, we try to pick the two or three systems that are risk-significant to look at and analyze. I recognize that we may have a flaw in our system, because we are not necessarily looking at as much detail in those systems that might get ignored. COMMISSIONER DIAZ: What you are really saying is . 31 that the entire process falls in the area of systematics. In other words, it is this system of risk you are going to be working with rather than the risk of any one component or area. MR. GILLESPIE: Yes. It's a compounding just because we have a complex machine and everything is interrelated. So it's not a simple solution. You might say that the entire QA process in a risk analysis isn't modeled. It is all included in this little beta factor that the risk people like to put in front because we don't understand it real well. Yet the failure of a complete QA program would be very risk-significant in our opinion. The limitations of the mathematics give limitations to the applicability, and I think we have got to keep that insight in the back of our mind as we approach inspection. CHAIRMAN JACKSON: It strikes me that all of this is tied together. Commissioner McGaffigan spoke of where you could have cumulative effects. You spoke of looking at what you think may be the most risk-significant things but if you ignore some others, you may be ignoring the effect or the implication of that. MR. GILLESPIE: Exactly. CHAIRMAN JACKSON: Commissioner Diaz spoke of having systematics. So the whole thing requires an . 32 integrated and systematic approach that has risk insights built into it, because that helps with the systematics. I think that is all they are really saying. Sometimes we fall into traps where we think we are talking about something here or something here, and it just doesn't work that way, because in the end the plant is an integrated entity itself. MR. GILLESPIE: Jumping to the short-term actions, quit honestly the short-term actions we have significant leveraging of the staff. We have three AE teams with five contractors on a team; two team leaders assigned from the NRC so that we can keep the paperwork going when the team goes out again and we can maximize our use of the contractors. I think we are leveraged to the maximum extent possible. We also have a significant program where we supply engineering specialists to the regions when they have a particular problem to do a safety system functional inspection, which is our traditional reactive procedure for doing a vertical slice or really digging into a specific system, which we are also supplying. This did not push off the more routine efforts. This was clearly an addition to what we were already doing. Regional inspections of engineering. We are trying to address now getting this incorporated rather than in a reactive way, which it was before. I am talking about . 33 the engineering aspects of design versus when we were looking at engineering more traditionally. We were looking at engineering in support of operations: were they being responsive when the operator said he needed assistance? It was the responsiveness to operations, not maintaining the integrity of the basic design itself and the basis for that design. So it's a different aspect of engineering that we are looking at now. We are not throwing out looking at the other piece, but we are looking at this right now in the short time. It's all an addition. COMMISSIONER DIAZ: Where are the senior specialists? We place two in each region and two in here? MR. GILLESPIE: That's our senior risk analysts. I do have a whole separate slide on that. COMMISSIONER DIAZ: All right. CHAIRMAN JACKSON: Let me raise one other thing, which is basically a comment on your comment. I recall, if I can paraphrase him, Commissioner Rogers always emphasizing the point that it's not just looking at design per se, but you want to look at how design and design changes get propagated into operations and plant changes and into procedures. I assume that when you speak about that that in fact is what you are looking at. MR. GILLESPIE: Exactly. That's why I said I don't want to throw away what we have been looking at . 34 before, but what we are now saying is we are not taking for granted the engineering solution or responsiveness to the operational problems. Now we are saying, well, when you propose that solution, where did you get your numbers? Where did you get the pressure requirements? Where did you get these other things? More goes into the robustness of the solution. So we are going kind of backwards down the stream flow of information. Longer term actions -- CHAIRMAN JACKSON: Before you go, I can't let you go by the 50.54(f) letter follow-up. You have already worked out who is going to do that and that the resources are available? Is it going to be in the headquarters where the reviews are? MR. MIRAGLIA: We are having a conference call. Hugh, did you get it set for tomorrow? MR. THOMPSON: It is tomorrow afternoon. We will reach a final position with respect to how we will integrate the review process between the NRR and the regions in order to assure that the right people look at this in a timely fashion. CHAIRMAN JACKSON: Have you worked out what any follow-on regulatory actions would be? MR. MIRAGLIA: We are looking at those elements as . 35 well. We just started to get some of the input from the licensees. I think the due date was the 9th of February, and the letters are coming in. But yes, one would have to say, given those letters, what appropriate action should there be for follow-up inspection questions or enforcement, and that all has to be part of our evaluation. CHAIRMAN JACKSON: And you have developed criteria for review, that is, what it is you are looking for? MR. MIRAGLIA: Yes. COMMISSIONER ROGERS: Before we leave this, if we could back up one half bullet or so to the new inspection procedure, what are the dollars and FTE resources that are going to be required in addition to what we have devoted to our inspection activities to carry out this new inspection procedure? MR. GILLESPIE: The reason this is short term is we are actually trying to achieve that within some limits that we placed upon ourselves. Our first question is those limits, and the limits that we have placed upon ourselves really are this would be done with existing staff. So I would have a diversion, and at a budget level dollar support-wise for contractor support at about the same that we have right now. Right now, for this element of it we are spending at the rate of about $2 million a year for individual contractor support. . 36 If you would like, I could outline what we are looking at. COMMISSIONER ROGERS: I don't need it right now, but if you could supply us with that, it would be interesting to look at it. MR. GILLESPIE: Yes. CHAIRMAN JACKSON: Are you comfortable with the trade-offs? MR. GILLESPIE: For the short term I'm comfortable with the trade-offs. I'm speaking from a prejudiced position, because I'm seeing immediate gratification from the results of the architect/engineer inspections, what you might call the safety severity or safety significance of the findings. I feel comfortable right now with the trade-offs in that I am not giving everything away and I am not diverting from some of the things that we are doing that are well. Current engineering, which we had just beefed up last year relative to the resources and the core inspection program in engineering, we had raised the number of hours and adjusted it. We are trying to work within that volume. It would be the kind of approach that says if you haven't had anyone look at a specific system in detail, in the vertical slice kind of sense, in the last two years you haven't had one of the architect/engineer inspections, then . 37 you should consider them for this procedure. If you have, then you should carry out the old core procedure. What we would do is end up like on an every second SALP cycle first looking at engineering one way and then looking at engineering from a more design detail way. We are exploring that. That is conceptually how we are going. Will everyone be satisfied with the pace that that is? I'm not sure. I can only say this is the kind of thing we are exploring, because you have to have a baseline for people to comment on. MR. MIRAGLIA: This is the first attempt. In fact it's out for regional comment. The goal would be to utilize the resources from the region, augment from headquarters if necessary, or use contractors. When you use the regional resources, perhaps they are not doing something else that they would have been doing. So those are the trade-offs. And to really assess what that impact is. Or if we use headquarters people. These are the trade-offs that are yet to be determined. We are going to have to try the procedure, apply the procedure. CHAIRMAN JACKSON: Within the overall head count numbers, you had people who were doing design reviews relative to design certification. Is there any fungibility of people? MR. MIRAGLIA: Yes. What has also happened in . 38 that area is we have been ramping down in terms of those resources over the last few years as design certifications for the evolutionary designs have come to conclusion. There is only one design under active review. So those resources have been coming down and we have been applying those resources that were there, as well as our staffing levels have come down. CHAIRMAN JACKSON: You have been shifting them? MR. MIRAGLIA: Yes, and that's how we have been able to keep up with lots of the initiatives that we have had. The increment that is left from that to roll over to that is much smaller than it was two years ago. CHAIRMAN JACKSON: I understand. Okay, Frank. MR. GILLESPIE: Long term actions. We are going to have a presentation next week on Millstone lessons learned. The major policy question that comes up here is the licensing basis, what it's called, is it tabulated. In the simplest sense it might be. If an inspector was going to go out and do an inspection on the auxiliary feedwater system at a plant, does he have a list of all of the commitments and requirements for that system easily available to him? Right now the list is not easily available to either us or the licensee. That is going to be a major focus of our discussion next week. So I am going to . 39 try to not do it today. The other one is long term, the balance between operations and design inspections as it applies to our own staff. Quite honestly, I think it's very well known. I just mentioned we are putting about $2 million a year into contractor support for engineering type support for the regions on a reactive basis. The A and E inspections are budgeted at about $4.5 million a year to keep three teams going. That's $6.5 million. That's a big resource, and at some point we have to step back and say can we continue to pay for that for contractors or do we have to bring our own staff talent to bear on it and recruit people with maybe different skills than we had before, and how do we do that and how do we factor them in. That's why it's long term; it's not something we would do overnight, and requires some planning out, and a different resource base, a different recruiting base, quite honestly. Our focus for recruiting has very much been the resident type person operations, former operators. [Slide.] MR. GILLESPIE: The next topic is inspector qualification and training. This came up. The basic question the IG was asked: Are we following our own rules on how we are documenting our training? . 40 There were weaknesses found in our training documentation. Management expectations for performance-based inspections were not clearly communicated. Let me go through the list, if I could, and I would like to address the difference between performance-based inspection and inspecting performance-based rules. I find that our own inspectors get confused, and I would like to take a shot at at least trying to distinguish between those two. A combination of NRC organizational changes. We eliminated 40 section chiefs in the regions. What those section chiefs provided was an integration function, in a sense, of some of the results coming out of their individual reactors. That integration function in some cases got pushed down on the resident inspectors. You will hear resident inspectors say, I'm doing administrative tasks. What we have got is inspectors, seniors particularly, who are fulfilling two roles. There is information collection and then there is diagnostics. The traditional inspector really enjoys information collection and he may not enjoy diagnostics, and that is preparing for PPR, preparing SALP packages, preparing for senior management meetings. So we are really going into that right now and trying to understand it, which gets us to one of the other actions down here. . 41 Short term actions. We have republished our manual chapter. You will hear in the staff the acronym 1245, like everyone in the world is supposed to understand it. That is our Inspection Manual Chapter which covers training. It reflects our current policy on training, meaning it added some new training courses in. In addition, a significant change which is now going around the halls, because I held a training session for 250 people in NRR yesterday on this, is that it makes inspector classifications generic. What we found in the regions was that the regions very much comply generally with our program; they have inspector qual journals that are signed off by different people; it is very structured, very disciplined. In NRR we turned around and we have evolved to a program office with many, many people supporting the regions in doing inspections. In fact, we at least equate to one if not two division equivalents for the engineering divisions in the regions in supporting inspections in pieces here and there and everywhere when you add them up: fire protection, maintenance inspections, PMs filling in for seniors that go on leave or take vacations. My people do special team inspections, the architect/engineer inspections. What we have done is taken the word "region" out of inspector and now we are looking at making sure that . 42 people who carry out functions have the same qualifications independent of their geographic location. We are creeping into this right now. We are working with our partnership and we are working with NTU. You will see our new postings going up. Yesterday two postings went up in NRR which now have a new page on a position description that says you are expected to be qualified to one of these categories in 1245, and you may be called upon to have as much as 10 percent of your duties as supporting inspection activities. So we are trying to be open and honest with people, and we are getting into that now. This is a very big change. It's a cultural change within the organization in integrating the two functions of licensing and inspection together. A new inspector training course was developed. And now let me address performance-based inspection before I try to address inspection of performance-based rules. Performance-based inspection and the way we were viewing it -- and we started this back in 1988 when we published our first NUREG on this -- is basically an inspector looking at the system, understanding first what the system's function is supposed to be and first looking at the system as actually a hardware system installed and functioning and saying, is this system carrying out or . 43 capable of carrying out its basic function, and what does it have to do that? That would lead you hopefully into things like if you have a pump and the tank and the suction of the pump is on the same level. Do you have a problem with that positive suction head, yes or no? Then it leads you into possibly looking at the paper and tests. In fact, I'll use a success. The senior resident at Fermi, who has recently moved to Zion, asking this type of question, found a tank that should have been three levels higher for net positive suction head. He approached it this way, and in fact they had to nitrogen load the tank to make sure that this surge tank would work. That is a techniques course, because we hire people in with the technical expertise of being a good mechanical or electrical engineer, but we have not necessarily hired someone with the conceptual idea of how to approach a system. So while we had a performance-based inspection course already there, it wasn't good enough, and this is our next step, to try to provide input insights and conceptual insights on how to approach inspecting a system and trying to get at the root cause of the problem. It's a first step. As you can see, it's scheduled for March 10, 1997. But it builds on the way the . 44 contractors we hire actually approach a system: Will it remove the heat? Well, the first question on a heater exchanger is, look at the nameplate data. I think this was something that came out of Millstone. They looked at the nameplate data and found out the Btu rating was significantly lower than the expectation. That is a performance-based inspection independent of what the rule says. Inspecting a performance-based rule is more like what we are doing now on the maintenance inspections. The first go-around on the maintenance inspections is more programmatic, quite honestly, to make sure the system is in place that can deal with this: How are they using PRA? How are they classifying system? After the first go-around, which is really looking at the paper, then we will hopefully use our performance-based approach to inspection techniques on the systems when we are inspecting the rule. I am trying to make sure that it is clear. Performance-based inspection is not superimposing anything different on a licensee; it's how the inspector actually approaches his day-to-day work. We are going to continue to emphasize this. We are putting a new training course in place, and it's really a thought process rather than a cookbook. You can apply . 45 this conceptually to an electrical system, to a mechanical system, to an instrumentation and control system. Now we are trying to work to that point. It is not something I can write an inspection procedure for and use a checklist. It really is a mental process to try to get into the program. COMMISSIONER ROGERS: I like what I am hearing you say here. It sounds like a very sensible way to proceed. How much of this is all written down someplace that somebody can take a look at? What you are saying seems to make a lot of sense, but I haven't seen it written down anywhere in anything that has come before my eyes. CHAIRMAN JACKSON: Do you have it in terms of any of your inspection modules? Is it part of some training course? MR. GILLESPIE: As it happens, I had that same feeling, because I've been working in this. My mind has been working this way since I was an inspector. It is not written down in a real visible way, but I went back and got our performance-based inspections report, which was a NUREG that was completed in March of 1988. The words are still valid today. Have we pulled it together the way I articulated here in a policy document? No. COMMISSIONER ROGERS: I think that would be very useful to do. MR. GILLESPIE: We have put the training courses . 46 in place. What I have just articulated is the inspection begins -- I am reading from the abstract -- "the inspection begins with a performance-based observation and then the inspectors let discrepancies or uncertainties lead to the inspection of other areas such as quality verification, organizational . . ." And you start looking at the system, and then you go on. CHAIRMAN JACKSON: I would put it in the realm of more explicit guidance for your own people and ensuring that there is consistency between what is in the training courses and what is in that guidance. MR. MIRAGLIA: As Frank said, back in 1988-89 we started to talk with the inspectors, that you need to inspect performance as opposed to programs. We were articulating that in the broad context in the training program and in broad context within fundamental of inspection. Perhaps not everyone had the same mental picture of what we were talking about. CHAIRMAN JACKSON: The way Frank just explained it is very well articulated, and I think that articulation perhaps needs to be written down somewhere. COMMISSIONER ROGERS: I wouldn't want to just see this buried down in inspection manuals. I think this is a very important point of view that ought to be very clear as to how we view our activities here. Your connection between . 47 inspecting for performance versus inspecting against a performance-based rule and drawing that distinction in the context, I think that is very important, because we have to be clear on what it is we are talking about. I thought what you said was very good. It sounded excellent to me. CHAIRMAN JACKSON: We may want to see it surface as this is the NRC approach. COMMISSIONER ROGERS: This is our concept. CHAIRMAN JACKSON: Right. This is the concept. I think he is saying to elevate it, and we can promulgate it. MR. MIRAGLIA: We will certainly do that. COMMISSIONER DIAZ: I do agree that it sounds very good. I just want to make one observation of your points. We are talking about training and you said that inspectors are more prone to do information collection than to do diagnostics. I take that as a value judgment on your part. That might very well be true, but in training, to be able to streamline the information process you have to introduce diagnostics at that very first level. The fact is it is diagnostics that allow him to go from one component to a system. So it is critically important in the training that, even if they like more to do information collection, diagnostics is an indispensable element, and maybe we should put an emphasis on that in our training programs. . 48 MR. GILLESPIE: And we do. I was drawing a more distinct diagnostic line, and I will go through that when I have a diagram here. You often hear residents when you visit the sites saying, I spend too much time at my computer. I'm drawing the line of diagnostics at preparing for senior management meetings, preparing for plant. That is performance-based inspection as you just described. Yes, is the answer. I was drawing a different line just from what we are finding in interviewing people and the job task analysis is also finding that we are going around and doing, which is coming close to an end. CHAIRMAN JACKSON: Let me ask you this question. It's kind of a judgment question. Given what you have outlined in terms of this increased focus on definition of the distinction between doing performance-based inspection versus inspecting performance-based rules, have you been able to do a survey or detect how much of a problem it is in terms of a lack of understanding of that by our own inspectors? MR. MIRAGLIA: I think Frank referred to an IG audit that looked at our training. The maintenance rule audit brought the performance-based issue not being well understood into focus. CHAIRMAN JACKSON: How comprehensive was that audit? . 49 MR. MIRAGLIA: I'd have to go back and look at that report, Madam Chairman, but they talked to lots of inspectors and regional people and headquarters people. CHAIRMAN JACKSON: Does that square now with your own judgment? MR. GILLESPIE: Yes, it does. The people were focusing on the phrase, and the phrase "performance-based" had "rule" after it when it was used once and "inspection" after it when it was used again. Then you had the question of, well, why are we looking at the inspector question? Why are we looking at paper with the maintenance rule on the first go-around? Well, the first go-around you have to make sure that the processes and systems are in place by which you can judge performance. If an inspector sees a component failing, he then says, okay, that component is failing. Are you treating it correctly? That way the inspector is not trying to second-guess the system and the paper. He doesn't have to reinvent the look at the paper. We had to tell the inspectors, yes, this is still performance based, because you have to make sure the performance process is in place so that when you do performance-based inspection you don't have to go back and do that all over again. MR. MIRAGLIA: With some of these concepts I think . 50 it would be preferable for us to be cartoon type characters, so when you use a phrase you have the little balloon over everyone's head and everyone can say, what is the mental picture? I think what we will find on some of these concepts is we have talked past one another; we are not fully communicating because the definitions are a little bit different. I think that is an element that is here. Even if we have this broad statement, Commissioner Rogers, that is the start. It has to permeate the training programs; it has to permeate the communications. COMMISSIONER ROGERS: It shouldn't be so broad that it leaves out what you are seeing. MR. MIRAGLIA: I understand. That's the starting point, and then it has to permeate. CHAIRMAN JACKSON: You are talking about different tiers of the same thing. MR. MIRAGLIA: Yes. CHAIRMAN JACKSON: You can have the overarching thing that is some broad-based kind of point of view, but in the end, if it isn't inculcated into everything that helps our inspectors do what they need to do, then we haven't accomplished anything. MR. GILLESPIE: I felt very good when I was at Fermi and met with the senior and he showed me this, and I . 51 said, aha, I at least now know we have reached one. Long term we have a job task analysis which will be completed this March by the contractor. One of our questions was this question of diagnostics in the sense that I had used it earlier. We have made some changes to the organization. We shifted a lot of things around. What has been the impact? Are we really getting what we think we are getting? If someone is doing something, do they have the necessary training to do what they are doing? The split of work: what is done at a site and what is done at a region. That is all we are looking at in this job task analysis. It is not one of specific positions. It gets to that, but it is one of the division of reactor projects in the regional organizations, and we have looked at all four regions, because all four regions carry out the same program slightly differently and functions get allocated differently even within a region between branches. We are looking for the most efficient, effective process through the whole thing to deal with it: How should we split functions? Where should they be? Who should be in charge of it? Signature authority. CHAIRMAN JACKSON: What kind of time line have you laid out for yourself with milestones or for getting the whole thing done? . 52 MR. GILLESPIE: The contractor effort will be completed in March. I've got some informal insights back from them. Just like any anxious person, we call them up and say, what are you generally finding? We are generally finding some of the things I have actually already said. The level of discussion about time at the computer, probably there is more discussion to it than there is the actual time. We went out and put expert panels together and said, how much time do you really spend that goes with this anecdote? What we were trying to do is take all this anecdotal information that everyone had gotten and put it into a total perspective so that we are not underinflating or overinflating what a particular problem might be in the whole system. So weighting is what we are hoping to get out of this report to pull it all together. All the regions participated; headquarters participated; there were many, many meetings. So it's going to be, I think, a very beneficial rebaselining. MR. MIRAGLIA: The output from this is going to be useful in many ways. As Frank said, it's a function, an assessment of the function. We have many changes in the program or contemplate in the program improvements. We'll have to also say, given those changes, where are we today and what does that mean to assessment of functions and . 53 training? CHAIRMAN JACKSON: I guess it still is important that we get some sense of when you are hoping to really see some of this reflected in the inspection program. MR. GILLESPIE: It will be. I have no problem coming back once I get the report in hand and we can deal with it and get people's opinions again. CHAIRMAN JACKSON: All right. Is a similar thing being done for the project management organization? MR. MIRAGLIA: That brings us to the next slide. The Chairman is moving us along. MR. GILLESPIE: Roy is going to cover this slide and I get to have a rest. [Slide.] MR. ZIMMERMAN: There are areas for noted improvement with regard to project manager training and qualifications. The Inspector General issued a number of reports over the past year indicating areas where we needed to improve our activities as well as our own self-assessments. Although there were some findings where individual staff performance could be improved, mostly the areas where we needed improvement were for management to provide clear expectations to the project managers on what we wanted them to do and to provide the necessary training and tools to . 54 accomplish that job. To that end, in the short-term actions we issued expectations on the interactions with the regional office. It became very clear from lessons learned from Millstone and from Maine that there needs to be a very close linkage between the project manager, the resident inspector and the region-based inspectors. They need to talk often about activities that they are both working on. The inspectors in the field benefit from the insights that the project manager has from the licensing matters that they have been involved with both of a plant-specific nature and also the benefit of being here in headquarters and having a generic knowledge. CHAIRMAN JACKSON: How in fact are you implementing this shared safety responsibility? How are you ensuring or enforcing this integration? MR. ZIMMERMAN: This is one of the first key messages that we have articulated a number of times. I personally have done it as well as my direct reports. We have had discussions in workshops in the auditorium to go over the issue that our job is early identification of problems at facilities and declining performance, trying to find those problems early. That is not just the region's responsibility; that's all of our responsibility. CHAIRMAN JACKSON: How do you ensure that this . 55 kind of interaction and interaction with the desired outcome occurs in fact? MR. ZIMMERMAN: We have put a feedback mechanism in place. We have indicated that these phone calls should take place with the region no less often than three times a week. It is documented in our PM handbook. We have in our process improvement plan a feedback mechanism for the project directors to be able to come back and say that they have monitored these phone calls, they have reason to believe and to support their reasons that in fact we are having success in this area. It is not enough to put out the expectation; there is a need for feedback to test it, to see if it is actually occurring, and we have built that into our process. CHAIRMAN JACKSON: Have you also built that into your accountability and your expectations in terms of people's actual performance? MR. ZIMMERMAN: To a degree. We are moving in that direction. We are working on our elements and standards. We see there are some areas that need to be improved so that it is in line with the expectations that I'm going over. We need to do the necessary activities in a coordinated way with our partnership group. That is where we are headed. We are headed to element and standard changes to reflect the points that currently aren't in the . 56 standards. COMMISSIONER ROGERS: How many times per year would a project manager actually visit the project? MR. ZIMMERMAN: It varies, depending on the particular site and geography to an extent. Typically it's about four times a year. If it's a plan that is having more difficulties with us and there are more management meetings, SALP meetings, PPR meetings, I would expect the project manager and his supervisor to be making more visits to the site and to the regional office. COMMISSIONER ROGERS: Okay. MR. MIRAGLIA: There is broad guidance out on that in terms of the numbers that Roy indicated. There are reasons why that may change because of specifics on the project. COMMISSIONER ROGERS: I understand. MR. MIRAGLIA: There are those kinds of guidelines out there. MR. ZIMMERMAN: One of the lessons that we clearly learned out of Millstone and the partial core offload is that we have blind spots, and we have to do our best to avoid those blind spots via the communications between the project manager, the resident and the region-based asking for input from the region that we are closing a licensing matter; we are ready to approve this; and talking to the . 57 region and to the inspector in the field about that. They've had a chance to think about it. They can provide us thoughts that maybe we haven't thought of. We can provide insights to them on where they can inform inspections, that we have been looking at some activities here that make us think that this particular system might be one that warrants sampling during your next inspection. That is the coupling that has been there in the past, but we are trying to make it front and center heavy emphasis that that needs to be done. That was a clear lesson over the past year. Another area has to do with the fact that we have also issued expectations on maintaining the FSAR current. We have put out an internal requirement that the project managers are to update their version of the FSAR with the latest revision within 30 days after they receive that revision, and that we expect that that FSAR will be used as they carry out their licensing matters and other readily available portions of the current licensing basis. Again, not all portions of the CLB are readily available, but there should be a reasoned attempt to add to the FSAR as they go through looking at licensing actions and licensing activities. Many of the lessons that we have been learning over the past year are process related. We have talked . 58 about a few of those. We also had some challenges over the past year in the area of handling of allegations, errors that we have had in identification of allegers' identities inadvertently. Sensitivity to the way we treat and deal with allegers is an area that we have also had significant emphasis in our training to not repeat some of the problems that we have had in the past. Other issues that have come up -- and we track them actively in our process improvement plan -- get into areas of how do we handle E-mail that we get in. In this day and age we get a lot more E-mail than we used to get from various stakeholders. We have put out guidance about if it's a professional record that needs to be maintained, how it needs to be maintained; does it need to go into the PDR? Answering questions about, is it like a phone call or is it like a letter? We have come a long way in providing guidance in areas like that. Another lesson learned that we had was if a licensing manager were to call up a project manager and indicate that there is a concern that they are working on and "we just want to let you know about it." What is the expectation of the project manager? We have gone over the fact that we would expect that project manager would let his or her supervisor know, . 59 get the word to the region so they're aware of it, and mark it down, that that is an item that maybe is owned right now by the licensee, but we need to follow up on it in a reasonable amount of time. That's to stay on our radar screen. This was really separate from the Maine Yankee and Millstone lessons learned, but we recognized the benefit of rotating project managers on about a five-year period similar to what we do with resident inspectors. A fresh set of eyes is always good to have over a certain period of time. We have seen benefits clearly in the regions when this has been done. We have been doing it in my organization for probably the last six to eight months, and I will expect the same dividends. Take your knowledge and expertise to another facility. It helps consistency as well. CHAIRMAN JACKSON: When was this implemented? MR. ZIMMERMAN: I believe it was about six to eight months ago. In our process improvement plan we have the closure date. I have that with me and I would be glad to provide it to you. CHAIRMAN JACKSON: Do you assign more experienced PMs to more challenging facilities? Do you factor that in? MR. ZIMMERMAN: Yes. We are trying to put our . 60 stronger, deeper individuals on the plants that we find most challenging. I think we have made headway in that regard, but there is still more that we can do. The process improvement plan, just to mention it again. I believe a few months ago we sent a draft of that document to the Commission. It's a very active document. We are going to continue to add issues to that for us to track, big and small. It seems to be working well. We have a project manager, project director advisory board. As we look at making changes we provide those changes to this group, almost like an internal partnering group, to give them an opportunity before we update procedures: is there something from their vantage point, from their perspective that maybe we haven't thought of that we need to consider before we make the change? Because changes can be a challenge for people to adapt to. So we have tried to have a grass roots movement that I think has been very successful. In the long-term actions, we spoke a moment ago about the job task analysis out in the field with the residents. We are getting ready to undertake a similar activity here with the project managers. We are working currently on a statement of work that should be issued shortly. It's approximately a six-month effort, and we will likewise be monitoring it very closely. . 61 We are aware that AEOD is working on a knowledge, skills and abilities developmental activity, and we have had dialogue with AEOD. We want to coordinate closely. We see how this could be a good fit between what we are planning on doing and the larger agency action, and we will incorporate those. CHAIRMAN JACKSON: It seems to me one has to be a part of the other. MR. ZIMMERMAN: Right. There is a natural feed that occurs. MR. MIRAGLIA: We had some activities underway that were related. So there is a clear understanding to make sure that they do fit and there is no unnecessary duplication. MR. ZIMMERMAN: We try to minimize duplication and build off it. I guess my closing message on this would be that we are not waiting for the job task analysis, we are not waiting for the knowledge, skills and ability review. We will work in parallel as items come up. We are going to be looking for ways to improve our performance. CHAIRMAN JACKSON: The jobs task analysis for the PMs has already begun? MR. ZIMMERMAN: No, ma'am. The statement of work should go out within the next week or two. Right now we are . 62 hopeful that it will begin in March. It's the same contractor. MR. MIRAGLIA: It's the same contractor. So when they are finished with the residents. CHAIRMAN JACKSON: How long do you think again? MR. ZIMMERMAN: Six months. CHAIRMAN JACKSON: Thank you. [Slide.] MR. GILLESPIE: Going to the next slide, PRA applications in the inspection program. Here I want to both discuss where we are and the limitations of where we are based on our own regulatory structure and how we are approaching it and how it fits into the overall PRA implementation plan. We do recognize that there are areas for increased improvement in using PRA in the inspection program. The basic use over the most recent past has been in the prioritization of what we look at, particularly in design inspections, or when you select a sample. If you select a sample within what safety system am I going to look at for maintenance, what system am I going to look at for this observation or test, then that is where that one chart that seems to come out of all the IPEs generally will get used in the resident offices and by our team inspections. That could be said to be a simplistic . 63 use, but that dominates how we do use it for right now. The evolving use gets me into the short-term actions. CHAIRMAN JACKSON: Can you give us some examples or at least an example of where you see an opportunity for increased use of PRA in the inspection program? MR. MIRAGLIA: I think we are going to get into it. Frank is going to talk about the senior reactor analyst and trying to get someone with a little deeper knowledge in risk assessment, having two individuals within each region. Those individuals now are assisting and participating in the maintenance team. CHAIRMAN JACKSON: That tells me people. You gave an example of how we are using it today in terms of selection of safety systems to review. What else would you be using such people for? MR. GILLESPIE: I think what we are going to need to do in the future is be significantly more articulate and, to use a term that we have used here before, transparent on how we use risk relative to the balance between support systems and primary systems; how do we allocate things. The primary source of our information for enforcement is the inspector. As risk gets more inculcated into our enforcement policy as a measure of severity or weighting factor, it's actually the inspector who is the . 64 primary person writing down those initial words on what his observation was. So it is going to be important in the future as we get risk into things like the enforcement policy past what we look at, get involved in how much we look at it. But the inspector is the source. He is the guy that pulls it together. It all comes together at the bottom, quite honestly, rather than coming together at the top. MR. ZIMMERMAN: That's happening. We had a meeting just yesterday on a particular plant that we are considering escalated action on. We had the region on the phone, Office of Enforcement involved, and we were talking about what we do know from the IPE in this particular case, what insights does that give to us. So it is something that is being actively used now. MR. GILLESPIE: A specific example is, as we change our own rules, it's going to be the maintenance rule. In fact, the SRAs were actively participating in what I will call the risk element of the maintenance rule, which has become kind of a critical element: which systems are classified in what class, how are they characterized, because that leads to what actions will be taken later if the reliability values are found to be deficient. The other thing is now the inspector, once that . 65 rule is out there in place, process approved, is going to have make judgments as to, is this component failing consistent or inconsistent with the guidelines in place? That's one rule, and in fact the inspector is now going to be inspecting against that. Our approach here. What you see in the slide is two years ago when we proposed the senior reactor analyst program we had also proposed in the risk implementation plan the three level diagram on training. The SRA is the middle level diagram. It's a practitioner who is expected to apply something but not expected to necessarily have the skills to go out and completely do a PRA himself, to get that analytic view in the regions. He coincidentally hit, if you will, or got in place at the same time the maintenance rule was happening. It came together very nicely and they participated on them. In addition, to make sure that they did have the analytic background, 29 of the 39 presentations on the IPE results were given by the SRA when they were doing their rotation through Research, when they were giving it to both the PMs and the regional staff, to make sure they were familiar with the facilities that were going to be in their regions. So we really through the training program have tried to elevate them to that second level. I will go to the last bullet, which is the PRA . 66 development course for inspectors on long term. Basically what we have got is a course we are putting together with TTC which is going to treat PRA as a basic technology that should be known by each inspector and reviewer. Much the same way we have a GE or Westinghouse or B&W basic technology course we expect inspectors to have gone through, we are looking at a two- to three-week course which takes the individual courses and lessons learned from the SRAs going through it and comments and puts them together in a two- to three-week comprehensive course that is intended to be the bottom level of what was on the risk implementation plan. That's a person who knows enough about it and how it's applied so as he sees it applied in a 50.59 review he might be doing, as he sees it used in a continuing categorization of a failed component, or he sees it as part of the justification for a change to the plant, he has the wherewithal to either make a judgment that he can say, yes, that looks okay, or "I need to call the SRA who is my backup back at the region who is at the next level in the thought process and the reference period." So what we are trying to do is have the people in place with the talents to keep pace with the change to what they are inspecting to, because, quite honestly, we still have to inspect to the rules that are in place today and the . 67 license that is in place today. But as that becomes more risk oriented or risk has an application of compliance or inspectability, we have to really have the people there with the tools. That is where our focus is now, getting them prepared. We were right there at the right time with the SRAs, which was a very nice coincidence to have occurred. I would like to say we planned it that way two years ago, but it worked out very, very well. CHAIRMAN JACKSON: There are some pilots going on on risk informed tech specs, ISI, IST. Do we expect those to have -- MR. MIRAGLIA: As those programs are developed, then we will have to say if the rules are changing in that area, then that has impact on how we have to train and how we have to prepare the people. MR. GILLESPIE: Absolutely. The engineering inspector in the region who traditionally inspects ISI now has to have this other tool in his background. Once you say the program is okay, the implementation of the program is left to that engineering inspector in the region. He is now inspecting a different technology, a different set of records, a different approach, and making judgments as to reduced frequencies because of lack of flaws found or increased frequencies: are they proportional? . 68 It's as much reliability engineering as PRA in the aspects that the inspector is looking at. So we think we can give the inspector that level of tools, although we are not making him a full-fledged analyst. That is where we are going with that, and we are trying to keep pace. The next slides really come together totally. We call it performance assessment process. [Slide.] MR. GILLESPIE: The short-term actions as we are looking at improving our performance assessment process. This has been a particular interest to the Commission for a time now. It has also been an interest to us. We would like to improve ourselves. We have issued some management directives which very much try to document what we are doing today. There is a diagram on one of these management directives which has lines going all over the place. My simplified view of that is actually the last picture on this whole package, and it's listed as backup slide. [Slide.] MR. GILLESPIE: It becomes very important only because what we have done is we have set up our system with the plant issues matrix to mesh the plant issues matrix report. Now what we are doing is identifying where do the . 69 facts of the situation of our process come from. This diagram tries to show that the inspection report is where everything really starts to come together; every LER is followed up on; AITs have results, and those are followed up on. Inspection reports become not the only source but a significant source of assessment information. After that -- and this gets to the root of what I was calling diagnostics before -- you are basically dealing with the same database but you are refining it into different levels. The PPR uses the same information; the senior management meeting uses the same information. Every three PPRs you do a SALP. It uses basically the same roll-up of information. This will introduce me to my second to last slide, which is, how could we use technology now to maybe help us in this as the agency comes out of an era of the early 1980s of technology or information management, which was really document management, into the technology of the late 1990s where we are going to try to use information management to help us do this and to get more diagnostic, more transparent, if you will, into the whole process? So the short-term action really here in these management directives was to document what we are doing today in a way that people could understand it. We also then have the Arthur Andersen report that . 70 needs to be considered and how does that factor in. I think what you can see is that leads to this idea that we have to be able to digest more information more rapidly to do it efficiently and not get inundated with the weight of our own paper in our inspection reports. We need a rapid indexing to the information. The next diagram, which is a pictorial, while it says future assessment model, this actually lays out the information sources that we currently have in the current system. The problem with our information in the current system is that they are basically disconnected. If I want to know what were all the items in our compliance found at a Westinghouse four loop plant, I've got one computer that tells me what are all the Westinghouse four loop plants, but I've got another computer that tells me what all the items on our compliance are, and while the docket number should be able to relate this information together, the two computers don't talk to each other. So you get a manual printout from here and a manual printout from here, and immediately you are in a system that is so awkward we don't do it. CHAIRMAN JACKSON: Is IRM in the room? MR. MIRAGLIA: No. MR. GILLESPIE: What this is trying to do is be a pictorial. What we are looking at in the inspection report . 71 half on the plant issues matrix that was developed and has evolved is as an index to the details in the reports. This will both be technology and program. We are going to be coordinating changes to our manual chapters and process to describe how we do things with the introduction of the technology because it is going to change what an inspection report looks like from today. For us this is a relatively high risk operation because we have never done this before. It is pulling an awful lot of information together from a lot of dislocated places right now to try to put it into the same useful form. Which gets me to the next slide, because the key is being able to access and sort information rapidly. [Slide.] MR. GILLESPIE: Reporting and information technology is, I think, going to be key to our success here. If I could describe this in a way that Commissioner Diaz explained it to me, he saw the diagram, and he said, I'm glad you got rid of the triangle on the backup diagram. Once you identify where is your basic factual base coming from, then you can get a sense that everything after that is a different sort on the same information. Now the question is, how do we want to sort the information? If I can relate the facility's characteristics to the docket number, which gets me a relationship then to the . 72 list, then I can do peer plants much the way AEOD does their performance indicators right now, and I can look for correlations; I can do a lot of diagnostics that I can't do today. Not because we don't have the information, but because we have so much information. The ability to diagnose, sort and filter the information is extremely tedious and manpower intensive. That is one of the complaints from the residents. They are doing what they are doing today manually. They are rehashing it for the PPR; they are rehashing the same stuff for the SALP. If we can come up with a way of doing the cutting and pasting in the electronic form rather than the manual form -- in fact, Region II even graphs these on graph paper with pencil and ruler manually. CHAIRMAN JACKSON: You've got to be kidding. MR. GILLESPIE: I have to automate it. We have to come out of docket management into information management, and it's a 20-year leap in technology for us. CHAIRMAN JACKSON: And you are going to do it in two years? MR. GILLESPIE: We are hoping to do it within two years. That is what our commitment is. We are hoping to prototype this in a full region within 12 months. I'm going to give it one heck of a try. CHAIRMAN JACKSON: Twelve months. I'm writing . 73 this down. MR. GILLESPIE: We have two what I will call small prototypes that we have actually tried in our own LAN environment already with regional participants. With time I am gaining more confidence that it's fairly doable. We found some glitches in the whole system, things you would think smart people -- we're all engineers -- should know. In the last two years we have created a number of databases and we were smart enough to use the same kind of off-the-shelf commercial software, but we weren't smart enough to label everything the same. So docket numbers in one system are zero five zero dash something; in another system it's five zero dash something; and they don't even talk to each other. CHAIRMAN JACKSON: This is where the CIO organization should be helping you out, because your jobs are one thing. The best way to handle information is another thing. MR. GILLESPIE: If I could jump into the conclusion on 14. [Slide.] MR. MIRAGLIA: I think, Frank, we ought to say we have been coordinating with IRM. MR. GILLESPIE: We have joint project managers with IRM and my organization and people working full time in . 74 both organizations to try to make this work. Actually we have no conflicts. The conflicts are technological ones. CHAIRMAN JACKSON: All right. MR. GILLESPIE: The conclusions. We have an evolving program, particularly as it applies to inspection. One of the conclusions and one of the strengths, I would say, of the inspection program which we have been talking about is the discipline of having an inspection manual in some detail and generally having a baseline to change. That includes detail in training, detail in how we do things. If we have an unusable procedure, at least we can point to it and know what we have to change. That is a strength, the discipline we have approached it with. Both self-assessments and external reviews have led to program improvements. I would not want to claim either one is stronger than the other. They tend to fit together and actually be self-supporting. Future direction will be closely coordinated and integrated. This is this tying of the diagnostic and the inspection and a recognition of how it's the same group of people doing it. So we have to give them the tools to do it as best we can. The balance between operations and design is going to be with us long term as we make decisions: Is it staff? Is it contractor? And out of the whole how much is it? . 75 What is the safety significance? How does it queue up? Performance assessment I've touched upon. Our real challenge is how do we use technology to try to get us this giant leap forward in the technology area. With that, I thank you. CHAIRMAN JACKSON: Thank you. Commissioner Rogers. COMMISSIONER ROGERS: I think this is very interesting. A lot of questions I had when I walked in here got answered, but let me touch on a couple of points. On your performance assessment backup slide, one of the problems that I have with this is that it looks like it's a totally event driven process. [Slide.] COMMISSIONER ROGERS: It starts with an occurrence observation or allegation. It looks like there is something external to what we do that starts a process. In other words, it looks like it's event driven and not initiated in a certain sense by us. MR. GILLESPIE: That's my drawing problem. The observation piece is our routine program. We have got 434 people out there who are for the most part working the observation aspect of it. What I was trying to do is get a recognition of . 76 even when we have an event and have an AIT, even the AIT report is basically an inspection report while we call it something else because it doesn't generally come through to compliance. But any unresolved issues or any compliance items or problems that come out of that event-driven end up right back in our system of observations, because it's the inspector where the rubber meets road who has to follow up on it. This was only intended to show that really we do have a way. If all the information comes together in enough of a bottleneck, we may actually be able to deal with this if we can deal with how we articulate our findings. We added in the PIM about a year and a half ago, and if I wrote an inspection report so that the list item was in the executive summary and every list item has detail in the background of the report and it's a relationship, and I index that item to SALP functional area, cause code -- in fact in the second part of the Arthur Andersen report there is one page on it which had, here's what we did to come up with what we have; we think you should consider these other things. But we didn't do it. If you look at those other things, they look very much like the LER cause codes; it's procedural compliance personnel over design. So those are actually things we already feed our . 77 computers but we can't get it out because it's not connected to anything, and that is called the IV system, which is something we had hoped to fix. So observation should probably be in much bigger letters than occurrence and allegation follow-up. Observation is what we do day to day. CHAIRMAN JACKSON: You should capitalize it. MR. GILLESPIE: Yes. COMMISSIONER ROGERS: We don't want to be controlled simply by events. MR. GILLESPIE: No. Absolutely not. COMMISSIONER ROGERS: Otherwise we will be chasing those all the time and not taking another kind of look. CHAIRMAN JACKSON: I just think it's presentation. COMMISSIONER ROGERS: The other point on performance assessment is that I was just curious as to what your view of performance assessment is compared to the one that the materials people have, particularly in looking at waste sites. They have a rather large performance assessment, or had a large performance assessment effort there. I wonder if the terminology here is really basically the same or different, and whether you had looked at their approach to performance assessment. MR. GILLESPIE: I have only got a sketchy background on high level waste. Performance assessment in a . 78 waste sense is the assessment of the repository? COMMISSIONER ROGERS: No. It's not just the high level. MR. THOMPSON: It's a modeling process, which ends up with a release rate into the environment. The words are the same, but they really are focusing on something different. COMMISSIONER ROGERS: It's the process that I am talking about. I'm not saying they should be the same. I am just asking whether you have looked at and understand any comparison between the two. MR. GILLESPIE: We are not trying right now to create a mathematical model. So in that sense it's different. In principle, my view would be that the PIM should actually be part of a report, because the licensee then gets to see it. If the licensee agrees with the fact that the inspector saw, now maybe I have a fact, and now I can index and classify the fact. MR. MIRAGLIA: I think the key is we are assessing a different type of performance. I am not that familiar with the performance assessment models used. COMMISSIONER ROGERS: I don't want to get into it too deeply here, because it might take us a little bit far afield. I just want to raise the point that we are using terminology that has kind of a superficial look about it . 79 that is trying to achieve something close to the same thing in one area versus another. I am just asking you to make sure that you understand anything that might be beneficial that comes from that. MR. THOMPSON: The more we go to performance-based regulations in the reactor side the better overlay there will be, that is, where you can actually do some information input into a model and you do calculations that come out to drive some aspect of your evaluation to something that matches your regulatory criteria. In those cases they will actually come closer as we get that way. COMMISSIONER ROGERS: As you use this approach to performance assessment that you sketched out, Mr. Gillespie, it does seem to me that there may be a little problem here in that we are moving back into the use of judgment. I'm not averse to that, but I think we ought to recognize that in looking at the total ability of a system to perform. You cited some examples which I think were very good, the tank problem. That came about because somebody exercised some engineering judgment in looking at it. That was a darn good thing. So it's not a bad thing to use engineering judgment in this performance assessment, but I think we ought to recognize that that is probably going to have to be a part of it. At any rate, I would ask you to think about . 80 that and address it as you proceed. The other point is that I am a little concerned about the resident inspectors. I wonder if you are not perhaps overloading them. They have got a heavy burden here, and I think particular care should be taken to see that as you add into their activities these new responsibilities that they get the kind of personal support that they need to be able to get through this initial period. MR. MIRAGLIA: I think the job task analysis will certainly give us a basis to evaluate that in a very conscientious kind of way. MR. GILLESPIE: Let me give you an example of the feedback from the job task analysis. There is real aggravation with the residents about preparing PPR packages, SALP packages, senior management meeting packages, and cutting and pasting. In fact, what we are trying to do with the technology is, if we could standardize that, I can relieve them of that burden of right now literally manual cutting and pasting and then Xeroxing it to make it look like an original piece of paper. One of the objectives is to actually reduce that burden through some sense of standardization and who does what function. In fact, to try to reduce the cutting and pasting or administrative burden of the whole thing all . 81 across the board. CHAIRMAN JACKSON: Frankly, I think one of the greatest achievements in addition to actually saving in terms of people spending their time on administrative things is that in fact it can help to ensure actual consistency of approach as you move up the line from PIMs to PPRs to SALPs to whatever. MR. MIRAGLIA: I was going to say something very similar. In terms of not only efficiency and technology, it will have us look at the functions: What functions are we performing? Are they clearly the expectations out there and are they understood? I think that is another thing that is going to be an outcome of that. It is going to be an iterative type process. We are going to learn as we go along. CHAIRMAN JACKSON: Commissioner McGaffigan. COMMISSIONER McGAFFIGAN: I don't want to get into things that we are going to do next week, but this last backup slide I think is one of the best things I've seen since I've been here in trying to get a lot of information down on one piece of paper. The plant issues matrix, when you get through this computerization that you are going to do, is it going to be the central document? It seems to be on this slide. If you have a good plant issues matrix, then everything else flows . 82 from it. So that's the heart of it? MR. GILLESPIE: Yes. COMMISSIONER McGAFFIGAN: We talked in my office the other day and you just mentioned some day the goal is to be able to -- at the moment, as I understand it, licensees every six months get an update to their inspection plant, and it is sort of like reading tea leaves or the Kremlin wall: Where is the NRC today? They are worried about X, Y and Z; there is a change there, therefore I better do something. But we don't convey any more information than that. If we could convey the plant issues matrix, would we be conveying a lot more information? MR. GILLESPIE: Yes. This is something that has kind of come up. It is a public document. If someone submitted an FOIA request, we would release it; we wouldn't fight it. So then the question came up among the staff, well, what is the best place to promulgate it since it's a critical element in going into the PPR relative to the summary document? What we have toyed with within my group is going to be a recommendation to the regional administrators that we just attach it to the letter that already goes out with the inspection schedule. MR. MIRAGLIA: That hasn't been run up the chain yet. Let's talk very candidly and frankly. It's not a . 83 secret. Everything in the PIM is going to be public. That's what our goal is. We just instituted this process last year. Each region approached it differently, and there are differences in consistency. I think our goal is we don't want to have one PIM in Region I looking very different from Region II. So we are evolving. We are going to get there. At a point in time I think we can share this information. That is our goal and objective. CHAIRMAN JACKSON: We may do something to hurry you along in the process as a motivator for driving toward the consistency faster. MR. MIRAGLIA: That's the goal. The PIM was a tool that we tried to use and we are learning. There are differences; they are getting closer together; and as we get to that point, as Frank says, I think their recommendation will probably be accepted. It's just to make sure that everybody is doing it in a consistent kind of manner. COMMISSIONER McGAFFIGAN: I would agree entirely on the consistency. From my perspective, this all leads into the senior management meeting and the watch list and all that sort of stuff. It would be nice to be conveying to all 108 operating plants where we think they stand every six months. I think somebody said the PPR drives resources . 84 around here, and since it drives so many resources, the outcome shouldn't perhaps just be the X number of plants that are on the watch list; it should be information conveyed to everyone. MR. MIRAGLIA: It's a tool to the whole process, and we have to make decisions on how to apply the resources to all 110 plants. This is the mechanism for doing that. Another issue with respect to the PIM besides consistency is we need to have a common understanding among all of the regions and ourselves as to what is the threshold for putting something on there. COMMISSIONER McGAFFIGAN: Right. When is it an issue? MR. MIRAGLIA: Yes, and that is another thing where there are differences and we are trying to sort out. I think we are striving to that same goal and objective, and the question is how soon can we get there. MR. GILLESPIE: Because we are in a public environment, we have a programmatic requirement that each PIM item actually have its reference. Nothing is really allowed to be on the plant issues matrix that doesn't have a reference. The reference will be either an inspection report or an inspection report which then gets you to like a licensee report, because what we are interested in is, is the licensee system working? Is he finding and correcting . 85 things also? So it's important to note that we give credit on that. The basic question is, if you find something wrong at a facility, that thing not showing up again is some evidence that the root cause has been fixed. If that thing shows up again even though it was fixed once, then you have to question whether it has been fixed. So there is a trending nature to why you would want put even things that we give them credit for in one report on it if they are safety significant. We are thinking out how to use the PIM. These are the thoughts that we had in place when we put it in place. It becomes an index then to what is in the reports, because it does have that link to a public document. The details are in there. COMMISSIONER McGAFFIGAN: Thank you. CHAIRMAN JACKSON: Your backup slide is a big hit, because what I want to tell you is that Commissioner Diaz asked me to, and I will just read it, commend the staff for transforming the pyramid into a functional line process that is conducive to implementation of pass/no pass filters and actual weighted feedback to decrease or prioritize information for decision-making. N. Diaz. So I would like to thank you for a very informative briefing. Obviously the Commission is very . 86 interested in this and obviously is closely monitoring it. I think you have begun to address it as a guide to a thought process. Your backup slide suggests this. As you continue to implement the improvements you talked about and others, what we are particularly interested in is how our NRC staff activities provide the inputs into the assessments that we make and the regulatory actions that follow from those assessments, and then how they lead ultimately to our attaining and providing to the public and the licensees a clear and coherent picture of their performance, the performance at operating reactors in this particular case. We are looking forward to the additional Commission meetings. Again, thank you very much. I would like to remind the Commissioners that we do have an affirmation session. [Whereupon, at 3:55 p.m., the briefing was adjourned.]
Privacy Policy |
Site Disclaimer |