[This Transcript is Unedited]

DEPARTMENT OF HEALTH AND HUMAN SERVICES

NATIONAL COMMITTEE ON VITAL AND HEALTH STATISTICS

Subcommittee on Privacy and Confidentiality

February 23, 2005

Hubert H. Humphrey Building
200 Independence Avenue, SW
Washington, D.C. 20201

Proceedings By:
CASET Associates, Ltd.
10201 Lee Highway, Suite 160
Fairfax, Virginia 22030
(703) 352-0091

TABLE OF CONTENTS


SUBCOMMITTEE ON PRIVACY AND CONFIDENTIALITY

MEMBERS

STAFF


P R O C E E D I N G S [9:11 a.m.]

Agenda Item: Introductions and Opening Remarks – Mark Rothstein Chair

MR. ROTHSTEIN: Good morning, everyone. We've got a full schedule today that should be very interesting, so I hope we can start on time, or slightly late, even though we are expecting more people to join us as we proceed.

My name is Mark Rothstein. I'm the Director of the Institute for Bioethics Health Policy and Law at the University of Louisville School of Medicine and Chair of the Subcommittee on Privacy and Confidentiality of the National Committee on Vital and Health Statistics.

The NCVHS is a Federal advisory committee consisting of private citizens that makes recommendations to the Secretary of HHS on health information policy. On behalf of the Subcommittee and its fine staff, I want to welcome you to today's hearing on the National Health Information Network.

We are being broadcast live over the Internet, and I want to also welcome our Internet listeners.

As is our custom, we will begin with introductions of the members of the Subcommittee, staff, witnesses and guests. I would invite Subcommittee members and full members of the NCVHS to disclose any conflicts of interest they may have. I will begin by noting that I have no conflicts of interest on today's issue, although I have taken many public positions on matters related to health privacy.

And so it's my great pleasure to introduce Maya Bernstein, the new lead staff to the Committee, and she will begin the introductions.

MS. BERNSTEIN: I'm Maya Bernstein. I've just joined the Department yesterday and –

[Laughter]

MS. BERNSTEIN: -- It's my second day and I'm looking forward to hearing what you all have to say and to mostly listening, but I will be the lead staff succeeding Kathleen, who will help me to make that transition, and I'm just looking forward to being here and to hearing what you have to say.

MS. FYFFE: Yes. And Maya has an excellent background, having worked at the Office for Management and Budget and also as the privacy czar at the Internal Revenue Service. And we are very delighted that she has joined our EEH staff.

MR. ROTHSTEIN: Kathleen, you might as well begin by –

[Introductions; no conflicts of interest stated.]

MR. ROTHSTEIN: Thank you, and welcome to everyone.

This afternoon from 3:15 to 3:45, members of the public may testify for up to five minutes on issues relating to the topic of today's hearings. If you're interested in testifying, please sign up with Marietta Squire at the registration table.

The first panel of witnesses have been asked to limit their initial remarks to 20 minutes. After all the witnesses on the panel have testified, we will have time for questions and discussion.

Witnesses throughout the day may submit additional written testimony to Marietta Squire if they desire within two weeks of the hearings. I would request that witnesses and guests turn off their cell phones. Also, during the hearing, if we all speak clearly into the microphones, those listening on the Internet will be very appreciative, I'm sure.

Today is the first of several rounds of hearings on the National Health Information Technology system. The second round of hearings is scheduled for March 30th and 31st in Chicago. Additional dates and locations will be announced as soon as they are scheduled.

As I'm sure everyone in this room knows, the United States is committed to adopting a system of electronic health records.

TELEPHONE OPERATOR: Excuse me; Simon Cohn joins.

MR. ROTHSTEIN: Good morning, Simon.

DR. COHN (on phone): Good morning, Mark. Sorry for breaking in in the middle here.

MR. ROTHSTEIN: Oh, that's okay. Welcome. Simon is Chair of the National Committee on Vital and Health Statistics, and it's very early in California and we appreciate your joining us.

DR. COHN: Thank you for having me.

MR. ROTHSTEIN: So let me just back up for a minute. I started to say that, as you all know, the country is committed to a system of electronic health records within the next decade.

There are many reasons why electronic health records are claimed to be better than paper-based records. These include greater access to records from remote locations, accurate and fast information from cognitively or otherwise impaired individuals, cost savings, increased safety, as well as research and public health benefits from such a system.

Senator Bill Frist wrote in The New England Journal of Medicine in January of this year: "Widespread adoption of electronic health records will reduce errors, improve quality, eliminate paperwork and improve efficiency. Once fully implemented, electronic records will dramatically reduce cost and improve quality."

It remains to be seen whether a National Electronic Health Records system will produce any or all of these claimed benefits. I would note that in a study of medication errors in 2003 which was released in the last few weeks and summarized in the American Medical News, computer entry and other electronic errors far outnumbered the medication errors caused by illegible or unclear handwriting on paper prescriptions.

So substantial implementation problems will still need to be addressed under any system.

Even assuming that the promised benefits from electronic health records will be realized in terms of safety and efficiency, there are very serious challenges for health privacy and confidentiality raised by creating a cradle-to-grave, comprehensive, longitudinal electronic health record for every person in the United States.

Some of the privacy protections that we now have are the direct result of the fragmented nature of health records. When an individual moves from one city to the next, his or her health records do not automatically follow the individual, and new records can be generated with new providers. In addition, patients who want to get a second opinion from a physician without the second physician knowing what the first physician recommended may now do this easily.

So the lack of coordination and integration, clearly a deficiency in paper-based systems, may also have positive unintended consequences.

I believe that realizing the benefits of electronic health records while protecting privacy and confidentiality is one of the greatest challenges to bioethics and health policy in recent years.

Numerous questions come to mind. For example, what level of patient control over the contents of the records will be permitted? If too little, patients will have insufficient privacy and may object to the system as being overly intrusive. If patients have too much control, then health care quality may be jeopardized.

Health care providers may believe it is necessary to supplement the electronic health record with additional questioning, and they may even be concerned about liability for medical errors that could have been avoided with greater knowledge of the patient's health history.

The Subcommittee on Privacy and Confidentiality fully recognizes the difficulty and importance of these issues. We have decided to begin our consideration by calling on some of the world's leading experts on health privacy and confidentiality to give us background and perspective and to help us frame the issues that we will need to be pursuing through our next sets of hearings.

On behalf of the Subcommittee, but especially personally, I want to thank all of you and for now the two members present of the first panel for joining us today and for your willingness to help us try to deal with these issues.

So without further introductory remarks, it's my great privilege to introduce Dr. Thomas Murray, who is the President of the Hastings Center.

Presentation – DR. THOMAS MURRAY

DR. MURRAY: Thank you, Mark. It's somewhat chastening to be talking about privacy, health privacy, in the presence of one of the world's great experts and champions of privacy, Alan Westin, Columbia, and I hope in the soon to be presence of Anita Allen, who is one of the great legal theorists about health privacy. But I'll do my best, do it in plain English.

Here we're talking about not all forms of privacy. Scholars distinguish among may different concepts of privacy. We're talking here about informational privacy. It's not the kind of privacy that was at issue in Roe v. Wade; this is privacy about health information.

And I know of no better definition than Mr. Westin's, who talked about privacy as the claim of individuals, groups or institutions to determine for themselves when, how and to what extent information about them is communicated to others. That will be my working definition.

A little background on the state of the electronic medical record in the United States, and I apologize if the Committee is boringly familiar with some of these numbers but I'm not sure that all the listeners or people in attendance will know them.

Among primary care physicians in the United States –

MR. ROTHSTEIN: I think – could we have his microphone turned up a little bit, please. Thank you.

DR. MURRAY: By 2002, 17 percent of U.S. primary care physicians were using electronic medical records. That compared to 58 percent in the United Kingdom and 90 percent in Sweden. So we are not the first adopters.

If one thinks broadly about e-health, you have a number of functional categories:

You have the category of medical records, and privacy is clearly at issue there.

Of communication between physicians, other health professionals and patients. Privacy is implicated there.

Of decision support where individual recommendations might be forthcoming electronically to help providers.

And of knowledge base management, where security is a big issue and if security were to fail, privacy could be pierced.

The electronic medical record is conceived of as being potentially useful in many different venues, in many different ways:

In patient care, obviously.

In population health, by alerting, for example, a health provider that a patient fitting these criteria and therefore this profile should be screened.

In public health, if one could get access to aggregate data.

In health services research and in quality improvement efforts.

In population-based research. For example, in genetic research which is linking genes to health outcomes in places with the enormous databases now available in Utah and the smaller database, but still an important one, in Iceland, for example.

In registry research, wherein particular outcomes, for example, cancers are noted. In registry research, by the way, it has become quite apparent that trustworthy intermediaries play an essential role in having a well-functioning registry, well-functioning in two senses:

First of all, that one can get complete or near complete data in the registry. If you simply leave it up to individuals whether or not they want, for example, their cancer noted in the cancer registry, you will get something less than full population agreement on that. But if you can have a fully trustworthy intermediate and assure the individual that no one will know who they are without sufficient protections, one can have a functioning registry.

There are important barriers to the electronic medical record. Among physicians, many of them find a very stressful learning curve in dealing with the electronic medical record. Primary care docs spend more time with electronic medical records than with paper charts, several studies show. Time spent answering patient email in at least one important study was actually greater than the time saved in phone and office visits.

So many of the miracles we're hoping for and are promising with the electronic medical record are rather like the miracles promised by many other technologies, not least gene therapy.

And I'm told that many physicians don't like to type, though I wonder if that's a cohort phenomenon. There are studies that show this, that younger physicians, being accustomed to using computers, are probably going to be more comfortable with this.

To put in place a full electronic medical record system could cost a practice $50,000 per doctor in the first year. That's a significant capital expense. In one study, at least 10 primary care residency programs purchased and tried to use and then abandoned electronic medical records systems, just finding them too cumbersome and trouble-making.

Other barriers include incompatible and the fear of ever-changing software. Now, I know that efforts to standardize data fields and such in medical records will go a long way towards overcoming that particular obstacle.

Lack of reimbursement. Some places are, I guess, reimbursing or crediting physicians with answering patient email but many people who work in a fee-for-service system, that's just an extra burden on the physician with no reimbursement.

And the last, but the one I will focus on, barrier is the problem of patient privacy.

So, philosophers don't have a lot of tools, but making distinctions is one of our most important ones. So let me lay before you three distinctions which I think are important. I don't know that they're the only ones, or the most important, but they're three that struck me as ones that you are wrestling with, or should be wrestling with.

First, we should distinguish between the issue of control over the content of the electronic medical record and the control over access to that content. It may seem fundamental and obvious, but it's an important distinction to bear in mind at all points, I think.

As we think about control over content, which I think for many health professionals is the more problematic issue because a doctor doesn't want to have less than full information from a patient if the physician is prescribing a drug or prescribing an appropriate treatment or trying to diagnose an illness. Physicians rightly have been taught that they should get as full a picture of all the relevant information as possible.

Giving a patient control over the content of that could result in a less than optimal, and therefore a less than good decision, diagnosis, treatment, drug prescription et cetera for that patient. The health professionals don't want that.

On the other hand, patients may not want certain things in their health record at all, and certainly – I mean, I would wonder if we should take a little survey here – is there anything that could possibly have been in your lifelong medical record that you would not want anybody to know, any physician to know, or any health professional et cetera? Most of us have items that we'd rather not have widely available. So control over content will be important, as will control over access.

I'm not here ready to recommend a particular view of control over content or over access, but keep the distinction in mind, and just bear in mind that you have my sympathy; I think this is one of the great challenges you will have to confront in your work.

The second distinction is to distinguish among purposes of access. For what purpose does an entity want access to all or some portion of your electronic medical record? It might be for your medical care, so in that case, an ER physician who's never met you before and your primary care physician who you've been seeing for 20 years both have the same purpose: They want to care for you in this episode of illness.

But others, of course, may want access to the same information – prospective or current employer, an insurance company, a Federal agency that is investigating potential health fraud, or researchers interested in questions that aspects of your medical record might help important questions, might help them answer. So that's the second set of distinctions. It's distinctions among the purposes for which people might seek the record.

And the third distinction would be the relationship between you, the subject of this electronic medical record, and the entity seeking the information. It could be, in that case, the primary care physician and the ER doctor are quite different. The primary care physician you presumably have learned to trust over these 20, 25 years, knows many of the most intimate details of your life, and that's okay with you. The ER doc you've never met before, will never see again, in fact may have been unconscious throughout the entire treatment. They're both treating you, so the purpose is the same but the relationship is quite different.

And then other relationships. Again we can talk about entities that may want access to your electronic medical record with your best interest at heart, your medical care, and others who have a different relationship with you, like your health insurer or your prospective life insurer or some vendor who may want to sell you a product and they'd be very interested to know if you've ever been treated for erectile dysfunction. Or your spouse's divorce lawyer who may see some interesting ways of improving her action against you if she could get access to your record.

So those are the three distinctions. The control over content versus control over access, the different purposes and the different relationships.

The data in electronic health records are distinctive in a couple of ways. For one, they're persistent, and Mark Rothstein mentioned this. Potentially, if you had an electronic health record, medical record, that began prenatally, that could follow you into the nursing home with every data element intact.

As Mark mentioned today, health records follow us quite imperfectly. If any of you have moved a couple of times and take your children to school and they want to see their immunization record, if you didn't keep that, good luck, because odds are it somehow got lost along the way. That would change potentially with an electronic medical record.

It's also ubiquitous, the electronic medical record, in a special sense, but here I don't mean that it's out there for anyone at any time for any purpose, but it's ubiquitous in the sense that it's not limited by time and place in the way that paper records have been limited. Access in principle is available at any time in any geographical location.

Another challenge that we all face in an effort to move towards a constructive use of electronic medical records is public trust. You probably read today's paper, the Palm Beach incident where the names of people being treated for HIV were emailed out. It's not the first time such a thing has happened; it's not the last time such a thing has happened.

It's not a distinctive problem with the electronic medical record. Back in the early 1980s, the Hastings Center became involved in the very earliest work on the – at that point it was just AIDS; we didn't know what the virus was or even if it was a virus – and the names of people, of gay men, were released in New York City and the interest group that represented those men who were all gay people in New York City urged no collaboration whatsoever with researchers because they didn't trust them anymore, and we were brought in to help come up with a way of protecting the privacy of individuals and yet allowing researchers to go forward with the necessary research which was going to benefit everyone, including that population. And that didn't require email, but email makes it easier information to be transmitted.

We're also going to have to assure people that the technology works reliably and well. I was just in Davos at this World Economic Forum, which was rather an interesting experience – quite a strange experience, if any of you have been there; I found it quite strange.

And Hewlett Packard provided everyone with a Davos Companion, a little IPAQ hand-held, wireless hand-held, so you could closely keep track of your schedule and send and receive emails and the like. And mine failed by the last day and I talked to another colleague and hers had failed, and so I think we need to be assured that the technology is going to be stable, easy to use, and reliable.

Now, I don't know that there's any connection between the failure of the Davos Companions and what happened to the head of HP a couple of weeks later, but I just note the coincidence.

So to give you a sense of how I think people might view the electronic medical record and why they might have privacy concerns, let me give you a hypothetical patient. Call her "Amy."

Amy had enurisis; she was a bed-wetter at age five. And so her parents took her to the pediatrician and the pediatrician examined her; you know, found no significant physical problems. Eventually that problem went away. But that's now on her electronic medical record.

At age 12, Amy's father discovers a pack of cigarettes in her room, marches her over to the doctor again; in fact, Amy's been taking up a little bit of smoking. They try the nicotine patch. That's all in her electronic medical record.

At age 15, an acquaintance of Amy pushes her to try – pick your drug of choice, illegal drug of choice. Amy becomes frightened and thinks I really – I tried this a few times, but I'm scared; I really don't want to get hooked on this drug. I want to go to my doctor. But if I go to my doctor, is that from now on permanently in my electronic medical record? What will that mean? And for a 15-year-old, you know, life is wrapped up in the next moment – I mean, it's hard. She may not be thinking ahead to her first job or later life choices, but she may still be frightened of going to the doctor, particularly if she believes, not knowing, if she believes that this record is permanent and indelible.

Let's say she goes. Now at age 23, she's been out of college for a year, knocking around, doesn't really have the kind of job she hoped for; she's a little discouraged and she has a bout of depression. So she goes to see her physician, who diagnoses a clinical depression, mild case, and prescribes the appropriate medication for it. That is now a part of Amy's electronic medical record.

Other people, Amy's friends, may be less willing to go to a doctor for fear that a diagnosis of depression would be entered into her record and that that might come back to hurt her later on in her life choices.

So I don't think I've solved anything here, but I've made the distinctions, and I hope they will be helpful, something that I think this group probably understands very well, but I'm not sure that everyone understands, is that the medical record, including the electronic medical record, is not a unitary thing.

We talk about the medical record as if it were one entity and everything in it is of an equal status. We can't think that way any longer. We have to simply disabuse ourselves of that over-simplification. The medical record, including the electronic medical record, would be composed of a welter of disparate types of information, some of it very time specific and of no continuing interest, most of it quite inane for most of life purposes, but some of it will be exquisitely sensitive.

The questions of how to understand that medical record, how to identify sub-categories within that record – and I would say that if you can do that, to the extent that that is accomplished, that would be an enormously valuable thing to do so that we have not an infinite number but a relatively small number of major categories of information that are attuned to the sensitivity and other major dimensions that would be of interest to the subject of the record, the patient, and to health providers.

The reason I'm hammering home this point is in the debate over Social Security, privatization of Social Security, one of the issues that's come up is that people don't deal well when they're given a huge variety of choices. They tend to tune it out.

So to find some way, and this would have to be actually studied empirically, coming up with a sort of good number of choices which would be a handful or a couple of handfuls probably at most of categories. You don't want people to judge all 400-plus categories in the medical record. You want them to have a handful. You can say this, I'm comfortable with this purpose but not for that one, this user but not for that purpose. That will be important.

So issues about control over content will be a challenge. The control over access will be a challenge. And finding those categories that will make whatever choices we decide are in the end in the interest of people, that will serve their medical interest but will also permit them to make sensible choices and not be overwhelmed and bewildered by those choices. Those are, I think, key challenges that you face.

Thank you.

MR. ROTHSTEIN: Thank you, Tom. Appreciate the comments, and we will go to questions after Bartha's talk.

I've been informed that Anita Allen will not be able to join us this morning. She is unfortunately ill. I don't know what's wrong with her, and even if I did, I wouldn't tell you.

[Laughter.]

MR. ROTHSTEIN: But I do with her well and we will certainly make use of her prolific writing in this field and perhaps even consult with her personally to get her input on this.

It's now my great pleasure to introduce another one of my good friends and colleagues from the University of Montreal, Dr. Bartha Maria Knoppers.

Agenda Item: Presentation – Dr. Bartha Maria Knoppers

DR. KNOPPERS: Good morning, everyone. Not being a techie, it's only proper that I'm having difficulties with – you all have the PowerPoint in front of you, so I will speak to it rather than read it out. And my background is largely in research ethics and the laws of different countries around the world with respect to medical data for research purposes.

I think the reason that Mark invited me here today is because there is an increasing number of international trials, increasing number of patients around the world involved in these trials, and the need for comparability of data and a flow of data between countries, researchers, and obviously, if participants are participating in research, they should be able to benefit from those trials. So we will look at these European approaches, see what lessons have been learned, and then

I'd like to make some concluding remarks which are not in the PowerPoint which you have in front of you.

Now, I need to remind you that the European Union on which I will largely concentrate now has 25 countries, and so we are talking about a legal entity that actually prepares directives or conventions and laws and so on and has an effect on a great number of countries, so look at it in a way as an example of the difficulties you would have with a national approach, considering the fact that the states usually have jurisdiction over privacy and health legislation within their own area, within their own geographical and political area.

So I begin with a few general remarks, look at health data, how it's treated, the effect member states have – we'll only look at one country, France, the problem with what we call "extra-territoriality" – what happens when you want to send data out of a given legal system to another country, specifically from the European Union to the United States or from the European Union to my own country, which is Canada? And then conclude, as I said, with personal remarks.

Now, all the countries that I'm dealing with have universal health care systems, but I wish to tell you that it's not because there's universal coverage or a two-tier system in some of the countries in question, that the issues of privacy or electronic e-records or the use of medical data is any different. It's surprising to see in genetics has shown us this, that the Europeans were much faster to regulate against genetic discrimination even in countries with a universal health care because they saw life insurance as being a socio-economic right and did not want insurers in countries such as Belgium and France and so on to have access to genetic data.

So in 1995 already in Europe we've had the directive which I state has legal force on the processes of personal data. And health data is seen under this directive as sensitive data but is included under a very broad notion of personal data.

And the effect of the directive was to force countries in their domestic national legislation to harmonize with the directive and to be in conformity with it so as to offer an equivalent level of protection in the countries that were members of the European Union. So that's 10 years ago, also very important because the ultimate goal, of course, is to have what we call portability, to be able to go from one country to the other within the European Union and have equivalent protection, equivalent rights, and equivalent treatment of medical data.

Now probably for me, but that's just – and it'll come back in my conclusion, the most important thing about the directive is that it incorporated the principles that were found in the OECD guidelines on the protection of privacy in trans-border flows.

And the reason that it's important is people often fail. You and your principles, you know, because I'm always working on ethical codes and principles and so on; it's just political, you know, decoration, window-dressing, whatever. And yet, if you look at the power of these principles over time, not only are they reflected in the European directive 15 years later, but the majority of countries within Europe obviously harmonized and incorporated the principles but other countries outside the European Union were equally inspired to adopt these principles. You can see the power and the flexibility, not limit to any one technology that principles have.

So, it's a negative principle, the one that deals with health data; it's not a "you may" – it's actually "you may not" unless specific and suitable safeguards are there. And this is a non-limited list. In other words, there can be additions made.

One obviously, which is now subject to contention – I'll get back to that – is the explicit consent except where you have laws that authorize, such as either national security, public emergencies, state surveillance programs and so on that constitute legitimate because they're already in the law's exceptions to the need for explicit consent.

For those who are weak or vulnerable or incapable and so on, it's their vital interests that need to be protected, and so consent is listed for that as well.

So you have the overriding power of the state, you have the protection of persons, and finally, very large exception as you can see, the processing for the purposes of preventive medicine, medical diagnosis, the provision of care or treatment, or the management of health care services, provided it's by a health professional subject to professional secrecy, and I'll get to professional secrecy in a minute.

You cannot run health care systems – I'm talking at the level of systems – without having to come in. I mean, they're going to do your blood type, they're going to do routine tests. A system cannot work, you cannot be cared for, if there aren't this basic exception, if you like, to explicit, written – or it doesn't say written, sorry – explicit consent by the individual.

And of course you have that additional protection of professional secrecy by those health professionals involved.

Now, France even prior to the directive had already adopted in 1978, so it's even much earlier, a loi – if there are no official English translations in titles, I don't usually put them down, but loi relative to informatique, its informatics, fichiers would be files and freedoms, or dossiers, let's say, and freedoms, recently modified in the 6th of August, 2004.

They already had the wording that eventually found itself to the European directive and applied it to public and private sectors, because in European countries you can still buy additional health insurance or buy additional kinds of private services and so on if you have the means. I think that is not allowed in my own country.

The prohibition of the Article 8 of the European directive 1995 was reiterated in the recent August 6, 2004, law in France. So again, it's a negative principle, and here they talk about express written agreement. The same provision to protect the vulnerable person, the same article about processing, so again as a state possibility of overriding and a systems one for overriding if there's a health professional subject to professional secrecy.

But what we have are two interesting new additions here. One, exception for processing for research in the health sector and the possibility of anonymizing data without express consent, now written, if you get the authorization of the CNIL. The CNIL is the national commission for informatics and freedom, and this was created way back in '78 with the earlier law.

Now, we'll have at a look at this processing in the health sector. It's obviously necessary for quality assurance in any given health care system; you have to have data to know whether you're cut off for different tests are accurate and so on. In a health care system, you have to have surveillance data and incidence data to know whether is HIV going up, is it going down, are there pockets in the country that are at higher risk for environmental or epidemiological or other reason. So there is a reason for including research in the list of exceptions.

One comment on anonymization, if I forget to do it in my conclusion: Anonymization is legally and ethically expedient; it gets rid of a lot of problems. It means that you use various – and there's a slate of about 32 different terms in documents unfortunately around the world that's irreversibly anonymized, i.e., enough identifiers have been removed that you cannot retrace an identifiable person.

Now, we all know that it only takes seven snips et cetera, et cetera; we all know that there's no such thing as impossible in the world of computers and snips and the like. So let's say in a practicable, reasonable context it works.

But it works to the detriment, one of science in the long run, because once you've anonymized, you cannot update that data because you can't trace it back to anyone. So if clinically something changes, whether a person dies earlier or never dies at all or never develops the disease, you won't know; the data is static in time.

And secondly, so scientifically your shortchanging yourself, after five years the data's good only for controls; it's not good for anything long-term.

And you may be even shortchanging the individuals whom we're trying to protect sometimes against themselves because their participation in research then does not achieve the goal which usually is to find some sort of therapy or at minimum a test for certain conditions. So it's easy to get it through the larger half – the ethics committees are happy; but think about the goals before you do it.

So when they added this research exception in France, they also created another committee, and this one is the consulting, or advisory committee, I should say, on the treatment of information in research in the health sector. And it was under the Research Ministry that this committee was created, because the CNIL is a big, large commission that they've had since 1978; they're not experts on health research. And just because they'd be doing informatics and personal data processing opinions for 25 years at least doesn't mean they understand the particularities of health research that uses patients' medical records and so on and research feeding back in member clinical trials, feeding back into the medical record.

So they've created the committee since August 8th – the 6th, sorry, 2004 – to examine the requests for processing of health data for research. And they give an opinion on the research methodology. So they'll build up a corpus of knowledge, if you like, that makes them particularly – well, hopefully – well informed about this domain.

And they give their opinion prior to going to the big body. You can get stopped right there where you're processing of health data as an exception to the explicit written consent or presence of a law.

So once you have this opinion and the authorization of the CNIL, notwithstanding the rules of professional secrecy, you still have certain conditions. If the data allow identification of any way, you must code the data, so obviously that's a minimalist protection, what I would call it; except obviously where people were taking drugs and you have to find them and follow them in pharmaco kinds of studies and so on.

But the results, once published – again, to me this is obvious – but they have to put it in to not allow the identification direct or indirect on the data subject. But they create a possibility of opposition.

Opposition is not the same thing as consent and refusal. Opposition is about at the level of assent and opposition. So you oppose, but they won't deliberately go out and deliberately and seek your consent if you don't want to have your data used in a research and any processing. Interesting to see what will come if anyone will oppose. And then you can obviously suffer the sanctions of the higher body if you fail to respect such an opposition.

So what's happened with other countries, then, because this is 1995. The science found its way obviously into many international trials. You could only export data from countries, members of the European Union, if the other countries offer an adequate level of protection. So how do you evaluate if they're offering an adequate level of protection?

Well, it's written in front of you – the nature of the data, the purpose, the duration, look at the rules of law in the countries – the general ones, the state ones, the security measures and so on. And every country in Europe in its own legislation has this sort of extra-territoriality. It's very rare that a law goes beyond their own borders and says "you can't do this other countries unless you follow these rules." You can imagine the chilling effect that this had in 1995.

Let's look at the United States then, a minute. What happened then in companies, international and others, or research projects that were international?

Well, there was the beginning of negotiation with respect to recognizing whether the United States or member states or whatever, states within, met the levels of the protection offered by the directive. So we waited till July, 2000, to get a decision from the European Commission on the adequacy of your privacy principles and since then obviously HIPAA has come into effect.

Just a brief reminder of what the Safe Harbor Privacy Principles are and we'll move right on then to the frequently asked questions that you can find on the website.

If personal data are collected in the European Union and they're transferred to the United States for pharmaceutical research, do member state laws apply or do the Safe Harbor Principles apply?

And the answer then on this FAQ from the Commission was that the member state law applies to the collection and processing prior to the transfer, so you have to meet your own laws before you transfer. The Safe Harbor Principles, however, apply to data once they've been transferred to the United States. And again they come up with the year 2000 -- data used for pharmaceutical research and other purposes should be anonymized when appropriate.

So what if you've been recognized in another country and an individual in Germany, let's say, wants to withdraw the data and the data's already sitting here with a research group at Johns Hopkins, so what happens to an individual's data if the person wants to withdraw? This is a universal, Helsinki declaration. If a person wants to withdraw from research, he should be able to do so.

Or the sponsor might want to withdraw someone for reasons by looking at the data that a certain group of participants are more vulnerable. So participants may decide, or sponsors may decide, to withdraw. Any data collected previous to withdrawal may still be processed. You can't really go back and start taking stuff out of data sets; it makes absolutely no sense and you probably can't do it. And along with any other data collected as part of the clinical trial. If this was made clear to the participant at the time that they participated that data collected prior to the request to withdrawal could not be removed but that you would stop using it at that particular time.

Final frequently asked question that comes up is: If you're key-coding at the country of origin and the country to which you're transferring health data cannot access the key either through a custodian or by contract or any other way, get back to unlock that code to identify the person.

So let's say a company here does not have the key; it stays in Germany. If the unique key code is held in the country of origin so that he or she could only identify research subject under special circumstances, i.e., maybe needed for medical care, some toxicity of a very small group or particular individual is discovered, does it transfer from the European Union to the United States of data coded in this way constitute a transfer of personal data? No. In other words, the whole directive would not apply to such data.

What about my own country then, Canada, where every province has jurisdiction under the Constitution over health?

Well, we asked the European Commission as well whether the new Canadian Personal Information Protection Electronic Document Act met the equivalent protection, just like the states needed to know in the year 2000.

And what's interesting is that personal health information under this Act includes not only information but you can see under C information concerning the donation by the individuals of any body part or bodily substance. So in addition to health services information, physical and mental information, or any kind of other, you'll see that samples, what we call "wet data," are now subject to this Act.

It's interesting about this Act -- it wasn't even supposed to apply to health; it was supposed to commercial activities in Canada. But commercial activities are sort of seeping in in the sense that with research, a lot of it's a public/private sponsorship. There's always a company there usually. It's very rare that you have a 100 percent publicly funded, non-commercial, whether it goes from buying servers or paying for communications or whatever, as soon as you have a commercial entity somewhere, it fall under commercial activity. So we end up with a law that wasn't meant for health data applying to health data.

Now, another interesting feature about this law is that to insure flexibility over time with additional technologies or new knowledge coming up, they attached to the law as an annex but as part of the law the Canadian standards association code on the protection of data.

Now, codes, as you know, as set by CSA or ISO, let's say, can be modified and changed over time to take into account new technologies or when knowledge changes, if you like, or new protections are required or old protections become redundant, which means that by changing the code over time, you don't have to go back to Parliament and start to read the whole legislative process. Annex is part, but it's "professionally" kept up to data.

We also have a tri-council policy statement covering all research involving all research – demographic, medical, historical and so on – involving humans, which is being updated, as you can tell, by the Canadian Institutes for Health Research draft Privacy Best Practice Guidelines.

The statements of ethical conduct in 1998 are increasingly seen as over-protectionist and paternalistic, i.e., since 1975, the twin pillars of privacy and autonomy have reached a point where health research, which is more for finding out things about certain conditions and not necessarily individually oriented, was seen as hampered by the need for explicit individual consent.

So it's going to be very interesting. All kinds of epidemiological research – leftover blood spots in newborn screening programs were batched together by the thousands so you couldn't figure out where they came from and we were checking for HIV prevalence. Well, because of the explicit consent, that whole program was dropped. So we have no idea if HIV is going up or down. You can see

how this privacy autonomy when taken to extreme individual autonomistic sort of extremes can harm, if you like, harm the individual and thwart the role of the state to protect, prevent and promote health.

So these are the principles, then; I won't go into them.

And finally, then, for Quebec. It's a province, well, let's say state, can show that their internal legislation meets the substantial equivalent – it's sort of like what the European Union did with their directive – then provincial legislation or state legislation does not have to follow. You can follow your own internal, once you've proven substantial equivalence. That's an interesting mechanism as well.

So in conclusion, in the 25 member countries of the European Union, health privacy is a fundamental human right. In countries of civilian tradition – so the countries that followed the Napoleonic Code and I suppose the Common Law tradition such as the U.K. or Australia or Canada except for Quebec – not only is privacy a fundamental human right, it is what they call a subjective right. It is found in the civil codes of most of these countries in the chapter on the rights of personality, distinct as rights concerning goods or objects or other chapters in civil codes.

If it's a subjective right, that means you don't have to show harm when you consider that there's been an infringement or a breach. You do not have to show any economic or other loss. So it's an automatically what we call actionable right. And this makes it very powerful. At the same time, privacy, like honor, which is in there, like reputation, which is in there, is amorphous, it's ambiguous, it's hard to define.

And it's also found in all the countries as a constitutional – in the German constitution, for instance, right, and under private law, in different statutes.

So for concluding remarks that are less on the European Union or on France or even on Canada, and Tom already mentioned some of these issues with transfer to e-records, concerns over content, concerns over access, and I look at it more concerns over the quality of research, for me, that's quality.

Use of data, data transferability, is extremely important. And we have several barriers to this. If we're looking at some of the large international studies or even if we look with comparison, let's say, between population studies that are emerging, population studies that involve large cohorts where let's say the U.K. bio-bank, 500,000, they want to build a resource, it's not a research project. They want to build a research infrastructure, a research tool, a research resource based on data that would be at the level of let's say normal genomic variation before you get different diseases, testing themselves against this background, you cannot build these infrastructures, you cannot build these resources that have no immediate benefit for the person, that are really longitudinal studies that construct research tools against which you can then later do your hypertension, your diabetes, your cancer studies. You can't build them with the current ethics and legal requirements surrounding research.

And we've run into a problem, that the protection of privacy is such that, to coin a phrase, we've run unto semantic inoperability. We cannot compare data because enter it differently, they protect it differently. What's called de-identification under your HIPAA, you find it under the international code on pharmaco genomics as meaning double-coding, which is totally different.

So people are using irreversibly anonymized, delaying anonymized, truly anonymous, de-identified, traceable and so on, and we can no longer know whether we've got equivalent levels of privacy and protection.

We need a concordance, we need a language that we can understand – double-coded means back there; we're not saying everybody has to use the same words. We need to know, to protect privacy, what we're talking about.

We can never validate, we can never gain statistical significance, if we can't understand how we're protecting privacy.

Secondly, the whole issue of consent, explicit consent. The OHRP here in the United States at the end of August put out a very revolutionary, if you like, statement on if certain conditions were met with respect to research with codes, with data holders, with contracts and so on, that an explicit consent was no longer required because a person would not be identifiable, and you're only a person under law, for privacy purposes, if you're identifiable.

So this whole notion of consent, broad consent, authorization, is major.

Thirdly, we're getting to the issue of portability in Europe and Canada as well. Smart cards or administrative – all the countries in Europe, if you're a Portuguese person traveling in Italy and you have a car accident, how to insure with the eventual harmonization that they're looking for that you can give your card as a Portuguese and get treated in an Italian hospital as a member of the European Union. And I'm not even talking about adding a few little basic medical data such as allergies and so on onto that; just talking at the level of an administrative health card.

And in closing, since there's a lot of people here from public health and from WHO, I'd like to make a plea in your privacy considerations not to forget the need for accessibility to certain minimal data not on persons but on populations for public health purposes.

This is extremely important. We saw with anthrax, we saw with SARS in Canada, we will see it again with the Avian flu, that unless we get used to a citizen providing the state, who's supposed to protect, promote and prevent, needs a minimum amount of health data to fulfill that role.

And people will turn to the state and say, "Why didn't you do this?" And you're going to say, "Well, we didn't get your explicit written consent. We couldn't follow you. We had no idea what was happening in the country. We couldn't think ahead. And we were waiting for WHO to come out with their international health regulations."

So there's a plea here to make sure that we do live in society, we do live in relationships and in communities, we do travel on airplanes, and there is that public health aspect that needs to be kept in mind.

Thank you.

Questions, Answers and Comments

MR. ROTHSTEIN: Thank you very much, Bartha. Both of you have raised so many questions, it's hard to know where to begin. But I will, and I'd like to ask one question first to Tom and then see what see Bartha's response as well is.

I think fairly over the last 30 or so years, one of the dominant themes in law and bioethics is autonomy in health care, that a competent adult has the right to decide whether to see a physician at all – and now I'm leaving aside public health issues – whether to see a physician at all, whether to comply with the advice given by the physician, and even to the extent of lifesaving therapies.

So my question is: How do we work in to an electronic health record system the notion that we have built up of patient autonomy so that an individual level one could say that you have a right not to go to the physician if you don't want to? You could also say that the physician has the right not to treat you unless he or she has sufficient data from which to base a diagnosis and treatment. But clearly there is going to be some level of autonomy in the individual relationship between a particular patient and a particular physician where they could in theory at least negotiate what is done with the results of the information; that happens all the time today.

So given that dominant ethnical principle at the individual level, how are we to give, or maybe oughtn't give, any effect to that at a societal level when our task is to try to make recommendations about the level of privacy and, I would add, inevitably autonomy, that individuals should have, or might have, in the electronic health record systems. So I'll just begin with that simple question.

DR. MURRAY: That's a lie; that is not a simple question.

[Laughter.]

DR. MURRAY: That goes right to the heart of the problems that I think you face.

You're certainly correct in identifying a major theme in the whole field of medical ethics over the past 35 years, at least dating back to the birth of the Hastings Center in 1969. And that is a concern for empowering patients in the relationship with physicians. The bete noir was the paternalistic physician and the hero was the autonomous patient.

That was a horse that people rode for about 20 years very successfully. It turned out to be a bit of an oversimplification of this complex relationship between patient and physician. It was an important corrective because at that time few people were standing up for the individual autonomy of patients.

How do you integrate that now into our understanding of what to do about health records?

We clearly need to be very sensitive to the interests and reasonable and informed preferences of patients. Those are important qualifiers.

It is true that a patient today could go into a physician and give minimal information and request treatment. It's also true that a physician could refuse to treat a patient on those grounds, right?

Back 25 year ago when we did the first study on drugs in sports at the Hastings Center, not how to use them but the ethics of performance-enhancing drugs in sport, I learned that the first question that a savvy trainer asked when you ran out on the football field after a player was injured was, "What have you taken today?" because you didn't want to give the player drugs that would interfere with drugs that would interact with drugs that they were already taking.

So, I mean, a good physician would probably want to say, "What are you taking? Before I prescribe anything to you, what are you taking?" That has to be negotiated, and there is a certain amount of – there's complexity there. There are issues about professional integrity of the health professional that I think are weighty and there are issues of the autonomy and privacy rights and interests of the patient. All of them are important.

Bearing that in mind, I think it goes back to the distinctions that I tried to make earlier. The challenge that you face is figuring out to what extent patients would have control over the content of their record and over who has access to sub-areas of content within that record.

I don't think a once-for-all answer except to say that the interests you've identified are in fact legitimate and significant interests.

MR. ROTHSTEIN: Bartha, would like you to comment?

DR. KNOPPERS: Yes, and I'll start with Tom's last comment and work backwards.

I totally agree with his concluding remarks, his previous concluding remarks that he just reiterated about the idea of breaking up levels of privacy protection into categories and selected – in other words, creating a hierarchy, if you like, within the medical record by broad categories, as opposed to saying "no access at all" or total control by the patient as to what's in or out, which of course would be a medical disaster.

I would also argue in favor of elaborating broad principles for electronic records rather than leaving the SOPs perhaps to the IT experts as to what kind of procedures best respect those principles.

Your statement about autonomy still being the overriding, if not prevailing –

MR. ROTHSTEIN: I didn't say "the," I said "a" –

DR. KNOPPERS: A, all right. And, as Tom said, empowering the patient, moving away from the sort of formal imperialist or authoritarian or paternalistic or whatever people called it.

We've had 35 years of this, and I'm still a believer in it, and as Tom said, provided that autonomy is exercised in a reasonable and informed way.

The new emerging glory of that, as you see mainly in Europe, is the right not to know, which has now found its way into charters and constitutions and so on. Difficult to know how much information you have to give someone to exercise the right not to know so that the giving of fully informed consent as to what they don't want to know without giving away what they don't want to know. So this is a quagmire in terms of legally trying to figure this out.

I would argue that autonomy has reached an apex in terms of individual preferences, individual exercise, individual choice. I now see a large margin of autonomy being taken away ironically by the overriding preoccupation with privacy. In other words, either legally or through ethics committees, we're starting to protect people against themselves.

If I want to give a broad consent to a particular longitudinal study and I say as long as there's continual ethics review, annual audit, and the research question, you know, is not something in one of these following areas – let's say bioterrorism or some areas that personally would go against my values, I want that consent to be valid over time. And it's not specific enough for most ethics committees and so on and so they will say, "To protect your privacy, we're not going to allow you that consent."

So we're now in an ironic situation where privacy which was once seen as a liberty interest, in some countries privacy is seen as a freedom close to autonomy, yet is being undermined by second guessers in the name of privacy. So – I'm not sure if I answered your question at all.

MR. ROTHSTEIN: No, but I'm not sure you did, either, but I liked your comments.

[Laughter.]

MR. ROTHSTEIN: And so let me recognize my colleagues who may have questions. Dr. Harding?

DR. HARDING: Well, thank you both for beginning this very good discussion.

During the last 10 years, we have discussed this sub-category issue, or sub-classification, in electronic medical records, and it's been called various things, especially sensitive information and so forth; you know, talking about OB-GYN, genetics, infectious disease, mental illness and so forth, those kinds of things.

Dr. Murray, you were mentioning sub-categories and then trying to limit choices of those sub-categories to a reasonable number, because if you have too many, it gets out of hand – or did I kind of extrapolate a little bit from what you were saying? But if you both could comment on what you see as the sub-categories and how they might be utilized, I would appreciate it.

DR. MURRAY: I don't think I could add much to your deliberations on the question of what are the appropriate ways of parsing out this data and creating the sub-categories except to say that whatever good ideas we may have about how this ought to be done. Say we end up with seven sub-categories of data – I don't know what they are. One's going to be psychiatric and psychiatric-related. Another may be reproductive health-related. But they may be just segmented off from other categories.

What I would suggest is that we should take whatever our guesses are how to chunk these, how to parse these, as a hypothesis to be tested empirically and bring it out to patients, the public, through various ways you could test things and look at opinion surveys and focus groups and other ways. Try it out with health providers; does it work for them? Does it turn out to be a sensible way to enter and process data and use data to do optimal medical care for them? I would take it as a hypothesis.

One lesson I think we've learned in bioethics is sometimes people have had wonderful ideas that look just great in principle and they flopped completely and sometimes even worked the other way. I mean, the idea of the so-called "required request" for organ procurement for transplant, we thought the barrier to getting more organs for transplantation was that doctors weren't asking the families of dying patients. It turns out that when they require doctors – we've passed laws requiring doctors – we got fewer organs, because doctors resisted and kicked back and it just didn't work.

So, any great ideas we come up with, we need to road test them. So that's one thing I'd say.

And my point about the number of categories, it's simply an acknowledgment of the limitations of human information processing. More choices, the menu gets very long, people's choices, if anything, may become less informed.

And I mean less informed by their own standards in the sense that in the end you say, "Here's what you did; my God, I didn't mean that to happen. What I really wanted was this." So we need to do is present an array of choices that are optimally designed so that when people exercise choice in these ways, they get the results they had hoped for. And if they had been wiser and could see ahead, they're the results that they would be pleased to live with.

And that's not a trivial task. It's not one philosophers are going to solve for you.

DR. KNOPPERS: If we do have categories, they would have to be reviewed. We'd have to make a concept on them automatically written in.

For instance, 50 years ago, if you had the "big C," you didn't talk about it. I mean, and in a lot of European countries still today, victims can hold back and cannot communicate such what they call "fatal prognoses" under their codes of ethics.

So what is sensitive today might actually be considered normal 10 years from now, and the hope is that genetic data and psychiatric data will just be normal data, part of the human condition, and a lot of the stigma that's associated with it will disappear.

So those sensitive and so on would have to be reviewed.

Now I remember what I wanted to say here – then I get off in another – in terms of privacy to sort of match without seeming paternalistic, a really neglected area is the de-entological, or professional, ethics, not just a code of code of conduct in a professional sense but actually legally actionable. In other words, if you have professional secrecy as in Europe, it's not only in their codes of psychiatrists, physicians, nurses and so on, all those handling health data in their own professional codes, so they're subject to disciplinary, but it's also found in other statutes as well.

In France, it's in the criminal code. In other countries, the right to professional secrecy, the obligation of the professional is actually in the constitution. They've raised it to a higher level. In other words, you'd better protect data not by asking the person all the time what they want but actually imposing on the health professional this very serious obligation of professional secrecy to avoid breach. I mean, I think that's a window that we haven't really sufficiently explored, how to up, if you like, the legal obligation surrounding professional secrecy.

MR. ROTHSTEIN: Mr. Reynolds?

MR. REYNOLDS: Excellent testimony. I'm not sure whether to run out of the room screaming before the rest of the people speak or what.

[Laughter.]

MR. REYNOLDS: That's not that I'm not comfortable with privacy.

A number of things that you mentioned, as you look at the privacy rule under HIPAA right now, privacy notices are a big deal there. They're complicated, they're lengthy.

One of the things you mentioned in the testimony was clear to the participant the notice at the time he or she plans to participate, and that's most every time somebody goes to a doctor in the United States when it's not a privacy notice.

And then you're talking about opting out, you talk about we have treatment, payment and health care operations which is another philosophy under which our privacy is dealt with. You mentioned parsing available data, the patient's choice to parse data in or out.

Then we try to get the good medicine and some of us have been adjudicating e-prescribing. You try to get the good medicine and then you anonymize versus de-identify, which is what we call it. And then you have the liability of the doctors involved and so on, based on treaty.

So you gave us a lot of information. With that framework, that's kind of the way we do things here right now, and I really enjoyed understanding what they do in Europe. Where do you start when you think of the electronic? What is the hierarchy? You know, so I mean, there's lots of subjects and one doctor could have a good privacy notice and the other one could have one that would allow them to do a whole lot more but because it's a privacy notice and because it's signed by the patient and everything, they could do it, could fall under certain –

So what is the hierarchy, as we look at electronic medical records, as we even look at what we have currently in the privacy, what's the hierarchy, in your opinion, really approach this? Because you've got to have some starting point as kind of the Holy Grail of what we deal with.

DR. KNOPPERS: Tom was mentioning the idea of field-testing. You might want to field-test one or two different approaches.

My preference would be, since the controllers both of the quality of data, the type of notices, the data that's in there that's correct and not – and Mark's opening statements reflecting the errors may be even greater than on written records which were in open drawers and hallways of hospitals and so on – the preference I think would be to begin with the health care providers themselves and ask them. Say, this is your chance to feed into a future system that you're going to have to live with and be subject to and liable for, for moving this way not only what you think would work; how would you best protect your patients' interests in terms of the confidence that is inherent in the physician/patient relationship?

If you don't have that trust in the – I think it was Tom again that mentioned it – you're not only going to get poor medical care, you're going to get poor medical data and the relationship itself will deteriorate.

So I would go back to starting with the framework of principles of which professional secrecy, or medical secrecy, whatever you call it, is reiterated, reborn, retooled, and becomes a reminder, but in the building of it under this new context of e-records becomes part of the profession's work before the AMA or whatever and actually get them involved in how this could work, because it's really at the family physician.

I mean, I've worked mainly with researchers, but I can see that treating physicians' patients who are eventually involved in the research and the data that comes back. So I would start with a framework of principles, two or three approaches that would translate those principles into different kinds of systemic applications and actually go out there and test them. If you don't get buy-in, if you don't get participation, then you need the consumer, the patients' organizations, involved. Sometimes they don't like what we come up with and it's awful. You've gone through five years of work and you really talked amongst each other but not exactly to the people who are at the front lines, i.e., the physicians and the patients.

And they want quality care. I think they'd worry more about a wrong prescription than taking out the fact that they live on this particular part of town. You'd be surprised what people are sensitive to. And yet if you don't know that they live on that part of town, right next to the iron smelter, that's a very important environmental data. Not a thing about the usual thing about abortion or psychiatry or all these sensitive genetics or whatever, you're missing a piece of vital data. You miss subjective – you know, opting in and selecting and parsing. I'd be very, very careful because there's data there that doesn't look medical but is equally as important.

Marital breakups, for instance. You have three in a row and you don't know it and you've just moved somewhere and you've got a patient in front of you. You won't even be able to diagnose where the stress comes from.

So I would involve patients' organizations as well.

And then actually check out some of these countries that have otonic records such as the U.K. and Sweden and say, "What are the lessons learned?" That would be the third sort of empirical data-gathering, as Tom called it, and say, "What works? What doesn't work? And why?" so that we really use their data for a project that you can't really sort of build yet because you're not there, but they've done it and maybe they've learned something that you need to know.

DR. MURRAY: With no disrespect meant towards the Chair or my friend and colleague testifying or any other attorney in the room, to me it would be a very sad thing if what we ended up with were detailed, legal descriptions that people didn't read but signed anyway. I mean, I don't know how many privacy notices I get in the mail these days from all the different entities with which I deal. How many of you read every single one, every word of every single one that comes in, and how many of you understand all the implications of everything you read now?

DR. KNOPPERS: Even the lawyers don't.

DR. MURRAY: Well, maybe they don't. So that wouldn't be a step forward empowering patient autonomy, right?

If you really are after respecting patients and respecting their liberty, promoting individuality, the kind of the moral foundations that lie behind our concerns for privacy, lawyering it up, as the expression goes, doesn't get us there. It gets the entities off the hook legally, maybe, but it doesn't actually accomplish the human moral goals that all of us here really are seeking.

There's a constant tension whenever you create a consent form particularly for clinical trials today between, you know, wanting to do it in such a way that specifies every possibility in there and so it goes on in boring and technical detail. Few people read it carefully; even fewer understand it.

Try to explain in as simple a language as you can what really is at stake if you agree to take part in this trial. And that's a tension – there's not an easy way to resolve it – and it's a tension that you will face in this, and I just hope that, A, you're aware of the tension, and, B, you will always push against the impulse to take recourse in exculpatory language, right? We want to be as clear, transparent and truly empowering as we can be, thus the shorter menu rather than the massive menu with many other choices.

DR. KNOPPERS: Totally agreeing in deference to all my legal colleagues. Consent has become like a notarized deed. It's absolutely pathetic, totally understandable, and it's just preemptive legal cover-up, and I think we should stick to principles and mechanisms and procedures that translate those principles in an understandable way with the participation of the patients themselves. I totally agree with my colleague on that.

MR. ROTHSTEIN: Other Committee member questions? Michael?

DR. FITZMAURICE: Health information seems to be special. When I apply for life insurance, I've got to sign away things that say they can get any information they want to in order to give an actuarial judgment about whether they're going to make money off of me or not.

Same thing when I get car insurance. Have you had any accidents, any tickets? Do your kids have any accidents or tickets? You have to reveal it and you have to sign away information.

But health insurance seems to be somewhat different. Many people try to work for a large employer where you don't have to reveal any pre-existing conditions in order to get coverage. And so I'm not sure what the range is. Is there a right of privacy for health insurance for health information that doesn't exist for life insurance and for auto insurance? And certainly, people are more resistant to giving out that information.

And my questions. When we were writing the privacy rule, it was very hard and it was hindered only at the last moment in the privacy rule if you have at least one case of the use of contracts between the holder of patient information and the use of that information to protect the confidentiality of that patient information. I guess one would be a business associate agreement.

But what I'm thinking of is a limited data set given to a researcher who in turn signs a data use agreement with the giver, usually a health provider, for the use of that information in research, and it has several different conditions in it – you won't try to contact the person, you won't try to re-identify the information.

And my question is: Is the use of contracts similar in the European Union and in Canada to protect the confidentiality, protect the personal health information, when the patient has not given authorization for its use?

DR. KNOPPERS: For that particular answer, you'd have to look at the laws of every country, unfortunately.

DR. FITZMAURICE: Is there a simple case then where they do have contracts?

DR. KNOPPERS: Yes. There are model contracts, but normally most states would already have legislation in place that would cover sensitive data such as health data and yet not use the language of limited data sets but actually use a body such as in France to approve the kind of arrangement needed for that particular type of research.

So those kind of personal contracts with physicians or entities and eventual users are not the way it's usually done because there's already a legal framework that would dictate – or a body in place to vet it.

DR. FITZMAURICE: Something like going to an IRB for research?

DR. KNOPPERS: But a health information IRB, yes, yes --

DR. FITZMAURICE: Okay.

DR. KNOPPERS: -- in the different countries.

I only gave the example of France but they exist in other countries as well.

A contract approach might work, provided that I think that somewhere there is a minimal content because the access to a limited data set would probably be the institution who would hold the data set that the person, the researcher, wanted. The user, the researcher, knows the particular needs of a protocol, what kind of data, doesn't need all the data, needs certain data.

And to be able to do that without any kind of patient consent or knowledge, or even a notification: "All ye who enter into this institute, we send out limited data sets for approved research protocols" and you presume to remember that, and so you've consented, might be problematic. I think it would be problematic in Europe unless it said, "Contractual arrangements will be permitted under law if they contain the following elements" and then, having passed a law on the public context, the legislature, i.e., the politicians who are in touch with the writings, would know and would have received this.

So I think there is a role for contracts to play, but not without any kind of minimal content required by law.

MR. ROTHSTEIN: Other questions? Oh – Michael, sure.

DR. FITZMAURICE: I want to follow up with one more, and that is, there's a presumption you make, at least in the United States, about opting in versus opting out. There are times when a patient has to opt in, that is, I have to give you my authorization for use of my information if you're going to use it for other than treatment, payment or health operations purposes.

Other times, you have to opt it where it's presumed that you've opted in, such as when you're being wheeled into a hospital and they say, "Is it all right if we put your name in a hospital registry?" You say, "Ah" – they've given you a medication and you're dopey – oh, we gave them the opportunity; that's all we have to do. So we assume that the person did not opt out.

Is there a principle such as maybe autonomy or self-determination that applies to whether you would use opt-in or opt=out, or is it a matter of balancing the benefits against the harm? Is it a weighing versus the application of a principle?

DR. KNOPPERS: Yes, I know it's a really important question. Tom mentioned the Icelandic health sector database. In Iceland, the health sector database law of 1998 was deemed unconstitutional because it presumed an opting in on the part of Icelanders unless they opposed. They weren't asked, but you could oppose, and citizens were supposed to know; it's a small country and there are 50,000 and so on. And it was seen as a way of creating a database of clinical data that was missing from what they already have, which is an excellent demographic genealogical database, and the nation's budding, growing genetic database with explicit consent. That was deemed to be unconstitutional and so the law was declared three years later invalid because of the opting in.

And I think the only kind of opting in that we can really presume, and this is totally personal, would be for routine tests. A hospital can't run if you can't, you know, have basic information for quality control using leftover samples without identifiers obviously for calibrating machines and checking that there isn't Aspergillis in the operating room and all the kinds of things that you need to know to run a quality hospital. Patients expect quality care; they presume you know how to take care of things and run things and they don't really care what their leftover urine is used for.

So that's a different level of opting in that's presumed from being in a dataset without knowledge.

If you are sufficiently anonymized, however, and you're no longer identifiable, then all the privacy law protections from a legal point of view – I'm not saying ethical, but from a legal point of view, would not apply. They're obliged, for instance, for incidence reports to see how many people still in the hallway or how many – so you end up being an aggregate in a dataset. Again, there, I think, you don't need an explicit opting in.

The only kind of opting in presumed might be at that level. I can't see other levels where you wouldn't need a consent or –

DR. MURRAY: That's a terrific question, I think. And I'm not sure that I follow your either/or as to whether it's going to be application of a principle or sort of balancing.

It's a sort of a principle balancing. The categories I was trying to provide earlier are an effort in a way to understand the structure that lies underneath that balancing.

So, for example, it matters what the purpose of getting the information is. It's really important if I show up unconscious in an ER that you know what my allergies are, right? And I can't tell you today, but that would be really important information to have before you start treating me.

So the information, you look at the sensitivity information and the utility of the information in the context of the purpose why people want the information – what's their purpose and what's their relationship with you? Are you in a care-giving relationship? Are you in potentially adversarial relationship with them, say as a life insurer?

So that's in a way the underlying moral structure, aspects of the information itself, purpose and relationship, I think.

Now, there may be more. That's a first pass at this. But that's the way I would think about it.

MR. ROTHSTEIN: Other questions? Yes, please, Bob?

MR. HUNGATE: Kind of an observation to start with. I'm perhaps dating myself, but when I hear the term "autonomy," I think of Robinson Crusoe on his island all by himself, fully autonomous, no issues at all. But most of us don't deem that the life we wish, so we choose to live in a society in a place in a community. But we don't, as individuals, any more understand what we've given up in that, and that's kind of where I'm coming from in this.

Your mention of simple choices made me wonder whether you're thinking in terms of within the system, the categories of information being a limited number of different degrees or whether that is the way in which the patient expresses maybe at the bottom of the HIPAA form the choices they would wish in terms of their degrees of autonomy.

Am I making any sense in relationship to the simple choices that you spoke of? Were those within society or were those at the individual interface?

DR. MURRAY: If the question is this first problem of to what extent and in what ways do individuals have control over the content – not access to the content of their medical records; it was a sort of threshold question that you're going to have to deal with, I think parsing out these, you know, handful or two handsful of relevant categories that pass muster with individuals and with health professionals as being meaningful and usable, useful, that's where you might make decisions about what sort of categories might individuals choose to have or not have in their medical record. And also whether to provide these under certain conditions of access, certain purposes.

It very quickly becomes an extremely complex problem. I don't mean this in a disrespectful way, but I want choices to be meaningful and not really formal, right? Presenting a huge array of choices with many boxes to tick off pretty quickly loses meaning for most of us.

So I would say, you know, let's road-test it, let's make it a meaningful set of finite, relatively small number of crucial choices, so that people really can exercise autonomy in a way that really serves their interest and is meaningful for them and doesn't turn out to be the case where they look back five years and say, "I didn't mean to deny this kind of information to these people. I didn't understand what I was doing." We want people to make choices that they can later take ownership of and feel good about.

MR. ROTHSTEIN: And I think the idea, to pick up on Bob's point, and this sort of goes back to the initial question I asked, is even more complicated technically, assuming that we decide what sort of stuff people will have a right to keep out of their records as to how to do that. So you could imagine not getting the information at all in the first place, having a right to excise that information once it already existed, having the right to modify that information in some way so that you have the diagnosis but not the explanation of a psychiatric condition or you have the effect of, say, domestic violence in the record in terms of broken bones or whatever but not the cause necessarily.

And so even reaching an agreement on what sorts of things, then we would have to deal with the issue of sort of technically how to do that and even making those choices is a reflection on what we consider objectively that someone subjectively might want to exclude. In other words, I mean, if we're really autonomous, we would have 280 million different choices as to what we want, but that's totally impractical. So we're going to have to, as a group, somebody is going to have to make a judgment call as to what it's reasonable to exclude if we want to do that weighed against the costs medically and so forth of doing so.

Steve?

MR. STEINDEL: I just wanted to add to your list. I think we need to consider the transfer of information out of your record to someone else, because I could see where if you're going to see someone for mental health purposes and they're keeping an electronic health record, you might want the information kept in that record but not necessarily transferred out.

DR. MURRAY: Right. To me, that would be the access issue: Who would get access to that particular category, right.

MR. ROTHSTEIN: And at our last hearings, which were not specifically on the National Health Information Network, we spent a considerable amount of time talking about third party access to health information which some would argue is much more of a problem than the unauthorized access to information where anyone with economic leverage over you can require as a condition of all sorts of things that you sign an authorization disclosing your records.

And, in fact, I'm in the process of trying to estimate how many compelled authorizations there are each year and we think it's – I mean, this is a very wide range – between 20 and 50 million. I hope to have a closer number in a few weeks. But, I mean, it's just extraordinary.

Marta?

DR. KNOPPERS: Isn't that where insurers – or let's just even take life insurers and say we're private contractors, we're doing a contract on your eventual death and it's, you know, a statistics game as to who's going to make money, lose money, and so on. Even there where you do sign that line at the bottom, by saying insurers for life insurance purposes will only have access to the following categories, a lot of the fear -- and the economic version, you can't get a mortgage or buy a car or get a loan if you don't have life insurance; you need life insurance in order to get another economic good. So it's an entry point for other economic goods in society but it's a forced disclosure of something that is really quite intimate to the person.

So you can say, okay, insurers, this is the game. We know you're private and premiums have to be, you know, no adversarial selection and so on and so on, then have them only get limited categories of access to data that they really need for their actuarial tables.

MR. ROTHSTEIN: Well, and the same thing could be argued for employers –

DR. KNOPPERS: Same thing, same thing.

MR. ROTHSTEIN: -- and so on, but the problem is that at the current time, with a paper-based system, there's certainly no practical way that this can be done. I mean, it just would be cost prohibitive, time prohibitive, and so even when they get them with limited authorization just as a matter of convenience, health care records holders tend to send everything because it's so much easier.

And unless we build into the electronic health record of the future the capacity to limit the fields of disclosure, something that is quite daunting, I would add, in many cases, we'll never have that capacity, and my understanding is that is not even currently being done.

DR. MURRAY: That seems to me a fundamentally important thing to attend to now. Again, maybe it's identical with this notion that I've been promoting of having, you know, a finite number of categories of information in terms of their sensitivity and utility or I would hope it's the same thing, but for example, a life insurer doesn't need to know if you and your wife needed to consult an infertility specialist in order to try to have children – right? – unless they can show good actuarial data that that is somehow of predictive in a way that they will use. Can we then find a way to construct, you know, data sets that limit? They would be limited data sets not in the sense that Dr. Fitzmaurice was talking about, but only the relevant data goes to you.

If we could do that, then we could actually improve some aspects of personal privacy with electronic medical records over the current paper record, which is actually as you've described it, the use of those records.

MR. ROTHSTEIN: Yes. Yes, I mean – and I'm very anxious to hear from Alan Westin after the break about public views on this, but I think a common misconception is that electronic health records pose this tremendous privacy threat when in the context that we're talking about, it represents a great hope or possibility of protecting privacy but only if we seize it, and I for one certainly advocate for doing so.

Other comments? Yes, please. Beverly?

MS. DOZIER-PEEPLES: Along the same lines as the life insurance, you mentioned in the beginning of your presentation that some European countries were early to get on board with privacy laws to protect the genetic information for this reason. Could you just kind of briefly describe what that looks like? I mean, is it an anti-discrimination type law or is it restrictions on what information insurers can obtain, or what exactly does that look like?

DR. KNOPPERS: Okay. There's three or four different approaches across Europe.

The first was Belgium in 1992 which put it in their chapter in their civil code on insurance, simply saying there will be no access to genetic data for life insurance purposes – stop. This was interesting because there were protests outside the Senate at that time by people complaining that their condition hadn't been labeled genetic, i.e., they were being discriminated against because insurers would have access to their data. There's one established. So you can legislate it simply with the problem that that leads to.

Plus, all of the approaches, and there's three more, that actually put genetics separate make it seem as though genetic data is not medical data, something different, which leads to exacerbations of stigmatization and discrimination and reinforces the idea that there's something spooky about genetic data.

A second approach is a moratorium. This is when insurance companies wanted to forestall – because once you get it into law, it's really hard; civil codes don't have sunset clauses, they're not statutes. And so voluntary moratoria. Sometimes time-limited because they wanted to open it up, but maybe they could change their mind in the future. But in France, for instance, until the recent laws, they had a moratoria; now it's in the law.

But there's still countries that have moratoria. Holland, for instance, has a moratoria, which means that they will test or ask for. But if it's in your medical record, they will still have access to it, which then leads the problem of people saying, "Well, I don't want to go participate in genetic research or go for a test when you're doing linkages or pedigrees to help my sister find out if she has it; if it goes in my medical record, then it's obviously so. There's a weakness to that as well.

And the last approach is one in the U.K. where they've set up a special commission which decides which conditions are so certain, genetically speaking, that of course a life insurer should have access to them. And the only one they've able to come up with, and even there they take it on and they put it back it and they off and so on, is Huntington. But they would have found that out from the family questionnaire in any event. I mean, they've been doing questionnaires – you've seen them; what did your aunt die of, your mother, your sister? – as a basic requirement for selecting what risk group you belong to. That would have been in your family questionnaire anyway.

So this commission who decides when there's sufficient actuarial evidence so that insurers can legitimately discriminate in spite of discrimination legislation is, you know, a work in progress, let's call it.

So it's legislated either in a code or in a human rights code or it's a moratorium or it's a commission.

The most interesting approach, though, and this is not a particular country and I'll stop there, is that the European Convention on Biomedicine and Human Rights in 1997, which countries as they sign in ratify their internal law has to be in conformity, all they said, and it's quite smart, all they said was no tests shall be done that are not done for medical reasons. That's it. They didn't mention insurance, employment; they didn't say who they were aiming this at. By saying that all tests should be done for medical reasons, that means you cannot do a test simply – a genetic test – for insurance purposes, because it's not a medical reason. That's an interesting approach.

MR. ROTHSTEIN: Tom, you wanted to comment?

DR. MURRAY: Well, just quickly. My conversion to heresy began in the early 1990s when I chaired a task force for the genome project on genetic information and insurance. And so we set out to make the case that genetic information ought to be and could be distinguished from other kinds of health-related information.

And over the two years of the life of this task force, we ultimately decided that was either dishonest or impossible. The great majority of conditions that people suffer from that affect your health, your longevity, that are causes of illness, that create expense, health care expenses, are very complex combinations. They're actions of genetics and a host of other environmental factors, developmental factors, and the like.

So to try to cleave off genetic information and treat it as distinctive, toxic, et cetera, in the larger context of, for your purposes, the electronic medical record, makes no sense whatsoever.

MR. ROTHSTEIN: With that final word, I want to thank the members of our panel and I also want to thank the questioners for helping to further flesh out these issues.

We are going to recess for 15 minutes, take a break, and then we'll back with Professor Alan Westin.

DR. MURRAY: Thank you.

[Break from 11:04 to 11:20 A.M.]

MR. ROTHSTEIN: We are back in session now. This is Day One of our hearings of the Subcommittee on Privacy and Confidentiality of the National Committee on Vital and Health Statistics.

And we have a unique privilege this morning not only to hear from one of the great experts in the country on privacy issues, and has been for many years, but also to get access to some very important new data that he has collected and will be sharing with us.

So I'm very pleased to introduce Professor Alan Westin.

Agenda Item: Presentation – Dr. Alan F. Westin

DR. WESTIN: Thank you, Mark. My name is Alan Westin and I am Professor Emeritus at Columbia University. My background is in law and in political science where I have my Ph.D. And I taught for 37 years on Upper Broadway in Columbia.

It's been about four decades that I have been working on privacy issues. I go to Japan a lot because of interest in what the Japanese are doing, and once when I got off the airport, there was a gentleman meeting me and he had a large sign and it said, "Father of Privacy."

[Laughter.]

DR. WESTIN: I like that, because especially in Japan I could even be Grandfather of Privacy and still be appreciated.

But the issues of information technology and privacy have been an occupation of mine ever since the 1950s and in my prepared testimony I give you a little background on the various studies and publications I have done specifically on the health area.

I kind of left this alone for the last three or four years as the HIPAA issues unfolded. I was an advocate of Federal health privacy legislation, but once it became the domain of the lawyers, even though I am a lawyer, I thought that I really didn't have that much to contribute to all of the very specifics about the privacy rule and so forth.

But now, with the movement toward electronic health records and technologically oriented health care system, I'm back in play, and as I'll mention, I'm the director of a new program on information technology, health records and privacy at the non-profit center that I head called the Center for Social and Legal Research.

I was interested in Tom's comment about being very careful to be aware of the limits and the challenges of technology approaches. I learned this very early, in the late 1960s when I was writing my first book on privacy and freedom. It was mail merge techniques were just developed in word processing, and so once I got an actual letter that was addressed to me and it said: "Mr. Alan F. Westin 1100." The second line said "Trafalgar Street." Third line said: "Teaneck, New Jersey," then the zip code. And then, using the mail merge, it started off: "Dear Mr. ll00."

[Laughter.]

DR. WESTIN: But the best thing was the first line of the letter: "You are not just a number to us."

[Laughter.]

DR. WESTIN: That's when I learned that technology can get things very wrong.

I'm here today to report to you on a survey that we have just done last week on how the public views the application of computerization to the health care system. Let me start, though, by saying that I view the electronic health record initiative as a very positive potential step in reshaping the nation's health care system, and the reasons that are usually given for doing it – enhancing patient care, reducing medical errors, reducing high paper handling costs – seem to me to be very, very worthy and important objectives.

And we all know that this is going to reshape the medical record as we know it and the flows of health information throughout not just the first sector, which is patient care, and the second sector, which is payment and quality assurance, but it's going to reach into the third sector, which are all the social uses of patient information – employers, licensing, insurance, research et cetera -- very important, but not themselves the givers and providers of health care.

And so I'm very pleased that you've taken the topic of sounding out how the public feels about this early in the game. And my program was able to sponsor with Harris Interactive a national survey on public views of application of computers and effects with HIPAA and also attitudes toward the computerization of health records.

Our survey was just in the field February 8th to the 13th. It was a telephone survey conducted by Harris Interactive. We had a sample of just over 1,000 respondents and they represent approximately 214 million adults. And a survey of this size has a margin of error, as the statisticians like to say, of plus or minus three percent.

Before I do that, though, let me just remind us all that long before we did our survey last week, there have been, by our count, 14 national surveys between 1978 and the present that have dealt either completely with health privacy issues or have had major sections of health privacy questions on them. And there are a series of sort of top line findings that are very well established.

When you ask people, here's a list of information about you; what do you consider to be the most sensitive, or which of these you would be most concerned if the information was released without your knowledge or consent, health information and financial information are always the top two scorers. And in many ways, I think personal health information these days would be slightly ahead of financial information in any kind of a rating.

The surveys show that people are concerned by very large majorities about the privacy and security implications of going to computers and having electronic collection and use of health information. This is because the public essentially views technology, rightly, I would say, as a two-edged sword – enormous benefits in many, many situations in our contemporary life, but also sharp problems and problems that often seem to be beyond effective control arising as a result of high technology settings.

Because of this, when people visit health websites, and we find that 80 percent of the online population report that they have visited a site that dealt with health and health conditions, they are often highly concerned about their privacy and security and they don't share their personal data or take full advantage of these sites because of that nervousness over the privacy and security dimension.

Especially we find in the surveys that consumers who have chronic or genetically based health conditions are particularly concerned about the flows of their health information into the zone three, all the social uses of personal information from the health sector.

So we began our survey by first repeating a question which we used in 1992 in the Harris Westin survey on health information privacy. We gave a list of people in the health care system and said: Do you believe that any of the following have disclosed your personal medical information in a way that you felt was improper?

When we asked it in 1993, 27 percent of the public, which represented then 50 million adults, said they believed that one of these five groups had released their personal medical information in which at the respondent considered an improper way. And this year, when we asked the same question, it dropped from the 27 percent down to 14 percent, which is really quite dramatic. It's almost halving of the number of people who have this feeling of improper release of their medical data.

In each one of the five categories, as you can see from the numbers there, the total was down. At the same time, we should remark that 14 percent of the current American population means 30 million people, so it isn't as if two people and a dog are complaining about this; we're talking about a very substantial part of the public that feels this way.

On the other hand, to put it in some perspective, in all the surveys that I've done over the years, we have about 25 percent of the public that says, I believe that business or government has invaded my privacy.

So you start off with a quarter of the American Population having a view that in general they have been the victims of privacy invasion. So a 14 percent figure is not outside the parameters that we'd expect when you have that kind of a general victimization perception on the part of roughly a quarter of the American public.

It was sort of interesting the health insurance company was the biggest drop – that was 15 percent in 1993 and it dropped down to eight, and I find it interesting that it's eight percent for the health insurers who were the dirty birds back in 1993 of getting to patients' information; now it's on a level with a clinic or hospital that treated you or a family member, which is sort of a rise one could think in the health insurer status and a drop in the clinic and hospital status.

We then turned to HIPAA because we asked ourselves: Is this drop in the perception that medical information has been released improperly an effect of the

privacy notice and the whole HIPAA roll-out system, which as we know dates from April of 2003? So in our question we said, In the past three years, have you ever received one of these HIPAA health privacy notices? We first described it, and the text of all of our questions is in the appendix of my testimony which you all have copies of so you can see exactly how we worded every single question and all the responses.

When I started out with this, you always ask yourself when you're developing a survey: What do you think the answers are going to be? And when I wrote this question, I said, geez, we'll get 90-plus percent of people who say they've received the privacy notice. You can't go to a doctor, a dentist, a pharmacy, you can't have health insurance without having had privacy notices thrust at you for the past two and a half years.

Astonishingly, 32 percent of the public, representing 68 million adults, say they can't remember ever receiving a HIPAA privacy notice.

Now, think about that for a minute. All of us in the room think that this is all-pervasive and the American public is seeing what's going on. It cautions you that many of the things that are done, especially with complicated lawyer-driven notices and the way in which this may be shown to people by pharmacist or a doctor et cetera, that 68 million adults, if they had to raise their hand and say, swearing on a Bible, have you ever received a privacy notice? "Don't believe so."

Let me make one other cautionary comment.

Sometimes in survey research, you get worried that people will give you what's known as a socially acceptable answer. For example, every study that asks people, Did you vote in the last election? Finds that millions and millions more people say they voted than actually voted because not voting is considered not socially acceptable. I don't think here that we had a socially acceptable phenomenon at all. I don't think if somebody said no, I don't remember getting a notice they were trying to present themselves in more socially acceptable ways.

On the other hand, it's always a question of which figure you like to look at. Two-thirds of the public does remember recalling a privacy notice, and that represents 158 million adults. Here, only one percent said they weren't sure, so the main figures are the ones that are interesting.

We then asked the people who said they remembered receiving a privacy notice the following question:

"Based on your experiences and what you may have heard, how much has this Federal privacy regulation and the privacy notices increased your confidence that your personal medical information is being handled today in what you feel is the proper way? And we got 67 percent who said it increased their confidence.

But please notice that only 23 percent chose "a great deal" as their answer and 44 percent said "only somewhat." That tells you that joy does not reign supreme in the nation over the HIPAA effect and that the verdict is out, you know, in terms of how people perceive their medical record being handled after the privacy notice phenomenon.

We then turn to the real intended focus of our survey, the electronic health record or electronic medical record situation. We described it in this fashion, and the way we described it I think it's very important for you to understand and you can make your judgment about whether this is the acceptable or best way it could have been done:

"The Federal government has called for medical and health care organizations to work with technology firms to create a nationwide system of patient electronic medical records over the next few years. The goal is to improve the effectiveness of patient care, lessen medical errors, and reduce the costs of paper handling. Have you read or heard anything about this program?"

Here, my prediction was sound. I said, gee, you know, this is not something that engages the American public yet in a deep way; yes, the President spoke about it in his State of the Union message. Yes, he went out to Cleveland and talked about it and made the television station. But I didn't think that this would have a majority of the American public shaking their head "yes, I've heard of the war in Iraq and, yes, I've heard about Social Security."

And so, 29 percent, which still represents 62 million adults, said they had read or heard about it.

And when I took a quick look at our demographics, it was, as you'd expect, the better educated, higher income, and technology using members of the public that were the ones that said they had read or heard.

We then developed six concerns that we said some people have about the effects of having an electronic health record system. We talked about leakage of sensitive health data, whether there'd be more data sharing without the patient's knowledge, whether there wouldn't be adequate security for health data stored on computers, whether this might lead to an increase rather than a decrease in medical errors, whether people would be less willing to provide necessary information to their health care providers because of their concern about computerization, and finally, that there might be a reduction of Federal health privacy rules in the name of efficiency in this kind of a system.

The bottom line, as I'll go into next, is that two-thirds of the American people say they're concerned about each of these possibilities.

Specifically, on the sensitive medical information being leaked because of weak data security, that was our number one choice. Seventy percent said they were concerned about this. And I give you the "very" column because it's often important to look at the people who choose "very" anything, and so 38 percent said they were "very concerned" about this.

More sharing was at 69 percent, and the highest "very" here, 42 percent, were worried about the sharing without their knowledge.

Inadequate data security, 69 percent, and 34 percent "very."
Increasing errors rather than decreasing them, 65 percent, and 29 percent saying "very."

That some people wouldn't disclose the information to their provider because of worries that it would go into computerized records, 65 percent, and 29 percent saying "very."

And that the Federal health privacy rules would be watered down, 62 percent, and 28 percent saying they were "very concerned."

It's typical in surveys like this, and I've been doing these for 30 years, that after you describe some kind of business practice or government program and you ask people how concerned they are about aspects of it that you develop what's often called a "tie-breaker question" – that is, you say, well, you've told us about the program, you told about your concerns; what's your balanced view then on this program?

And so our question was phrased:

"Overall, do you feel that the expected benefits to patients in society, which we already mentioned in our non-question, outweigh potential risks to privacy, which we had just probed in our concerns, or do you feel that the privacy risks outweigh the expected benefits?

And the winner was: No one. The public is deeply divided – 48 to 47 percent on their view whether the benefits outweigh the privacy risks or the privacy risks outweigh the benefits.

And I was able to look at the demographics before I came down but not by the time I was preparing this, and I'd say the most important thing is that the people who believe the privacy risks outweigh the expected benefits are very widely distributed across the demographic categories. That is, it's not concentrated in Democrats and liberals and African-Americans and so forth.

It is so widely distributed that you would be absolutely sound in saying that this impacts at that level men and women, upper income and lower income, high education, low education, et cetera.

And we will be publishing in a couple of weeks all of the demographic information, the factor analysis and so forth, and I'll make sure that your Committee gets this.

One of the things I've done over the years has been to create what's called a segmentation of the public on privacy issues. The way we do this is to create three or four trend questions that tap fundamental attitudes and then we see in the public how many people take the strong privacy view on all of the trend questions, how many take it on some of the trend questions, and some of them that don't take the privacy view on any of them.

And that enables us to create a high, medium and low segmentation of the public and puts some numbers on how many people fall into each category and then to look at the demographics for each of those segments to say: Who are the people then who are not concerned about privacy, or highly concerned?

And we created this here by taking the six concern questions that you have heard me describe, and if somebody chose their concern in five or six statements, we call those high electronic medical record privacy concern. And what I think is a striking finding, 56 percent of the public scores high in their privacy concern related to electronic health records.

Sixteen percent fell into the medium category, meaning that they expressed concern in three or four statements. Fourteen percent chose one or two. And no statement was chosen by 14 percent of the public.

So, we have a solid, national majority in a high electronic medical record privacy concern camp. And that compares, and I think this is a very meaningful comparison, with our studies that show only 35 percent of the public when you deal with consumer privacy issues score in this high or fundamentalist orientation. So we have almost double the number of people in an intense privacy view in the health area as in the general consumer privacy area.

And this is about what we found – it's a little more intense, but it's close to what we found in 1993 when we did a similar segmentation when President Clinton was promoting the national health insurance plan.

We wanted to test what I have always thought was gong to be one of the most critical issues in this whole electronic health record development, which is: What's the role of the patient going to be here? Not the providers, not the insurers, not the health data analysts, but the end user, the patient.

So we framed a question that read:

"Since most adults now use computers, the new patient electronic medical record system could arrange ways for consumers to track their own personal information in the new system and exercise the privacy rights they were promised. How important do you think it is that individual consumer tools be incorporated in the new patient electronic medical record system from the start?"

Eighty-two percent of the public believes that this is important, and here, 45 percent rated this as "very important." Only 17 percent did not see this as important.

That was a little bit of a socially acceptable answer here. If you don't have to pay any price, if you're not worried about the tradeoffs between the patient's access or the patient content control and giving the patient these tools we described, what does it cost to respond to this, to say "Yup, that's very important"?

Having said that, though, I still think that this is an extremely important finding. And the way I view it is that this is a public mandate for what I call a privacy design specification for any electronic health record system. That is, from the start, the public is saying: Program me in in my privacy choices, my privacy access, my technology access. Otherwise, I am not going to be confident that this system serves my interest and is good for me and for society.

And so I think everybody who is an advocate and manager and participant in the building of this system as the decade unfolds really has to say what laws, what rules, what practices, what technology arrangements, what education about privacy and what kind of building of positive patient experiences will it take to get that 47 percent of the total public to feel that the privacy risks are not outweighed.

Let me turn to some conclusions and recommendations I draw from the study. Incidentally, the study is being released today, so you are the first to hear it, but I do believe you'll read something about it in the media and in health publications and so forth since it's now out there for general public discussion.

My first premise is that an electronic medical record or electronic health record system does hold enormous promise for patients, health care delivery, for breakthrough research, and for the interests of the whole society. I also think that probably the system is more likely to proceed now than at any time in the past, and I'm sure all of you here remember that in the ‘60s and the ‘70s and the ‘80s and the ‘90s there were major efforts to go to computerization in the health care system and enormous sums of money were spent with I think one has to say limited result, if not often complete failure of some of those health records.

I remember when Dr. Weed's problem-oriented record was seen as the great gateway by which we would computerize the record and change the whole way in which the medical record would be used in the system. Didn't happen.

On the other hand, I think it could happen now. Step one, medical professionals are now pretty much technology conversant. They've got laptops, they've got cell phones, they're used to going into databases. And obviously this is something of an age-related phenomenon, but I think the generations of current health care professionals are now more open and ready to using technology than has been the case in previous decades.

Secondly, the technology is much more powerful. We now have data mining and data linkage techniques and we have software power, a whole host of tools which major technology firms have been developing and university research has developed which I think hold much more promise to achieve the cost effectiveness and the reduction of medical error problems and so forth than previously.

It's true, of course, that technology has been growing steadily ever since the computer came along, but up until fairly recently, I did not myself see the technology tools as having the sophistication and the depth and the reliability that I think is needed. And so I'm more optimistic about the technology opportunity that I would have been five or 10 years ago.

On the other hand, I hope the survey results remind us all that no matter how good the technology and no matter how ready the medical practitioners are to embrace the technology, this system will not succeed if public concerns over privacy are not understood and addressed.

So, what do we need?

I think there needs to be an institutionalized privacy-by-design working group, and the best analogy I think is the excellent LC program with the human genome project where major money, very talented people and institutional support was given to examining the kinds of issues that the breakthroughs in genetics rush before society. So it has to be active, well-funded and impressively staffed.

It may be that such an organization is government supported but not government run, that it calls for a kind of consortium of government, private sector, consumer and patient advocates and so forth. But it's charter is: How do you design privacy, from the start, into an electronic health record system?

Secondly, I would be very worried if privacy becomes a sub-topic of what is being discussed now as an electronic health record standards board. I think the standards board is extremely important in terms of interoperability issues and regional system linkage issues, et cetera, certainly, the medical record issue itself.

But if privacy is consigned inside that board, I'm afraid that it will not have the right kind of pressure, the right kind of poise, and so forth.

So I would like to see an independent privacy standards board that sits alongside the larger technology and record standards board.

The kinds of things I would see a privacy-by-design working group to carry out would be, first, take the excellent materials that over two or three decades we've developed on how to do privacy risk assessment and threat assessment and apply it into the concrete development of electronic health record systems. There are many, many organizations that do this kind of privacy risk assessment. There are auditing firms, there are law firms, there are university firms and so forth. It is a specialty and we know how to do it well. I think it has to be a continuing privacy risk assessment, not a one-time.

Secondly, I see the group looking to identify the kind of system design elements that would enhance rather than defeat privacy interest. For example, I think there's probably broad agreement that creating one national health record system organized nationally and under the Federal administration no matter how benign is a disaster for privacy and therefore, regional systems with linkages and interoperability standards and so forth seems to me an initial major design component that is privacy oriented.

There are many other things that I could talk about here but in the interest of time I just want to say that across all of the technology and organizational design choices that are coming up, I think this privacy-by-design group should have the charge to say: What are the privacy implications if we do it this way compared to if we do it that way?

Third, I think that identifying anonymization techniques would facilitate research and data trend analysis is absolutely essential. This was mentioned earlier today, that if we pursue privacy at the expense of fundamental epidemiological and health system research, it will be a heavy casualty. And it doesn't have to be, I don't think. There'll be some problems that we'll find it very, very difficult to work out, but in general, I think we can apply anonymization techniques in ways that will still allow important and socially valuable research to go forward.

In order to do this, though, I think we will have to – and I'll talk about this in a minute – try to conceptualize a segmented medical record that will have in it parts that are for identification and use and other parts which from the start are identified as the kind of things that are subject to anonymization and therefore will be organized differently in the medical record. And I'll explain that in just a minute.

As far as the legal and policy rules are concerned, I think most people would agree that you can't just take the current HIPAA privacy rule, slap it on something called the electronic health record, and think you've done your job. It's going to take a lot of very thoughtful consideration as to what the policies and the legal rules should be for the kind of systems that will be rolling out with electronic health records. And I think this privacy-by-design working group should pay a lot of attention to just that kind of issue.

Obviously, lots of others will be doing this. For example, the Markle Foundation has a find project on connectivity in which they've got excellent people looking at the privacy and security issues. So even for the beginning I'm not suggesting that this kind of a function is going to be the only one. It's going to have many, many parallels and competitors and so forth and to me, that's fine.

Fifth, I think that the privacy-by-design working group should try to identify and test procedures that would, responding to the 82 percent of the public, empower individual patients to access systems directly so they can see certain kind of information that's there and so they can carry out with all the power of the computer technology the privacy rights they are given.

Today, we have a paper-based, almost a ballpoint pen-based, patient access system. I think that as we move the medical record into high computerization, we've got to move the patient access and patient control functions into equal technology driven opportunity.

A hundred and sixty-five million Americans are now online. We have become a society which more and more has people comfortable in using information technology in the online world. I think that should be a major understanding as we think about this system – not the patient coming to the doctor's office, sitting in a chair, being thrust a notice, but being able to sit in their kitchen or their study and have access to the system under the right defined rules and so forth and to be able to exercise their privacy rights wherever they are, sitting in a airport terminal with their wireless and coming and looking at something in their medical record. I think that's a tremendous opportunity that we've got to grasp right away.

Finally, as these approaches are done, it seems to me we have some real test beds we can think about. As the regional programs unfold in the electronic health record systems, those are the beta sites for looking at the privacy design world.

My experience in doing a lot of empirical studies is that you want to go to the place where the pioneers are putting forward new technologies, changing the way things are done, altering balances of rights and responsibilities. You want people there who are going to do objective, empirical research into what difference does it make that this is happening now in this clinic or in this hospital or this doctor's office in this IMS Health data set. You want really to have people studying hard the actual impact of technology in the organization, on the patients, et cetera.

That's why my organization has created the program that I said I'd say a little more about. We see ourselves as a not-for-profit research organization continuing to conduct public opinion surveys on how the public and various health care leadership groups and others feel about more and more and more specific aspects of an electronic health record system.

We'd like to do some of those empirical case studies that I mentioned of how the programs are actually working as they roll out. We would like to help develop the legal and policy rules that are necessary for privacy, confidentiality, subject access, due process and so forth. We think this will require going quite beyond HIPAA.

I was very glad to hear the discussions of what the Europeans are doing and what is going on with electronic health records in other countries. There's much to be shared and learned in those countries, and one member of my staff is an expert in this and we would like to see not just what the legal rules are but what the actual experiences and patient reactions are in countries that are also experimenting and moving forward with electronic health records.

We are going to be publishing in a couple of weeks a white paper in which we take a broad look at computers, health records and privacy in the 21st century and we will have a variety of other reports and we expect to publish a quarterly electronic newsletter and as always to organize seminars and hold conferences on program themes.

You can go to our website, www.pandav.org. You'll find posted there today the top line results of our survey, a report that we've done on how the public views health privacy, survey findings from 1978 to 2005. I think many of you will find a lot of interesting specifics there, not just like the top line couple of comments that I made.

My testimony from today's hearing will be there and this PowerPoint will be there as well.

We will publish an expanded survey report with all the demographics and factor analysis, and the white paper that I mentioned is listed there.

Let me just add one thing that was not in my prepared testimony because I was stimulated by some of the conversation earlier today to share this with you.

If we think about a patient medical record as a one thing, all unified and all there, I think we'll be making a privacy-by-design mistake. Rather, I think we could imagine a six- or seven-segmented and formatted medical record which the technology is perfectly capable of storing and retrieving in that kind of segmentation.

And how might it be divided?

First, a segment on personal identifiers – name and address and Social Security number and all the stuff the hackers will try and get to.

Second could be a medical transaction segment – came in, complained of knee problems, probably arthritis, going to put him on whatever the latest acceptable Cox-2 is or isn't.

Third would be a prescription history, something that systematically listed the pharmaceutical agents that had been used by the patient in their medical record history.

Fourth would be anything that had mental health or psychological or psychiatric components.

Fifth would be life style information, all the terribly sensitive stuff about sexual life and drugs and alcohol and bungee jumping and everything where life style as we know can affect the medical system.

Finally, what I'll call anonymized data, data that from the beginning is seen to be important for research purposes and which is stored in the medical record ready to use in anonymized form so that we institutionalize some of the research function right from the start – we don't wait for a research protocol to be done and then go back and scratch our heads and say, you know, what do we need from the medical record? Even though, of course, that'll always be necessary with highly customized research.

But for a great deal of epidemiological research, pharmaceutical utilization research and so forth, I think we could create from the beginning a set of patient data that would be what we see as high value for research and therefore under the proper research access would be acceptable from the record.

And just to show you where this goes, the psychiatric segment obviously requires the highest level, or one of the highest levels, of access power. It probably should be kept in encrypted form because we really must pay attention to data security.

As you know, many health systems today only store the psychiatric data in encrypted form and I think that's exactly the kind of requirement that would have to be set, whereas the anonymized data, there would be no need for patient consent, opt in, opt out, or anything. From the beginning, explain to the patient there would be a set of data that was going to be useful for public health purposes, for research purposes, and so forth.

Now, I'd be the last one to say that I've just given you the Lord's work on a segmented medical record, but I think it's interesting to think about and to say, since the technology is capable of giving people access to 1, 2 and 4 but not to 3, 5 and 7, that if we have the right rules as to patient access and the right rules as to third party access and provider access, we could think about a medical record as being a set of records, not a record, in which rules of privacy and access and consent and disclosure would be customized for the nature and sensitivity and functions of the different types of information.

With that, just to show you there's a lot of work to do, let me stop and invite comments and questions.

Questions, Answers and Comments

MR. ROTHSTEIN: Thank you very much. We greatly appreciate your sharing the new survey data with us and I'm sure that will generate some questions. But I'm equally, or perhaps more fascinated, by some of your comments and conclusions and recommendations.

Let me just ask one question before we sort of open things up. And I'm sort of taken by your suggestion that there should be some sort of external group that you call the privacy-by-design working group that would be tasked with helping to design a system for electronic health records in which privacy would be a key element.

I don't disagree with the aim, but my personal observation is a very practical one, and that is, it seems to me that the electronic health record train is zooming down the track and the thought is that those of us concerned about privacy, our job is just to make sure that the train doesn't leave the track, whereas what I hear you suggesting is that the privacy element is so fundamental that it really needs to be worked out in advance of the system itself.

So perhaps you could explain your thinking some more.

DR. WESTIN: I guess the first thing I should do is say "amen." I think you said it just right. I think that the privacy issues are so central to whether this will succeed – incidentally, we can be very concrete and suggest that Congress and the state legislatures will take their cue from how the public feels about this whole electronic health records system. You want get appropriations? You want to get Congressional committees to give this the kind of support that the human genome project was given in the LC appropriation? You've got to convince the legislators that this an acceptable privacy system.

So I don't think the train can go down the track; it's not going to have any fuel when you get to it if there is not this kind of clear mandate for privacy. Now, I'd be the first one to say, being a political scientist, but how you institutionalize this, where you locate it, and public/private, and funding and so forth, are all very important questions.

But however it's done, I'm looking for there to be a free-standing, high prestige, well-funded and well-staffed entity that is like a privacy impact assessment group is inside Federal agencies if you know that was required, a privacy impact assessment in Federal agencies now when they are in e-government programs and so forth.

You know, I'm looking for those kinds of functions to be institutionalized, and I think you've got it right. Some people believe the train is going and it's too late already. I don't think so. I think that the train is gathering steam but one of the other things that I have in my prepared testimony that I get on to the laptop was that unlike some situations where a business program or a government program is in fundamental collision with the consumer and privacy groups, and you really have to have a confrontational, dragged out kind of battle, that is not the way I read the situation here. I think the health care community is privacy oriented. I think the technology groups that are building many of the software tools also accept the importance and centrality of privacy.

But I think there's more community of interest and intention in this area than you find in some other areas, homeland security or telemarketing and other kinds of collision areas for privacy versus the other interests. If that's the case, then I think there's more possibility of creating this kind of an institution and getting the right kind of support for it than there would be in some other privacy area.

MR. ROTHSTEIN: Thank you. And now, questions from my colleagues? Mr. Reynolds?

MR. REYNOLDS: Excellent survey and excellent coordination of this information – thank you.

You mentioned on Slide 14 that for any national EMR system, as you think of the philosophy of segments, as you think of the philosophy of privacy, as you think of the philosophy of structure, and then you think of regional things, one of the things everybody ran into in HIPAA was everybody had done their own thing for so many years and set things up and then we tried to come up with a standard and it was a bit of a fist fight getting it all done, whereas do you see any of these categories or any of these segments or any of these other things that could be put in place so that as regionals do their work, they are basing it on some kind of a foundation that when you try to tie multiple regionals together, if you don't go through a singular EMR system, you have some kind of a structure that allows you to play off of the benefits of doing it regionally but then be able to transfer that information because more and more with the special centers of excellence and everything else that goes on, people are going to be moving around to get care? It may be the more significant only in the future than they are now, so any comments you can make on that?

DR. WESTIN: I have to start by saying I'm not a technologist though I watch technology. But my understanding of some of the things that are happening among the technologists is that, first of all, they are working hard on interoperability. They understand the need for operating standards that will cut across various technology approaches and systems.

And secondly, that they're looking at linkage techniques rather than uniformity techniques.

So I would myself assume that you should take a look at those kinds of studies and offers and see whether they're going in the right direction and if not, some pressure might be needed to make sure that the people who are developing the whole new system see the need for that kind of interoperability and so forth.

But I want to start by saying that though I think there are answers out there, I'm not a technologist.

On the other hand, I think that probably the moment is right, given the technology, to move into much more uniformity in medical record formats. I don't think every hospital's way of doing it and every clinic's way of doing it must be saluted and preserved.

And to the extent that we are able to come up with highly refined and correct estimates of language and the techniques, I think that we are going to move in the next decade to a much more uniform system of reporting and formatting and so forth.

A couple years ago, I helped a company that was developing some small medical record software. They asked me to come in and deal with the privacy issues. And what they were doing was hoping to move the physician from a pad with a piece of paper to a hand-held data device which was all formatted so that if there was a diagnosis or there was a prescription, it was uniform throughout their entire system. That seems to me to be easy to do and we're ready to do. So there are some approaches that already are pretty well tested that I think will now move into more and more use.

MR. ROTHSTEIN: Mr. Hungate?

MR. HUNGATE: A question. Going back to your segmented record, in thinking about consent by a patient, then one level of consent might be the linking of the personal identification information to the autonomized data. An example might be that your genomic information would be in your personal identifier and the PO coding information would be in the anonymized data. Is that a correct conclusion?

DR. WESTIN: Yes, that's very promising. In other words, I was taken by the comments earlier that the dilemma with anonymization is that when you truly anonymize without any preserved linkage file, you lose the ability to update the file or to add relevant information to it.

And so one solution, as I'm sure everybody knows, is the trusted keeper solution. For example, some years ago when anti-war demonstrations were at their height, the American Council on Education, which did annual surveys of college students and would ask people, "Are you using marijuana? Are you against the war?" and all kinds of questions of attitude, decided that they would promise anonymity and they would take the linkage file and move it to Canada and promise that if a subpoena was ever given by a Congressional committee, people would thumb their nose from across the border and would never give the linkage file.

So I think there's a range of ways that we could go at this that have to do with both the linkage of the personal information to the anonymous data and also maybe we are going to need these trusted organizations, and I'm sure you know that there are organizations that are now promising to be your trusted agent for purposes of your creating your own medical record, your personal health record, and so forth.

MR. HUNGATE: Right.

DR. WESTIN: And I think there's a lot of promise in that because the trusteeship concept, if it's the right people institutionalized in the right way with the right legal sanction behind them so you don't have to run to Canada, I think that's very promising.

MR. HUNGATE: I agree. My sense, though, is that I ought to worry about it. I believe in chaos theory.

But a friend of mine is one of those people that's developing that trusted organization and he assures me that he can take care of all the privacy issues through what he's doing. I have reason to doubt from this discussion that that could really happen.

But given that there is strong incentive and strong commercial interest behind these kinds of efforts, I wonder if there isn't some way to get an understanding within that group, the developers of those specialized personal information systems, that they, in order to be a trusted source, are going to have to have a way of dealing with this privacy issue and whether that's not an interested party, too.

DR. WESTIN: I think that's very important.

MR. HUNGATE: The next level.

DR. WESTIN: I'm glad you mentioned it.

MR. ROTHSTEIN: Dr. Steindel?

MR. STEINDEL: Thank you, Mark. Thank you for this really fascinating and very, very timely survey.

I like the idea of this national privacy data board in getting involved in the design of EHRs early et cetera; it's a very strong point with a lot of people and I think very necessary. What concerns me about that, though, is expressing privacy in EHR systems is done usually through security. Privacy is a concept and you have to somehow express that in the software.

And yet we see in one of your bullets where you ask the five or so different questions that the consumers do not have high confidence in computer security. So what should we be addressing? If we put this board in place and the board says, you know, the EHR should be designed with this, this and this, whatever it comes out to be, how do we assure the public that we can design computer systems that express the needs of the board?

DR. WESTIN: I thought about that when I was putting a label on that, and I thought about privacy and security design function. My problem is I think that the security area probably belongs deep in the technology sector much more than it does in what I think of as the policy orientation of the privacy.

When people ask me what's the difference, my favorite way of saying it is: Data security is the way you keep your promises of privacy and confidentiality. It doesn't define what is privacy, it doesn't set the confidentiality, but it enables you to have confidence.

On this, though, let me express some deep reservations.

Many of you have seen recently that one private data supplier just was hacked into by a Nigerian ring which

set up 50 false customer accounts and got into 150,000 to 500,000 records and used them for identity theft. All over the world, identity theft through a variety of techniques, sometimes employees inside being corrupted, sometimes the hacking from outside, it's a very insecure world, and I think anybody who runs a data system would say that it's next to impossible to provide truly 100 percent data security.

So we're dealing here in how close can we get to a system that will give confidence to the public that in fact there's adequate data security? That's why in my full testimony I talked about the fact that I think we're going to need a biometric identifier for the public and I think we're going to have one by the end of the decade primarily for homeland security purposes, but I think this electronic health record system will be another driver of the creation of a biometric system. And that would have an enormous damper on identity theft because it would have a much more secure way of authenticating who people are.

And of course that raises its own privacy issues and we'll have to address them, but if you take a combination of biometrics – for example, a finger image and a retinal scan – and you add to it a smart card chip, you've got about as secure a way of authenticating people as you could want. And the ability for somebody to corrupt all three of those is going to be extremely small.

So I think that when we look at how technology can enhance privacy, not just press against it, let's keep in mind that there are technologies that are going to enable people to be much more secure in access to their data and other people's access to the data via technology solutions.

MR. ROTHSTEIN: Dr. Harding?

DR. HARDING: Well, I very much appreciated your testimony, and I think most of us know that Mr. Westin was the research coordinator for the National Commission on Confidentiality in Health Records in 1980 or so, around that time, and really led out in this area, kind of was a precursor of this group and we really appreciated your work through the years.

Let me ask you a political question. Your Slide Number 11 said that it's 48-48 on the issue of privacy versus the benefits and that there's no red state/blue state kind of thing; it's not a political – there's a little bit of education involved.

If you were talking about the Congress votes when they are influenced by the electorate, how do you go about doing that --

DR. WESTIN: I think probably –

DR. HARDING: -- in such a split, non-demographic kind of way.

DR. WESTIN: I welcome your question. Let me try to give you a reaction.

First of all, the people who believe the benefits outweigh the privacy risks are not hostile to privacy. It's just that when they look back on it, they see the benefits as being quite significant. So it isn't even just the 47 percent that you have to address. You have to address that segment of the 48 percent that even though they think the benefits outweigh the risks, they're going to be responsive to privacy.

My feeling is that we're going to need some champions in Congress. We need some Senators and Representatives who will say, gee, this is a good issue. Same way that it took genetic information legislation and other kinds of legislation to need a champion, I think we need to identify and stimulate some leading Senators and Congresspersons to say, this is an issue of the decade; it's good for me, it's good for the country, because that's the way things get done.

When I remember how the LC program was put in, that was done because one member said we ought to put some money into these ethical, legal and social issues instead of just assuming that it'll happen. And that whole program really was the result, first, of a staffer and then of the

Senator putting it in and not being opposed because the money was not up to the national debt level or something like that.

So politically, I think that Congress is an important place.

Now, the Administration obviously is important, and HHS is important. But from a larger political sense, I think there should be political leadership of the privacy campaign, and I could think right off the bat of some people I would love to see lead it and probably all of you, too. We should make it happen.

MR. ROTHSTEIN: Final question of the morning. Ms. Wattenberg?

MS. WATTENBERG: Yes. You said before that electronic health records goes beyond HIPAA, and I just wanted to get a little bit more of a read from you on –

MR. ROTHSTEIN: Sarah, a little closer to the mike, please.

MS. WATTENBERG: Oh – sorry. You said before that electronic health records goes beyond HIPAA, and I just wanted to get more of a read from you if you have any sort of more specific thinking on that.

DR. WESTIN: Not really, not yet. I took a list of the HIPAA mandate and laid it next to the electronic health records. It didn't seem to me that there was a good fit yet because these whole issues about control and access as they were mentioned this morning, the whole concept that you would have a patient participation that is technologically enhanced and so forth, isn't there yet.

And so I think that's where we would have to start thinking. What will it take?

And, of course, one of the problems is the whole liability system, and everybody's aware that as you build these records, practitioners are going to say: What duties do I have in relationship to these records not to be brought up on malpractice and not find myself disciplined et cetera, et cetera?

And so I think we have to rethink some of our liability system if we're going to use the medical record in as positive a way as we'd like to. So that's just something.

MS. WATTENBERG: Can I ask just a follow-up question --

MR. ROTHSTEIN: Certainly.

MS. WATTENBERG: -- since there's so many attorneys in the room?

I mean, is it true that it's a liability issue that if a patient says, no, you can't have access to a certain kind of information and treatment is prescribed based on what they have, is that still a liability issue for the physician? This doesn't make sense to me, but –

MR. ROTHSTEIN: Let the Internet listeners appreciate that there were several nods of the affirmative on the question.

I want to thank Professor Westin for his typically expert and provocative comments and we appreciate very much your coming here to spend some time with us.

We will now stand in recess for our lunch break until 1:15 and then we'll hear from Panel 2 on privacy in health care and in society.

[Lunch break from 12:22 P.M. to 1:25 P.M.]

MR. ROTHSTEIN: Good afternoon. We are back with the afternoon of Day One of the hearings of the Subcommittee on Privacy and Confidentiality of the National Committee on Vital and Health Statistics.

Before we begin Panel 2, I just want to mention that we have at the moment no public testimony scheduled from 3:15 to 3:45, so we will just move forward the rest of the afternoon agenda and so we should be adjourning approximately 4:15 this afternoon.

For those of you who were with us this morning in person or on the Internet, I'm sure you will agree that it was a very fascinating discussion and either you could look at this in a either positive or negative way but it certainly raised more questions than it answered, so it was provocative, and that's the positive side of it, and also daunting, in the negative side that there's so many issues that we need to deal with.

And I'm sure that this afternoon's panel is going to be equally provocative, and I appreciate very much the folks who have joined us today for Panel Number 2. So without further ado, I will recognize Panel Number 2, reminding you to please limit your initial remarks to 20 minutes and then we will have at least 45 minutes for questions and answers with the Committee and staff.

So with that, I would like to welcome and recognize Dr. Bernard Lo.

Agenda Item: Presentation – Dr. Bernard Lo

DR. LO: Thanks very much, Mark. It's a pleasure for me to be here. I know that this morning's panel is a very tough act to follow, and since then we've had our healthy NIH lunches, so there may be a bit of a post-prandial slump. I will take literally, Mark, your exhortation to be provocative, and I'll be provocative.

So let me start by asking: How many of you in the room have had back pain, knee pain, shoulder pain that was so great that you thought about taking a medicine for it?

MR. ROTHSTEIN: Have or currently have?

DR. LO: Ever have had?

[Laughter.]

DR. LO: Okay. So then you have been very interested in the news about all these new miracle arthritis drugs that turned maybe not to be so much of a miracle. And we all know that one of the large manufacturers actually voluntarily withdrew one of the Cox-2 inhibitors, Vioxx, because of reports that it actually increased cardiac problems.

Well, let me tell you first about what one integrated health care system clinic did in response to this, really taking advantage of electronic health records. And they actually wrote about this in a publication which is referenced in your hand-outs.

What they did is they notified by patients by mail within 24 hours of Merck's recall. They immediately withdrew the drug from the pharmacies so you could not get a refill prescribed. They notified every provider, all the patients for whom she had prescribed the medication, and they also used their electronic record so that the next time the patient came to clinic, an alert, a flag, went up to the physician reminding her to talk with the patient about how to manage this new information.

So this strikes me as a very innovative, effective use of the sophisticated electronic health record really to respond quickly to changing information all in the best interest of the patient.

Now, of course, once the original drug, Vioxx, was withdrawn, everyone then started to say: Well, what about other drugs? And if you were taking another drug – Celebrex is another cousin Cox-2 inhibitor – you naturally asked: Well, is it safe for me to take that drug? And I think those of you in practice probably had your emails and phones ringing off the hook from patients who were very worried.

And the news picked this up and this headline reads: "Cardiologists Question Safety of Vioxx-Like Painkillers/Doctors to Avoid Prescribing" two others that he named.

There's another headline, again, raising questions that were on every patient's mind, I think.

Well, given that you really only had the existing data to go on and that it was impossible on the spur of the moment to design and carry out and analyze a large, definitive, randomized clinic trial, how could you provide relevant information?

Well, I'm going to skip that and just say one example – this was carried out by Kaiser of Northern California – was to use, again, a comprehensive, sophisticated, electronic medical record to look through and identify patients who had been on various drugs, follow them out through the medical record to track outcomes. And this study was done very quickly, as soon as the question was raised, using pre-existing data that already existed within a sophisticated, integrated electronic medical record system. And this was published in a fast-track publication through the Lancet because it's such timely news of vital health importance.

Well, let's think for a minute about database studies. If we're going to have the opportunity to answer questions quickly using existing data on questions that have real sort of health import for many people in the country, what kind of database do you use?

Well, first I would argue you need comprehensive data. You need to pull together lots of different types of information which may not be integrated in existing medical record systems. You need to integrate pharmacy data, outpatient visit data, hospitalizations, laboratory tests and deaths.

And I think the point I want to sort of put in front of you is to do this in systems that aren't totally integrated, you need to maintain individual identifiers to cross-link all these different types of data.

You also, I would argue, need complete follow-up. What that means is you have to be able to access care outside the system.

So if a patient in Cleveland Clinic or in Kaiser has chest pain, gets taken by the ambulance to the nearest emergency room and has a heart attack there, that data, that information, may or may not be captured in their home base electronic record, so you need to be able to integrate care outside this, and again, you need an individual identifier for that.

You also need to have very few refusals or dropouts. If you allow people to say, "I don't want you to use my personal information in this kind of research," I would argue that you're not going to get a scientifically valid answer to the question: Does Drug X cause greater or fewer heart problems compared to any other drug you might choose?

The reasons are that you would, first of all, lose statistical power if a lot of people didn't allow their data to be used. But more importantly, you could well have selection bias, that the people who don't allow their records to be used may be different from those who do in a way that actually makes the results come out differently than the way that the scientific relationship actually is.

Let me also parenthetically say there are other uses you might make of an electronic medical record, again, in an ideal system, to respond to this kind of breaking news of great health impact. And this, I think, would fall into the rubric of quality improvement. For many of these drugs, and this is one of them, the current data showed that it's only at the higher doses, highest doses, that you see this adverse association between use of the drug and adverse cardiac influence.

So one thing you might do is alert physicians who are prescribing above the recommended dose that once they cross that threshold, they may be entering into the realm of undesirable side effects.

Secondly, again, as part of a theoretical quality improvement mechanism, you might say that I certainly have patients who love these drugs because they haven't responded to other drugs, they have stomach problems that have precluded the use of the standard arthritis drugs, for them the advantages of these drugs might outweigh the benefits, but only because we know that they have other contraindications to other drugs. And again, you could put a sort of tickler, a reminder, in the electronic record system if the doctor prescribes one of these Cox-2 inhibitors, to say, having tried other drugs, that may be safer.

Okay, let me switch gears a minute and say, of course, this is not the only example that's been in the news recently about an important health issue where there was real uncertainty based on the existing, randomized clinical trials and the existing data where the drug's benefits outweighed the risks, or vice versa.

So here's an example from psychiatry, as it turns out. "Prozac," this headline reads, "Linked to Child Suicide Risk. Study Finds 50% Greater Chance. Companies Defend Anti-Depressant."

So again – front page news, a huge issue in terms of depression in children, adolescents, a serious public health problem and the concern on the one hand that effective drugs might be withheld because of unsubstantiated concerns about suicide versus the countervailing concern that these drugs might actually increase suicide risk.

So again, one way to try and get at more data on this question is to go back to large databases that have the comprehensive data I just talked about.

Well, when you're dealing with a sensitive condition, and certainly depression and suicide are very sensitive, we need to think about the potential for benefit and the potential for harm. The potential for benefit from database might be to do database research, to use the electronic medical record to inform patients and doctors of concerns about, or uncertainty or controversy about, the use of drugs, and again, to enforce quality control in prescribing.

But on the other side, I don't think very people are concerned that the fact that they have severe back pain or knee pain is very sensitive information like psychiatric data, but with sensitive conditions, concerns about privacy and confidentiality obviously are heightened.

And as you well know, medical health records may have special protection under HIPAA and under various state laws and in fact it's not uncommon in our health care system to have separate mental health and medical providers and records. At least in California, much of mental health care is a carve-out totally separate from the medical system.

Lest you think I'm sort of a wild person sort of ripping apart privacy and confidentiality, let me say that there clearly are very important reasons why privacy and confidentiality are important in this. You know, we believe it encourages people first to seek medical information and secondly to disclose sensitive information.

And I would actually posit, suggest, to you, that's not just psychiatric information, substance abuse, things like that that are sensitive. But just by going to a doctor for an ordinary exam, you get a routine check-up, you get asked questions about all kinds of things –- sexuality, for example; you're asked to take your clothes off; as part of cancer screening, the doctor may probe your body in ways that would be unthinkable in the non-medical context. So for many patients, just going to the doctor is very sensitive, let alone special topics like mental illness or genetic information as we talked about this morning.

We think that confidentiality prevents stigma and discrimination and we think it also is morally and ethically important because it respects patients as persons, so there's good reason for starting with a very strong presumption of confidentiality in the medical care system.

But, and again, this is something that you've all worked on with your work on HIPAA, confidentiality is not an absolute ethical goal and actually I would suggest perhaps dual policy goals – and again, this is part of the sort of supporting language in HIPAA – on the one hand, we want to protect confidentiality, but on the other hand, we want to have access to information for clinical care, for public health and research.

You know, what's interesting is everyone would agree, I think, that in SARS, anthrax, avian flu, things like that, the public health system needs access to individualized medical records for contact tracing, epidemiological source outbreak research, and the issue is not does the patient have to consent; the issue is really notifying the patient in a compassionate way and carrying out that investigation in a way that respects their privacy and confidentiality to the greatest extent possible.

Research we ordinarily think of, I would argue, as being somewhat more elective, that wouldn't it be nice to do research but it's not a moral necessity, and certainly our common rule, our Federal regulations for research going back to the Delmar(?) Report suggests that it's optional – it's morally desirable, but optional.

But I would turn it around and say on issues like Cox-2 inhibitors for arthritis where, you know, these drugs really, two of these drugs at least, are in the top 10 prescribed drugs in the country, blockbuster drug; the question of depression and suicidality in children and teenagers is again a pressing public health problem, I would argue to you or I would suggest to you that certain types of outcome research may be more like public health than research that we typically think of (?).

So if we are going to really do certain types of databased research on really important topics of great public health import, what are some of the issues?

Well, first I see the technical challenges, and you talked about some this morning. You need compatibility between different organizations that have very different data formats.

In terms of how myocardial infarction is recorded in one medical care provider electronic record system may be very different than the way it's recorded in another. And you need to be able to protect identifying links if you're going to merge different databases that have different types of information you need to answer the research question.

Patient authorization. As you know, under HIPAA, the starting presumption is you need patient authorization and then there's certain exceptions or waivers. The empirical data are very clear: If you ask patients about authorization to use their personal health information for research, they say it's important. And in fact, in studies where attempts were made to enroll patients into a database where just their data will be collected and pooled, about half the patients do not give permission.

Now, I must say that one problem with all these studies is it's not clear what the patients were told about the database, why it's important, how it might help other people like them, so it's not clear how well they're informed. But they certainly gave a preference in terms of not sending back the consent, the authorization.

What's, I think, even more disturbing is that there's clear evidence of a selection bias. One very nice published study – it's actually a Canadian database on stroke – showed there was clear selection bias, that the sickest patients, the most complicated patients, did not agree to allow their data to be used. Now, whether it's they really refused, they were just too sick to fill out the form, we don't know.

But my concern is that incomplete data may be misleading and it could conceivably even be worse than no database at all because you might find associations that really are spurious.

So the issue of patient authorization and how it might actually serve as a deterrent to the kinds of research we would like to see done in certain conditions I think is a tricky condition that I hope you can sort out.

Okay, and finally there's the oversight regulations which I think is really in your bailiwick. I think we need to distinguish between what the regulations literally say and how they're implemented, interpreted, on the front lines by IRBs, by privacy boards, by researchers.

As you know, there's provisions under the current HIPAA regulations for de-identification of data, a waiver of authorization.

First, let me point out that de-identified data will not suffice to do the kind of databased research I was talking about because you need identifiers to link these different databases. Waiver of authorization is permissible, is possible, under HIPAA. I think that the impression you get talking to IRB chairs and researchers is that IRBs are confused about this and they're not allowing waivers to be granted in situations where it would seem from a, you know, straightforward reading of the protocol in HIPAA that it would fall under that.

Now, obviously IRBs, privacy boards are allowed to be stricter than the regulations themselves, but this, I would suggest to you, has the impact of making research more difficult.

So I think at the very least there needs to be guidance for IRBs and privacy boards and I actually put in researchers as well as what kind of outcome databased research is permissible under HIPAA.

And I also think there's a problem at the other end, that even if the IRB approves a study involving existing data under a waiver of authorization, the provider may be reluctant. The hospital, the doctor, the clinic may be reluctant to provide the data.

In California, we've seen that where there's actually a state-mandated cancer registry where access to that squarely falls into one of the HIPAA black letter provisions. Hospitals are not submitting the data to the cancer registry as they had been before HIPAA and as HIPAA clearly permits because of their concerns about their liability and their concerns about patients' concerns about privacy.

So I think this at least needs to be clarified, but I would argue even if the regulations work on the ground as they were intended to on paper, it still may make it difficult or impossible to do the kinds of studies we talked about in the beginning of my presentation.

Okay, so let me be provocative and throw out some things for you to chew on.

First, I think there needs to be a lot of public education about the value of databased research, both its value and of course its limitations, but also I think about the tradeoffs between confidentiality and patient benefit as evidenced by the possible usefulness of outcomes research.

Secondly, I would suggest we might want to think about some research as being similar to public health, and the implication, I would suggest, might be that the ethical issue might be notification rather than consent or authorization as the kind of the entrée into the data.

And finally, I think there needs to be a focus on confidentiality as well as privacy. Professor Westin and others have clearly documented the public's concerns about leakage of their personal data to people who really shouldn't be seeing it. So how personal health information is protected and could be protected and making that as airtight as possible I think is a real challenge.

So let me stop there, and I hope I stimulated you to think about these things. And I'm going to now disconnect so the third speaker can get –

MR. ROTHSTEIN: Thank you, Bernie, and I can tell you the answer is: Yes, you did stimulate lots of sort of synapses firing and I'm other sure others will have questions as well when we have our panel discussion.

It's now my pleasure to recognize our second speaker on this panel, someone who has shared her expertise with us many times in the past, and we're always very grateful for her comments, Joy Pritts.

Agenda Item: Presentation – Dr. Joy Pritts

MS. PRITTS: Good afternoon. I'd like to thank the Committee for inviting me back to speak with them.

And I've been asked to speak about patient interest in health information technology, and I have to say, after listening to Dr. Lo, I'm really torn between doing a full rebuttal of everything he said and going on with my presentation as planned, but I think there'll be time for that during the question and answer period.

Where I would like to start is a period almost 10 years ago I think it was, with a quote from the former

Secretary of Health and Human Services, Donna Shalala. And she was talking about health information technology and people and where it might lead us, and she posed a question which I think is as valid today and will continue to be valid in the future as it was 10 years ago. She asked:

"When all is said and done, will our health records be used to heal us or to reveal us?"

And I think this is a question that we have to continue to ask as we continue in the process of developing health information technology.

Essentially what she's asking is a point which has been brought up by other speakers during the day, which is this balancing between the benefits and the risks, and particularly what I'm going to speak of is how patients perceive those benefits and risks.

We'll start with the benefits, because clearly there are benefits to patients to having some information in electronic form. Patients aren't all about just protecting their own information and keeping it quiet. There are some valid reasons for having information in electronic form and developing health information technology.

A lot of these have been mentioned earlier in the day, so I'll just briefly touch on some of them. The improved quality of care from doctors having complete set

of records, from being able to read the records between different health care providers where illegible records result in unfortunate errors. The records are more complete. Theoretically, at some point in our lives, we may have a longitudinal record of our health from before we were born until the current time, which gives a doctor a full picture of your whole medical condition.

Electronic records are more readily accessible to providers. They could be more accessible to patients if the system is set up in a certain manner.

They can eliminate duplicative tests which for any patient who's had to undergo more than one test for a condition can really be a large improvement.

And they can streamline the administrative process. Any human being who's ever been in a hospital system knows what it's like when you go from department to department and you're asked the same questions every new department. And with an electronic medical record, that should be no longer a problem.

Now, that might seem to be a minor point, but when you're in the hospital and you're upset, it becomes a very major inconvenience, and in fact, a major interruption to the health care system.

But for every one of these benefits, you can kind of flip the coin and see also the risks that are associated with it. These risks are real, and they must be taken into account when we look forward in developing health information technology.

Yes, the records are more accessible. They're accessible to providers in their office, at home, on wireless networks, and it raises the concern with a lot of people as to how secure the records are. Are they being protected? As has been told to me, somebody could potentially sit in a cafeteria with a WIFI and pick up some information that they probably should not be picking up because it's being transferred within the hospital.

The records are also potentially more accessible to others outside of the health care system. As interesting as the National Health Information Infrastructure is kind of bubbling along and we're talking about how it's developing, there are different players who are coming into the system that I'm not really sure were originally anticipated being central core players in the system, including, as I'm sure you've heard me on this course before, banks and financial institutions.

And we see that this is happening right now. There are financial institutions who are administering health savings accounts, so they have access to very detailed health information. They are not covered by the privacy rule, and we have yet to see any of the regulations come out from the banking authorities as to how they may use that information under the FACCT act – that's the acronym and I can't remember what all the letters stand for right now, but it's dealing with financial information.

So we're still not sure, even though it's moved to that point, it's happening now; we don't know if that information is covered. We know it's not covered under HIPAA, and right now actually it's not covered under anything because the regulations are out.

There's also more information accessible, as I said here. You could have this longitudinal record of everything that's happened to you from birth, and one hand that's really good, but a lot of people have things happen to them at times in their lives, often when they are younger, that they would just as soon not have subject to review perhaps when they are older, and that's something that some people are a little bit concerned about: Are the sins of their youth always going to be around to haunt them?

Overall, there is also seeming to be kind of a loss of control over your health information, and this just comes from technology in general, that people believe that once it gets in the system, it's beyond their control; they don't know who has access to it, they don't know what information they're seeing. And I would say that the notice of privacy practice unfortunately has not really helped alleviate this concern.

One of the other benefits I mentioned was that it's possible that the information could be more accessible to the patient. But it's also possible that it will be less accessible to the patient. If it's written in code, some uniform codes so that it can be transferred easily between providers and health plans, are patients going to be able to access this information and understand what it says?

It's a problem that's not too much different from how it is now. Many people can't understand what their health records say when they get access to them. They're often written in medical jargon or shorthand. But there are resources that you can go to that explain what that means.

So there is some concern that you'll get some obscure computer code somewhere down the line and you'll have no idea what it means. Some people who get explanations of benefits have certainly already experienced this. And there's no requirement under HIPAA that any of this information be translated into plain language or English.

There's also a possibility that the movement towards electronic records is going to leave some patients behind. There's a real potential here to widen the health care gap because not everybody has access to a computer. In fact, some of the most vulnerable populations health-wise do not.

And as we move forward, it's a privacy concern in the sense that people should have access and be able to control their information. It's also just a general issue of fairness and equity that, as we move along, people should be brought up so that everybody has an even playing field.

So these risks really can result in harm. And the harm that comes from the downside of the electronic medical records includes the stigma that's attached to certain medical conditions being spread. The consent that most people that are at the top of almost everybody's list, they're afraid they're going to lose their job and they're afraid they're going to lose their insurance, and that's because of the way our health care system is.

Most people have their health care insurance through their employer and people are concerned that if their employer finds out they have certain health conditions that they're going to lose your job and if you lose your job, you lose your health insurance, and if you lose your health insurance, you're in big trouble.

The majority of bankruptcies in this country today involve health care costs and health care debt.

There's also a concern about police power and how authorities are going to access information if it's on this electronic database. Now, I believe Dr. Westin talked this morning about how opposed people would be to a national database, and I can guarantee you, every time I've been in a meeting or a conference and anybody mentions that word, you almost see the helicopters circling the room; it really does provoke a very adverse reaction from people.

And the unique identifier has been mentioned as something that is necessary in order to transfer information between providers, and it also an item which provokes very adverse reaction among many people. And I think there are a couple of reasons why people are really adverse to the unique identifier.

One is the experience we've had with Social Security numbers. They ended up being used for a purpose that they never really were intended to be used for, and people are afraid that the national health unique identifier would fall in the same category.

I think it also probably increases the potential risk for identity theft. Having all this information in an electronic form, you now not only have the person's name, their address, their date of birth, their Social Security number, you probably also know their mother's middle name

if you have this longitudinal record, so you have information all in one place that is very tempting for people who have possible bad motives.

A recent example of this is the case that happened out in Seattle at the end of last year where an employee of a cancer clinic obtained the patient identification of a cancer patient. He had his name, the Social Security number and his date of birth, and he obtained credit cards in this patient's name and he charged $9,000 under the patient's name and I think four or five cards, all while the man was undergoing chemotherapy. And I think that we need to be aware of this because people who are sick are vulnerable and they do not have the time or the strength or the energy to be dealing with these types of issues.

Now, the person in Seattle actually pled guilty and he was sentenced under HIPAA for a criminal violation of HIPAA, and I believe that's the first one in the country.

But I will also say that many legal experts came out afterwards and said that they believed that HIPAA didn't even apply to this gentleman because he isn't a covered entity. He's not a health plan, he's not a health care clearinghouse, and he's not a health care provider under HIPAA.

So there was a lot of discussion in the legal community saying that, you know, people who hack into medical systems to get this kind of information probably are not subject to the HIPAA civil and criminal provisions, and that is a gaping hole that needs to be fixed.

When you're balancing these interests, I think that you would find when you talk to different patient groups, you'd find that they give very different weight to the benefits versus the risks.

And the question pretty much comes down to, well, what do you think you have to lose? If you're healthy, if you're young, if you're kind of in the background – nobody really knows about you, you probably don't have a lot to lose. But if you're sick or somebody in your family's been sick, then the ratio changes pretty quickly.

It also changes depending on in some ways what your background is. We've heard a lot that people who are in different ethnic groups have different perceptions of these major databases, and a lot of this runs back to Tuskegee and the mistrust that that engendered in the African-American community, and that needs to be repaired, even though it's not clinical trials going on or things of that sort, but the trust in the clinical community as to what you can do with health information comes from how you've been treated in the past.

Now, there are other groups of people who also have a lot to lose. If you have the fortune, I guess, of being famous, your information is probably a lot more vulnerable than somebody who's kind of anonymous like most of us are.

When President Clinton was hospitalized, they found 17 people who tried to access his health information while he was in the hospital who weren't supposed to have any access to it. They ranged from doctors to clerks. Not one of them was fired, I would like to note. They were suspended.

I think that's what's noted here is – I'd like to quote from the newspaper article. They talked to some of the employees. They interviewed one who's leaving and she said, "I'm not surprised. People are nosy. It happens all the time." So it's not only President Clinton, ex-President Clinton, who has to be worried about it.

This curiosity factor also applies to people in small communities where everybody knows everybody else. It happens to people who work within the health system – you go into a hospital; everybody knows you. People are curious, they want to see your records. People know you because they're related to you. You're involved in a divorce action or something of that nature and people have an interest. So it's not just people who are famous, but there's a large kind of curiosity factor that applies to many just everyday people.

Also, people who have diseases that are still largely stigmatized in the community have a lot to lose. Here in Washington, we saw this a number of years ago with a local hospital where a gentleman went to a local hospital and he knew the clerk who was checking him in. As a matter of fact, they both worked at another job together in the evenings.

And the clerk was curious as to why the gentleman was there, and she found out he had HIV. She went back to her workplace, the other workplace, and she spread the rumor all around that the man had that "alphabet disease." And they made his life pure, unmitigated hell.

He sued the hospital and he won $250,000, even though the hospital had all these great policies in place about who could have access to the information. And the reason he won was because even though they had those policies, nobody was really enforcing them.

There is a recent survey that's done by HIPAA Advisory which came out a few months ago which showed that a lot of people are not following up on the privacy policies that they have in place. They're not monitoring. And that just is not following through the way people should.

On the other hand, there are a lot of people who have chronic diseases for whom following through with having electronic medical records and following through with their care would provide an unbelievable benefit to how they can manage their care. Being able to transmit their health symptoms on a daily basis to their doctor could really remove a lot of doctor visits and improve their care.

Another thing that we've seen – I've heard this at a lot of conferences and in my older age I'm fairly amused by it – there are some of younger people who say, "Well, you know, we surf all the net all the time. You know, we don't really care about privacy that much. We get information all the time."

But I'd also like to say that this is the same group of people where, you know, they get on these Internet sites and they're not giving necessarily the accurate information. You know, the 25-year-old male is now an 18-year-old female who's in college or something.

So, you have a group here that's used to giving information but they're also used to subverting the system, and patients have always tried to subvert the system and we should be aware that there just may be new ways of doing it.

I think some of the unifying things here are that, no matter which end of the continuum you're on, people want to have some ability to control how their health information is used and who it's shared it and they want to know who has access to it. They want to be able to trust that it's being used properly, as Dr. Lo said, that it's being used for health care purposes and not other purposes.

And they want accountability. If somebody violates their privacy, they want the person held accountable. And I've heard this time and again.

We did a series of focus groups with veterans dealing with how they felt about having their health information used for research purposes, but that's a whole other discussion, but one of the items that came up repeatedly was: What happens if someone violates the rule and actually discloses the information? So this is a concern that people have. They want to see accountability.

And – which leads to, well, does HIPAA do this? And I would say it's a good start, but it doesn't do it all. In some ways, HIPAA was outdated the day it was written because the health care system continues to evolve so quickly that it's hard to keep up.

One of the areas that really needs to be addressed is that HIPAA does not directly cover everybody who will have access to health information, to a health information technology. This is something that has to be resolved by Congress; it's not something that HHS can solve.

But as we see, you have many more players coming into the picture and we don't know how they are covered, or whether they will be covered. And the system is leaving the station and the privacy protections are back on the platform.

In HIPAA, as they are written now, there are at least many people who believe the penalties apply only to covered entities. That also needs to be resolved by Congress. Penalties should apply directly to anybody who improperly accesses health information. And I think that's maybe what they were intended to do, but there are a substantial number of people out there who think that they do not do that.

And I also believe on a more practical note that that the notices of privacy practice need to be improved. They aren't doing the job, and it would be, I think, good for everybody to go back and revisit them and look at why we want them and what information they should have in them and why they're not serving the purpose, because time and again we have heard that people don't read them or if they do read them, they do not understand them.

Thank you.

MR. ROTHSTEIN: Thank you very much, and I know we'll have questions for you at the conclusion of the panel presentations.

And the third witness of this panel is Mr. Thomas McLellan.

Agenda Item: Presentation – Dr. A. Thomas McLellan

MR. McLELLAN: Thank you very much.

I enter the debate on the other side, I guess. I do think the electronic health record train is leaving and my pitch is really quite a simple one. That is, that addiction treatment information should be part of the electronic health record.

And there are really two reasons that I take this position. One is that it's necessary for public health and public safety and two, because you can do it, okay? And I'll try to be brief and direct in talking about this.

I should say I'm a researcher in the substance abuse treatment field. I'm a professor of psychiatry in the Department of Psychiatry at the University of Pennsylvania and I have a small research institute. But I'm not an advocate; I don't represent a particular treatment perspective or organization.

I think it's very important to have addiction information, substance use and substance abuse information, in an electronic health record because addiction itself is a chronic illness that requires treatment. There are at least two million people in specialty care alone, but many more receive unrecorded care through primary care and other mechanisms of mental health.

Moreover, literally every month, new medications, therapies, interventions are entering the scene, which means that there's going to be more access, more availability, more options. It's going to be a much more mainstream event.

So, for no other reason, it's time to acknowledge that addiction's an illness, that it's being treated in the health care system, and that it needs the same kind of information to manage it as any other illness.

But wait, as they say on the game shows, there's more. And that is, that addiction is part of addiction or substance use. Even problematic or excessive substance use, sub-diagnostic, is a very important part of the management of lots of other chronic illnesses. Diabetes, hypertension, asthma, breast cancer, sleep disorders, chronic pain, they all are affected by substance use. Too often, these illnesses now are badly treated because of poorly disclosed, unavailable information.

Now, a lot of people think you don't really need it; this is a completely segregated system, the addiction treatment, and it's a very simple one. You basically take a substance abuser, you put him in a box called – I don't know if you have these slides if you're looking for them – but you put him a box called an addiction treatment program. They stay there for 28 days and like a washing machine, they come out a non-substance abuser.

Well, that's the old days. Now, addiction treatment is much more like the treatment of other illnesses. There's an acute care phase, usually in hospitals, for brief purposes to transfer often to specialty care, but sometimes directly to continued care. And all of the options that are available in other areas of medicine are becoming, or are already, available in the treatment of addiction.

The point here is that like the rest of medicine, like the rest of health care, addiction also is affected now.

So those are the reasons why I think addiction treatment, addiction information, substance use information, should be included.

Now I want to tell you why I think it's possible, and I take this broadly within the rubric of the Institute of Medicine Crossing the Quality Chasm principles of patient-centered care. So all the points that I would make are within that rubric.

First, you might say, no, you can't really do this with substance abusers because they have diminished capacity or they have loss of control; they don't have the wherewithal to make informed decisions about the use of their information – and let's remember, it is their information. That's not so.

Yes, people in the throes of withdrawal are under temporary incapacitation as are patients who have strokes or who have terrible pain. It's not qualitatively different. And it is temporary.

What we're recommending – I'm recommending – is that the sharing of the information within the confines of the medical health system be the standard, but as Dr. Lo was talking about, we think within the boundaries of patient-centered care, the patient definitely has to be notified of this and definitely has the right to deny that.

Too often now, the standard is not for the benefit of the patient, just for the sake of reduced workload and hassle, not sharing simply to hide behind that. Not to say that there aren't important issues – of course there are – but my point, again, is that the same issues as lots of other – of the ones that you'll be confronting.

Now, unlike other areas, I think there are special provisions that are going to have to happen if this is to be done the right way, which is what I think we all want.

One, there are codes for most of the contemporary medications, therapies, interventions that are presently under delivery. They're not widely disseminated, they're not widely used, and that prevents accurate communication of things. And I think that's very remediable.

Finally, however, the addiction treatment specialty care sector has special problems due to decades of under-funding and segregation really. There aren't the capacity for computer integration information management specialty that there are in the rest of health care, and I think special provisions are going to have to be made.

So in summary, I think for the sake of patient safety and public health safety, it is necessary and wise to include addiction information into the developing electronic health record. It's good for both the patients and for the rest of those who are affected by addiction and substance use in mainstream health care.

Moreover, I think that within the confines of patient-centered care and existing statutes, it is possible and practical to integrate this information, and that would be my call.

So – thank you.

MR. ROTHSTEIN: Thank you very much.

Questions, Answers and Comments

Well, those three statements were certainly provocative and they've provoked me to ask several questions, but I will keep my initial questions short so my colleagues will also get a chance.

I'd like to begin with Dr. Lo and ask you a question that relates both to your testimony and testimony that we heard this morning. And I think it's fair to say that you suggested that certain kinds of sensitive information should be treated distinctly, is that fair to say?

DR. LO: I'm not sure I'd say that. I just noted that certain types are currently treated.

MR. ROTHSTEIN: Okay. So is it your position that that should be changed or that should be continued in some degree in the –

DR. LO: Well, let me go back to the topic.

MR. ROTHSTEIN: Okay.

DR. LO: To the extent that certain types of information are singled out for special treatment, it makes outcomes research using large databases on those conditions harder to do because you don't have access to comprehensive, integrated information.

So I think as you decide whether genetics or HIV or psychiatric therapy should be separate or not, keep in mind what you may be giving up in terms of research that I would argue is close to public health, obviously.

MR. ROTHSTEIN: Okay. Well, instead of asking you the question then – I'm still determined to ask this question –

[Laughter.]

MR. ROTHSTEIN: -- so maybe I'll start with Joy and then ask the others to comment.

We have heard several people clearly say at least that there should be separate rules for certain unspecified classes of medical information that's considered to be sensitive. And the question I have, and I think it's a very important one, is whether those should be inclusion rules or retrieval rules, which are quite different.

In other words, let's suppose we decided we wanted to treated Condition X separately because it's very sensitive. Does that mean that X does not get into medical records in any place, or does that mean in retrieving those records, certain classes of people who have access to the records don't get that?

And I think that's an important distinction, and so let me ask whether – I'm not saying you've said anything about that; I'm just asking whether you've got a comment about that point.

MS. PRITTS: It is my understanding that there are electronic health record systems that are designed the way that you spoke of, which is that all the information is in somebody's record, but it's not necessarily retrievable by all health care providers in a system.

MR. ROTHSTEIN: Yes, I think in Alan Westin's comment when he was talking about a segmented health record, he wasn't talking about six different health records or six components. I think he was really talking about a sort of a retrieval algorithm, and so –

MS. PRITTS: I'm sorry – I didn't see his, but I know I'm thinking of – I believe it's Duke University or somewhere; I may have the wrong –

MR. ROTHSTEIN: Well, he said that the elements of what he called a segmented health record were personal identifications, medical transactions, drug history, mental health which had to be encrypted, life style information and anonymous data that could be used for research purposes.

And my take on this is that what he was suggesting was a retrieval algorithm where the information where the information would be in the record. Yet, an argument could be made that in promoting patient autonomy there should be certain classes of information that the patient should have the ability or right to not get into the record at all. And so I'm trying to see if you buy that and if you do, is it an inclusion rule or is it a retrieval rule?

MS. PRITTS: It's a hard question to answer. I think it depends on who – I mean, even within the patient community you'll get different answers on that because there are some patient privacy advocates who very firmly believe that the patient should have the right to say what gets in the record and what doesn't get in the record. And that's kind of on the extreme edge of things.

I would say the group that's kind of moved one over from that says that the patient at least gets to say who gets to see what's in the record. And I find it a little disturbing that as we're talking about the electronic health record and how consumer driven this is, I keep hearing that term used with it, that it's consumer driven, that really the consumer, the way things are now, doesn't really have a whole lot of say in who the record goes to and how they can use it.

The retrieval system helps in that it gives the patient the option. It kind of fits within the privacy rule because it's like the patient can request a restriction on how the information is used for treatment and payment in saying, "Well, yes, it's in the record, but I'd prefer that this information only be available to my treating physicians."

MR. ROTHSTEIN: So you used the example in your testimony about things that you did many years ago, right?

MS. PRITTS: Right.

MR. ROTHSTEIN: But under your discussion of a retrieval rule, you wouldn't have the right to expunge that now; you would just have the right to limit who could see it, is that what you're saying?

MS. PRITTS: You know, on the one hand I think it makes perfect sense if you want your medical record to be complete to have it all there.

On the other hand, you know, a lot of us would like to go back and erase things that we did in the past. And they may not really be relevant now.

And I'm not a physician, so I don't feel like I'm qualified to say how relevant some of that information is.

But I do remember the DES cases when they had to go back quite a bit of time to find out information about mothers who took DES and were able to make the connection with I believe it was cancer in their daughters.

So, you know, I think from a medical point of view it's kind of hard to say that you should be able to really just expunge your medical record at some point.

MR. ROTHSTEIN: Okay. Bernie, now that the statement has not been attributed to you, would you like to comment?

DR. LO: No, no – I like your second question better than the first.

[Laughter.]

DR. LO: I would think that, as you put it, retrieval rules are a lot more flexible than exclusion rules. So what happens all the time, I think, in medicine is that it depends on the situation. If I'm coming in for a vaccination, I think a lot of information really isn't relevant to that encounter. If I'm coming in because I suddenly fall down in the street and I'm comatose, a lot of information, all the drugs I've ever been prescribed and taking become very relevant.

So I think if the data are there, then they're potentially usable for the direct benefit of the patient for clinical care decisions and presumably a treating physician in an emergency or acute situation should have access to a pretty broad spectrum of information. But if the information never got in in the first place because of an exclusion rule, then that would be absolutely irretrievably lost.

MR. ROTHSTEIN: MR. McLELLAN?

MR. McLELLAN: Yes, I think the principles are safety and efficacy first, patient preference second, because they have the rights to deny. But with those rights come unavoidable consequences.

If you don't make pertinent information of yours accessible for your medical care, you cannot expect complete health care.

I still think that's the right of the patient, because after all – and it's quite consistent with the Institute of Medicine's rules. But it's like full disclosure to the patient – yes, you may keep this sensitive information out of your record, but it may complicate the treatment of your other condition down the road.

The one that I can't quite figure out is where public health and public safety information – the patient has cholera and doesn't want that put in his record, is that permissible? What level does that occur?

The other point I would make is that if you go down the list of things that are seemingly not very sensitive and those that are very sensitive, you're going to quickly find that the sensitive ones are way more prevalent than the not sensitive ones. Even something like a vaccination, you have a terrible reaction to it two weeks later is going to be sensitive.

MR. ROTHSTEIN: I have one question from your testimony. You mentioned – I don't believe you gave a figure, but you said there were many people who were in drug treatment programs who receive "unrecorded care," right?

MR. McLELLAN: Yes.

MR. ROTHSTEIN: How many of those would forego treatment altogether if it had to be recorded?

MR. McLELLAN: An excellent question. It's an issue that we're looking for all the time. But you can imagine if you don't know how – I don't even know how to ask that question.

The answer is I don't know. But it is widely reported. Not just for substance abuse but for mental health diagnoses of all types. Physicians prescribe medication under a different diagnosis. They don't record them at all. The primary care physicians give brief interventions for substance use that aren't recorded at all.

MR. ROTHSTEIN: Yes. I mean, that's one of the things that does give me pause. If we adopted an inflexible, non-patient-controlled system, are we going to have a proliferation of Mark's drive-through, no-record, no-research, no-questions-asked health care that would cater to people who are very privacy conscious and would we be, you know, making things worse not only for their treatment but for all sorts of other things that we're concerned about – public health and research and so on?

Let me recognize other members of the Subcommittee with questions. Mr. Reynolds.

MR. REYNOLDS: In one of the previous presentations, it talks about the public being divided equally between – 48 percent say that the benefits outweigh the risks to privacy and 47 say the reverse. As I listened to the presentations – we had a class last week on how to look at things maybe differently, and I heard "no, because" and "yes, because."

So as you hear presentations, some people say, "No, don't do it because of this" and others say, "Yes, do it because of this." So think about how you would answer it if you said "yes, if" or "no, if." In other words, "yes" – so, Joy, for example, in your testimony, you obviously have a lot of good points of what the concerns are. But if this were to happen, what would be the list of stipulations that would absolutely need to be, you know, foundation agreements, and the same thing with Dr. Lo and the same thing with Tom.

So, you know, that's because we continue to hear testimony where it's very clear that there are keen points of view. And those points of view all have merit. But as you try to drive to if this should or shouldn't happen and if it does, what would be the stipulations under which you could take the 48 and 47 and try to make somebody – you know, make that a much larger number that accepts it and then the ones that don't, they would have some way to do something. So –

And I'm not asking you to re-testify everything, but just – you know, those are the kind of things that I'd just love your opinion as to whether those approaches would be more of a way to really get to the bottom line on what ought to happen.

MS. PRITTS: I see your point, and I think that there are some basic issues that – first of all, I think that it's going to happen whether you want it to or not because people are moving into electronic health records. They'll do it in the medical field just like they've done it in every other field. So I don't think that's a question. It will happen.

And so, when it is happening, I think that it's essential that – I mean, this kind of goes back to really the basic. I think that if people trusted the system a little bit more, then you would weigh more towards people who saw the benefits. But people don't trust the system right now, and so I think the question is, well, how do you make them trust the system?

And one way you make them trust the system is you make sure that everybody who has access to the information is subject to penalties if they use the information incorrectly. I mean, and in an ideal world, what you would also do is you'd say, you're going to have health insurance no matter what your medical condition is.

And since you asked, I'm going to go there, because I think that one of the reasons people don't like their medical information being kept in electronic form where it's, you know, shared a lot between health care providers and insurers, they don't like their insurance companies to know what's wrong with them. And the reason they don't want them to know what's wrong them is because they're afraid they're going to lose their health insurance.

And I don't know that you're really going to solve this issue until you solve the general just health care issue of how people get health care in this country. I know that's not really where we wanted to go with this, but I think it's a macro level issue that really needs to be addressed because I think fears would be much reduced if people weren't afraid they were going to lose their job and they were going to lose their insurance and then they're going to lose their house if they get sick.

MR. REYNOLDS: I understand when you asked a "yes, if" you expect –

MS. PRITTS: Okay!

[Laughter.]

MR. REYNOLDS: No, it brings the issues right – well, maybe a little more clearly.

MR. McLELLAN: I'll throw my two cents in. I don't think there's anybody who wouldn't say yes if, just as you were saying, the system were secure and under their control.

I'm on an Institute of Medicine panel right now looking at quality of care issues, a follow-up to the Crossing the Quality Chasm. We're debating this issue because there are tremendous errors in medicine at all levels because of lack of information. At the same time, there are dangers.

But nobody, I think, would take the position that it is inherently bad to have that information in the hands of people that can do you good. The question is how, not whether, how to keep it out of the hands of people that can do you harm, and who has authorization.

The second question, I think, was answered by the Institute of Medicine in their last report, and that it's patient's information. It's not the health system's information, it's the patient's, and that person has the right to ultimate decision with the attendant responsibilities for what occurs in the course of their care.

Now, that's quite glib for me to say it's simply a question of how. There are lots of thorny issues. Financial systems have been very good – not perfect, but very good at handling this kind of information. Perhaps you can learn lessons from them.

MR. ROTHSTEIN: Dr. Harding? Did you – I'm sorry –

MS. PRITTS: I disagree with something you said. I think there are some patients, there is a group of patients, that even if the information were secure, there would be a large difference of opinion as to who should have access to that, even within like researchers.

Some people feel it's my information and even though they'd be in the public good, I don't want researchers to have information to it unless they ask me. And it goes back to what you're saying about asking, but it's not that it's secure. It's almost like the sense of ownership, that it is my information, it's part of me, and, you know, I want people to at least ask me if they can have it.

MR. ROTHSTEIN: Richard?

DR. HARDING: You, Mr. McLellan, have raised a unique thing of testimony that we have had for some time where you have suggested that we should have substance abuse, or one of the sensitive issues, made a part of the general record instead of being separated out in some way. I think you're the first person who has made that recommendation to us.

Now, is that because you're saying that the bar should be raised for all medical issues and therefore high enough that sensitive issues can be covered. That's what you're saying today.

MR. McLELLAN: I'm saying there's nothing qualitatively different about venereal disease, mental health illnesses, infectious diseases, substance use, that make it so distinct that it shouldn't be a part of it.

I don't think it has anything to do with the information or its relevance to health care. I think it's merely a historical fact that some illnesses have been segregated financially.

DR. HARDING: Okay. And because of prejudice or whatever, okay. So then your issue is with access to that information.

MR. McLELLAN: Yes.

DR. HARDING: You want it all in there.

MR. McLELLAN: Yes, sir.

DR. HARDING: But then who accesses it, or how it's accessed?

MR. McLELLAN: My point specifically is if you made a system that would protect a person with venereal disease or who had abortion or anything else, that system would easily accommodate substance use information, mental health information, Alzheimer's information, things like that.

DR. HARDING: Does anybody else have a comment about that?

DR. LO: Well, is it pertinent that substance abuse is often linked with illegal activities and the legal implications are more than what occurs with these other medical conditions?

MR. McLELLAN: Not within the confines of the health care. It doesn't matter to the doctor that's treating it whether –

MR. ROTHSTEIN: But it does matter to the patient when the police --

DR. HARDING: The crack mother --

MR. ROTHSTEIN: -- come in and subpoena the records. I'm sorry –

DR. HARDING: The crack mother or something like that.

MR. ROTHSTEIN: Yes. Ms. Wattenberg, you wanted to say something?

MS. WATTENBERG: Yes, and I'll just sort of add on some comments here.

You know, for those of you who know me, and I work over at SAMHSA, I think I have finally sort of really gotten substance abuse in the protection of those records and that we have this other existing Federal statute – it's

the Part II regulation – I think I'd like to pat myself on the back, sort of fairly successful in that. And here I, you know, invite Tom to come, knowing that he was sort of in direct opposition to that.

You know, I do think that –

MR. ROTHSTEIN: Your open-mindedness is much appreciated.

MS. WATTENBERG: Another pat on the back, right?

[Laughter.]

MS. WATTENBERG: You know, but there are issues about, you know, avoidance of help-seeking behaviors, and I do think that there is this overlay of the legal issues that HIPAA does not protect records from law enforcement in the same way that Part II does.

So I'm not sure that that's sort of what SAMHSA would be advocating or that, you know, a large portion of what the substance abuse community would be advocating. And yet, as I pointed out in one of the previous hearings, right now substance abuse information, health care information, is subject to sort of this very myriad implementation where if you're doing screening a brief intervention in an emergency room because you don't meet the definition of a program, some information is part of the general health care record. And yet if you're from a program like a Betty Ford Center, then that information is not part of the health care record.

So, you know, in preparing and sort of talking to people for this particular hearing, I now have a list of like 20 things that we in the substance abuse community really need to think about.

That's all I want to say.

MR. ROTHSTEIN: I'd like to recognize Bob Hungate now.

MR. HUNGATE: Thank you. I was just trying to get my arm in front of Richard.

[Laughter.]

MR. HUNGATE: I take some of my information from The New Yorker instead of from health services researches and listened to Dr. Groopman and Dr. Atuguaday(?) in their recent articles about medical information.

Groopman talked about the fact that there is little research on pediatrics in many medications. In your comments on Celebrex, I think wouldn't it have been nice if somebody took the database that you referred to and parsed it and found out which of those patients were particularly susceptible and published that two years ago instead of now?

Patients don't understand how much is lost by not doing that. I learned it when I worked at Hewlett Packard and sold medical devices and began to understand the silo that we drew around our product and the silo the profession did around their product.

And with electronic health records, we can cut across that and use database research to do some more things. But patients don't have the understanding of that. And so I think it's different if there were some organization that were doing that for my benefit and I knew it, I would be more willing to divulge it than to this broad spectrum of researchers who are nameless people that I know nothing about. It just seems to me there's something there that's not done that might help the understanding.

DR. LO: Again, let me try and tie that back with Dr. Reynolds's "what if" question. I mean, I think what you're suggesting I would certainly agree with is that patients really need to understand the potential benefit of the uses of health information that are not the classic that doctor and nurse taking care of you looks at in the medical record.

But I think that when we've talked about how this all has to be patient-centered and a patient has to have the final say, we want that to be an informed decision. I think this morning you heard a lot of testimony that the current notification under HIPAA where you get this piece of paper and some people don't even remember getting it doesn't do that job.

So I think there needs to be some effort, and I think you're right. If people really understood what certain types of outcomes research might do, how it might benefit them or their families, and the protection that would be taken to make sure the data doesn't leak out, then I think they would make an informed decision.

And I think most people would say, sure, if you can tell me whether Drug A works and doesn't from this data, not a final answer but better than just sort of my doctor's personal opinion, let's go for it.

MR. HUNGATE: My sense is that that would have to be an independent entity, a UL lab kind of a thing that's separate that looks across the system and has obligations, a vetting process that guarantees that it is a kind of an entity that's different than the vested entities people are accustomed to dealing with.

DR. LO: And that entity needs to be very transparent and very accountable back to the public.

Yes. I have things like that in public health. I mean, we have public health departments that try and achieve that amount of trust by sort of really going the extra mile at least in crisis situations to explain, to be credible, to address concerns and fears.

MR. HUNGATE: But it's not well understood.

MR. ROTHSTEIN: Yes, Mr. McLellan?

MR. McLELLAN: I've come to think of it, again, as part of the deliberations in our IOM panel as the same kind of a disclosure and conversation that you would get in if you were starting to talk with the patient about surgery.

You have a recommendation and it's based, if you're a physician or you're part of the health field, you have a recommendation based on the research, you have a recommendation based on prior care. It's ultimately the patient's decision. So that's why you're trying to inform them.

And I see it the same way with regard to the collection and utilization of information. I think it is a fact that if all information that's pertinent to health care is not available within the health care system, there are risks, definable risks, to safety and benefit.

Ultimately, the patient must weigh those risks versus what he/she is willing to share. And that's how I've come down on it.

MR. ROTHSTEIN: Well, I'd like to follow that up because I think Dr. Lo's provocative suggestion regarding automatic access to database research is perhaps more radical than people think at first hearing because with the exception of the kinds of waivers that an IRB or a privacy board could give with regard to authorization or informed consent to researcher access to certain database information, we have not since Nuremberg recognized the dichotomy between clinical intervention and other forms of research.

And I think what I hear you suggesting is some sort of carve-out justified by the sort of the public health/research value of doing that. I mean, you might be able to make an argument that is persuasive to some people, but it clearly is, it seems to me, a major shift from the way we currently view research. And I'll give you a chance to respond to that.

Even if one could support that on its own merits in terms of public health in the broadest sense, I think we would need to be concerned about the slippery slope to other areas including anti-terrorism where Homeland Security might want real time monitoring of all the people who come in with fevers et cetera, et cetera, et cetera.

And so that we appreciate that privacy costs, and part of the cost is clinical care and part of the cost is public health and part of the cost is research and so on. I would be much more comfortable with your educational model where people were better informed about the value of broad participation in research studies certainly that don't have any physical risk to them and the value of the security protections that would be built in and so forth than this change from the way we've viewed research for the last 50 years.

So I wonder if you could comment on that.

DR. LO: Well, first I would want to try and make a difference between – there's a spectrum of outcomes, databased research, and I think some of it frankly is of the trivial who cares and some of it, I think, starts to get very close to other areas where there's evidence that there is some public policy that supports access without patient consent.

So states that have cancer registry, you don't consent to have your case included in the cancer registry; your legislature voted to have that. And you can go to the legislatures and repeal the law but that was a public policy geared for a specific condition with certain kind of qualifications.

MR. ROTHSTEIN: And all sorts of infectious disease reports. Sure.

DR. LO: And again, I think the most recent sort of attempts to balance the need for public health access to data and protections for privacy and individualized more generally have tried to sort of say both values are very, very important. We would like to have both, but there are situations in which the individual may not have the right to say I'm not giving you information. Certainly in public health, that's most apparent because there's immediate, serious, visible harm to individuals which you'll be able to identify.

With databased research, it's statistical harm, but the number of people potentially affected – you know, again, people can sort of extrapolate numbers, but some people are saying as many as 10,000 people may have had heart attacks, fatal heart attacks, due to use of these Cox-2 inhibitors when in fact, had it been known, they would not have been on the drug.

So I think you have a tremendously important task to sort this all out and as you all know, the sort of directness of the threat and the threat to identify individuals makes it more pressing. But I think we do need to sort of at least be open to the possibility that there's some types of research that may be so beneficial for the public, so serious, so widespread, that we may choose to say that if we can get the safeguards strong enough and have the public understand it well enough to agree, we may be willing to proceed without individual –

MR. ROTHSTEIN: And how would you make that determination of whether it's so important, and who would make the determination of whether that's so important that it could be accessed without consent or authorization?

DR. LO: Again, I think it really comes down – I mean, I think you were getting at that question. It needs to be someone who's perceived as trustworthy, someone who's accountable, and someone who is transparent in terms of stating actually what's being done, why it's important, and, you know, the ability to sort of have that decision reviewed.

And again, in public health now we're sort of thinking about what are the parameters under which we will allow a governor or a state department of public health to declare an emergency with quite broad powers but having some sort of due process procedures that makes the public feel that they're not overstepping their bounds. But I think there are models.

MR. ROTHSTEIN: Joy?

MS. PRITTS: Yes. I'd like to announce that many of the states that have these disease registries like cancer registries also have very stringent rules that are imposed directly on the people who have access to those registries as to what they can do with the information and how they can share it. A little different than it is currently at the Federal level where you only have – it's kind of bootstrapped in where the provider is the one who really bears the responsibility for insuring that the researcher have their ducks in a row so that if you were to move to any system such as is suggested by Dr. Lo, I think it would be imperative that the laws address the researchers directly, not through this indirect fashion. And that would, I think, really go a long way to help people trust a little bit more if they knew they are directly responsible and they are directly accountable if they do something wrong with the information.

DR. LO: I think you could make that even stronger. You could say that you may not pass that information on to a third party.

I think the other thing that must be kept in mind is that you only need the initial identifier to link the data. Once you have your database to do the Cox-2 study, you can turn it over to the researcher in de-identified format.

Now again, someone with enough computer time and savviness might be able to identify some people, but you could make that as de-identifiable as you'd like once you've got all the information in place. Now, you lose the chance to go back. You have to reassemble the data to do a, you know, five-year follow-up down the road, but those are the sorts of tradeoffs I think that you have to look at.

MR. ROTHSTEIN: So maybe you're saying that if we did that, besides the things that Joy suggested, we might consider a kind of minimum necessary and least identifiable form consistent with the uses requirement to add on?

DR. LO: Yes – well, I would be willing to go further, that this isn't sort of an open-ended, here's a great database; see what you can mine it for. It's that we have a really pressing problem; let's create a database tailored to that and then we don't want you to really use it for anything other than this extremely pressing research question which we think really is so close to public health investigation that we really want to treat it similar to that. And this is what public health officials are used to doing.

MR. ROTHSTEIN: And maybe some sort of board to make those –

Bob?

MR. HUNGATE: I'm an old marketing guy, and you got to look for somebody that wants something. And Vioxx is up in lights. We don't have all the data we need to know what's right. If Vioxx comes back on the market, why shouldn't the piece of paper that you get at the druggist when you get your drug say, "This is going to be treated as data in a public health way and these are the safeguards and we're going to do this with this information. And a condition of taking this medication is that you're part of that trial." Now, why couldn't we do that?

DR. LO: Well, I think you could, and yet there's a precedent for that in the FDA with warnings for Accutane for women who might become pregnant.

So I think, again, to the extent that you tailor it and you offer people alternatives and have ways to educate them, I think that's an option you would want to look at.

MR. ROTHSTEIN: But I think you would have to do it on a case-by-case basis. You couldn't automatically turn everyone who took a prescription into a research subject.

DR. LO: No. But I'm thinking that we have a public privacy issue here and a need to communicate a way of doing things that's different. And we have a medication that is up in lights. And so there's a way of physicians I think a teachable moment, public, and limited to that specific medication, to test.

MR. ROTHSTEIN: Well, it may be going forward that people are on notice of that, but we've needed to look back.

When Vioxx came on the market, there wasn't the assumption that==

DR. LO: No, I understand. But I'm thinking that why not create an entity which is the kind of entity we think would be needed to guarantee the privacy around a specific example and do a test? That's kind of what I'm arguing here.

MR. McLELLAN: My only point on that is Vioxx is the most recent and visible example, but how do you know that the next, there's not a drug tomorrow that's going to come out?

That's why I'm arguing that the standard ought to be sharing but the thing that I would disagree with you about is that I wouldn't make it mandatory. I would give the patient the ultimate right to not participate.

DR. LO: Well, I think I would agree with that. But all I'm trying to argue is that rather than go off and study this approach for some time and think about what the best way to do it is, an alternative is to also test. And that's the part that I'm trying to articulate as a way of also trying to grapple with this.

MR. ROTHSTEIN: Joy, did you want to comment?

DR. PRITTS: I find this discussion pretty disturbing. I think that at the state level when you have cancer registries, they've at least been enacted by your state legislature and if you don't like it, you vote them out. So there's accountability there.

The idea of having some kind of nebulous entity deciding when something becomes an issue enough that you're going to do this, I just find it disturbing.

MR. ROTHSTEIN: Well, there is I think clearly a sense of potential abuse or slippery slope or whatever and whether it can be saved is an open question and what the price in privacy we would pay for putting that into effect is remains to be seen. I mean it's now Issue Number 87 for us to deal with.

Other questions or comments for this panel?

Thank you very much for your testimony. We appreciate your being here.

We will take a 15-minute break and then proceed with the rest of the afternoon's agenda.

[Break from 2:58 P.M. to 3:16 P.M.]

MR. ROTHSTEIN: We are back with our hearing, the National Committee on Vital Health Statistics Subcommittee on Privacy and Confidentiality.

Agenda Item: Statements from the Public

We have two individuals who have signed up for public testimony, and so we are happy to recognize each of them for five minutes. The first one is Kathryn Serkes.

MS. SERKES: Thank you. I'm Kathryn Serkes. Today I'm speaking on behalf of the Medical Privacy Coalition, and that is a coalition of 29 groups, a nonpartisan coalition including the American Conservative Union, the American Mental Health Alliance, American Policy Center, Americans for Tax Reform, Citizens for Health, Foundation for Health Choice – I'm not going to read all 29, I promise – Free Congress, Public Citizen's Congress Watch project, the Republican Liberty Caucus, American Psychoanalytical Association, the World Privacy Forum, and Citizen Health Advocacy Group and the Association of American Physicians and Surgeons. Also, the Pain Relief Network and the Drug Policy Alliance, which I mention in light of the testimony on the addiction.

I have just a couple of brief comments for you today. Particularly this morning, you heard about the attitudes in the survey that were presented, attitudes on privacy, attitudes on the success or problems with HIPAA. And I'd like to give you a little bit of information to tell you about privacy in action and the application.

A survey of AAPS physicians show that physicians believe that third parties ask for information that they believe to violate confidentiality, overwhelming number of physicians. Fifty-one percent of physicians surveyed report such requests from government agencies that they believe stretch the bounds of privacy, and 70 percent believe that requests from health plans exceed what is necessary and violate privacy.

Now, on the patient side, 87 percent of physicians surveyed report that a patient had asked that information be kept out of the record, and 78 percent of physicians said that they had indeed withheld information from a patient's record due to privacy concerns.

While only 19 percent would admit to lying to protect a patient privacy, which may have been a loaded word, 74 percent state that they have withheld information for that reason.

Okay, so that's what we have in practice.

The Medical Privacy Coalition objects to the standard in general of the way the privacy is enforced now in that it is based on the assumption that there is a compelling need for the individual's medical information and the position that the public health usurps the individual rights. That is a concern of the Coalition.

Patients who fear disclosure of their sensitive information to government agencies have no choice but to withhold the information from the physicians. That's the perception now of patients.

Another area of concern is law enforcement. The biggest concern – you talked about research; you've heard from people talking about the payment, et cetera – but the biggest concern for members of the Medical Privacy Coalition is government access to the medical records. And as you heard Joy Pritts talk about patients subverting the system, that's what we have now, is we have increased ways for a patient to try to subvert the system by asking that information be withheld from the record.

I've given you the survey information. Anecdotally, I can tell you that physicians tell us that the requests are on the increase to withhold information.

I'd like to also respond to a couple of things that we heard in the testimony. Dr. Westin's presentation about the privacy notice – he said 32 percent say they had not received a HIPAA privacy notice. We would contend that people may or may not have received that privacy notice, that people don't really know whether they're getting it or not, because as some of you remember, when you go in to see a new physician, you're signing a lot of papers. So we're not sure about that. The problem is that they don't know what they're signing when they're signing. They think they're signing a consent form for privacy as opposed to – they still don't get it – that there is a difference between the consent and being advised of the information under HIPAA.

The Chairman talks about the issue of retrieval versus inclusion in the record and the problem with that. This brings up the issue and it has been mentioned around, the difference between the clinical use of information and the others – payment, research et cetera. And there is a need for clinical use of the information, there is a need for inclusion, because that is the concern now, is that physicians are getting incomplete medical records because of the privacy concern.

And the area of addiction that we brought up is that patients foregoing treatment, and it was suggested that patients would forego treatment if everything had to be included in the record. The difference between addiction and other illnesses or infectious diseases is the issue of law enforcement because patients are concerned about law enforcement access to that information.

And the difference here is that because of the way that HIPAA was written, nothing in the rule, if you remember, nothing in the rule permits covered entities from avoiding disclosures required by other laws. And there are limited restrictions on law enforcement. So we believe that the information from law enforcement, that the Fourth Amendment should apply to this information as well. Medical records should be at least as well protected as the papers in one's home.

Do you follow me on the difference with addiction? You're talking about the illegal use of substances. So patients fear the criminal prosecution when they put that into their record, not just whether they will be stigmatized because they have AIDS or because of cancer.

The Chair also asked the question whether we would end up with a whole new class of drive-through type of medical treatment where there are no questions asked.

I can say that we have that now. We've seen a large increase in the number of cash-based practices. We call them patient-doctor direct practices where there is no third party payment and patients frequently are doing this to avoid privacy intrusions or because of their concern for privacy. They are preferring to pay a doctor directly and they may or may not file an insurance claim, depending on their concern for privacy.

What we're seeing now is that the benefit of any IT advances will be lost unless patients and physicians can be guaranteed privacy. I think that's been a recurring theme in what you've heard today.

And I think I agree with what Joy Pritts mentioned, that HIPAA was almost obsolete when it came into effect because we now need a new set of rules. HIPAA was really geared to moving towards a national database, moving towards a centralized idea of medical records and centralized electronic records.

And what we're seeing now is a push away from that. We're hearing patients say, we want electronic medical records, but we don't particularly want them to be in a government nationalized database of medical records. We can see the savings, we can see the advantages of having electronic medical records. I can see the advantage of having my physician be able to share with another physician and having an electronic medical record.

What's really ringing with patients now is the personal health record. Again, back to the issue of patient-centered and controlled, where the patient controls the record. And in fact the Association of American Physicians and Surgeons is working with a company named WorldDoc that has a personal health record – exactly that, controlled by the patient. The patient chooses whom to give the information and what parts of the records to give the information, which gives the patient the confidence that that is being protected.

And just as a sidelight on that, of the physicians who have started using this personal health record, eight percent of the patients are actually paying an annual fee to be able to use that type of system. So there is a demand for that.

I would like to sum up by saying that we would like to see the personal health record as the first step so we can try this out and move towards that so that the patients are in control. And the bottom line is that Joy Pritts asked, how do we make patients trust the system again? And there is a simple, though not particularly easy, answer, and that is to reinstate the issue of consent. Rather than advising patients of how their records may be used and their information may be used, but reinstituting the actual provision of consent, once again allowing patients to decide and make the decisions how their medical records will be used.

Thank you very much.

MR. ROTHSTEIN: Thank you. If you could stay there for just one second, I have one quick comment and a short question as well.

On your statement about the notice of privacy practices, I went to a new physician about two weeks ago and I filled out all the papers and the last thing that I was given was the acknowledgment form to acknowledge that I had received the notice of privacy practices, which I had not been given.

And so I felt compelled to –

PARTICIPANT: They didn't know who you were!

[Laughter.]

MR. ROTHSTEIN: -- initiate a discussion as to the obligations of a covered entity under the privacy rule. And I said, "Don't you think it would be a good idea if you would give me a copy of the document that you're asking me to acknowledge that I have received?"

And the response was, "Nobody wants to see them, nobody reads them when we give them the copies anyhow; why would you want to read them?"

[Laughter.]

MR. ROTHSTEIN: And so that was a bad idea to ask me a "why" question. So then we critiqued line by the –

[Laughter.]

MR. ROTHSTEIN: Okay.

MS. SERKES: The question is, did you ask them for a restriction to see what they did?

MR. ROTHSTEIN: I'm seeking new medical care now.

[Laughter.]

MR. ROTHSTEIN: Okay. The question that I have is: Do you have a copy of that survey that you referred to that we could –

MS. SERKES: Yes, it was a very short – essentially the questions that I mentioned were the questions that were asked.

MR. ROTHSTEIN: Okay, but can we have that in some sort of form?

MS. SERKES: Certainly.

PARTICIPANT: Is it on your website?

MS. SERKES: No. I will email that to you and – I'll email it to Marietta --

MR. ROTHSTEIN: If you would email it to Marietta, that would be great.

MS. SERKES: -- so you have it for the record.

MR. ROTHSTEIN: Okay. We have another question.

DR. FITZMAURICE: I wanted to follow up on what our esteemed Chairman was asking about the survey. You mentioned a large number of government agencies' requests for information to doctors, that they reported receiving a large number of requests.

It could be disturbing, but I wonder how many of those were Medicare and Medicaid information requests to make decisions about the reasonableness of claims, that it's something dealing with a payment, for example, versus another purpose such as law enforcement or public health. Did the questionnaire get at any of that information?

MS. SERKES: No, that's why I say the questions were as simple as I stated them, and so I guess that reflects that you're correct; some of the questions that may have been asked may have been for treatment, but at the same time that the physicians felt that they overstepped what was necessary, I guess if we want to go back to the issue of the minimum necessary, that was necessary for the treatment, payment or operations into violating confidentiality.

So the short answer is I can't answer how much of that was treatment and how much of that was law enforcement et cetera.

MR. ROTHSTEIN: Thank you very much.

MS. SERKES: Thank you. We'll be happy to work with the Committee any way we can to advance the personal health records.

MR. ROTHSTEIN: Our next public witness is Sue Blevins.

MS. BLEVINS: Good afternoon. I'm actually speaking on behalf of Robin K. first who has prepared a one-minute statement and then I'll speak on behalf of the Institute for Health Freedom.

Robin K. is a private citizen and I'm going to read her statement:

"My name is Robin K., an attorney who has tracked medical privacy since 1996 as a concerned private citizen. Thank you for allowing the opportunity for Sue Blevins to present my comments today. My comments are as follows:

"Some things in life are obviously not good ideas. Some things usually have demonstrable risks associated with them, like driving without your seat belt, smoking in bed, leaving your wallet out where strangers can take it, or throwing sensitive personal health information out in the trash without taking precautions to protect against identity theft.

"In the rush to embrace technology, well-intentioned Federal officials and consultants are ignoring hard evidence that putting ultra-sensitive medical information into electronic format and exchanging such information between health care providers and entities will expose sensitive medical information to being hacked into and wrongfully disseminated.

"The danger of wrongful access or human error resulting in wrongful dissemination has been demonstrated again and again, yet the Federal government is contemplating requiring every citizen to have his medical information placed into an electronic medical record that tracks him from birth to death. In a free country, shouldn't it be up to each American citizen whether he wants to accept such a risk of exposure of his sensitive medical information?

"I have testified in front of this Committee that if electronic medical records are the wave of the future, regardless of the inherent risks involved, each citizen should be able to opt in or opt out of such a system.

"My central premise that private information stored electronically is wrongfully exposed again and again continues to manifest almost daily. Just this month alone, it was reported that a confidential list of 4,500 persons with AIDS and 2,000 others who are HIV positive, mostly living in Florida, was inadvertently emailed to more than 800 Palm Beach County health care workers.

"Also this month it was reported that ChoicePoint Inc., a company in Georgia that gathers private information

on nearly everyone in the United States, transmitted personal data on as many as 145,000 persons to thieves using stolen identities to create what appeared to be 50 legitimate businesses were found to be fake companies.

"And last month, the public learned that a hacker in California was able to read Secret Service emails and files after he breached the cellular network of T-Mobile. Ironically, he did the hack-in during the Secret Service's ongoing investigation targeting underground hacker organizations. This is a classic example of how vulnerable an electronic system can be. In this system, it was a hacker who was being pursued by the Secret Service who breached their own electronic communications system.

"Time and time again, these example show that no electronic database system is foolproof and failsafe.

"Therefore, with such obvious risks of improper access of information, a resulting invasion of privacy, each and every American citizen should be given freedom of choice whether he wants his sensitive information to be stored electronically.

"In conclusion, in view of these seemingly unending and inappropriate electronic disclosures and violations, no citizen should be compelled to risk a similar disclosure of his most private and personal medical information.

"Thank you for this opportunity to share my thoughts on this issue."

And this was written by Robin K., who is an attorney and really, really cares about this issue.

Now I'll read a statement on behalf of the Institute for Health Freedom, and I'm speaking as Sue Blevins now.

Good afternoon. My name is Sue Blevins, and I am founder and President of the Institute for Health Freedom, a Washington, DC-based think tank that studies and reports on individuals' freedom to make their own health care choices and to maintain their health privacy, including genetic privacy.

It is clear from thousands of public comments submitted to the U.S. Department of Health and Human Services and public opinion polls that Americans highly value and expect medical privacy. Citizens want to exercise the right to give or withhold consent before their personal health information is shared with others.

Unfortunately, however, the Federal medical privacy rule, which was released in December, 2000, and modified in August, 2002, eliminated the precious right to give or withhold consent before one's personal health information could be accessed by many others.

Thus, until the right to give or withhold consent is restored, individuals do not, and I repeat do not, have control over who has access to their personal health information. And I need to interject there and say the Office for Civil Rights really needs to make that clear. The public is still confused, and in fact that Office's website states the exact opposite, so the public really needs to be told the truth, that they do not have control over the flow of their personal health information.

Additionally, moving toward adopting electronic health records is a recipe for privacy invasions, and here's why. It is clear that combining the lack of consent with adopting electronic medical records would lead to a greater number of persons accessing patients' medical records without their permission.

The U.S. Department of Health and Human Services acknowledges concerns about electronic health records, and I want to read three quotes from HHS's own analyses that were released with the Federal medical privacy rule regarding electronic information. The first quote:

"The electronic information revolution is transforming the recording of health information so that the disclosure of information may require only a push of a button. In a matter of seconds, a person's most profoundly private information can be shared with hundreds, thousands, even millions of individuals and organizations at a time."

And I'd be happy to provide references for these quotes.

The second quote, and this one is very, I think, important to health care providers:

"In short, the entire health care system is built upon the willingness of individuals to share the most intimate details of their lives with their health care providers. The need for privacy of health information in particular has long been recognized as critical to the delivery of needed medical care. More than anything else, the relationship between a patient and a clinician is based on trust. The clinician must trust the patient to give full and truthful information about their health, their symptoms and medical history. The patient must trust the clinician to use that information to improve his or her health and respect the need to keep such information private. In order to receive accurate and reliable diagnosis and treatment, patients must provide health care professionals with accurate, detailed information about their personal health, their behavior, and other aspects of their lives. The provision of health information assists in the diagnosis of an illness or condition and the development of a treatment plan and in the evaluation of the effectiveness of that treatment. In the absence of full and accurate information, there is a serious risk that the treatment plan will be inappropriate to the patient's situation. Individuals cannot be expected to share the most intimate details of their lives unless they have confidence that such information will not be used or shared inappropriately. Privacy violations reduce consumers' trust in the health care system and institutions that serve them. Such a loss of faith can impede the quality of health care they receive and can harm the financial health of health care institutions."

There's one last short quote:

"Patients who are worried about the possible misuse of their information often take steps to protect their privacy. Recent studies show that a person who does not believe his privacy will be protected is much less likely to participate fully in the diagnosis and treatment of his medical condition. One in six Americans reported that they have taken some sort of evasive action to avoid the inappropriate use of their information by providing inaccurate information to a health care provider, by changing physicians, or by avoiding care altogether."

So in conclusion, I think it's very important for us to consider that basically, unless we reinstate consent and uphold true rights to privacy, we're putting all citizens in the position of choosing between three options:

One, they can seek care and have information shared without their permission.

Two, they can lie to their health care providers and others if they really want their privacy.

Or three, they can forego care altogether in order to maintain their privacy.

And I have to add that for some people maybe privacy isn't an issue and they aren't being squeezed or pressed to choose between those three options, but I'm sure, as you probably heard today, many, many people care about privacy.

So I want to thank you all for the opportunity to comment today and for considering these comments and I think for your work on trying to help privacy become a reality in this country. Thank you.

MR. ROTHSTEIN: Thank you. A couple of quick questions.

Do we have a copy of that study that you referred to where you said one in six took evasive action?

MS. BLEVINS: I'd be happy to get that for you.

MR. ROTHSTEIN: Thanks.

MS. BLEVINS: That was cited in the HHS's – when they released the privacy rule, but I've cited it in a paper and I'll make that available to the Committee.

MS. WASSERMAN: That actually, Mark, was, I think, the health privacy project – sorry. The California HealthCare Foundation had that data.

MR. ROTHSTEIN: If we can get copies of that, though –

MS. BLEVINS: Sure. HHS is cited, but I'd be happy to get that for you, too.

MR. ROTHSTEIN: Yes, I'd like to see the original. And can we get copies of the statements that you read?

MS. BLEVINS: Absolutely. Those I can send them on electronic format so that –

MR. ROTHSTEIN: Okay.

MS. BLEVINS: -- they're easier to transfer.

MR. ROTHSTEIN: And now I have one question about consent, and I really don't want to spend too much time on this, but it's a point that I think has to be made.

The Committee has spent hours and hours on the issue of consent, as most of you know. I would support consent only because it's a traditional prerequisite of medical care and is, I think, a very important symbolic value.

I don't think it has any practical value whatsoever because if you go back to my visiting this new physician two weeks ago and they put a bunch of papers in front of me, if the paper had been called at the top not "acknowledgment of receipt of blah-blah-blah-blah" but "consent to treatment and uses and disclosure of medical information which says I consent to the disclosure of information for reimbursement purposes et cetera, et cetera" and I don't get to see the doctor unless I sign that form, how is that any different from the form that they gave me to sign based on the assumption that they'd given me the information that they didn't give me?

So I don't think we're advancing the interest of privacy by getting hung up on this, whether it's, you know, an acknowledgment or a "consent form." I think what we need to do, if we're interested in privacy, sort of deal with the substantive questions of who gets access to the information and how much and under what circumstances and how is it disclosed et cetera, because under any regimen, the patient is going to have to sign something in order to get to see the doctor, assuming that they are conscious. And what we label that, unfortunately, in my judgment, is not the sticking point.

MS. BLEVINS: Can I comment?

MR. ROTHSTEIN: Sure.

MS. BLEVINS: I have to say I respectfully disagree, and let me tell you why. And I say that respectfully because this isn't for me; I don't personally feel like I have a need that I'm hearing from a lot of people who quite frankly are telling me, A, I lie; I don't tell my doctor I drink anymore. You heard about the case in Pennsylvania where a man lost his license because it became routine to report people if they had a – I don't know the specifics, but a DUI, DWI, and anyway, later it was reported by his doctor and he lost his license.

Again, I know there are legitimate reasons for why there has to be reporting at times. But if you just hear from people – and quite frankly, I think people who really care about privacy, they're not going to contact your Committee and they're not even probably going to report things to the Federal government, and that has nothing to do against the view of the government but just that people that are privacy conscious call me up and they want to tell me things and they don't want anyone else to know. When reporters ask me for cases and I say I can't even tell you the case, because if I tell you the case, I'm giving away people's very sensitive information.

So I say I come here on behalf of those people who will not come before a government committee. They will not talk to you. They probably wouldn't even write or send in comments to HHS.

And I differ. I say the difference between consent and notification is if you just imagine we change the rule regarding who has access to your home and instead of asking for your consent before someone enters, they now have permission, if they're a legitimate person, to do something in your home. They just to notify you that they're coming in. There's a huge difference there.

And where I agree with you strongly is that what we have as consent is coercive consent, and the doctor or the nurse and everyone says in order to treat you, I need to be able to share your information. And at that point, that's where the individual gets to sit down and say, okay, who gets it, what happens? That's a whole different scenario and even legal. It's a different ethic than notification.

MR. ROTHSTEIN: I agree with you, but I think it's a sort of a symbolic value and I don't think it has that much difference in practice if it's –

MS. BLEVINS: I think it does. I mean, I won't spend a lot of time, but I'm a former nurse; I worked at Johns Hopkins Hospital. I took care of a lot of "VIPs." And I can just give you one example.

There was one time – and I can't even get too detail oriented – there was one time a gentleman who was hospitalized who had a very big name on Wall Street, and this was way back in the 1980s, and he asked – and that was before we had – I don't even know what they use in hospitals now; I'm assuming everything's digital, nobody gets paged. This was when we had the old intercom system and they would page nurses and doctors and even patient's names over an intercom.

And this man was like "Please do not page my name." And it wasn't that he was guilty and had something to hide, but he was so "humanly powerful" that if the word got out about his disease and his prognosis that that would affect his company and their bottom line.

So there's so many reasons why people care about privacy. And I agree with you that most people – the reason why I think it would be so easy to reinstate consent and make it work is that the majority of people are going to say "fine," and it will work. But when you take it away, when you basically say, okay, we're going to take away that right to decide who can come into your house, most people, if they're calling an exterminator or an electrician or somebody, they're glad to have them just come into their house and they're glad to have a neighbor have a key, go in, open it up for them.

But there is a small minority who really care, and I just feel so strongly that for those people, I'd love to see that ethic maintained.

And anyway, I agree with you strongly that it's symbolic and it's coercive, but it's still a fundamental ethic that we've had for many years.

MR. ROTHSTEIN: Okay, thank you.

MS. BLEVINS: Thank you.

Agenda Item: Subcommittee Discussion

MR. ROTHSTEIN: Let me alert our Internet listeners as well as those of us who are here about some schedule changes for tomorrow that may affect the timing.

We will begin tomorrow at 9 A.M. as scheduled and there are no changes in Panel 3, disease and health advocacy groups.

After lunch, there is a change. We will not have a witness from the Electronic Privacy Information Center; they are not able to be with us. Neither will we have anyone from Patient Advocates in Research. So Panel 4 will only have two individuals.

And Panel 5, Dr. Marshall will not be able to be with us and WebMD will submit written testimony instead.

There are two session marked tomorrow for Subcommittee discussion, 11 to 11:45 and then again 4 to 4:30. We're going to have a brief Subcommittee discussion this afternoon and I think that will obviate the need the second of our Subcommittee discussions scheduled from 4 to 4:30.

There is some threatening weather tomorrow, I understand, so it may well be that we will be able to finish earlier in the afternoon, although I'm not sure exactly what time that will be yet. That will depend on some factors that we will get to tomorrow.

For today's Subcommittee discussion –

DR. HARDING: Mr. Chairman?

MR. ROTHSTEIN: Yes, please – Dr. Harding?

DR. HARDING: They're talking about three to four inches of snow tomorrow morning. Do things work? I mean, do we start at 9 o'clock here? I mean, you know, I'm from Ohio. That doesn't stop many people but in South Carolina it would paralyze the city for the next two weeks.

MR. ROTHSTEIN: Well, I would assume that the rugged folks at HHS laugh at three to four inches of snow.

DR. HARDING: Okay.

[Laughter.]

DR. HARDING: So we'll make the assumption it's 9 o'clock, no matter what the weather.

MR. ROTHSTEIN: Yes, we will come, rain or snow. We will start at 9 o'clock because if we don't, we're going to get all messed up with flights going out and so on. If people can't make it, we will make arrangements to hear from them at some other time at one of our later hearings. That's okay.

Now, you should have been distributed a two-sided hand-out. On side it says "Draft 11/21/04" and the other side , "Observations and Recommendations." Let me remind you that the side that says "Draft 11/21," this is the document that the Subcommittee approved as a result of our

November 18th, 2004, hearing on issues related to e-prescribing, and the purpose of this was to draft some language that would be sent to the Subcommittee on Standards and Security to be incorporated into their overall statement which will be coming before the full NCVHS at our next meeting, which is March 3rd and 4th.

So at this time, what I would like to do is note for you the revisions that were made – you can see on sort of the side, and they are, I think, primarily editorial changes in the language that was used by the Standards and Security Subcommittee perhaps to make this similar to other kinds of phrasing in the other document.

So at this point, the question that I would have for the members of the Subcommittee is whether there are any changes that you think, you know, changed our meaning or that we need to change back or whether we can accept the revisions and express that to the other Subcommittee.

So let's put it this way. I'll give you a minute because perhaps you haven't seen it.

DR. FITZMAURICE: Mark, while we're reading, could I ask a question about the work "implicate?" Is that the right use of the word "implicate?"

MR. ROTHSTEIN: Can you tell me where you are, Michael?

DR. FITZMAURICE: This first line of the second paragraph: "E-prescribing regulates implicate other –" It's like pointing to a criminal and saying, "He was in it, too." But maybe that's a proper legal use of the term; I just don't know.

MR. ROTHSTEIN: Well, we could make it "relate to," if that bothers you.

DR. FITZMAURICE: I would understand it better.

MR. ROTHSTEIN: It was not meant as a term of art, just "raise issues involving" – I mean there are a variety of ways that we can say that that I'm sure would be acceptable. Do you have one in mind?

DR. FITZMAURICE: No, I don't. I just was wondering about the word "implicate." I'd not seen it used in that context before, but maybe it's all right.

MR. ROTHSTEIN: Okay, so we'll leave it for now and then we'll see whether the full Committee is similarly troubled by the implications of that word.

[Laughter.]

MR. ROTHSTEIN: Richard?

DR. HARDING: Could you read the last paragraph in its final form to me?

MR. ROTHSTEIN: Okay, I'll try.

"The NCVHS recommends that any e-prescribing pilot project initiated in 2006 by HHS include measures to identify and address the privacy interests of consumers and

the inclusion of substance abuse, mental health and –" To address – to identify and address the privacy interests of consumers – I don't understand why they did that.

Harry, can you help?

MR. REYNOLDS: Where are we going?

MR. ROTHSTEIN: We're in the last paragraph of the 11/21 draft. It now is sort of odd.

DR. FITZMAURICE: Could I suggest that it might be "and to include substance abuse, mental health and HIV providers" – replace "the inclusion" with "to include." I think that's the sense of it.

MR. ROTHSTEIN: Okay. "...to identify and address the privacy interests of consumers and to include substance abuse, mental health and HIV/AIDS providers?"

DR. FITZMAURICE: Yes. So that their privacy concerns will be studied in the pilot project.

MR. REYNOLDS: Where did you use –

DR. FITZMAURICE: "To include."

MR. ROTHSTEIN: Richard, does that –

DR. HARDING: That's better. And there's a period after "meaningful way?" Is that –

MR. ROTHSTEIN: Yes.

DR. HARDING: And then –

MR. ROTHSTEIN: It's still not great, but it's better. It was great before and then Simon got a hold of it.

[Laughter.]

MR. ROTHSTEIN: He's not listening.

MR. REYNOLDS: We won't defend him until – we won't jump as a Committee –

[laughter.]

MR. ROTHSTEIN: No, I think it's okay.

So are there any other questions about the language? There being none, then I think we're comfortable with that.

Now, if you would flip over the page, and this is language that we did not draft. This is in particular – especially the recommended action item.

MS. FYFFE: Who drafted it?

MR. ROTHSTEIN: This was drafted by the Subcommittee on Standards and Security.

MR. REYNOLDS: Taking input from the letter.

MR. ROTHSTEIN: So they said – you know, the letter is at the top, and then they have come up with – fortunately, Harry's on both subcommittees – drafted these two recommended actions, 10.1 and 10.2. Do you want to discuss, Harry, explain the thinking?

MR. REYNOLDS: On the wording?

MR. ROTHSTEIN: Well, on why you did what you did.

MR. REYNOLDS: Basically just trying to keep it in the same wording, the same framework, the same other that we have done on the rest of the letter. So we have come up with observations; we have a list of observations.

The full letter is 18 pages.

MR. ROTHSTEIN: Right.

MR. REYNOLDS: We have a list of observations ranging all the way from new testimony to things that have been covered in the past, giving updates on those, as well as pulling in the information from this Committee.

So we turned this into one of our observations, tried to keep pretty much the content, and then, in line with the rest of the letter, actually came up with recommended action.

MR. ROTHSTEIN: Okay.

MR. REYNOLDS: And I would recommend that unless there's anything dramatic missing above the recommended action or you see something that you felt was omitted from the original letter, well, that would be good to include that. Otherwise, I think it would be good for the Committee to look at the recommended actions and make sure that those are actions that are amenable to the Committee as it goes forward to the Secretary.

MR. ROTHSTEIN: Well, I have a concern about Action 10.2.

MR. REYNOLDS: Okay.

MR. ROTHSTEIN: I think 10.1 accurately reflects the opinion of the Privacy and Confidentiality Subcommittee with regard to what should be done during the 2006 pilot tests. I think that's fine.

But 10.2 then goes beyond the pilot phase and says "HHS should use experience gained from the pilot tests to develop and communicate guidance to the industry."

And so what this suggests to me is a choice of sort of soft regulation. Based on the pilot test, it seems to me that HHS could develop regulations dealing with e-prescribing and it might call for something that is stronger, more prescriptive, in terms of what covered entities have to do and what they can't do, et cetera, et cetera, et cetera, whereas Action 10.2 says "based on the pilot test results," all that HHS is going to do is develop and communicate guidance on handling privacy issues rather than developing, you know, privacy rules. Am I misreading that?

I am misreading that? Okay. Wouldn't be the first time.

MS. FRIEDMAN: This is Maria Friedman, and I have a different view. And the view is that the pilot tests are really supposed to inform the implementation of the Medicare Part D benefit and the pilot, we're in the process

of developing an RFP for pilot participants, so there's concern not about the replicability on the larger scale necessarily, but I think you're going a lot farther downstream than the results of the pilots would take us.

MR. ROTHSTEIN: Okay. I wonder if you could explain that distinction between sort of the post-pilot phase and something further downstream.

MS. FRIEDMAN: Okay. The pilots are supposed to be conducted during calendar year 2006. There's an evaluation in 2007. The final rules will come out – and again, the pilot tests were standards to help inform the implementation of Part D benefits, okay? The final rule comes out in 2006 to be implemented a year later, sometime in 2009. So that's the timeline.

MR. ROTHSTEIN: So, I mean, here's my concern. Maybe this is misplaced and then you can help me again.

I'm reading this to say that CMS is not going to be enforcing anything other than to give suggestions and guidance to the industry and there may be things – I don't know what they are yet; we haven't done the pilot – that we need sort of stronger rules on.

Am I way off on that?

Sue McAndrew and then –

MS. McANDREW: If I could also comment here. I mean, it seems to me that the way the pilots are being structured, their main focus is going to be to prove both the effectiveness of the foundation standards that are being adopted to the rule-making this year as well as the other initial sets of standards for e-prescribing that will then, following the evaluation, be formulated into regulations like the foundation standards are.

But the focus, and the focus of the evaluation, is on the standards for the e-prescribing system, and unless one of the standards included in the pilot is actually going to be a privacy-based standard –

MR. REYNOLDS: HIPAA privacy-based standard.

MS. McANDREW: -- then there is nothing for the pilots to test or evaluate with regard to privacy and it's not going to be a function of the pilot tests to evaluate other operative law such as the HIPAA privacy standards.

MS. FRIEDMAN: And, of course, pilot participants will have to adhere to HIPAA privacy and security laws anyway. I mean, that's a given.

MS. McANDREW: But it's not a moving component, you know; it's not a standard that is being tested through the pilot. And so the outcome of the evaluation is not going to be focused on, or in a position to direct, changes, regulatory changes, in standards that aren't part of any of these pilot-tested systems.

That's not to say that the experience gained through the pilots would not otherwise inform us as the Department and the regulator of either the security standard or the privacy standard that in this context X needs to happen. And then the determination is, can we get to X by guidance or do we get to X by a rule change?

But I think that is something that right now is outside of the standards that are being scoped up for the pilot tests.

And it's also outside of the scope of the pilot tests.

MR. ROTHSTEIN: Okay, so let me explain – I'm going to try one more time.

I could, I think, live with deleting 10.2. I could live with revising 10.2 to point out that we don't know what the various regulatory options are or what regulatory options we or the Department might want to pursue as a result of the pilot in terms of the privacy rule and e-prescribing rule. But the choice that the other Subcommittee made seems as if they made sort of a substantive determination that all that is needed as a result of this is guidance, and I don't think we've made that determination yet. That's what it seems to me. Michael, maybe you can clear this up.

DR. FITZMAURICE: I agree with everything that Sue said, and my take is that the 10.1 says as long as you're doing these pilots, try to learn something about privacy; so, identify any privacy issues that come up. And then 10.2 says, if you learn something useful, be sure to tell the industry.

Now, you could quarrel maybe a little bit with guidance and say should "report out to the industry," but I don't think anybody has a sense that we're going to go change the privacy rule on the basis of these pilots.

MR. ROTHSTEIN: Kathleen?

MS. FYFFE: My superficial reading of this implies more regulation, which is what you're saying.

I mean, HHS should identify and address any privacy issues within the context of the HIPAA privacy rule that arise during these tests. I read that and I say, uh-oh, HHS is going to come out with more regs.

DR. FITZMAURICE: No, no, they're looking for burdens and benefits.

MR. ROTHSTEIN: It's interesting. I read it just the opposite way.

MS. McANDREW: I mean, I read the parenthetical within the context of the privacy rule is you run these pilots, the rule is extant.

DR. FITZMAURICE: Applies.

MS. McANDREW: It's there. It applies as is.

MR. REYNOLDS: Yes. Throughout the document, I think we continually reference we're setting standards that are not already in place. HIPAA security and HIPAA privacy are in fact in place.

MR. ROTHSTEIN: Right.

MR. REYNOLDS: And so as this is discussed – but also we felt it was important that if you are going to do e-prescribing to take another look at privacy to see if the implementation of e-prescribing brings up other issues that were not dealt with because the difference in e-prescribing – I'm not saying this just to make the distinction – most of the HIPAA that has gone in before has been claims-related, eligibility-related, kind of non-care related, you know, non-actual giving somebody care.

So as we're looking at putting in a process that would actually be at the point of care, would actually be involved in other things, we have more information that would be downloaded from a PBM, then, you know, as part of the medical history, different than what you get – a claim is a claim is a claim – and some of the other stuff are individual transactions.

We want to make sure that at least it gets looked at. And then, we had recommended guidance. And I understand your concern that either we should recommend that if there is significant findings that something happens and whether that comes back to NCVHS or whatever.

Or, we thought about things like they've used an excellent job in using frequently asked questions, they've done an excellent job in doing some of these other things based on the other HIPAA transactions and code sets which would allow the industry to truly understand the privacy rule as it relates to e-prescribing and nothing more than a frequently asked question as to how you would address that, using that current rule, are the kind of things we also thought about.

MR. ROTHSTEIN: I'm sorry, Harry.

MR. REYNOLDS: Yes, so that's kind of it.

Let me ask Harry –

MR. REYNOLDS: Is that a fair statement?

MR. ROTHSTEIN: Let me ask Harry and Maria and others what you would think of, at the end of 10.2, just adding these words: "Communicate guidance to the industry on handling privacy issues or take other regulatory action as necessary."

DR. FITZMAURICE: I would think that by saying taking other regulatory action, you're saying that's regulatory action – putting guidance out is regulatory action. I think it is.

MR. ROTHSTEIN: Yes, or other regulatory –

DR. FITZMAURICE: I don't think that's regulatory action. I think that's informing the industry. I get the sense that this is to inform the industry. If the Committee wishes to recommend regulatory action back to HHS once they see what is learned from the pilots, that could be another set of recommendations.

MR. ROTHSTEIN: Well, all that I'm – I guess I've made my point poorly too many times already.

MR. REYNOLDS: No, you haven't. Keep going.

MR. ROTHSTEIN: I don't want to prejudge the issue before we do the pilots on what may be necessary in 2007 or whatever, and it may be that we need to do something besides issue guidance to the industry. It may be FAQs, it may be more outreach, it may be different – you know, revisions to the privacy rule or what have you. I don't know what. But as I read this, it says what we find out has just got to be used to guide industry and that seems to be too soft.

MS. FRIEDMAN: That's presuming what you find.

DR. FITZMAURICE: Depends on what you find.

MS. FRIEDMAN: Yes, I was just going to say we don't know what we're going to find.

MR. ROTHSTEIN: We may not need to do anything, and I don't want to just say this is what we're going to do.

MS. FRIEDMAN: That's why there's some benefit in leaving it "some guidance." Guidance can run the range from FAQs to regulatory action and leaving that door kind of open till we see what comes out of the pilots.

MR. ROTHSTEIN: Well, then, how about if we just say "HHS should use experience gained from the e-prescribing pilot tests to develop appropriate responses" or something like that?

MS. McANDREW: Appropriate actions.

MR. ROTHSTEIN: Appropriate actions or innovations or something. I'm happy with that. I'm happy to leave it open-ended. I don't want to pick "guidance" beforehand and make it look like we're wed to guidance and only guidance.

MR. REYNOLDS: So then what we're doing is we're leaving whatever that is up to CMS and NCVHS, right?

MR. ROTHSTEIN: Yes, exactly.

MR. REYNOLDS: But it's not just saying it's guidance.

DR. FITZMAURICE: And it doesn't urge them report out what they found, either.

MR. REYNOLDS: Well, the process, I think, is set up that the pilot tests are to be reported back to NCVHS.

MS. McANDREW: The evaluation.

MR. REYNOLDS: Do you have a problem with those words?

DR. FITZMAURICE: I don't have a problem with that at all.

MR. ROTHSTEIN: Okay, so what words are you happy with? I forgot.

MR. REYNOLDS: Your last ones.

[Laughter.]

MR. ROTHSTEIN: The clerk will repeat my –

[Laughter.]

DR. FITZMAURICE: Whatever is appropriate, I think.

MR. REYNOLDS: HHS should use experience gained from the e-prescribing pilots –

MR. ROTHSTEIN: To develop –

MS. McANDREW: -- appropriate action.

MR. ROTHSTEIN: -- to develop appropriate actions?

MS. FRIEDMAN: On handling privacy issues?

MR. REYNOLDS: Right.

MR. ROTHSTEIN: Yes.

MS. WATTENBERG: I have an additional comment. I don't know if it's welcome at this time or not.

MR. ROTHSTEIN: But I just need to make a statement. I want to celebrate my victory!

[Laughter.]

MR. ROTHSTEIN: I get so few victories, even the teensy-weensy ones I'm very happy to –

MR. REYNOLDS: We have a conference call from 5 to 6:30 at which your praises will be sung.

[Laughter.]

MR. ROTHSTEIN: Sarah?

MS. WATTENBERG: Yes. My concern is about 10.1 and the parenthetical material about identifying the privacy issues within the context of the HIPAA privacy rule. I'm sure many of you will know what I'm going to say, which is – [laughs] – is are we also going – I talked previously with Maria about the issue of the Part II and that it has sort of a higher level of restrictions on some information.

DEA also has restrictions on information about I think Category I drugs that can be electronically transmitted and whether or not state HIV laws on electronic information are also sort of regulatory issues that the e-prescribing needs to account for.

MR. REYNOLDS: One thing – again, as I said, it's an 18-page letter. That discussion is dealt with in a significant way in numerous places throughout this, because it is – I mean, obviously since we were the standards and security. I mean, the whole idea, and the DEA testified, and the whole idea of whether or not e-prescribing as it exists now, if the DEA comes in and puts significantly other monitoring on the controlled substances, what that would or wouldn't do to adoption and what that does or doesn't mean and how they need to work together. And we recommended HHS works with them.

So that's predominant throughout because that is a significant issue because you've got non-controlled – 85 percent of the drugs, I think it was, were non-controlled, in the testimony had and then 12 percent were Schedules III to V and two to three percent were Schedule IIs, okay? So we address that in significant –

MR. ROTHSTEIN: Harry, one way I think clearing that up perhaps to Sarah's satisfaction, if you take a look at the end of the second paragraph under Observation 10, because we talk about drug abuse, the last sentence says any e-prescribing regulations must consider these – that is, the ones mentioned in that paragraph – and other health records laws.

One thing that we could do and let me see what you think of it is in that parenthetical, in 1, say HHS should identify and address any privacy issues, paren, within the context of HIPAA privacy rule and other health records laws which tracks the prior language. Would that be okay?

MR. REYNOLDS: Yes.

MS. CHAPPER: And other health records laws?

MR. ROTHSTEIN: Well, yes. That's the language that we used at the end of the second paragraph.

MS. McANDREW: Now is that your triumph or mine?

MR. ROTHSTEIN: No, I'll give you credit for that –

[Laughter.]

MR. REYNOLDS: We only have an hour and a half conference call, so if all of you are going to need accolades, write them short, nice and short.

MR. ROTHSTEIN: Okay. Michael, did you have –

DR. FITZMAURICE: No, no.

MR. ROTHSTEIN: -- something else? Are you okay with that?

DR. FITZMAURICE: Yes.

MR. ROTHSTEIN: Other comments? So I would entertain a motion to endorse the revisions as amended. Dr. Harding?

DR. HARDING: To endorse the paragraphs as amended?

MR. ROTHSTEIN: Correct. Any opposed? I mean – a second? Any discussion? All in favor say "aye."

PARTICIPANTS: Aye.

MR. ROTHSTEIN: Opposed? Okay, so we are on the record as supporting that.

MR. REYNOLDS: We would like it noted in the record that it was unanimous vote and camaraderie amongst the Committee.

MR. ROTHSTEIN: Yes, absolutely.

Okay, it's now 4:15, and do I hear any requests for additional business to consider? Hearing none, I am very pleased to adjourn nearly a half an hour early, and we will start tomorrow rain or shine, snow or sleet, at 9 o'clock. Thank you on the Internet for hanging with us. And we'll see you tomorrow.

[Hearing adjourned at 4:18 P.M.]